Technology that knows what you're feeling | Poppy Crum

138,587 views ・ 2018-07-10

TED


請雙擊下方英文字幕播放視頻。

譯者: Clare Wong 審譯者: Karen Lo
00:12
What happens when technology knows more about us than we do?
0
12881
4456
當科技比我們更了解自己時, 會發生甚麼事呢?
00:17
A computer now can detect our slightest facial microexpressions
1
17992
3672
現今的電腦,
可以偵測到我們的微表情──
00:21
and be able to tell the difference between a real smile and a fake one.
2
21688
3611
它分辨得出真笑和假笑。
00:25
That's only the beginning.
3
25323
1734
但這只是個開始。
00:27
Technology has become incredibly intelligent
4
27466
2865
科技日益進步,
00:30
and already knows a lot about our internal states.
5
30355
3400
它知道很多我們的內心世界。
00:34
And whether we like it or not,
6
34085
2286
不管我們喜歡與否,
00:36
we already are sharing parts of our inner lives
7
36395
3499
我們已將自己部分內心世界分享出去,
00:39
that's out of our control.
8
39918
1733
這不在我們的掌控之中。
00:43
That seems like a problem,
9
43413
1421
那似乎是個問題,
00:44
because a lot of us like to keep what's going on inside
10
44858
3246
因為我們當中很多人,
喜歡隱藏內心想法,
讓旁人無從察覺。
00:48
from what people actually see.
11
48128
1647
00:50
We want to have agency over what we share and what we don't.
12
50323
4420
我們想要掌握自主權去選擇 什麼可被分享、什麼不可以。
00:55
We all like to have a poker face.
13
55473
2321
我們都想要張撲克臉。
00:59
But I'm here to tell you that I think that's a thing of the past.
14
59584
3346
但我在這裡告訴你們, 我認為這種想法已經過時。
01:03
And while that might sound scary, it's not necessarily a bad thing.
15
63347
4770
儘管這件事好像頗為嚇人,
但未必是件壞事。
01:09
I've spent a lot of time studying the circuits in the brain
16
69030
2770
我花了很多時間研究腦內的迴路,
01:11
that create the unique perceptual realities that we each have.
17
71824
3693
它創造了我們獨特的感知。
01:16
And now I bring that together
18
76110
1405
現在我把它和當今科技結合,
01:17
with the capabilities of current technology
19
77539
2062
01:19
to create new technology that does make us better,
20
79625
2537
創造新技術,令我們更好、
01:22
feel more, connect more.
21
82186
1600
感受更多、聯繫更多。
01:24
And I believe to do that,
22
84482
1786
而我認為要達到這個目的,
01:26
we have to be OK losing some of our agency.
23
86292
2749
我們要接受失去一些自主權。
01:30
With some animals, it's really amazing,
24
90149
2523
和某些動物在一起時,真的很奇妙。
01:32
and we get to see into their internal experiences.
25
92696
3474
我們可知道牠們的內心經歷。
01:36
We get this upfront look at the mechanistic interaction
26
96649
3722
我們可直接得知
牠們對周圍環境的反應
01:40
between how they respond to the world around them
27
100395
2817
和生理系統狀態之間的相互影響。
01:43
and the state of their biological systems.
28
103236
2008
01:45
This is where evolutionary pressures like eating, mating
29
105268
3809
這就是來自進化的壓力, 像是進食、交配
和確保我們不會被吃掉,
01:49
and making sure we don't get eaten
30
109101
1762
01:50
drive deterministic behavioral responses to information in the world.
31
110887
4157
促使確定性行為 回應我們身邊的資訊。
01:55
And we get to see into this window,
32
115806
2794
而我們透過這個窗口,
01:58
into their internal states and their biological experiences.
33
118624
3636
透視牠們的心理狀態及生理變化。
02:02
It's really pretty cool.
34
122284
1642
這真的很酷。
02:03
Now, stay with me for a moment -- I'm a violinist, not a singer.
35
123950
4103
現在請你們忍受我一會兒。
我是個小提琴手,不是歌唱家。
02:08
But the spider's already given me a critical review.
36
128077
3590
但這隻蜘蛛已挑剔地批評我。
02:16
(Video) (Singing in a low pitch)
37
136907
2060
(低音唱歌)
02:19
(Singing in a middle pitch)
38
139868
2888
(中音唱歌)
02:23
(Singing in a high pitch)
39
143800
2505
(高音唱歌)
02:27
(Singing in a low pitch)
40
147069
1421
(低音唱歌)
02:29
(Singing in a middle pitch)
41
149236
1600
(中音唱歌)
02:31
(Singing in a high pitch)
42
151403
1777
(高音唱歌)
02:33
(Laughter)
43
153204
1150
(笑聲)
02:36
Poppy Crum: It turns out, some spiders tune their webs like violins
44
156387
3198
由此可知, 有些蜘蛛會調節自己的蜘蛛網,
02:39
to resonate with certain sounds.
45
159609
2158
像小提琴一樣調至特定的音階。
02:41
And likely, the harmonics of my voice as it went higher
46
161791
2771
同樣地,我唱歌的音調
隨分貝數拉高,
02:44
coupled with how loud I was singing
47
164586
1730
02:46
recreated either the predatory call of an echolocating bat or a bird,
48
166340
4467
重新創造類似捕獵者, 蝙蝠或鳥的回聲定位。
02:50
and the spider did what it should.
49
170831
1881
而蜘蛛做了牠應該做的事。
02:53
It predictively told me to bug off.
50
173300
2817
牠預言性地叫我離開。
02:56
I love this.
51
176824
1150
我愛這樣。
02:58
The spider's responding to its external world
52
178546
3309
蜘蛛回應外在環境時,
03:01
in a way that we get to see and know what's happening to its internal world.
53
181879
4350
我們藉此看到並了解牠的內心世界。
03:07
Biology is controlling the spider's response;
54
187069
2206
生理控制著蜘蛛的反應,
03:09
it's wearing its internal state on its sleeve.
55
189299
2776
牠把自己最真實的情感流露出來。
03:13
But us, humans --
56
193768
1655
但我們,人類 ──
03:16
we're different.
57
196184
1150
我們不一樣。
03:17
We like to think we have cognitive control over what people see, know and understand
58
197899
5735
我們自認能夠控制自己的感知能力。
像是讓別人怎樣看待、知道和理解
03:23
about our internal states --
59
203658
1409
我們的內心世界、
03:25
our emotions, our insecurities, our bluffs, our trials and tribulations --
60
205091
4303
我們的情感、我們的局促不安、
我們的虛張聲勢、 我們的麻煩和艱難⋯⋯
03:29
and how we respond.
61
209418
1267
及我們的反應。
03:31
We get to have our poker face.
62
211927
2282
我們需要一張撲克臉。
03:35
Or maybe we don't.
63
215799
1200
但或者,我們其實並不需要。
03:37
Try this with me.
64
217728
1182
跟我一起嘗試這樣做。
03:38
Your eye responds to how hard your brain is working.
65
218934
2690
你的眼睛能反映出大腦使用的程度。
03:42
The response you're about to see is driven entirely by mental effort
66
222363
3230
接下來你看到的反應
完全出於心理因素。
03:45
and has nothing to do with changes in lighting.
67
225617
2635
和亮度沒有關係。
03:48
We know this from neuroscience.
68
228276
1650
我保證這有神經科學的根據。
03:49
I promise, your eyes are doing the same thing as the subject in our lab,
69
229950
4560
你的眼睛和實驗室受試者的 眼睛會做相同的事,
03:54
whether you want them to or not.
70
234534
1734
無論你願意與否。
03:56
At first, you'll hear some voices.
71
236292
2173
首先你會聽到一些聲音。
03:58
Try and understand them and keep watching the eye in front of you.
72
238489
3278
嘗試去了解, 而且定睛於你面前的眼睛。
04:01
It's going to be hard at first,
73
241791
1498
一開始會有點難。
要放棄真的很容易。
04:03
one should drop out, and it should get really easy.
74
243313
2391
從瞳孔的直徑 你可以看見心理因素的變化。
04:05
You're going to see the change in effort in the diameter of the pupil.
75
245728
3325
04:10
(Video) (Two overlapping voices talking)
76
250140
2567
(兩種聲音重疊) 智能科技仰賴個人數據。
04:12
(Single voice) Intelligent technology depends on personal data.
77
252731
2963
(一種聲音)智能科技仰賴個人數據。
04:15
(Two overlapping voices talking)
78
255718
2446
(兩種聲音重疊) 智能科技仰賴個人數據。
04:18
(Single voice) Intelligent technology depends on personal data.
79
258188
2976
(一種聲音)智能科技仰賴個人數據。
04:21
PC: Your pupil doesn't lie.
80
261680
1326
你的瞳孔不會說謊。
04:23
Your eye gives away your poker face.
81
263030
2400
你的眼睛拆穿了你的撲克臉。
04:25
When your brain's having to work harder,
82
265990
1913
當腦袋愈努力地想事情,
04:27
your autonomic nervous system drives your pupil to dilate.
83
267927
2785
自主神經系統驅使瞳孔擴大。
04:30
When it's not, it contracts.
84
270736
1555
在相反的情況下,瞳孔縮小。
04:32
When I take away one of the voices,
85
272680
1691
當只有一種聲音時,
04:34
the cognitive effort to understand the talkers
86
274395
2262
理解說話內容需要的精力減少。
04:36
gets a lot easier.
87
276681
1158
04:37
I could have put the two voices in different spatial locations,
88
277863
3000
我把兩種聲音分別放在兩個地方,
04:40
I could have made one louder.
89
280887
1666
如果我將其中一種聲音調大聲,
04:42
You would have seen the same thing.
90
282577
1738
出來的效果也是一樣。
04:45
We might think we have more agency over the reveal of our internal state
91
285006
4786
我們可能自認為比蜘蛛有更多自主權
能控制披露自己的內心狀態與否,
04:49
than that spider,
92
289816
1579
04:51
but maybe we don't.
93
291419
1266
但或許不如我們所想像的。
04:53
Today's technology is starting to make it really easy
94
293021
2969
現今科技很容易就讀懂
04:56
to see the signals and tells that give us away.
95
296014
2690
我們內心深處釋出的信號。
04:59
The amalgamation of sensors paired with machine learning
96
299109
3294
傳感器連同機器學習,
05:02
on us, around us and in our environments,
97
302427
2413
與我們、周遭、環境相結合,
05:04
is a lot more than cameras and microphones tracking our external actions.
98
304864
4653
遠遠不止追蹤我們行動的 相機和麥克風而已。
05:12
Our bodies radiate our stories
99
312529
2818
我們身體的生理溫度變化
05:15
from changes in the temperature of our physiology.
100
315371
2666
輻射出我們的故事。
05:18
We can look at these as infrared thermal images
101
318546
2261
看看我身後的紅外線溫度圖。
05:20
showing up behind me,
102
320831
1160
05:22
where reds are hotter and blues are cooler.
103
322015
2070
紅色代表較高溫,藍色代表低溫。
05:24
The dynamic signature of our thermal response
104
324458
3183
動態的身體溫度徵象
05:27
gives away our changes in stress,
105
327665
3031
讓我們對壓力的反應露了餡,
05:30
how hard our brain is working,
106
330720
2008
我們有多努力用腦想問題,
05:32
whether we're paying attention
107
332752
1936
我們和別人聊天時是否專注,
05:34
and engaged in the conversation we might be having
108
334712
2627
05:37
and even whether we're experiencing a picture of fire as if it were real.
109
337363
4095
甚至我們是否像身歷其境遭火炙。
05:41
We can actually see people give off heat on their cheeks
110
341482
2643
人們的臉頰真的會散發出熱力
05:44
in response to an image of flame.
111
344149
2200
來回應火焰的影像。
05:48
But aside from giving away our poker bluffs,
112
348013
2929
除了假裝冷漠露了餡,
05:50
what if dimensions of data from someone's thermal response
113
350966
4746
如果反應的人體溫度數據
05:55
gave away a glow of interpersonal interest?
114
355736
2659
也讓人際關係露了餡呢?
05:58
Tracking the honesty of feelings in someone's thermal image
115
358966
3532
用溫度圖追蹤人的真誠感受
06:02
might be a new part of how we fall in love and see attraction.
116
362522
3626
可能是部分墜入愛河 和偵測吸引力的新穎方式。
06:06
Our technology can listen, develop insights and make predictions
117
366172
3693
我們的科技可以聆聽、洞察和預測
06:09
about our mental and physical health
118
369889
2095
我們心理和生理的健康,
06:12
just by analyzing the timing dynamics of our speech and language
119
372008
4000
只需分析麥克風收到的 語言節奏變化。
06:16
picked up by microphones.
120
376032
1443
06:18
Groups have shown that changes in the statistics of our language
121
378038
3880
資料顯示,透過搭配
人們的語言變化數據和機器,
06:21
paired with machine learning
122
381942
1420
06:23
can predict the likelihood someone will develop psychosis.
123
383386
3161
可以預測一個人有精神病的可能性。
06:27
I'm going to take it a step further
124
387442
1751
我將更深入地探討,
06:29
and look at linguistic changes and changes in our voice
125
389217
2587
在不同的情況下,
06:31
that show up with a lot of different conditions.
126
391828
2239
我們語言和聲音上的改變。
06:34
Dementia, diabetes can alter the spectral coloration of our voice.
127
394091
4367
老年癡呆和糖尿病 可改變聲音中的光譜顏色。
06:39
Changes in our language associated with Alzheimer's
128
399205
3119
與阿茲海默症有關的語言變化
06:42
can sometimes show up more than 10 years before clinical diagnosis.
129
402348
4365
能在臨床診斷的前十年就被發現。
06:47
What we say and how we say it tells a much richer story
130
407236
3960
我們表達的內容和怎樣表達
比我們想像中透露出更多資訊。
06:51
than we used to think.
131
411220
1254
06:53
And devices we already have in our homes could, if we let them,
132
413022
4047
家中已有的儀器,如果我們允許,
06:57
give us invaluable insight back.
133
417093
2134
可以讓它們把珍貴的數據傳送回來。
06:59
The chemical composition of our breath
134
419998
2978
我們呼吸中的化學成分,
07:03
gives away our feelings.
135
423959
1354
能透露我們的感覺。
07:06
There's a dynamic mixture of acetone, isoprene and carbon dioxide
136
426363
4478
由丙酮、異戊二烯和二氧化碳 所組成的混合物成分,
07:10
that changes when our heart speeds up, when our muscles tense,
137
430865
3384
在心跳加速、肌肉緊張時不斷變化。
07:14
and all without any obvious change in our behaviors.
138
434809
2897
但我們的行為不會有明顯的改變。
07:18
Alright, I want you to watch this clip with me.
139
438268
2738
請你們跟我一同看這段影片。
07:21
Some things might be going on on the side screens,
140
441030
3119
有些東西可能出現在兩旁的螢幕,
07:24
but try and focus on the image in the front
141
444173
3777
但嘗試把注意力集中在前面的影像,
07:27
and the man at the window.
142
447974
1463
和窗戶旁的男人。
07:31
(Eerie music)
143
451633
2658
(怪異的音樂)
07:39
(Woman screams)
144
459767
1437
(女人的尖叫)
07:50
PC: Sorry about that. I needed to get a reaction.
145
470692
2395
抱歉讓你們經歷這些, 我需要你們的反應。
07:53
(Laughter)
146
473111
1785
(笑聲)
07:55
I'm actually tracking the carbon dioxide you exhale in the room right now.
147
475412
4972
我正追蹤你們在這房間 呼出的二氧化碳。
08:01
We've installed tubes throughout the theater,
148
481903
3293
我們在整間講廳安裝了管道,
08:05
lower to the ground, because CO2 is heavier than air.
149
485220
2595
在地下,
因為二氧化碳比空氣重。
08:07
But they're connected to a device in the back
150
487839
2667
這些管道連接後面的儀器,
08:10
that lets us measure, in real time, with high precision,
151
490530
3287
可以實時準確地計算
08:13
the continuous differential concentration of CO2.
152
493841
2922
一連串的二氧化碳濃度變化。
08:17
The clouds on the sides are actually the real-time data visualization
153
497246
5508
旁邊的雲圖顯示出這裡
08:22
of the density of our CO2.
154
502778
1998
二氧化碳濃度。
08:25
You might still see a patch of red on the screen,
155
505374
3699
你可能見到螢幕上有一小塊紅色,
08:29
because we're showing increases with larger colored clouds,
156
509097
3705
因為它顯示有顏色的雲團增多,
08:32
larger colored areas of red.
157
512826
2196
紅色的區域增多。
08:35
And that's the point where a lot of us jumped.
158
515046
2559
這正是顯示我們被嚇到的時候。
08:38
It's our collective suspense driving a change in carbon dioxide.
159
518173
4915
是我們集體的焦慮 導致二氧化碳變化。
08:43
Alright, now, watch this with me one more time.
160
523649
2722
現在再看一次這段影片。
08:46
(Cheerful music)
161
526395
2238
(快樂的音樂)
08:54
(Woman laughs)
162
534553
2137
(女人的笑聲)
09:05
PC: You knew it was coming.
163
545344
1349
你們已經知道會發生什麼事。
09:06
But it's a lot different when we changed the creator's intent.
164
546717
3363
當我們改變了創作者的意圖, 差別會很大。
09:10
Changing the music and the sound effects
165
550776
2769
轉換了音樂和音效
09:13
completely alter the emotional impact of that scene.
166
553569
3603
完全改變那場景帶來的情緒影響。
09:17
And we can see it in our breath.
167
557196
2134
從我們的呼吸可以知道,
09:20
Suspense, fear, joy
168
560196
2262
焦慮、恐懼、歡樂,
09:22
all show up as reproducible, visually identifiable moments.
169
562482
4507
都顯示可重複性, 視覺上可以辨識的時刻。
09:27
We broadcast a chemical signature of our emotions.
170
567473
4151
這些化學特質傳達了我們的情緒。
09:35
It is the end of the poker face.
171
575249
2133
撲克臉的時代將就此終結。
09:38
Our spaces, our technology will know what we're feeling.
172
578582
3566
我們的空間、科技 將會知道我們的情感。
09:42
We will know more about each other than we ever have.
173
582736
2785
我們將前所未有地更了解彼此。
09:45
We get a chance to reach in and connect to the experience and sentiments
174
585911
4307
我們有機會在情感上與社交上,
09:50
that are fundamental to us as humans
175
590242
1742
交換那些身為人類
09:52
in our senses, emotionally and socially.
176
592008
2410
基本擁有的經驗與觀點。
09:55
I believe it is the era of the empath.
177
595482
2540
我相信這是一個同理心的時代。
09:58
And we are enabling the capabilities that true technological partners can bring
178
598046
5222
我們正在實現那些能力,
是真正的技術合作夥伴 能夠帶來的能力,
10:03
to how we connect with each other and with our technology.
179
603292
3047
它能夠連結我們彼此,
也能連結我們與技術。
10:06
If we recognize the power of becoming technological empaths,
180
606363
3389
如果意識到我們能用技術 達到同理心的力量,
10:09
we get this opportunity
181
609776
1936
我們就掌有機會,
10:11
where technology can help us bridge the emotional and cognitive divide.
182
611736
4424
用科技幫助我們跨越 情感和認知的鴻溝。
10:16
And in that way, we get to change how we tell our stories.
183
616680
2723
如此一來,我們就能改變敘事的方法。
10:19
We can enable a better future for technologies like augmented reality
184
619427
3580
我們能為「擴增實境」等技術 創造更美好的未來,
10:23
to extend our own agency and connect us at a much deeper level.
185
623031
4193
擴大我們的自主權,
並令彼此更深入地連結。
10:27
Imagine a high school counselor being able to realize
186
627625
2547
試想像高中的輔導老師
能注意到表面歡樂的學生 其實正經歷著艱難的日子。
10:30
that an outwardly cheery student really was having a deeply hard time,
187
630196
3826
10:34
where reaching out can make a crucial, positive difference.
188
634046
3180
給予這名學生支持, 藉此得到關鍵且正向的結果。
10:37
Or authorities, being able to know the difference
189
637766
3230
或是讓權威人士有能力分辨
10:41
between someone having a mental health crisis
190
641020
2325
一個人是有心理健康障礙
10:43
and a different type of aggression,
191
643369
1826
還是具有侵略性,
10:45
and responding accordingly.
192
645219
1800
並根據情況作出回應。
10:47
Or an artist, knowing the direct impact of their work.
193
647609
3273
或者讓藝術工作者 能知曉其作品對人們的影響。
10:52
Leo Tolstoy defined his perspective of art
194
652173
2643
列夫·托爾斯泰將藝術定義為
10:54
by whether what the creator intended
195
654840
1785
創造者透過作品 把自己的想法傳達給觀賞者。
10:56
was experienced by the person on the other end.
196
656649
2586
10:59
Today's artists can know what we're feeling.
197
659259
2566
如今藝術工作者能知道我們的感受。
11:02
But regardless of whether it's art or human connection,
198
662204
3005
但無論是藝術或人際的連結,
11:06
today's technologies will know and can know
199
666608
2802
當今科技將會知道,能夠知道
11:09
what we're experiencing on the other side,
200
669434
2048
我們不為人知的經歷。
11:11
and this means we can be closer and more authentic.
201
671506
2618
這代表我們之間會更接近、更真誠。
11:14
But I realize a lot of us have a really hard time
202
674498
4293
但我明白有很多人抗拒
11:18
with the idea of sharing our data,
203
678815
2267
分享與自身相關數據的做法,
11:21
and especially the idea that people know things about us
204
681673
3111
特別是別人能洞悉
11:24
that we didn't actively choose to share.
205
684808
2321
我們不想與之分享的想法。
11:28
Anytime we talk to someone,
206
688728
2216
無論何時當我們和人交談、
11:31
look at someone
207
691946
1555
看著別人、
11:33
or choose not to look,
208
693525
1468
或不看時,
11:35
data is exchanged, given away,
209
695017
2647
數據已被交換並傳遞出去。
11:38
that people use to learn,
210
698533
2205
人們利用這些數據了解我們
11:40
make decisions about their lives and about ours.
211
700762
3267
來藉此做了和我們相關的決定。
11:45
I'm not looking to create a world where our inner lives are ripped open
212
705469
3968
我不是想創造一個 內心生活要被剖開的世界──
11:49
and our personal data and our privacy given away
213
709461
2348
把我們的私人數據和隱私
11:51
to people and entities where we don't want to see it go.
214
711833
2713
給我們不想給的人看。
11:55
But I am looking to create a world
215
715117
2762
我想創造一個世界
11:57
where we can care about each other more effectively,
216
717903
3408
可以讓我們更有效地關心彼此。
12:01
we can know more about when someone is feeling something
217
721335
3060
當有人需要被關心時,
12:04
that we ought to pay attention to.
218
724419
1872
我們可以獲得更多資訊。
12:06
And we can have richer experiences from our technology.
219
726800
3335
科技讓我們有更豐富的經驗。
12:10
Any technology can be used for good or bad.
220
730887
2357
任何科技都可用來行善或犯罪。
12:13
Transparency to engagement and effective regulation
221
733268
2412
參與的透明度和有效的規範,
12:15
are absolutely critical to building the trust for any of this.
222
735704
3120
這兩項要素能幫助科技建立信任。
12:20
But the benefits that "empathetic technology" can bring to our lives
223
740106
4834
同理心技術帶給生活的好處,
12:24
are worth solving the problems that make us uncomfortable.
224
744964
3891
值得讓我們解決令人不安的問題。
12:29
And if we don't, there are too many opportunities and feelings
225
749315
4025
如果我們不解決的話, 會有太多的機會和感受
12:33
we're going to be missing out on.
226
753364
1695
被我們忽略。
12:35
Thank you.
227
755083
1175
謝謝!
12:36
(Applause)
228
756282
2479
(掌聲)
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隱私政策

eng.lish.video

Developer's Blog