Technology that knows what you're feeling | Poppy Crum

138,587 views ・ 2018-07-10

TED


请双击下面的英文字幕来播放视频。

翻译人员: Ruilai Xue 校对人员: jacks peng
00:12
What happens when technology knows more about us than we do?
0
12881
4456
当科技比我们还了解我们自己,会发生什么?
00:17
A computer now can detect our slightest facial microexpressions
1
17992
3672
目前的电脑已经可以解读 我们最精细的微表情,
00:21
and be able to tell the difference between a real smile and a fake one.
2
21688
3611
能够区分假笑和真笑。
00:25
That's only the beginning.
3
25323
1734
这仅仅只是开始。
00:27
Technology has become incredibly intelligent
4
27466
2865
科技已经变得极度智能,
00:30
and already knows a lot about our internal states.
5
30355
3400
而且已经知道很多我们的内心状态。
00:34
And whether we like it or not,
6
34085
2286
不管喜欢与否,
00:36
we already are sharing parts of our inner lives
7
36395
3499
我们已经在分享我们的一部分私生活了,
00:39
that's out of our control.
8
39918
1733
但这一过程却不受我们控制。
00:43
That seems like a problem,
9
43413
1421
这似乎是个麻烦,
00:44
because a lot of us like to keep what's going on inside
10
44858
3246
因为很多人只想展示自己的表面,
00:48
from what people actually see.
11
48128
1647
而隐藏自己的内心。
00:50
We want to have agency over what we share and what we don't.
12
50323
4420
我们希望对我们愿意和不愿意 分享的东西拥有控制权。
00:55
We all like to have a poker face.
13
55473
2321
我们都喜欢有一张不露声色的脸。
00:59
But I'm here to tell you that I think that's a thing of the past.
14
59584
3346
但我来是想告诉你,这个想法已经过时了。
01:03
And while that might sound scary, it's not necessarily a bad thing.
15
63347
4770
虽然听起来可怕,但这不一定是坏事。
01:09
I've spent a lot of time studying the circuits in the brain
16
69030
2770
我花了很多时间研究那些
01:11
that create the unique perceptual realities that we each have.
17
71824
3693
创造出我们感官世界的脑回路。
01:16
And now I bring that together
18
76110
1405
现在我将这些研究成果
01:17
with the capabilities of current technology
19
77539
2062
和当前的科技水平结合在一起,
01:19
to create new technology that does make us better,
20
79625
2537
来创造出一种全新的科技,帮助扩展
01:22
feel more, connect more.
21
82186
1600
我们的感受,和建立连接的能力。
01:24
And I believe to do that,
22
84482
1786
而且我相信要实现这一点,
01:26
we have to be OK losing some of our agency.
23
86292
2749
我们需要接受失去部分控制力的后果。
01:30
With some animals, it's really amazing,
24
90149
2523
动物实验的研究结果很令人惊叹,
01:32
and we get to see into their internal experiences.
25
92696
3474
我们可以看到它们的内在感受。
01:36
We get this upfront look at the mechanistic interaction
26
96649
3722
我们得以全面了解
01:40
between how they respond to the world around them
27
100395
2817
它们回应世界的方式,
01:43
and the state of their biological systems.
28
103236
2008
以及它们的生物系统状态。
01:45
This is where evolutionary pressures like eating, mating
29
105268
3809
在这里,像进食,求偶 以及保护自己不被吃掉
01:49
and making sure we don't get eaten
30
109101
1762
这些进化压力
01:50
drive deterministic behavioral responses to information in the world.
31
110887
4157
驱使着决定性的行为, 对周遭世界的信息进行回应。
01:55
And we get to see into this window,
32
115806
2794
我们可以通过这个窗口,
01:58
into their internal states and their biological experiences.
33
118624
3636
看到它们的内心状态与生理感受。
02:02
It's really pretty cool.
34
122284
1642
这非常有意思。
02:03
Now, stay with me for a moment -- I'm a violinist, not a singer.
35
123950
4103
现在,认真听——虽然我 不是歌手,我是拉小提琴的。
02:08
But the spider's already given me a critical review.
36
128077
3590
但这个蜘蛛已经给了我批判性的建议。
02:16
(Video) (Singing in a low pitch)
37
136907
2060
(视频)(低音歌唱)
02:19
(Singing in a middle pitch)
38
139868
2888
(音调升高)
02:23
(Singing in a high pitch)
39
143800
2505
(高音歌唱)
02:27
(Singing in a low pitch)
40
147069
1421
(低音歌唱)
02:29
(Singing in a middle pitch)
41
149236
1600
(音调升高)
02:31
(Singing in a high pitch)
42
151403
1777
(高音歌唱)
02:33
(Laughter)
43
153204
1150
(笑声 )
02:36
Poppy Crum: It turns out, some spiders tune their webs like violins
44
156387
3198
波比 · 克拉姆:事实上,很多蜘蛛 会把它们的网调得像小提琴一样,
02:39
to resonate with certain sounds.
45
159609
2158
来和某些声音产生共振。
02:41
And likely, the harmonics of my voice as it went higher
46
161791
2771
相似的,当我的声调随着
02:44
coupled with how loud I was singing
47
164586
1730
音量变大而提高时,
02:46
recreated either the predatory call of an echolocating bat or a bird,
48
166340
4467
就重造出了蝙蝠或鸟类 等捕食者靠近的警告,
02:50
and the spider did what it should.
49
170831
1881
然后蜘蛛做出了本能的反应。
02:53
It predictively told me to bug off.
50
173300
2817
它在警告我走开。
02:56
I love this.
51
176824
1150
我很喜欢这个现象。
02:58
The spider's responding to its external world
52
178546
3309
蜘蛛在对外在世界进行回应,
03:01
in a way that we get to see and know what's happening to its internal world.
53
181879
4350
在某种程度上,我们可以借此看见 并了解它内心世界的样子。
03:07
Biology is controlling the spider's response;
54
187069
2206
生物学控制着蜘蛛的回应;
03:09
it's wearing its internal state on its sleeve.
55
189299
2776
它的内心状态体现了出来。
03:13
But us, humans --
56
193768
1655
但是我们,人类,
03:16
we're different.
57
196184
1150
我们不一样。
03:17
We like to think we have cognitive control over what people see, know and understand
58
197899
5735
我们希望可以控制他人看到, 认识以及理解到的
03:23
about our internal states --
59
203658
1409
我们的内心状态——
03:25
our emotions, our insecurities, our bluffs, our trials and tribulations --
60
205091
4303
我们的情感,不安全感, 我们的虚张声势,我们的磨难,
03:29
and how we respond.
61
209418
1267
以及我们回应的方式。
03:31
We get to have our poker face.
62
211927
2282
我们必须面不改色。
03:35
Or maybe we don't.
63
215799
1200
或者也许,我们不用这样做。
03:37
Try this with me.
64
217728
1182
和我一起试试。
03:38
Your eye responds to how hard your brain is working.
65
218934
2690
你的眼睛会反映你的用脑情况。
03:42
The response you're about to see is driven entirely by mental effort
66
222363
3230
你即将看到的反应 完全是由心理影响的,
03:45
and has nothing to do with changes in lighting.
67
225617
2635
与光线的改变毫无关系。
03:48
We know this from neuroscience.
68
228276
1650
我们从神经学了解到这一信息。
03:49
I promise, your eyes are doing the same thing as the subject in our lab,
69
229950
4560
我保证,你们的眼睛会发生 与我们实验室的受试者一样的变化,
03:54
whether you want them to or not.
70
234534
1734
无论你想不想这么做。
03:56
At first, you'll hear some voices.
71
236292
2173
最初,你会听到一些声音。
03:58
Try and understand them and keep watching the eye in front of you.
72
238489
3278
试着理解它们,并注视着你面前的眼睛。
04:01
It's going to be hard at first,
73
241791
1498
一上来可能很难。
一个声音消失后,就会简单很多。
04:03
one should drop out, and it should get really easy.
74
243313
2391
你们会看到瞳孔的直径发生了改变。
04:05
You're going to see the change in effort in the diameter of the pupil.
75
245728
3325
04:10
(Video) (Two overlapping voices talking)
76
250140
2567
(视频)(两个重叠的声音在讲话)
04:12
(Single voice) Intelligent technology depends on personal data.
77
252731
2963
(一个声音在说话) 智能科技取决于个人数据。
04:15
(Two overlapping voices talking)
78
255718
2446
(两个重叠的声音在讲话)
04:18
(Single voice) Intelligent technology depends on personal data.
79
258188
2976
(一个声音在说话) 智能科技取决于个人数据。
04:21
PC: Your pupil doesn't lie.
80
261680
1326
波比 · 克拉姆: 你的瞳孔不会撒谎。
04:23
Your eye gives away your poker face.
81
263030
2400
你的眼睛泄露了你真实的心理活动。
04:25
When your brain's having to work harder,
82
265990
1913
当你的大脑需要更快地运转时,
04:27
your autonomic nervous system drives your pupil to dilate.
83
267927
2785
自动神经系统会导致瞳孔扩张,
04:30
When it's not, it contracts.
84
270736
1555
反过来则会导致收缩。
04:32
When I take away one of the voices,
85
272680
1691
当我移除其中任何一个声道时,
用来理解说话者的认知努力
04:34
the cognitive effort to understand the talkers
86
274395
2262
04:36
gets a lot easier.
87
276681
1158
变得轻松很多。
04:37
I could have put the two voices in different spatial locations,
88
277863
3000
我可以把两个声音放在不同位置,
04:40
I could have made one louder.
89
280887
1666
或把其中一个调大,
04:42
You would have seen the same thing.
90
282577
1738
你会看到同样的结果。
04:45
We might think we have more agency over the reveal of our internal state
91
285006
4786
我们或许认为,我们掌控展现 自己内心情况的能力
04:49
than that spider,
92
289816
1579
比蜘蛛更强,
04:51
but maybe we don't.
93
291419
1266
但或许并不是这样。
04:53
Today's technology is starting to make it really easy
94
293021
2969
现在的科技水平使得
04:56
to see the signals and tells that give us away.
95
296014
2690
识别这些信号然后暴露 我们自己变得非常容易。
04:59
The amalgamation of sensors paired with machine learning
96
299109
3294
在我们身上,在我们四周与环境中
05:02
on us, around us and in our environments,
97
302427
2413
与机器学习相结合的传感器,
05:04
is a lot more than cameras and microphones tracking our external actions.
98
304864
4653
远不止监控着我们外在 动作的摄像头与麦克风。
05:12
Our bodies radiate our stories
99
312529
2818
我们的身体通过生理温度的改变
05:15
from changes in the temperature of our physiology.
100
315371
2666
辐射着我们的故事。
05:18
We can look at these as infrared thermal images
101
318546
2261
看一下我背后的
05:20
showing up behind me,
102
320831
1160
这些红外图像,
红色代表温度高,蓝色代表温度低。
05:22
where reds are hotter and blues are cooler.
103
322015
2070
05:24
The dynamic signature of our thermal response
104
324458
3183
我们热响应的动态信号
05:27
gives away our changes in stress,
105
327665
3031
暴露了我们的压力变化,
05:30
how hard our brain is working,
106
330720
2008
大脑的负荷程度,
05:32
whether we're paying attention
107
332752
1936
我们是否在关注
05:34
and engaged in the conversation we might be having
108
334712
2627
并参与我们可能正在进行的对话,
05:37
and even whether we're experiencing a picture of fire as if it were real.
109
337363
4095
甚至我们是否如实地 在感受一张火焰的图片。
05:41
We can actually see people give off heat on their cheeks
110
341482
2643
我们可以看见人们脸颊的温度上升,
05:44
in response to an image of flame.
111
344149
2200
这是对火焰图片做出的回应。
05:48
But aside from giving away our poker bluffs,
112
348013
2929
但除了泄露我们的情绪伪装,
05:50
what if dimensions of data from someone's thermal response
113
350966
4746
如果某人的热响应数据
05:55
gave away a glow of interpersonal interest?
114
355736
2659
还暴露了其人际交往的兴趣呢?
05:58
Tracking the honesty of feelings in someone's thermal image
115
358966
3532
追踪某人热学图像里的 真挚情感,或许会为理解
06:02
might be a new part of how we fall in love and see attraction.
116
362522
3626
我们如何坠入爱河 或被吸引提供新的信息。
06:06
Our technology can listen, develop insights and make predictions
117
366172
3693
仅仅通过分析麦克风接收的
06:09
about our mental and physical health
118
369889
2095
关于我们话语和语言的时序动态,
06:12
just by analyzing the timing dynamics of our speech and language
119
372008
4000
我们的科技就可以倾听,形成理解, 并对我们的心理和生理健康
06:16
picked up by microphones.
120
376032
1443
做出预测。
06:18
Groups have shown that changes in the statistics of our language
121
378038
3880
小组研究表明,我们语言中的变化
06:21
paired with machine learning
122
381942
1420
与机器学习相结合,
06:23
can predict the likelihood someone will develop psychosis.
123
383386
3161
可以预测某人患上精神病的几率。
06:27
I'm going to take it a step further
124
387442
1751
我会更近一步,
06:29
and look at linguistic changes and changes in our voice
125
389217
2587
去观察我们声音中的 语言变化与音调变化,
06:31
that show up with a lot of different conditions.
126
391828
2239
它们展现出了许多不同的状况。
06:34
Dementia, diabetes can alter the spectral coloration of our voice.
127
394091
4367
痴呆,糖尿病会改变 我们声音中的光谱颜色,
06:39
Changes in our language associated with Alzheimer's
128
399205
3119
与阿尔茨海默症关联的言语变化
06:42
can sometimes show up more than 10 years before clinical diagnosis.
129
402348
4365
有时在临床诊断十年前就会出现。
06:47
What we say and how we say it tells a much richer story
130
407236
3960
我们说什么以及说话的方式, 其实比我们曾经以为的
06:51
than we used to think.
131
411220
1254
透露了更多信息。
06:53
And devices we already have in our homes could, if we let them,
132
413022
4047
而且只要我们允许,我们家里的设施
06:57
give us invaluable insight back.
133
417093
2134
也可以给我们珍贵的反馈。
06:59
The chemical composition of our breath
134
419998
2978
我们呼吸中的化学构成
07:03
gives away our feelings.
135
423959
1354
能够暴露我们的情感。
07:06
There's a dynamic mixture of acetone, isoprene and carbon dioxide
136
426363
4478
丙酮,异戊二烯以及 二氧化碳的动态混合物
07:10
that changes when our heart speeds up, when our muscles tense,
137
430865
3384
会随我们心率的提升或肌肉的 紧张状态而改变,
07:14
and all without any obvious change in our behaviors.
138
434809
2897
尽管我们在行为上并没有明显变化。
07:18
Alright, I want you to watch this clip with me.
139
438268
2738
好了,我想让你们 和我一起看看这个片段。
两侧的屏幕上可能会出现一些东西,
07:21
Some things might be going on on the side screens,
140
441030
3119
07:24
but try and focus on the image in the front
141
444173
3777
但尝试关注中间的图像
07:27
and the man at the window.
142
447974
1463
和窗边的人。
07:31
(Eerie music)
143
451633
2658
(阴森的音乐)
07:39
(Woman screams)
144
459767
1437
(女人的尖叫)
07:50
PC: Sorry about that. I needed to get a reaction.
145
470692
2395
波比 · 克拉姆:不好意思, 我需要得到一个反应。
07:53
(Laughter)
146
473111
1785
(笑声)
07:55
I'm actually tracking the carbon dioxide you exhale in the room right now.
147
475412
4972
事实上,我正在检测 你们在房间中呼出的二氧化碳。
08:01
We've installed tubes throughout the theater,
148
481903
3293
我们在场馆内装了很多管道,
08:05
lower to the ground, because CO2 is heavier than air.
149
485220
2595
在贴近地面的位置, 因为二氧化碳比氧气重。
08:07
But they're connected to a device in the back
150
487839
2667
它们连接着后台的一个设备,
08:10
that lets us measure, in real time, with high precision,
151
490530
3287
允许我们实时并高精准度地测量
08:13
the continuous differential concentration of CO2.
152
493841
2922
二氧化碳浓度的持续变化。
08:17
The clouds on the sides are actually the real-time data visualization
153
497246
5508
两边的云朵实际上是二氧化碳浓度的
08:22
of the density of our CO2.
154
502778
1998
可视数据。
08:25
You might still see a patch of red on the screen,
155
505374
3699
你们或许还会看到屏幕上的一片红色,
08:29
because we're showing increases with larger colored clouds,
156
509097
3705
因为我们正在展示大型有色云朵的增加,
08:32
larger colored areas of red.
157
512826
2196
红色的部分变得更多。
08:35
And that's the point where a lot of us jumped.
158
515046
2559
这是我们被吓到的时刻。
08:38
It's our collective suspense driving a change in carbon dioxide.
159
518173
4915
我们对将要发生什么的共同期待 导致了二氧化碳的变化。
08:43
Alright, now, watch this with me one more time.
160
523649
2722
好了,我们再看一遍。
08:46
(Cheerful music)
161
526395
2238
(轻松愉悦的音乐)
08:54
(Woman laughs)
162
534553
2137
(女人的笑声)
09:05
PC: You knew it was coming.
163
545344
1349
波比 · 克拉姆: 你们已经知道会发生什么了。
09:06
But it's a lot different when we changed the creator's intent.
164
546717
3363
当我们改变制作者的意图时, 这就变得很不一样。
09:10
Changing the music and the sound effects
165
550776
2769
改变音乐与声效
09:13
completely alter the emotional impact of that scene.
166
553569
3603
完全改变了那个镜头的情感冲击。
09:17
And we can see it in our breath.
167
557196
2134
这一点可以在我们的呼吸中观察到。
09:20
Suspense, fear, joy
168
560196
2262
悬念,恐惧,快乐,
09:22
all show up as reproducible, visually identifiable moments.
169
562482
4507
全都是可重现的,视觉上可见的。
09:27
We broadcast a chemical signature of our emotions.
170
567473
4151
我们在“广播”我们情感的化学信号。
09:35
It is the end of the poker face.
171
575249
2133
这意味着“扑克脸”的终结。
09:38
Our spaces, our technology will know what we're feeling.
172
578582
3566
我们的空间和科技将能够了解我们的感受。
09:42
We will know more about each other than we ever have.
173
582736
2785
我们会远比以前更了解彼此。
09:45
We get a chance to reach in and connect to the experience and sentiments
174
585911
4307
我们有机会去触碰并连接那些
09:50
that are fundamental to us as humans
175
590242
1742
在我们的感官,情感与社交中
作为人类基础的体验与情绪。
09:52
in our senses, emotionally and socially.
176
592008
2410
09:55
I believe it is the era of the empath.
177
595482
2540
我相信这是移情的时代。
09:58
And we are enabling the capabilities that true technological partners can bring
178
598046
5222
我们正在实现真正的科技伴侣所带来的
10:03
to how we connect with each other and with our technology.
179
603292
3047
如何通过科技让彼此相连的各种能力。
10:06
If we recognize the power of becoming technological empaths,
180
606363
3389
如果我们认识到了 成为科技移情者的能力,
10:09
we get this opportunity
181
609776
1936
就将得到这个机会,
10:11
where technology can help us bridge the emotional and cognitive divide.
182
611736
4424
让科技帮助我们把情感 与认知分歧联系起来。
10:16
And in that way, we get to change how we tell our stories.
183
616680
2723
在这种方式下,我们就可以 改变讲故事的方式。
10:19
We can enable a better future for technologies like augmented reality
184
619427
3580
我们可以创造一个拥有 像现实增强技术的未来,
10:23
to extend our own agency and connect us at a much deeper level.
185
623031
4193
去延伸我们的媒介,并让我们 在更深的层面进行连接。
10:27
Imagine a high school counselor being able to realize
186
627625
2547
想象一位高中学生顾问,能够了解
10:30
that an outwardly cheery student really was having a deeply hard time,
187
630196
3826
一个表面很快乐的学生实际上处境艰难,
10:34
where reaching out can make a crucial, positive difference.
188
634046
3180
这样一来,交流就可以 促成一个显著,积极的改变。
10:37
Or authorities, being able to know the difference
189
637766
3230
我们的权威机构,能够区别
10:41
between someone having a mental health crisis
190
641020
2325
一个人是存在心理健康危机,
10:43
and a different type of aggression,
191
643369
1826
还是怀有一种不同的侵略行为,
10:45
and responding accordingly.
192
645219
1800
并对此做出相应的回应。
10:47
Or an artist, knowing the direct impact of their work.
193
647609
3273
或者让艺术家可以了解 他们艺术作品的直接影响。
10:52
Leo Tolstoy defined his perspective of art
194
652173
2643
列夫 · 托尔斯泰将他对艺术的理解定义为
10:54
by whether what the creator intended
195
654840
1785
创造者的意图
10:56
was experienced by the person on the other end.
196
656649
2586
是否被观看者感受到。
10:59
Today's artists can know what we're feeling.
197
659259
2566
如今的艺术家可以了解我们的感受。
11:02
But regardless of whether it's art or human connection,
198
662204
3005
但无论是艺术还是人类连接,
11:06
today's technologies will know and can know
199
666608
2802
如今的科技都将,且能够知道
11:09
what we're experiencing on the other side,
200
669434
2048
我们在另一端的感受。
11:11
and this means we can be closer and more authentic.
201
671506
2618
这意味着我们彼此可以更接近和更真实。
11:14
But I realize a lot of us have a really hard time
202
674498
4293
不过,我发现我们很多人不喜欢
11:18
with the idea of sharing our data,
203
678815
2267
分享我们的数据,
11:21
and especially the idea that people know things about us
204
681673
3111
尤其当知道人们会了解那些
11:24
that we didn't actively choose to share.
205
684808
2321
我们没有主动选择分享的数据。
11:28
Anytime we talk to someone,
206
688728
2216
每当我们与别人说话,
11:31
look at someone
207
691946
1555
看着某人,
11:33
or choose not to look,
208
693525
1468
或选择避开目光接触时,
11:35
data is exchanged, given away,
209
695017
2647
数据都会被交换,被分享,
11:38
that people use to learn,
210
698533
2205
被人们用于学习,
11:40
make decisions about their lives and about ours.
211
700762
3267
对他们以及我们的生活做出决定。
11:45
I'm not looking to create a world where our inner lives are ripped open
212
705469
3968
我并不想创造一个人们的 内心生活完全暴露的世界,
11:49
and our personal data and our privacy given away
213
709461
2348
我们的数据和隐私都被暴露给
11:51
to people and entities where we don't want to see it go.
214
711833
2713
给我们不愿分享的人和实体。
11:55
But I am looking to create a world
215
715117
2762
我想创造的,
11:57
where we can care about each other more effectively,
216
717903
3408
是一个我们可以更好地关心彼此,
12:01
we can know more about when someone is feeling something
217
721335
3060
更加了解何时有人产生了某种感觉,
12:04
that we ought to pay attention to.
218
724419
1872
而这种感觉也是我们应该关注的。
12:06
And we can have richer experiences from our technology.
219
726800
3335
我们可以利用科技获得更丰富的经验。
12:10
Any technology can be used for good or bad.
220
730887
2357
任何科技都像一面双刃剑。
12:13
Transparency to engagement and effective regulation
221
733268
2412
参与的透明和有效的监管
12:15
are absolutely critical to building the trust for any of this.
222
735704
3120
对在其中建立信任非常重要。
12:20
But the benefits that "empathetic technology" can bring to our lives
223
740106
4834
但是 "技术移情" 可以为我们带来的好处,
12:24
are worth solving the problems that make us uncomfortable.
224
744964
3891
值得我们去解决那些 让我们感到不舒服的问题。
12:29
And if we don't, there are too many opportunities and feelings
225
749315
4025
而且如果不这么做,我们就将错过
12:33
we're going to be missing out on.
226
753364
1695
太多的机会与感受。
12:35
Thank you.
227
755083
1175
谢谢大家。
12:36
(Applause)
228
756282
2479
(掌声)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隐私政策

eng.lish.video

Developer's Blog