请双击下面的英文字幕来播放视频。
翻译人员: Yip Yan Yeung
校对人员: suya f.
00:04
Since 2001, I have been working
on what we would now call
0
4334
3503
自 2001 年起,我就一直
在研究我们现在所说的
00:07
the problem of aligning artificial
general intelligence:
1
7879
3629
通用人工智能的对齐问题:
00:11
how to shape the preferences and behavior
2
11550
2085
如何塑造强大的人工智能
00:13
of a powerful artificial mind
such that it does not kill everyone.
3
13677
4838
所具备的偏好和行为,
以防它杀死所有人?
00:19
I more or less founded the field
two decades ago,
4
19224
3128
二十年前,我可以说是
开辟了这个领域,
00:22
when nobody else considered it
rewarding enough to work on.
5
22352
2794
所有人都认为这个领域
没什么吸引力,不值得干。
00:25
I tried to get this very important
project started early
6
25146
2711
我试图早点开启
这一非常重要的项目,
00:27
so we'd be in less
of a drastic rush later.
7
27899
2544
这样我们之后就不需要
手忙脚乱地奋起直追。
00:31
I consider myself to have failed.
8
31069
2044
我觉得我失败了。
00:33
(Laughter)
9
33154
1001
(笑声)
00:34
Nobody understands how modern
AI systems do what they do.
10
34197
3337
没有人知道现在的 AI 系统
是怎么做到这一切的。
00:37
They are giant, inscrutable matrices
of floating point numbers
11
37576
2961
它们是无法参透的
大型浮点数矩阵,
00:40
that we nudge in the direction
of better performance
12
40579
2460
我们在朝着更好的性能
试探着前进,
00:43
until they inexplicably start working.
13
43039
2044
直到它们莫名其妙地运行起来。
00:45
At some point, the companies
rushing headlong to scale AI
14
45083
3378
有朝一日,稀里糊涂
快速扩张 AI 的公司
00:48
will cough out something
that's smarter than humanity.
15
48461
2670
会带出一些比人类更聪明的东西。
00:51
Nobody knows how to calculate
when that will happen.
16
51131
2544
没有人知道如何计算
这一刻何时会到来。
00:53
My wild guess is that it will happen
after zero to two more breakthroughs
17
53675
3712
我大胆猜测,会在
Transformer 模型大小
00:57
the size of transformers.
18
57429
1710
再突破零至两次之后到来。
00:59
What happens if we build
something smarter than us
19
59139
2335
如果我们做出了比我们更聪明,
01:01
that we understand that poorly?
20
61516
2377
但不甚了解的东西,会怎么样?
01:03
Some people find it obvious
that building something smarter than us
21
63935
3170
有人认为做出比我们更聪明
01:07
that we don't understand might go badly.
22
67105
2294
又无法理解的东西
显然会招致恶果。
01:09
Others come in with a very wide range
of hopeful thoughts
23
69441
3879
也有人心怀各种希望,
01:13
about how it might possibly go well.
24
73361
2044
期盼它还有可能走上正轨。
01:16
Even if I had 20 minutes for this talk
and months to prepare it,
25
76281
3337
即使我这场演讲有 20 分钟,
花了几个月准备,
01:19
I would not be able to refute
all the ways people find to imagine
26
79659
3087
我也无法否认人们
01:22
that things might go well.
27
82746
1710
幻想一切都好的各种方式。
01:24
But I will say that there is no
standard scientific consensus
28
84456
4546
但我想说,就如何走上正轨
01:29
for how things will go well.
29
89044
1668
还没有达成一个标准的科学共识。
01:30
There is no hope
that has been widely persuasive
30
90712
2252
还没有广泛为人所接受、
01:33
and stood up to skeptical examination.
31
93006
1877
经得起推敲的希望。
01:35
There is nothing resembling a real
engineering plan for us surviving
32
95592
4421
还没有出现一个我愿意置评的、
接近真正的求生工程计划的东西。
01:40
that I could critique.
33
100013
1710
01:41
This is not a good place
in which to find ourselves.
34
101765
2752
这不该是我们的立足之地。
01:44
If I had more time,
35
104517
1210
如果我有多余的时间,
01:45
I'd try to tell you
about the predictable reasons
36
105727
2294
我想告诉你
一些可预见的原因,
01:48
why the current paradigm will not work
37
108063
1960
解释为什么以现在的水平
01:50
to build a superintelligence
that likes you
38
110065
2460
是做不出一个喜欢你、
01:52
or is friends with you,
or that just follows orders.
39
112567
3378
与你交朋友或
乖乖听从指令的超级智能的。
01:56
Why, if you press "thumbs up"
when humans think that things went right
40
116613
4379
为什么当人类觉得不错的时候,
你按下了“赞”,
02:01
or "thumbs down" when another AI system
thinks that they went wrong,
41
121034
3712
有一个 AI 系统觉得出错了的时候,
你按下了“踩”,
02:04
you do not get a mind
that wants nice things
42
124788
3503
你得到的并不会是
一个希冀好东西的大脑,
02:08
in a way that generalizes well
outside the training distribution
43
128291
4004
它不可能会超出训练范围,
02:12
to where the AI is smarter
than the trainers.
44
132337
2461
达到 AI 比训练者更聪明的境界。
02:15
You can search for "Yudkowsky
list of lethalities" for more.
45
135674
5130
更多详情请搜索《尤德考斯基
(Yudkowsky)的致命清单》。
02:20
(Laughter)
46
140804
1793
(笑声)
02:22
But to worry, you do not
need to believe me
47
142597
2169
但是别担心,你不用相信
02:24
about exact predictions
of exact disasters.
48
144808
2878
我对某场灾难的精确预测。
02:27
You just need to expect that things
are not going to work great
49
147727
2962
你只需要做好心理准备,
在第一次认真严肃的尝试时,
02:30
on the first really serious,
really critical try
50
150730
3045
结果不会如你所愿,
02:33
because an AI system
smart enough to be truly dangerous
51
153817
3503
因为聪明到相当危险的 AI 系统,
02:37
was meaningfully different
from AI systems stupider than that.
52
157362
3503
与比它更笨的系统
是有天壤之别的。
02:40
My prediction is that this ends up with us
facing down something smarter than us
53
160907
5005
我的预测是,我们最终
会直面比我们更聪明的东西,
02:45
that does not want what we want,
54
165954
1752
它不会想要我们想要的东西,
02:47
that does not want anything we recognize
as valuable or meaningful.
55
167706
4170
不会想要我们眼中
有价值、有意义的东西。
02:52
I cannot predict exactly how a conflict
between humanity and a smarter AI would go
56
172252
4588
我无法预测人类与更智慧的
AI 之间的冲突会如何发展,
02:56
for the same reason I can't predict
exactly how you would lose a chess game
57
176881
3838
同样,我也无法预测
你到底会如何在一盘国际象棋中
03:00
to one of the current top
AI chess programs, let's say Stockfish.
58
180760
3671
输给当今最顶尖的 AI
国际象棋程序,比如 Stockfish。
03:04
If I could predict exactly
where Stockfish could move,
59
184973
3545
如果我可以预测出
Stockfish 的每一步棋,
03:08
I could play chess that well myself.
60
188560
2544
我自己也能下一手好棋。
03:11
I can't predict exactly
how you'll lose to Stockfish,
61
191146
2502
我无法预测你会
具体如何输给 Stockfish,
03:13
but I can predict who wins the game.
62
193648
2044
但我可以预测谁会赢。
03:16
I do not expect something actually smart
to attack us with marching robot armies
63
196401
4421
我不认为会有聪明到
让眼里放着红光的机器人步兵
03:20
with glowing red eyes
64
200864
1543
袭击我们的东西,
03:22
where there could be a fun movie
about us fighting them.
65
202407
3086
再拍一部英勇抗击的有趣电影。
03:25
I expect an actually smarter
and uncaring entity
66
205535
2502
我认为会有那么一个
更智慧、更无情的东西,
03:28
will figure out strategies
and technologies
67
208079
2002
找到了迅速杀死我们的策略和技术,
然后把我们杀光。
03:30
that can kill us quickly
and reliably and then kill us.
68
210123
3545
03:34
I am not saying that the problem
of aligning superintelligence
69
214461
2919
我没有说超级智能的对齐问题
03:37
is unsolvable in principle.
70
217422
1710
是根本无法解决的。
03:39
I expect we could figure it out
with unlimited time and unlimited retries,
71
219174
4879
我相信如果我们有无限的时间、
无限的重试机会,是可以解决的,
03:44
which the usual process of science
assumes that we have.
72
224095
3796
在传统的科学研究过程中
就会有这样的假设。
03:48
The problem here is the part
where we don't get to say,
73
228308
2836
我们面临的问题是
我们不可能说:
03:51
“Ha ha, whoops, that sure didn’t work.
74
231186
2294
“哈哈,哎呀,这肯定行不通的。
03:53
That clever idea that used to work
on earlier systems
75
233521
3837
这个好点子在以前的系统上行得通,
03:57
sure broke down when the AI
got smarter, smarter than us.”
76
237358
3504
如果 AI 越来越聪明,
比我们更聪明,肯定搞不定。”
04:01
We do not get to learn
from our mistakes and try again
77
241571
2544
我们没有从错误中
吸取教训、再来一次的机会,
04:04
because everyone is already dead.
78
244157
1960
因为那时我们已经死光了。
04:07
It is a large ask
79
247076
2586
在第一次正经尝试时
04:09
to get an unprecedented scientific
and engineering challenge
80
249704
3003
就搞定这史无前例的
科学和工程挑战,
04:12
correct on the first critical try.
81
252707
2002
是个难为人的要求。
04:15
Humanity is not approaching
this issue with remotely
82
255084
3045
人们处理此事的态度
04:18
the level of seriousness
that would be required.
83
258171
2711
远不及它所需的慎重程度。
04:20
Some of the people leading these efforts
84
260924
1918
有些持这样态度的人
04:22
have spent the last decade not denying
85
262884
2502
在过去的十年里,不是在否认
04:25
that creating a superintelligence
might kill everyone,
86
265386
2920
创造出超级智能
可能会杀死所有人,
04:28
but joking about it.
87
268306
1376
而是戏谑。
04:30
We are very far behind.
88
270183
1877
我们落后太多了。
04:32
This is not a gap
we can overcome in six months,
89
272101
2294
这不是一个我们可以在
六个月内弥合的鸿沟,
04:34
given a six-month moratorium.
90
274437
1919
如果我们就这么停摆六个月。
04:36
If we actually try
to do this in real life,
91
276689
2503
如果我们真的
在现实生活中这么干了,
04:39
we are all going to die.
92
279234
1668
那我们都得死。
04:41
People say to me at this point,
what's your ask?
93
281402
2586
这时候人们就想问了:
你想干啥呢?
04:44
I do not have any realistic plan,
94
284697
1627
我没有任何实际的计划,
04:46
which is why I spent the last two decades
95
286324
1960
这也就是为什么我在过去的二十年里
04:48
trying and failing to end up
anywhere but here.
96
288326
2586
摸爬滚打,但这里才是我的归宿。
04:51
My best bad take is that we need
an international coalition
97
291913
3670
我觉得再不济
我们也得成立一个国际联盟,
04:55
banning large AI training runs,
98
295625
2252
禁止大型 AI 训练,
04:57
including extreme
and extraordinary measures
99
297877
3379
包括采取一些极端、特殊的措施,
05:01
to have that ban be actually
and universally effective,
100
301256
3253
让这项禁令确确实实
在全球范围内有效,
05:04
like tracking all GPU sales,
101
304509
2377
比如追踪所有 GPU 的销售记录,
05:06
monitoring all the data centers,
102
306928
2127
监控所有的数据中心,
05:09
being willing to risk
a shooting conflict between nations
103
309055
2711
愿意冒着国家间
发生武装冲突的风险,
05:11
in order to destroy
an unmonitored data center
104
311808
2711
摧毁未签署国
未被监控的数据中心。
05:14
in a non-signatory country.
105
314561
1793
(注:仅代表演讲者原文意见,
不代表译者立场)
05:17
I say this, not expecting
that to actually happen.
106
317480
3170
我这么说不是因为
我预见了这真的会发生。
05:21
I say this expecting that we all just die.
107
321109
2919
我这么说是因为
预见了我们都会死。
05:24
But it is not my place
to just decide on my own
108
324779
3295
但是自说自话认为
人类会选择去死,
05:28
that humanity will choose to die,
109
328074
2211
05:30
to the point of not bothering
to warn anyone.
110
330326
2461
甚至懒得警告任何人,
不是我的风格。
05:33
I have heard that people
outside the tech industry
111
333204
2336
我听说有非科技行业的人
05:35
are getting this point
faster than people inside it.
112
335582
2627
已经比科技行业的人
更快达到了这个境界。
05:38
Maybe humanity wakes up one morning
and decides to live.
113
338251
3712
也许人类某天早上醒来,
还是选择了要活下去。
05:43
Thank you for coming to my brief TED talk.
114
343006
2043
谢谢大家聆听我简短的 TED 演讲。
05:45
(Laughter)
115
345049
1585
(笑声)
05:46
(Applause and cheers)
116
346676
6965
(掌声、欢呼声)
05:56
Chris Anderson: So, Eliezer, thank you
for coming and giving that.
117
356102
4421
克里斯·安德森(Chris Anderson):
埃利泽(Eliezer),谢谢你的演讲。
06:00
It seems like what you're raising
the alarm about is that like,
118
360523
3962
听起来你是在敲响警钟,
06:04
for this to happen, for an AI
to basically destroy humanity,
119
364527
3962
认为这就会发生,AI 会毁灭人类,
06:08
it has to break out, escape controls
of the internet and, you know,
120
368489
5172
肯定会爆发,逃过互联网的控制,
06:13
start commanding actual
real-world resources.
121
373661
2920
开始掌控真实存在的资源。
06:16
You say you can't predict
how that will happen,
122
376581
2210
你说你无法预测它会如何发生,
06:18
but just paint one or two possibilities.
123
378833
3462
但还是请你给出一两个可能性吧。
06:22
Eliezer Yudkowsky:
OK, so why is this hard?
124
382337
2794
埃利泽·尤德考斯基:
嗯,那为什么这是件难事呢?
06:25
First, because you can't predict exactly
where a smarter chess program will move.
125
385131
3837
首先,因为你无法精确预测
更智慧的象棋程序的下一步棋是什么。
06:28
Maybe even more importantly than that,
126
388968
1877
有可能更重要的一点是,
06:30
imagine sending the design
for an air conditioner
127
390887
2794
想象一下把一台空调的设计
06:33
back to the 11th century.
128
393723
1877
发回 11 世纪。
06:35
Even if they -- if it’s enough
detail for them to build it,
129
395642
3128
就算他们有足够的细节把它做出来,
06:38
they will be surprised
when cold air comes out
130
398811
2753
冷气跑出来的时候,
他们还是会感到惊讶,
06:41
because the air conditioner will use
the temperature-pressure relation
131
401564
3670
因为空调会利用温度和压强的关系,
06:45
and they don't know
about that law of nature.
132
405276
2461
而他们还不知道这个自然定律。
06:47
So if you want me to sketch
what a superintelligence might do,
133
407779
4629
如果你想让我描绘出
超级智能会做些什么,
06:52
I can go deeper and deeper into places
134
412450
2336
我可以不断深入、深入,
06:54
where we think there are predictable
technological advancements
135
414827
2962
到那些我们还一无所知的
可预见的技术进步出现的地方。
06:57
that we haven't figured out yet.
136
417789
1626
06:59
And as I go deeper, it will get
harder and harder to follow.
137
419415
2836
我越深入,就越难跟上。
07:02
It could be super persuasive.
138
422251
1961
这是非常有道理的。
07:04
That's relatively easy to understand.
139
424253
1877
也比较容易理解。
07:06
We do not understand exactly
how the brain works,
140
426130
2545
我们不能彻底搞明白
大脑是如何运作的,
07:08
so it's a great place to exploit
laws of nature that we do not know about.
141
428716
3504
所以大脑就是探索我们
不了解的自然定律的好地方。
07:12
Rules of the environment,
142
432261
1502
环境的规则,
07:13
invent new technologies beyond that.
143
433805
2669
发明超出环境的新技术。
07:16
Can you build a synthetic virus
that gives humans a cold
144
436808
3879
你能做出一种合成病毒,
让人类感冒,
07:20
and then a bit of neurological change
and they're easier to persuade?
145
440728
4213
再动一动神经系统,
让人类更容易被说服吗?
07:24
Can you build your own synthetic biology,
146
444941
3336
你能做出你自己的合成生物,
07:28
synthetic cyborgs?
147
448319
1585
合成改造人吗?
07:29
Can you blow straight past that
148
449946
2044
你可以突破
07:31
to covalently bonded
equivalents of biology,
149
451990
4004
生物上的共价键物质,
07:36
where instead of proteins that fold up
and are held together by static cling,
150
456035
3754
蛋白质不再折叠、
由静电吸附聚合,
07:39
you've got things that go down
much sharper potential energy gradients
151
459789
3378
做出一个势能梯度更陡,
07:43
and are bonded together?
152
463209
1376
还互相联结在一起的东西吗?
07:44
People have done advanced design work
about this sort of thing
153
464585
3546
人们已经为这样的物质
做了一些先进的设计,
07:48
for artificial red blood cells
that could hold 100 times as much oxygen
154
468172
3879
可以多携带一百倍氧气
的人工红细胞,
07:52
if they were using tiny
sapphire vessels to store the oxygen.
155
472093
3754
如果红细胞可以用上
扩张导管储存氧气。
07:55
There's lots and lots
of room above biology,
156
475888
2586
生物学之上有很多空间,
07:58
but it gets harder and harder
to understand.
157
478474
2211
但越来越难理解。
08:01
CA: So what I hear you saying
158
481519
1543
CA: 你说到的是
08:03
is that these terrifying
possibilities there
159
483104
2294
这些可怕的可能性,
08:05
but your real guess is that AIs will work
out something more devious than that.
160
485440
5130
但你真正的猜测是
AI 会走上一条更曲折的道路。
08:10
Is that really a likely
pathway in your mind?
161
490570
3837
你脑海中有没有一条
比较有可能的道路?
08:14
EY: Which part?
162
494407
1168
EY: 你说的是哪一部分?
08:15
That they're smarter
than I am? Absolutely.
163
495616
2002
它们会比我更聪明吗?当然了。
08:17
CA: Not that they're smarter,
164
497660
1418
CA: 不是它们更聪明,
08:19
but why would they want
to go in that direction?
165
499078
3045
而是它们为什么会想走上这条道路?
08:22
Like, AIs don't have our feelings of
sort of envy and jealousy and anger
166
502123
5422
AI 没有我们这种嫉妒、
眼红、愤怒等感受。
08:27
and so forth.
167
507587
1168
08:28
So why might they go in that direction?
168
508755
2460
那它们为什么要这么发展下去呢?
08:31
EY: Because it's convergently implied
by almost any of the strange,
169
511215
4296
EY: 因为它们想要的这些奇怪、
08:35
inscrutable things that they
might end up wanting
170
515511
3379
难以捉摸的东西
通向了一个共同的结果,
08:38
as a result of gradient descent
171
518931
1710
来自 AI 内部的“赞”和“踩”
得出的梯度下降。
08:40
on these "thumbs up"
and "thumbs down" things internally.
172
520683
2711
08:44
If all you want is to make tiny
little molecular squiggles
173
524604
3920
如果你想要的只是
随意修改一下分子结构,
08:48
or that's like, one component
of what you want,
174
528566
2461
或者你想要一个成分,
08:51
but it's a component that never saturates,
you just want more and more of it,
175
531069
3628
一个不会饱和的成分,
那你就会越要越多,
08:54
the same way that we would want
more and more galaxies filled with life
176
534697
3337
就像我们想要
越来越多充满生命的星系,
08:58
and people living happily ever after.
177
538034
1793
人们从此过上幸福的生活。
08:59
Anything that just keeps going,
178
539869
1627
如果有一件不断运转的东西,
09:01
you just want to use more
and more material for that,
179
541537
3045
你就会想为它使用越来越多的资源,
09:04
that could kill everyone
on Earth as a side effect.
180
544624
2669
也会产生把地球上的
所有人杀光的副作用。
09:07
It could kill us because it doesn't want
us making other superintelligences
181
547293
3545
它把我们杀光是因为
它不想让我们做出与其竞争的
09:10
to compete with it.
182
550880
1168
其他超级智能。
09:12
It could kill us because it's using up
all the chemical energy on earth
183
552048
4379
它把我们杀光是因为
它要用完地球上的所有化学能量了,
09:16
and we contain some
chemical potential energy.
184
556469
2586
而我们体内蕴含着一些
潜在的化学能量。
09:19
CA: So some people in the AI world worry
that your views are strong enough
185
559055
6256
CA: AI 世界里的一些人
担心你的观点太强势了,
09:25
and they would say extreme enough
186
565311
1585
他们说太极端了,
09:26
that you're willing to advocate
extreme responses to it.
187
566896
3170
你都愿意支持极端的回应。
09:30
And therefore, they worry
that you could be, you know,
188
570066
3462
所以他们担心你会是
09:33
in one sense, a very destructive figure.
189
573528
1918
某种意义上极具破坏性的人物。
09:35
Do you draw the line yourself
in terms of the measures
190
575488
3170
你自己会划下我们
为了阻止这一切的发生
09:38
that we should take
to stop this happening?
191
578699
2753
采取的措施应有的界限吗?
09:41
Or is actually anything
justifiable to stop
192
581494
3253
或者说要阻止你刚谈到
将要发生的情景,
09:44
the scenarios you're talking
about happening?
193
584747
2211
有没有一些正当的措施?
09:47
EY: I don't think that "anything" works.
194
587834
3253
EY: 我不认为这些“措施”会有用。
09:51
I think that this takes state actors
195
591087
4379
我认为这涉及了国家政府
09:55
and international agreements
196
595508
2753
和国际协定,
09:58
and all international agreements
by their nature,
197
598302
3254
而所有的国际协定从本质上
10:01
tend to ultimately be backed by force
198
601556
1876
说到底都会受到签署国
和非签署国势力的支持,
10:03
on the signatory countries
and on the non-signatory countries,
199
603474
3379
10:06
which is a more extreme measure.
200
606894
1794
这就会是一个更极端的措施了。
10:09
I have not proposed that individuals
run out and use violence,
201
609856
2961
我没有提议个人
挺身而出,实施暴力,
10:12
and I think that the killer argument
for that is that it would not work.
202
612859
4337
就这一点,我的关键论点是
这是没有用的。
10:18
CA: Well, you are definitely not
the only person to propose
203
618489
2961
CA: 你肯定不是唯一一个提议
10:21
that what we need is some kind
of international reckoning here
204
621450
4004
我们得达成一些国际上的共识,
10:25
on how to manage this going forward.
205
625496
2044
如何处理它未来的走向。
10:27
Thank you so much
for coming here to TED, Eliezer.
206
627540
2419
感谢你来到 TED,埃利泽。
10:30
(Applause)
207
630001
2085
(掌声)
New videos
Original video on YouTube.com
关于本网站
这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。