Will Superintelligent AI End the World? | Eliezer Yudkowsky | TED

309,707 views ・ 2023-07-11

TED


请双击下面的英文字幕来播放视频。

翻译人员: Yip Yan Yeung 校对人员: suya f.
00:04
Since 2001, I have been working on what we would now call
0
4334
3503
自 2001 年起,我就一直 在研究我们现在所说的
00:07
the problem of aligning artificial general intelligence:
1
7879
3629
通用人工智能的对齐问题:
00:11
how to shape the preferences and behavior
2
11550
2085
如何塑造强大的人工智能
00:13
of a powerful artificial mind such that it does not kill everyone.
3
13677
4838
所具备的偏好和行为, 以防它杀死所有人?
00:19
I more or less founded the field two decades ago,
4
19224
3128
二十年前,我可以说是 开辟了这个领域,
00:22
when nobody else considered it rewarding enough to work on.
5
22352
2794
所有人都认为这个领域 没什么吸引力,不值得干。
00:25
I tried to get this very important project started early
6
25146
2711
我试图早点开启 这一非常重要的项目,
00:27
so we'd be in less of a drastic rush later.
7
27899
2544
这样我们之后就不需要 手忙脚乱地奋起直追。
00:31
I consider myself to have failed.
8
31069
2044
我觉得我失败了。
00:33
(Laughter)
9
33154
1001
(笑声)
00:34
Nobody understands how modern AI systems do what they do.
10
34197
3337
没有人知道现在的 AI 系统 是怎么做到这一切的。
00:37
They are giant, inscrutable matrices of floating point numbers
11
37576
2961
它们是无法参透的 大型浮点数矩阵,
00:40
that we nudge in the direction of better performance
12
40579
2460
我们在朝着更好的性能 试探着前进,
00:43
until they inexplicably start working.
13
43039
2044
直到它们莫名其妙地运行起来。
00:45
At some point, the companies rushing headlong to scale AI
14
45083
3378
有朝一日,稀里糊涂 快速扩张 AI 的公司
00:48
will cough out something that's smarter than humanity.
15
48461
2670
会带出一些比人类更聪明的东西。
00:51
Nobody knows how to calculate when that will happen.
16
51131
2544
没有人知道如何计算 这一刻何时会到来。
00:53
My wild guess is that it will happen after zero to two more breakthroughs
17
53675
3712
我大胆猜测,会在 Transformer 模型大小
00:57
the size of transformers.
18
57429
1710
再突破零至两次之后到来。
00:59
What happens if we build something smarter than us
19
59139
2335
如果我们做出了比我们更聪明,
01:01
that we understand that poorly?
20
61516
2377
但不甚了解的东西,会怎么样?
01:03
Some people find it obvious that building something smarter than us
21
63935
3170
有人认为做出比我们更聪明
01:07
that we don't understand might go badly.
22
67105
2294
又无法理解的东西 显然会招致恶果。
01:09
Others come in with a very wide range of hopeful thoughts
23
69441
3879
也有人心怀各种希望,
01:13
about how it might possibly go well.
24
73361
2044
期盼它还有可能走上正轨。
01:16
Even if I had 20 minutes for this talk and months to prepare it,
25
76281
3337
即使我这场演讲有 20 分钟, 花了几个月准备,
01:19
I would not be able to refute all the ways people find to imagine
26
79659
3087
我也无法否认人们
01:22
that things might go well.
27
82746
1710
幻想一切都好的各种方式。
01:24
But I will say that there is no standard scientific consensus
28
84456
4546
但我想说,就如何走上正轨
01:29
for how things will go well.
29
89044
1668
还没有达成一个标准的科学共识。
01:30
There is no hope that has been widely persuasive
30
90712
2252
还没有广泛为人所接受、
01:33
and stood up to skeptical examination.
31
93006
1877
经得起推敲的希望。
01:35
There is nothing resembling a real engineering plan for us surviving
32
95592
4421
还没有出现一个我愿意置评的、 接近真正的求生工程计划的东西。
01:40
that I could critique.
33
100013
1710
01:41
This is not a good place in which to find ourselves.
34
101765
2752
这不该是我们的立足之地。
01:44
If I had more time,
35
104517
1210
如果我有多余的时间,
01:45
I'd try to tell you about the predictable reasons
36
105727
2294
我想告诉你 一些可预见的原因,
01:48
why the current paradigm will not work
37
108063
1960
解释为什么以现在的水平
01:50
to build a superintelligence that likes you
38
110065
2460
是做不出一个喜欢你、
01:52
or is friends with you, or that just follows orders.
39
112567
3378
与你交朋友或 乖乖听从指令的超级智能的。
01:56
Why, if you press "thumbs up" when humans think that things went right
40
116613
4379
为什么当人类觉得不错的时候, 你按下了“赞”,
02:01
or "thumbs down" when another AI system thinks that they went wrong,
41
121034
3712
有一个 AI 系统觉得出错了的时候, 你按下了“踩”,
02:04
you do not get a mind that wants nice things
42
124788
3503
你得到的并不会是 一个希冀好东西的大脑,
02:08
in a way that generalizes well outside the training distribution
43
128291
4004
它不可能会超出训练范围,
02:12
to where the AI is smarter than the trainers.
44
132337
2461
达到 AI 比训练者更聪明的境界。
02:15
You can search for "Yudkowsky list of lethalities" for more.
45
135674
5130
更多详情请搜索《尤德考斯基 (Yudkowsky)的致命清单》。
02:20
(Laughter)
46
140804
1793
(笑声)
02:22
But to worry, you do not need to believe me
47
142597
2169
但是别担心,你不用相信
02:24
about exact predictions of exact disasters.
48
144808
2878
我对某场灾难的精确预测。
02:27
You just need to expect that things are not going to work great
49
147727
2962
你只需要做好心理准备, 在第一次认真严肃的尝试时,
02:30
on the first really serious, really critical try
50
150730
3045
结果不会如你所愿,
02:33
because an AI system smart enough to be truly dangerous
51
153817
3503
因为聪明到相当危险的 AI 系统,
02:37
was meaningfully different from AI systems stupider than that.
52
157362
3503
与比它更笨的系统 是有天壤之别的。
02:40
My prediction is that this ends up with us facing down something smarter than us
53
160907
5005
我的预测是,我们最终 会直面比我们更聪明的东西,
02:45
that does not want what we want,
54
165954
1752
它不会想要我们想要的东西,
02:47
that does not want anything we recognize as valuable or meaningful.
55
167706
4170
不会想要我们眼中 有价值、有意义的东西。
02:52
I cannot predict exactly how a conflict between humanity and a smarter AI would go
56
172252
4588
我无法预测人类与更智慧的 AI 之间的冲突会如何发展,
02:56
for the same reason I can't predict exactly how you would lose a chess game
57
176881
3838
同样,我也无法预测 你到底会如何在一盘国际象棋中
03:00
to one of the current top AI chess programs, let's say Stockfish.
58
180760
3671
输给当今最顶尖的 AI 国际象棋程序,比如 Stockfish。
03:04
If I could predict exactly where Stockfish could move,
59
184973
3545
如果我可以预测出 Stockfish 的每一步棋,
03:08
I could play chess that well myself.
60
188560
2544
我自己也能下一手好棋。
03:11
I can't predict exactly how you'll lose to Stockfish,
61
191146
2502
我无法预测你会 具体如何输给 Stockfish,
03:13
but I can predict who wins the game.
62
193648
2044
但我可以预测谁会赢。
03:16
I do not expect something actually smart to attack us with marching robot armies
63
196401
4421
我不认为会有聪明到 让眼里放着红光的机器人步兵
03:20
with glowing red eyes
64
200864
1543
袭击我们的东西,
03:22
where there could be a fun movie about us fighting them.
65
202407
3086
再拍一部英勇抗击的有趣电影。
03:25
I expect an actually smarter and uncaring entity
66
205535
2502
我认为会有那么一个 更智慧、更无情的东西,
03:28
will figure out strategies and technologies
67
208079
2002
找到了迅速杀死我们的策略和技术, 然后把我们杀光。
03:30
that can kill us quickly and reliably and then kill us.
68
210123
3545
03:34
I am not saying that the problem of aligning superintelligence
69
214461
2919
我没有说超级智能的对齐问题
03:37
is unsolvable in principle.
70
217422
1710
是根本无法解决的。
03:39
I expect we could figure it out with unlimited time and unlimited retries,
71
219174
4879
我相信如果我们有无限的时间、 无限的重试机会,是可以解决的,
03:44
which the usual process of science assumes that we have.
72
224095
3796
在传统的科学研究过程中 就会有这样的假设。
03:48
The problem here is the part where we don't get to say,
73
228308
2836
我们面临的问题是 我们不可能说:
03:51
“Ha ha, whoops, that sure didn’t work.
74
231186
2294
“哈哈,哎呀,这肯定行不通的。
03:53
That clever idea that used to work on earlier systems
75
233521
3837
这个好点子在以前的系统上行得通,
03:57
sure broke down when the AI got smarter, smarter than us.”
76
237358
3504
如果 AI 越来越聪明, 比我们更聪明,肯定搞不定。”
04:01
We do not get to learn from our mistakes and try again
77
241571
2544
我们没有从错误中 吸取教训、再来一次的机会,
04:04
because everyone is already dead.
78
244157
1960
因为那时我们已经死光了。
04:07
It is a large ask
79
247076
2586
在第一次正经尝试时
04:09
to get an unprecedented scientific and engineering challenge
80
249704
3003
就搞定这史无前例的 科学和工程挑战,
04:12
correct on the first critical try.
81
252707
2002
是个难为人的要求。
04:15
Humanity is not approaching this issue with remotely
82
255084
3045
人们处理此事的态度
04:18
the level of seriousness that would be required.
83
258171
2711
远不及它所需的慎重程度。
04:20
Some of the people leading these efforts
84
260924
1918
有些持这样态度的人
04:22
have spent the last decade not denying
85
262884
2502
在过去的十年里,不是在否认
04:25
that creating a superintelligence might kill everyone,
86
265386
2920
创造出超级智能 可能会杀死所有人,
04:28
but joking about it.
87
268306
1376
而是戏谑。
04:30
We are very far behind.
88
270183
1877
我们落后太多了。
04:32
This is not a gap we can overcome in six months,
89
272101
2294
这不是一个我们可以在 六个月内弥合的鸿沟,
04:34
given a six-month moratorium.
90
274437
1919
如果我们就这么停摆六个月。
04:36
If we actually try to do this in real life,
91
276689
2503
如果我们真的 在现实生活中这么干了,
04:39
we are all going to die.
92
279234
1668
那我们都得死。
04:41
People say to me at this point, what's your ask?
93
281402
2586
这时候人们就想问了: 你想干啥呢?
04:44
I do not have any realistic plan,
94
284697
1627
我没有任何实际的计划,
04:46
which is why I spent the last two decades
95
286324
1960
这也就是为什么我在过去的二十年里
04:48
trying and failing to end up anywhere but here.
96
288326
2586
摸爬滚打,但这里才是我的归宿。
04:51
My best bad take is that we need an international coalition
97
291913
3670
我觉得再不济 我们也得成立一个国际联盟,
04:55
banning large AI training runs,
98
295625
2252
禁止大型 AI 训练,
04:57
including extreme and extraordinary measures
99
297877
3379
包括采取一些极端、特殊的措施,
05:01
to have that ban be actually and universally effective,
100
301256
3253
让这项禁令确确实实 在全球范围内有效,
05:04
like tracking all GPU sales,
101
304509
2377
比如追踪所有 GPU 的销售记录,
05:06
monitoring all the data centers,
102
306928
2127
监控所有的数据中心,
05:09
being willing to risk a shooting conflict between nations
103
309055
2711
愿意冒着国家间 发生武装冲突的风险,
05:11
in order to destroy an unmonitored data center
104
311808
2711
摧毁未签署国 未被监控的数据中心。
05:14
in a non-signatory country.
105
314561
1793
(注:仅代表演讲者原文意见, 不代表译者立场)
05:17
I say this, not expecting that to actually happen.
106
317480
3170
我这么说不是因为 我预见了这真的会发生。
05:21
I say this expecting that we all just die.
107
321109
2919
我这么说是因为 预见了我们都会死。
05:24
But it is not my place to just decide on my own
108
324779
3295
但是自说自话认为 人类会选择去死,
05:28
that humanity will choose to die,
109
328074
2211
05:30
to the point of not bothering to warn anyone.
110
330326
2461
甚至懒得警告任何人, 不是我的风格。
05:33
I have heard that people outside the tech industry
111
333204
2336
我听说有非科技行业的人
05:35
are getting this point faster than people inside it.
112
335582
2627
已经比科技行业的人 更快达到了这个境界。
05:38
Maybe humanity wakes up one morning and decides to live.
113
338251
3712
也许人类某天早上醒来, 还是选择了要活下去。
05:43
Thank you for coming to my brief TED talk.
114
343006
2043
谢谢大家聆听我简短的 TED 演讲。
05:45
(Laughter)
115
345049
1585
(笑声)
05:46
(Applause and cheers)
116
346676
6965
(掌声、欢呼声)
05:56
Chris Anderson: So, Eliezer, thank you for coming and giving that.
117
356102
4421
克里斯·安德森(Chris Anderson): 埃利泽(Eliezer),谢谢你的演讲。
06:00
It seems like what you're raising the alarm about is that like,
118
360523
3962
听起来你是在敲响警钟,
06:04
for this to happen, for an AI to basically destroy humanity,
119
364527
3962
认为这就会发生,AI 会毁灭人类,
06:08
it has to break out, escape controls of the internet and, you know,
120
368489
5172
肯定会爆发,逃过互联网的控制,
06:13
start commanding actual real-world resources.
121
373661
2920
开始掌控真实存在的资源。
06:16
You say you can't predict how that will happen,
122
376581
2210
你说你无法预测它会如何发生,
06:18
but just paint one or two possibilities.
123
378833
3462
但还是请你给出一两个可能性吧。
06:22
Eliezer Yudkowsky: OK, so why is this hard?
124
382337
2794
埃利泽·尤德考斯基: 嗯,那为什么这是件难事呢?
06:25
First, because you can't predict exactly where a smarter chess program will move.
125
385131
3837
首先,因为你无法精确预测 更智慧的象棋程序的下一步棋是什么。
06:28
Maybe even more importantly than that,
126
388968
1877
有可能更重要的一点是,
06:30
imagine sending the design for an air conditioner
127
390887
2794
想象一下把一台空调的设计
06:33
back to the 11th century.
128
393723
1877
发回 11 世纪。
06:35
Even if they -- if it’s enough detail for them to build it,
129
395642
3128
就算他们有足够的细节把它做出来,
06:38
they will be surprised when cold air comes out
130
398811
2753
冷气跑出来的时候, 他们还是会感到惊讶,
06:41
because the air conditioner will use the temperature-pressure relation
131
401564
3670
因为空调会利用温度和压强的关系,
06:45
and they don't know about that law of nature.
132
405276
2461
而他们还不知道这个自然定律。
06:47
So if you want me to sketch what a superintelligence might do,
133
407779
4629
如果你想让我描绘出 超级智能会做些什么,
06:52
I can go deeper and deeper into places
134
412450
2336
我可以不断深入、深入,
06:54
where we think there are predictable technological advancements
135
414827
2962
到那些我们还一无所知的 可预见的技术进步出现的地方。
06:57
that we haven't figured out yet.
136
417789
1626
06:59
And as I go deeper, it will get harder and harder to follow.
137
419415
2836
我越深入,就越难跟上。
07:02
It could be super persuasive.
138
422251
1961
这是非常有道理的。
07:04
That's relatively easy to understand.
139
424253
1877
也比较容易理解。
07:06
We do not understand exactly how the brain works,
140
426130
2545
我们不能彻底搞明白 大脑是如何运作的,
07:08
so it's a great place to exploit laws of nature that we do not know about.
141
428716
3504
所以大脑就是探索我们 不了解的自然定律的好地方。
07:12
Rules of the environment,
142
432261
1502
环境的规则,
07:13
invent new technologies beyond that.
143
433805
2669
发明超出环境的新技术。
07:16
Can you build a synthetic virus that gives humans a cold
144
436808
3879
你能做出一种合成病毒, 让人类感冒,
07:20
and then a bit of neurological change and they're easier to persuade?
145
440728
4213
再动一动神经系统, 让人类更容易被说服吗?
07:24
Can you build your own synthetic biology,
146
444941
3336
你能做出你自己的合成生物,
07:28
synthetic cyborgs?
147
448319
1585
合成改造人吗?
07:29
Can you blow straight past that
148
449946
2044
你可以突破
07:31
to covalently bonded equivalents of biology,
149
451990
4004
生物上的共价键物质,
07:36
where instead of proteins that fold up and are held together by static cling,
150
456035
3754
蛋白质不再折叠、 由静电吸附聚合,
07:39
you've got things that go down much sharper potential energy gradients
151
459789
3378
做出一个势能梯度更陡,
07:43
and are bonded together?
152
463209
1376
还互相联结在一起的东西吗?
07:44
People have done advanced design work about this sort of thing
153
464585
3546
人们已经为这样的物质 做了一些先进的设计,
07:48
for artificial red blood cells that could hold 100 times as much oxygen
154
468172
3879
可以多携带一百倍氧气 的人工红细胞,
07:52
if they were using tiny sapphire vessels to store the oxygen.
155
472093
3754
如果红细胞可以用上 扩张导管储存氧气。
07:55
There's lots and lots of room above biology,
156
475888
2586
生物学之上有很多空间,
07:58
but it gets harder and harder to understand.
157
478474
2211
但越来越难理解。
08:01
CA: So what I hear you saying
158
481519
1543
CA: 你说到的是
08:03
is that these terrifying possibilities there
159
483104
2294
这些可怕的可能性,
08:05
but your real guess is that AIs will work out something more devious than that.
160
485440
5130
但你真正的猜测是 AI 会走上一条更曲折的道路。
08:10
Is that really a likely pathway in your mind?
161
490570
3837
你脑海中有没有一条 比较有可能的道路?
08:14
EY: Which part?
162
494407
1168
EY: 你说的是哪一部分?
08:15
That they're smarter than I am? Absolutely.
163
495616
2002
它们会比我更聪明吗?当然了。
08:17
CA: Not that they're smarter,
164
497660
1418
CA: 不是它们更聪明,
08:19
but why would they want to go in that direction?
165
499078
3045
而是它们为什么会想走上这条道路?
08:22
Like, AIs don't have our feelings of sort of envy and jealousy and anger
166
502123
5422
AI 没有我们这种嫉妒、 眼红、愤怒等感受。
08:27
and so forth.
167
507587
1168
08:28
So why might they go in that direction?
168
508755
2460
那它们为什么要这么发展下去呢?
08:31
EY: Because it's convergently implied by almost any of the strange,
169
511215
4296
EY: 因为它们想要的这些奇怪、
08:35
inscrutable things that they might end up wanting
170
515511
3379
难以捉摸的东西 通向了一个共同的结果,
08:38
as a result of gradient descent
171
518931
1710
来自 AI 内部的“赞”和“踩” 得出的梯度下降。
08:40
on these "thumbs up" and "thumbs down" things internally.
172
520683
2711
08:44
If all you want is to make tiny little molecular squiggles
173
524604
3920
如果你想要的只是 随意修改一下分子结构,
08:48
or that's like, one component of what you want,
174
528566
2461
或者你想要一个成分,
08:51
but it's a component that never saturates, you just want more and more of it,
175
531069
3628
一个不会饱和的成分, 那你就会越要越多,
08:54
the same way that we would want more and more galaxies filled with life
176
534697
3337
就像我们想要 越来越多充满生命的星系,
08:58
and people living happily ever after.
177
538034
1793
人们从此过上幸福的生活。
08:59
Anything that just keeps going,
178
539869
1627
如果有一件不断运转的东西,
09:01
you just want to use more and more material for that,
179
541537
3045
你就会想为它使用越来越多的资源,
09:04
that could kill everyone on Earth as a side effect.
180
544624
2669
也会产生把地球上的 所有人杀光的副作用。
09:07
It could kill us because it doesn't want us making other superintelligences
181
547293
3545
它把我们杀光是因为 它不想让我们做出与其竞争的
09:10
to compete with it.
182
550880
1168
其他超级智能。
09:12
It could kill us because it's using up all the chemical energy on earth
183
552048
4379
它把我们杀光是因为 它要用完地球上的所有化学能量了,
09:16
and we contain some chemical potential energy.
184
556469
2586
而我们体内蕴含着一些 潜在的化学能量。
09:19
CA: So some people in the AI world worry that your views are strong enough
185
559055
6256
CA: AI 世界里的一些人 担心你的观点太强势了,
09:25
and they would say extreme enough
186
565311
1585
他们说太极端了,
09:26
that you're willing to advocate extreme responses to it.
187
566896
3170
你都愿意支持极端的回应。
09:30
And therefore, they worry that you could be, you know,
188
570066
3462
所以他们担心你会是
09:33
in one sense, a very destructive figure.
189
573528
1918
某种意义上极具破坏性的人物。
09:35
Do you draw the line yourself in terms of the measures
190
575488
3170
你自己会划下我们 为了阻止这一切的发生
09:38
that we should take to stop this happening?
191
578699
2753
采取的措施应有的界限吗?
09:41
Or is actually anything justifiable to stop
192
581494
3253
或者说要阻止你刚谈到 将要发生的情景,
09:44
the scenarios you're talking about happening?
193
584747
2211
有没有一些正当的措施?
09:47
EY: I don't think that "anything" works.
194
587834
3253
EY: 我不认为这些“措施”会有用。
09:51
I think that this takes state actors
195
591087
4379
我认为这涉及了国家政府
09:55
and international agreements
196
595508
2753
和国际协定,
09:58
and all international agreements by their nature,
197
598302
3254
而所有的国际协定从本质上
10:01
tend to ultimately be backed by force
198
601556
1876
说到底都会受到签署国 和非签署国势力的支持,
10:03
on the signatory countries and on the non-signatory countries,
199
603474
3379
10:06
which is a more extreme measure.
200
606894
1794
这就会是一个更极端的措施了。
10:09
I have not proposed that individuals run out and use violence,
201
609856
2961
我没有提议个人 挺身而出,实施暴力,
10:12
and I think that the killer argument for that is that it would not work.
202
612859
4337
就这一点,我的关键论点是 这是没有用的。
10:18
CA: Well, you are definitely not the only person to propose
203
618489
2961
CA: 你肯定不是唯一一个提议
10:21
that what we need is some kind of international reckoning here
204
621450
4004
我们得达成一些国际上的共识,
10:25
on how to manage this going forward.
205
625496
2044
如何处理它未来的走向。
10:27
Thank you so much for coming here to TED, Eliezer.
206
627540
2419
感谢你来到 TED,埃利泽。
10:30
(Applause)
207
630001
2085
(掌声)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隐私政策

eng.lish.video

Developer's Blog