Will Superintelligent AI End the World? | Eliezer Yudkowsky | TED

284,020 views ・ 2023-07-11

TED


请双击下面的英文字幕来播放视频。

翻译人员: Yip Yan Yeung 校对人员: suya f.
00:04
Since 2001, I have been working on what we would now call
0
4334
3503
自 2001 年起,我就一直 在研究我们现在所说的
00:07
the problem of aligning artificial general intelligence:
1
7879
3629
通用人工智能的对齐问题:
00:11
how to shape the preferences and behavior
2
11550
2085
如何塑造强大的人工智能
00:13
of a powerful artificial mind such that it does not kill everyone.
3
13677
4838
所具备的偏好和行为, 以防它杀死所有人?
00:19
I more or less founded the field two decades ago,
4
19224
3128
二十年前,我可以说是 开辟了这个领域,
00:22
when nobody else considered it rewarding enough to work on.
5
22352
2794
所有人都认为这个领域 没什么吸引力,不值得干。
00:25
I tried to get this very important project started early
6
25146
2711
我试图早点开启 这一非常重要的项目,
00:27
so we'd be in less of a drastic rush later.
7
27899
2544
这样我们之后就不需要 手忙脚乱地奋起直追。
00:31
I consider myself to have failed.
8
31069
2044
我觉得我失败了。
00:33
(Laughter)
9
33154
1001
(笑声)
00:34
Nobody understands how modern AI systems do what they do.
10
34197
3337
没有人知道现在的 AI 系统 是怎么做到这一切的。
00:37
They are giant, inscrutable matrices of floating point numbers
11
37576
2961
它们是无法参透的 大型浮点数矩阵,
00:40
that we nudge in the direction of better performance
12
40579
2460
我们在朝着更好的性能 试探着前进,
00:43
until they inexplicably start working.
13
43039
2044
直到它们莫名其妙地运行起来。
00:45
At some point, the companies rushing headlong to scale AI
14
45083
3378
有朝一日,稀里糊涂 快速扩张 AI 的公司
00:48
will cough out something that's smarter than humanity.
15
48461
2670
会带出一些比人类更聪明的东西。
00:51
Nobody knows how to calculate when that will happen.
16
51131
2544
没有人知道如何计算 这一刻何时会到来。
00:53
My wild guess is that it will happen after zero to two more breakthroughs
17
53675
3712
我大胆猜测,会在 Transformer 模型大小
00:57
the size of transformers.
18
57429
1710
再突破零至两次之后到来。
00:59
What happens if we build something smarter than us
19
59139
2335
如果我们做出了比我们更聪明,
01:01
that we understand that poorly?
20
61516
2377
但不甚了解的东西,会怎么样?
01:03
Some people find it obvious that building something smarter than us
21
63935
3170
有人认为做出比我们更聪明
01:07
that we don't understand might go badly.
22
67105
2294
又无法理解的东西 显然会招致恶果。
01:09
Others come in with a very wide range of hopeful thoughts
23
69441
3879
也有人心怀各种希望,
01:13
about how it might possibly go well.
24
73361
2044
期盼它还有可能走上正轨。
01:16
Even if I had 20 minutes for this talk and months to prepare it,
25
76281
3337
即使我这场演讲有 20 分钟, 花了几个月准备,
01:19
I would not be able to refute all the ways people find to imagine
26
79659
3087
我也无法否认人们
01:22
that things might go well.
27
82746
1710
幻想一切都好的各种方式。
01:24
But I will say that there is no standard scientific consensus
28
84456
4546
但我想说,就如何走上正轨
01:29
for how things will go well.
29
89044
1668
还没有达成一个标准的科学共识。
01:30
There is no hope that has been widely persuasive
30
90712
2252
还没有广泛为人所接受、
01:33
and stood up to skeptical examination.
31
93006
1877
经得起推敲的希望。
01:35
There is nothing resembling a real engineering plan for us surviving
32
95592
4421
还没有出现一个我愿意置评的、 接近真正的求生工程计划的东西。
01:40
that I could critique.
33
100013
1710
01:41
This is not a good place in which to find ourselves.
34
101765
2752
这不该是我们的立足之地。
01:44
If I had more time,
35
104517
1210
如果我有多余的时间,
01:45
I'd try to tell you about the predictable reasons
36
105727
2294
我想告诉你 一些可预见的原因,
01:48
why the current paradigm will not work
37
108063
1960
解释为什么以现在的水平
01:50
to build a superintelligence that likes you
38
110065
2460
是做不出一个喜欢你、
01:52
or is friends with you, or that just follows orders.
39
112567
3378
与你交朋友或 乖乖听从指令的超级智能的。
01:56
Why, if you press "thumbs up" when humans think that things went right
40
116613
4379
为什么当人类觉得不错的时候, 你按下了“赞”,
02:01
or "thumbs down" when another AI system thinks that they went wrong,
41
121034
3712
有一个 AI 系统觉得出错了的时候, 你按下了“踩”,
02:04
you do not get a mind that wants nice things
42
124788
3503
你得到的并不会是 一个希冀好东西的大脑,
02:08
in a way that generalizes well outside the training distribution
43
128291
4004
它不可能会超出训练范围,
02:12
to where the AI is smarter than the trainers.
44
132337
2461
达到 AI 比训练者更聪明的境界。
02:15
You can search for "Yudkowsky list of lethalities" for more.
45
135674
5130
更多详情请搜索《尤德考斯基 (Yudkowsky)的致命清单》。
02:20
(Laughter)
46
140804
1793
(笑声)
02:22
But to worry, you do not need to believe me
47
142597
2169
但是别担心,你不用相信
02:24
about exact predictions of exact disasters.
48
144808
2878
我对某场灾难的精确预测。
02:27
You just need to expect that things are not going to work great
49
147727
2962
你只需要做好心理准备, 在第一次认真严肃的尝试时,
02:30
on the first really serious, really critical try
50
150730
3045
结果不会如你所愿,
02:33
because an AI system smart enough to be truly dangerous
51
153817
3503
因为聪明到相当危险的 AI 系统,
02:37
was meaningfully different from AI systems stupider than that.
52
157362
3503
与比它更笨的系统 是有天壤之别的。
02:40
My prediction is that this ends up with us facing down something smarter than us
53
160907
5005
我的预测是,我们最终 会直面比我们更聪明的东西,
02:45
that does not want what we want,
54
165954
1752
它不会想要我们想要的东西,
02:47
that does not want anything we recognize as valuable or meaningful.
55
167706
4170
不会想要我们眼中 有价值、有意义的东西。
02:52
I cannot predict exactly how a conflict between humanity and a smarter AI would go
56
172252
4588
我无法预测人类与更智慧的 AI 之间的冲突会如何发展,
02:56
for the same reason I can't predict exactly how you would lose a chess game
57
176881
3838
同样,我也无法预测 你到底会如何在一盘国际象棋中
03:00
to one of the current top AI chess programs, let's say Stockfish.
58
180760
3671
输给当今最顶尖的 AI 国际象棋程序,比如 Stockfish。
03:04
If I could predict exactly where Stockfish could move,
59
184973
3545
如果我可以预测出 Stockfish 的每一步棋,
03:08
I could play chess that well myself.
60
188560
2544
我自己也能下一手好棋。
03:11
I can't predict exactly how you'll lose to Stockfish,
61
191146
2502
我无法预测你会 具体如何输给 Stockfish,
03:13
but I can predict who wins the game.
62
193648
2044
但我可以预测谁会赢。
03:16
I do not expect something actually smart to attack us with marching robot armies
63
196401
4421
我不认为会有聪明到 让眼里放着红光的机器人步兵
03:20
with glowing red eyes
64
200864
1543
袭击我们的东西,
03:22
where there could be a fun movie about us fighting them.
65
202407
3086
再拍一部英勇抗击的有趣电影。
03:25
I expect an actually smarter and uncaring entity
66
205535
2502
我认为会有那么一个 更智慧、更无情的东西,
03:28
will figure out strategies and technologies
67
208079
2002
找到了迅速杀死我们的策略和技术, 然后把我们杀光。
03:30
that can kill us quickly and reliably and then kill us.
68
210123
3545
03:34
I am not saying that the problem of aligning superintelligence
69
214461
2919
我没有说超级智能的对齐问题
03:37
is unsolvable in principle.
70
217422
1710
是根本无法解决的。
03:39
I expect we could figure it out with unlimited time and unlimited retries,
71
219174
4879
我相信如果我们有无限的时间、 无限的重试机会,是可以解决的,
03:44
which the usual process of science assumes that we have.
72
224095
3796
在传统的科学研究过程中 就会有这样的假设。
03:48
The problem here is the part where we don't get to say,
73
228308
2836
我们面临的问题是 我们不可能说:
03:51
“Ha ha, whoops, that sure didn’t work.
74
231186
2294
“哈哈,哎呀,这肯定行不通的。
03:53
That clever idea that used to work on earlier systems
75
233521
3837
这个好点子在以前的系统上行得通,
03:57
sure broke down when the AI got smarter, smarter than us.”
76
237358
3504
如果 AI 越来越聪明, 比我们更聪明,肯定搞不定。”
04:01
We do not get to learn from our mistakes and try again
77
241571
2544
我们没有从错误中 吸取教训、再来一次的机会,
04:04
because everyone is already dead.
78
244157
1960
因为那时我们已经死光了。
04:07
It is a large ask
79
247076
2586
在第一次正经尝试时
04:09
to get an unprecedented scientific and engineering challenge
80
249704
3003
就搞定这史无前例的 科学和工程挑战,
04:12
correct on the first critical try.
81
252707
2002
是个难为人的要求。
04:15
Humanity is not approaching this issue with remotely
82
255084
3045
人们处理此事的态度
04:18
the level of seriousness that would be required.
83
258171
2711
远不及它所需的慎重程度。
04:20
Some of the people leading these efforts
84
260924
1918
有些持这样态度的人
04:22
have spent the last decade not denying
85
262884
2502
在过去的十年里,不是在否认
04:25
that creating a superintelligence might kill everyone,
86
265386
2920
创造出超级智能 可能会杀死所有人,
04:28
but joking about it.
87
268306
1376
而是戏谑。
04:30
We are very far behind.
88
270183
1877
我们落后太多了。
04:32
This is not a gap we can overcome in six months,
89
272101
2294
这不是一个我们可以在 六个月内弥合的鸿沟,
04:34
given a six-month moratorium.
90
274437
1919
如果我们就这么停摆六个月。
04:36
If we actually try to do this in real life,
91
276689
2503
如果我们真的 在现实生活中这么干了,
04:39
we are all going to die.
92
279234
1668
那我们都得死。
04:41
People say to me at this point, what's your ask?
93
281402
2586
这时候人们就想问了: 你想干啥呢?
04:44
I do not have any realistic plan,
94
284697
1627
我没有任何实际的计划,
04:46
which is why I spent the last two decades
95
286324
1960
这也就是为什么我在过去的二十年里
04:48
trying and failing to end up anywhere but here.
96
288326
2586
摸爬滚打,但这里才是我的归宿。
04:51
My best bad take is that we need an international coalition
97
291913
3670
我觉得再不济 我们也得成立一个国际联盟,
04:55
banning large AI training runs,
98
295625
2252
禁止大型 AI 训练,
04:57
including extreme and extraordinary measures
99
297877
3379
包括采取一些极端、特殊的措施,
05:01
to have that ban be actually and universally effective,
100
301256
3253
让这项禁令确确实实 在全球范围内有效,
05:04
like tracking all GPU sales,
101
304509
2377
比如追踪所有 GPU 的销售记录,
05:06
monitoring all the data centers,
102
306928
2127
监控所有的数据中心,
05:09
being willing to risk a shooting conflict between nations
103
309055
2711
愿意冒着国家间 发生武装冲突的风险,
05:11
in order to destroy an unmonitored data center
104
311808
2711
摧毁未签署国 未被监控的数据中心。
05:14
in a non-signatory country.
105
314561
1793
(注:仅代表演讲者原文意见, 不代表译者立场)
05:17
I say this, not expecting that to actually happen.
106
317480
3170
我这么说不是因为 我预见了这真的会发生。
05:21
I say this expecting that we all just die.
107
321109
2919
我这么说是因为 预见了我们都会死。
05:24
But it is not my place to just decide on my own
108
324779
3295
但是自说自话认为 人类会选择去死,
05:28
that humanity will choose to die,
109
328074
2211
05:30
to the point of not bothering to warn anyone.
110
330326
2461
甚至懒得警告任何人, 不是我的风格。
05:33
I have heard that people outside the tech industry
111
333204
2336
我听说有非科技行业的人
05:35
are getting this point faster than people inside it.
112
335582
2627
已经比科技行业的人 更快达到了这个境界。
05:38
Maybe humanity wakes up one morning and decides to live.
113
338251
3712
也许人类某天早上醒来, 还是选择了要活下去。
05:43
Thank you for coming to my brief TED talk.
114
343006
2043
谢谢大家聆听我简短的 TED 演讲。
05:45
(Laughter)
115
345049
1585
(笑声)
05:46
(Applause and cheers)
116
346676
6965
(掌声、欢呼声)
05:56
Chris Anderson: So, Eliezer, thank you for coming and giving that.
117
356102
4421
克里斯·安德森(Chris Anderson): 埃利泽(Eliezer),谢谢你的演讲。
06:00
It seems like what you're raising the alarm about is that like,
118
360523
3962
听起来你是在敲响警钟,
06:04
for this to happen, for an AI to basically destroy humanity,
119
364527
3962
认为这就会发生,AI 会毁灭人类,
06:08
it has to break out, escape controls of the internet and, you know,
120
368489
5172
肯定会爆发,逃过互联网的控制,
06:13
start commanding actual real-world resources.
121
373661
2920
开始掌控真实存在的资源。
06:16
You say you can't predict how that will happen,
122
376581
2210
你说你无法预测它会如何发生,
06:18
but just paint one or two possibilities.
123
378833
3462
但还是请你给出一两个可能性吧。
06:22
Eliezer Yudkowsky: OK, so why is this hard?
124
382337
2794
埃利泽·尤德考斯基: 嗯,那为什么这是件难事呢?
06:25
First, because you can't predict exactly where a smarter chess program will move.
125
385131
3837
首先,因为你无法精确预测 更智慧的象棋程序的下一步棋是什么。
06:28
Maybe even more importantly than that,
126
388968
1877
有可能更重要的一点是,
06:30
imagine sending the design for an air conditioner
127
390887
2794
想象一下把一台空调的设计
06:33
back to the 11th century.
128
393723
1877
发回 11 世纪。
06:35
Even if they -- if it’s enough detail for them to build it,
129
395642
3128
就算他们有足够的细节把它做出来,
06:38
they will be surprised when cold air comes out
130
398811
2753
冷气跑出来的时候, 他们还是会感到惊讶,
06:41
because the air conditioner will use the temperature-pressure relation
131
401564
3670
因为空调会利用温度和压强的关系,
06:45
and they don't know about that law of nature.
132
405276
2461
而他们还不知道这个自然定律。
06:47
So if you want me to sketch what a superintelligence might do,
133
407779
4629
如果你想让我描绘出 超级智能会做些什么,
06:52
I can go deeper and deeper into places
134
412450
2336
我可以不断深入、深入,
06:54
where we think there are predictable technological advancements
135
414827
2962
到那些我们还一无所知的 可预见的技术进步出现的地方。
06:57
that we haven't figured out yet.
136
417789
1626
06:59
And as I go deeper, it will get harder and harder to follow.
137
419415
2836
我越深入,就越难跟上。
07:02
It could be super persuasive.
138
422251
1961
这是非常有道理的。
07:04
That's relatively easy to understand.
139
424253
1877
也比较容易理解。
07:06
We do not understand exactly how the brain works,
140
426130
2545
我们不能彻底搞明白 大脑是如何运作的,
07:08
so it's a great place to exploit laws of nature that we do not know about.
141
428716
3504
所以大脑就是探索我们 不了解的自然定律的好地方。
07:12
Rules of the environment,
142
432261
1502
环境的规则,
07:13
invent new technologies beyond that.
143
433805
2669
发明超出环境的新技术。
07:16
Can you build a synthetic virus that gives humans a cold
144
436808
3879
你能做出一种合成病毒, 让人类感冒,
07:20
and then a bit of neurological change and they're easier to persuade?
145
440728
4213
再动一动神经系统, 让人类更容易被说服吗?
07:24
Can you build your own synthetic biology,
146
444941
3336
你能做出你自己的合成生物,
07:28
synthetic cyborgs?
147
448319
1585
合成改造人吗?
07:29
Can you blow straight past that
148
449946
2044
你可以突破
07:31
to covalently bonded equivalents of biology,
149
451990
4004
生物上的共价键物质,
07:36
where instead of proteins that fold up and are held together by static cling,
150
456035
3754
蛋白质不再折叠、 由静电吸附聚合,
07:39
you've got things that go down much sharper potential energy gradients
151
459789
3378
做出一个势能梯度更陡,
07:43
and are bonded together?
152
463209
1376
还互相联结在一起的东西吗?
07:44
People have done advanced design work about this sort of thing
153
464585
3546
人们已经为这样的物质 做了一些先进的设计,
07:48
for artificial red blood cells that could hold 100 times as much oxygen
154
468172
3879
可以多携带一百倍氧气 的人工红细胞,
07:52
if they were using tiny sapphire vessels to store the oxygen.
155
472093
3754
如果红细胞可以用上 扩张导管储存氧气。
07:55
There's lots and lots of room above biology,
156
475888
2586
生物学之上有很多空间,
07:58
but it gets harder and harder to understand.
157
478474
2211
但越来越难理解。
08:01
CA: So what I hear you saying
158
481519
1543
CA: 你说到的是
08:03
is that these terrifying possibilities there
159
483104
2294
这些可怕的可能性,
08:05
but your real guess is that AIs will work out something more devious than that.
160
485440
5130
但你真正的猜测是 AI 会走上一条更曲折的道路。
08:10
Is that really a likely pathway in your mind?
161
490570
3837
你脑海中有没有一条 比较有可能的道路?
08:14
EY: Which part?
162
494407
1168
EY: 你说的是哪一部分?
08:15
That they're smarter than I am? Absolutely.
163
495616
2002
它们会比我更聪明吗?当然了。
08:17
CA: Not that they're smarter,
164
497660
1418
CA: 不是它们更聪明,
08:19
but why would they want to go in that direction?
165
499078
3045
而是它们为什么会想走上这条道路?
08:22
Like, AIs don't have our feelings of sort of envy and jealousy and anger
166
502123
5422
AI 没有我们这种嫉妒、 眼红、愤怒等感受。
08:27
and so forth.
167
507587
1168
08:28
So why might they go in that direction?
168
508755
2460
那它们为什么要这么发展下去呢?
08:31
EY: Because it's convergently implied by almost any of the strange,
169
511215
4296
EY: 因为它们想要的这些奇怪、
08:35
inscrutable things that they might end up wanting
170
515511
3379
难以捉摸的东西 通向了一个共同的结果,
08:38
as a result of gradient descent
171
518931
1710
来自 AI 内部的“赞”和“踩” 得出的梯度下降。
08:40
on these "thumbs up" and "thumbs down" things internally.
172
520683
2711
08:44
If all you want is to make tiny little molecular squiggles
173
524604
3920
如果你想要的只是 随意修改一下分子结构,
08:48
or that's like, one component of what you want,
174
528566
2461
或者你想要一个成分,
08:51
but it's a component that never saturates, you just want more and more of it,
175
531069
3628
一个不会饱和的成分, 那你就会越要越多,
08:54
the same way that we would want more and more galaxies filled with life
176
534697
3337
就像我们想要 越来越多充满生命的星系,
08:58
and people living happily ever after.
177
538034
1793
人们从此过上幸福的生活。
08:59
Anything that just keeps going,
178
539869
1627
如果有一件不断运转的东西,
09:01
you just want to use more and more material for that,
179
541537
3045
你就会想为它使用越来越多的资源,
09:04
that could kill everyone on Earth as a side effect.
180
544624
2669
也会产生把地球上的 所有人杀光的副作用。
09:07
It could kill us because it doesn't want us making other superintelligences
181
547293
3545
它把我们杀光是因为 它不想让我们做出与其竞争的
09:10
to compete with it.
182
550880
1168
其他超级智能。
09:12
It could kill us because it's using up all the chemical energy on earth
183
552048
4379
它把我们杀光是因为 它要用完地球上的所有化学能量了,
09:16
and we contain some chemical potential energy.
184
556469
2586
而我们体内蕴含着一些 潜在的化学能量。
09:19
CA: So some people in the AI world worry that your views are strong enough
185
559055
6256
CA: AI 世界里的一些人 担心你的观点太强势了,
09:25
and they would say extreme enough
186
565311
1585
他们说太极端了,
09:26
that you're willing to advocate extreme responses to it.
187
566896
3170
你都愿意支持极端的回应。
09:30
And therefore, they worry that you could be, you know,
188
570066
3462
所以他们担心你会是
09:33
in one sense, a very destructive figure.
189
573528
1918
某种意义上极具破坏性的人物。
09:35
Do you draw the line yourself in terms of the measures
190
575488
3170
你自己会划下我们 为了阻止这一切的发生
09:38
that we should take to stop this happening?
191
578699
2753
采取的措施应有的界限吗?
09:41
Or is actually anything justifiable to stop
192
581494
3253
或者说要阻止你刚谈到 将要发生的情景,
09:44
the scenarios you're talking about happening?
193
584747
2211
有没有一些正当的措施?
09:47
EY: I don't think that "anything" works.
194
587834
3253
EY: 我不认为这些“措施”会有用。
09:51
I think that this takes state actors
195
591087
4379
我认为这涉及了国家政府
09:55
and international agreements
196
595508
2753
和国际协定,
09:58
and all international agreements by their nature,
197
598302
3254
而所有的国际协定从本质上
10:01
tend to ultimately be backed by force
198
601556
1876
说到底都会受到签署国 和非签署国势力的支持,
10:03
on the signatory countries and on the non-signatory countries,
199
603474
3379
10:06
which is a more extreme measure.
200
606894
1794
这就会是一个更极端的措施了。
10:09
I have not proposed that individuals run out and use violence,
201
609856
2961
我没有提议个人 挺身而出,实施暴力,
10:12
and I think that the killer argument for that is that it would not work.
202
612859
4337
就这一点,我的关键论点是 这是没有用的。
10:18
CA: Well, you are definitely not the only person to propose
203
618489
2961
CA: 你肯定不是唯一一个提议
10:21
that what we need is some kind of international reckoning here
204
621450
4004
我们得达成一些国际上的共识,
10:25
on how to manage this going forward.
205
625496
2044
如何处理它未来的走向。
10:27
Thank you so much for coming here to TED, Eliezer.
206
627540
2419
感谢你来到 TED,埃利泽。
10:30
(Applause)
207
630001
2085
(掌声)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7