The Transformative Potential of AGI — and When It Might Arrive | Shane Legg and Chris Anderson | TED

208,779 views

2023-12-07 ・ TED


New videos

The Transformative Potential of AGI — and When It Might Arrive | Shane Legg and Chris Anderson | TED

208,779 views ・ 2023-12-07

TED


Please double-click on the English subtitles below to play the video.

00:04
Chris Anderson: Shane, give us a snapshot of you growing up
0
4042
2878
00:06
and what on Earth led you to get interested in artificial intelligence?
1
6920
3462
00:10
Shane Legg: Well, I got my first home computer on my 10th birthday,
2
10423
4630
00:15
and I --
3
15095
1710
00:16
this was before the internet and everything.
4
16846
2128
00:18
So you couldn't just go and surf the web and so on.
5
18974
2544
00:21
You had to actually make stuff yourself and program.
6
21560
2544
00:24
And so I started programming,
7
24145
1585
00:25
and I discovered that in this computer there was a world,
8
25730
3462
00:29
I could create a world,
9
29234
1418
00:30
I could create little agents that would run around
10
30652
2753
00:33
and chase each other and do things and so on.
11
33446
2128
00:35
And I could sort of, bring this whole universe to life.
12
35574
2585
00:38
And there was sort of that spark of creativity that really captivated me
13
38201
4338
00:42
and sort of, I think that was really the seeds
14
42539
2919
00:45
of my interest that later grew
15
45500
1460
00:46
into an interest in artificial intelligence.
16
46960
2377
00:49
CA: Because in your standard education, you had some challenges there.
17
49379
3795
00:53
SL: Yeah, I was dyslexic as a child.
18
53174
3587
00:56
And so they were actually going to hold me back a year
19
56803
3796
01:00
when I was 10 years old,
20
60599
1418
01:02
and they sent me off to get my IQ tested to sort of,
21
62058
2586
01:04
you know, assess how bad the problem was.
22
64686
2210
01:07
And they discovered I had an exceptionally high IQ.
23
67230
3337
01:10
And then they were a little bit confused about what was going on.
24
70609
3128
01:13
And fortunately, at that time,
25
73737
1626
01:15
there was somebody in the town I lived in
26
75405
2336
01:17
who knew how to test for dyslexia.
27
77782
1752
01:19
And it turns out I wasn't actually of limited intelligence.
28
79576
3378
01:22
I was dyslexic, and that was the issue.
29
82996
2419
01:25
CA: You had reason from an early age to believe
30
85457
2210
01:27
that our standard assumptions about intelligence might be off a bit.
31
87709
3712
01:31
SL: Well, I had reason, from an early age, to sometimes doubt authority.
32
91421
4004
01:35
(Laughter)
33
95425
1001
01:36
You know, if the teacher thinks you're dumb, maybe it's not true.
34
96468
3086
01:39
Maybe there are other things going on.
35
99554
2169
01:41
But I think it also created in me an interest in intelligence
36
101765
4629
01:46
when I sort of had that experience as a child.
37
106436
2794
01:49
CA: So you're credited by many
38
109272
1460
01:50
as coining the term “artificial general intelligence,” AGI.
39
110774
3587
01:54
Tell us about 2001, how that happened.
40
114402
2711
01:57
SL: Yeah, so I was approached by someone called Ben Goertzel
41
117113
3671
02:00
who I'd actually been working with,
42
120784
1710
02:02
and he was going to write a book,
43
122494
2127
02:04
and he was thinking about a book on AI systems
44
124663
3503
02:08
that would be much more general and capable,
45
128166
2085
02:10
rather than focusing on very narrow things.
46
130251
2461
02:13
And he was thinking about a title for the book.
47
133046
2210
02:15
So I suggested to him, "If you're interested in very general systems,
48
135256
3295
02:18
call it artificial general intelligence."
49
138551
1961
02:20
And so he went with that.
50
140553
1210
02:21
And then him and various other people started using the term online
51
141763
3170
02:24
and the internet,
52
144933
1209
02:26
and then it sort of became popularized from there.
53
146142
2336
02:28
We later discovered there was someone called Mike Garrod,
54
148520
2711
02:31
who published a paper in a security nanotech journal in '97.
55
151231
4296
02:35
So he is actually the first person to have used the term.
56
155568
2711
02:38
But it turns out he pretty much meant the same thing as us anyway.
57
158279
3295
02:41
CA: It was kind of an idea whose time had come,
58
161616
2419
02:44
to recognize the potential here.
59
164035
2211
02:46
I mean, you made an early prediction that many people thought was bonkers.
60
166246
4212
02:50
What was that?
61
170458
1168
02:51
SL: Well, in about 2001,
62
171668
2711
02:54
a similar time to when I suggested this term artificial general intelligence,
63
174421
4838
02:59
I read a book by Ray Kurzweil, actually, "Age of Spiritual Machines,"
64
179259
3670
03:02
and I concluded that he was fundamentally right,
65
182971
3795
03:06
that computation was likely to grow exponentially for at least a few decades,
66
186808
5714
03:12
and the amount of data in the world would grow exponentially
67
192564
2836
03:15
for a few decades.
68
195442
1167
03:16
And so I figured that if that was going to happen,
69
196609
2670
03:19
then the value of extremely scalable algorithms
70
199279
3545
03:22
that could harness all this data and computation
71
202824
3295
03:26
were going to be very high.
72
206161
1835
03:27
And then I also figured that in the mid 2020s,
73
207996
3462
03:31
it would be possible then,
74
211499
1460
03:33
if we had these highly scalable algorithms,
75
213001
2711
03:35
to train artificial intelligence systems
76
215712
4254
03:40
on far more data than a human would experience in a lifetime.
77
220008
3420
03:43
And so as a result of that,
78
223470
1418
03:44
you can find it on my blog from about 2009
79
224929
3838
03:48
I think it's the first time I publicly talked about it,
80
228808
2586
03:51
I predicted a 50 percent chance of AGI by 2028.
81
231436
3920
03:55
I still believe that today.
82
235815
2336
03:58
CA: That's still your date.
83
238151
1668
04:00
How did you define AGI back then, and has your definition changed?
84
240320
4045
04:04
SL: Yeah, I didn't have a particularly precise definition at the beginning.
85
244824
4171
04:09
It was really just an idea of systems that would just be far more general.
86
249037
4588
04:13
So rather than just playing Go or chess or something,
87
253666
2503
04:16
rather than actually be able to do many, many different things.
88
256211
2961
04:19
The definition I use now is that it's a system
89
259214
2544
04:21
that can do all the cognitive kinds of tasks
90
261758
2627
04:24
that people can do, possibly more,
91
264385
2545
04:26
but at least it can do the sorts of cognitive tasks
92
266930
2419
04:29
that people can typically do.
93
269349
1877
04:31
CA: So talk about just the founding of DeepMind
94
271267
3003
04:34
and the interplay between you and your cofounders.
95
274270
4046
04:38
SL: Right. So I went to London to the place called the Gatsby Unit,
96
278316
4004
04:42
which studies theoretical neuroscience and machine learning.
97
282362
5046
04:47
And I was interested in learning the relationships
98
287450
2378
04:49
between what we understand about the brain
99
289828
2043
04:51
and what we know from machine learning.
100
291871
1877
04:53
So that seemed like a really good place.
101
293748
1960
04:55
And I met Demis Hassabis there.
102
295708
1544
04:57
He had the same postdoc supervisor as me,
103
297252
2002
04:59
and we got talking.
104
299295
1252
05:00
And he convinced me
105
300588
2628
05:03
that it was the time to start a company then.
106
303258
2252
05:05
That was in 2009 we started talking.
107
305510
2669
05:08
And I was a little bit skeptical.
108
308221
1668
05:09
I thought AGI was still a bit too far away,
109
309889
3545
05:13
but he thought the time was right, so we decided to go for it.
110
313434
3421
05:16
And then a friend of his was Mustafa Suleyman.
111
316896
3170
05:20
CA: And specifically, one of the goals of the company
112
320733
2503
05:23
was to find a pathway to AGI?
113
323236
1835
05:25
SL: Absolutely.
114
325113
1334
05:26
On our first business plan that we were circulating
115
326447
3879
05:30
when we were looking for investors in 2010,
116
330368
3045
05:33
it had one sentence on the front cover and it said,
117
333413
2586
05:36
"Build the world's first artificial general intelligence."
118
336040
2711
05:38
So that was right in from the beginning.
119
338793
2669
05:42
CA: Even though you knew
120
342046
1502
05:43
that building that AGI might actually have
121
343590
4337
05:47
apocalyptic consequences in some scenarios?
122
347969
2628
05:50
SL: Yeah.
123
350930
1210
05:52
So it's a deeply transformative technology.
124
352181
4672
05:57
I believe it will happen.
125
357437
1585
05:59
I think that, you know,
126
359856
1251
06:01
these algorithms can be understood and they will be understood at the time.
127
361149
3754
06:04
And I think that intelligence is fundamentally
128
364944
2962
06:07
an incredibly valuable thing.
129
367906
1876
06:10
Everything around us at the moment -- the building we’re in,
130
370241
2836
06:13
the words I’m using, the concepts we have, the technology around us --
131
373077
3671
06:17
you know, all of these things are being affected by intelligence.
132
377540
3087
06:20
So having intelligence in machines
133
380668
2878
06:23
is an incredibly valuable thing to develop.
134
383588
3879
06:27
And so I believe it is coming.
135
387467
2002
06:29
Now when a very, very powerful technology arises,
136
389510
4130
06:33
there can be a range of different outcomes.
137
393640
2627
06:36
Things could go very, very well,
138
396309
1918
06:38
but there is a possibility things can go badly as well.
139
398227
2586
06:40
And that was something I was aware of also from about 20 years ago.
140
400855
4963
06:46
CA: So talk about, as DeepMind developed,
141
406319
3253
06:49
was there a moment where you really felt,
142
409572
4296
06:53
"My goodness, we're onto something unbelievably powerful?"
143
413910
3628
06:57
Like, was it AlphaGo, that whole story, or what was the moment for you?
144
417580
4296
07:01
SL: Yeah, there were many moments over the years.
145
421918
2502
07:04
One was when we did the Atari games.
146
424420
2211
07:07
Have you seen those videos
147
427131
1252
07:08
where we had an algorithm that could learn to play multiple games
148
428424
3420
07:11
without being programmed for any specific game?
149
431844
2253
07:14
There were some exciting moments there.
150
434138
3295
07:17
Go, of course, was a really exciting moment.
151
437433
2586
07:20
But I think the thing that's really captured my imagination,
152
440937
3754
07:24
a lot of people's imagination,
153
444732
1460
07:26
is the phenomenal scaling of language models in recent years.
154
446192
3253
07:29
I think we can see they're systems
155
449821
1960
07:31
that really can start to do some meaningful fraction
156
451781
3670
07:35
of the cognitive tasks that people can do.
157
455493
2377
07:37
CA: Now, you were working on those models,
158
457912
2002
07:39
but were you, to some extent, blindsided
159
459914
1960
07:41
by OpenAI's, sort of, sudden unveiling of ChatGPT?
160
461874
5923
07:47
SL: Right.
161
467839
1168
07:49
We were working on them and you know,
162
469007
1835
07:50
the transformer model was invented in Google,
163
470842
2127
07:53
and we had teams who were building big transformer language models and so on.
164
473011
5296
07:58
CA: Google acquired DeepMind at some point in this journey.
165
478683
3170
08:01
SL: Yeah, exactly.
166
481894
1335
08:03
And so what I didn't expect
167
483229
4046
08:07
was just how good a model could get training purely on text.
168
487275
4713
08:11
I thought you would need more multimodality.
169
491988
2711
08:14
You'd need images, you'd need sound, you'd need video and things like that.
170
494741
3712
08:18
But due to the absolutely vast quantities of text,
171
498494
3671
08:22
it can sort of compensate for these things to an extent.
172
502206
2753
08:25
I still think you see aspects of this.
173
505376
2628
08:28
I think language models tend to be weak in areas
174
508004
3170
08:31
that are not easily expressed in text.
175
511174
3211
08:34
But I don’t think this is a fundamental limitation.
176
514385
2419
08:36
I think we're going to see these language models expanding into video
177
516804
4964
08:41
and images and sound and all these things,
178
521809
2211
08:44
and these things will be overcome in time.
179
524020
2044
08:46
CA: So talk to us, Shane,
180
526564
1335
08:47
about the things that you, at this moment,
181
527940
2962
08:50
passionately feel that the world needs to be thinking about more cogently.
182
530902
4754
08:55
SL: Right.
183
535698
1168
08:56
So I think that very, very powerful,
184
536866
2628
08:59
very intelligent artificial intelligence is coming.
185
539535
3504
09:03
I think that this is very, very likely.
186
543498
2669
09:06
I don't think it's coming today.
187
546209
1585
09:07
I don't think it's coming next year or the year after.
188
547794
2544
09:10
It's probably a little bit further out than that.
189
550338
2419
09:12
CA: 2028?
190
552799
1293
09:14
SL: 2028, that's a 50 percent chance.
191
554092
2293
09:16
So, you know, if it doesn't happen in 2028,
192
556385
2044
09:18
I'm not going to be surprised, obviously.
193
558429
2002
09:20
CA: And when you say powerful,
194
560473
1460
09:21
I mean there's already powerful AI out there.
195
561933
2127
09:24
But you're saying basically a version
196
564060
1793
09:25
of artificial general intelligence is coming.
197
565853
2127
09:28
SL: Yeah.
198
568022
1126
09:29
CA: So give us a picture of what that could look like.
199
569190
2711
09:31
SL: Well, if you had an artificial general intelligence,
200
571901
2669
09:34
you could do all sorts of amazing things.
201
574570
2419
09:37
Just like human intelligence is able to do many, many amazing things.
202
577031
3253
09:40
So it's not really about a specific thing,
203
580326
2002
09:42
that's the whole point of the generality.
204
582370
1960
09:44
But to give you one example,
205
584330
1752
09:46
we developed the system AlphaFold,
206
586082
2335
09:48
which will take a protein and compute, basically, the shape of that protein.
207
588459
5881
09:54
And that enables you to do all sorts of research
208
594382
2627
09:57
into understanding biological processes,
209
597009
1919
09:58
developing medicines and all kinds of things like that.
210
598970
2586
10:01
Now, if you had an AGI system,
211
601556
1459
10:03
instead of requiring what we had at DeepMind,
212
603057
2586
10:05
about roughly 30 world-class scientists
213
605685
2919
10:08
working for about three years to develop that,
214
608604
2670
10:11
maybe you could develop that with just a team
215
611315
2127
10:13
of a handful of scientists in one year.
216
613442
2211
10:16
So imagine these, sort of, AlphaFold-level developments
217
616154
3753
10:19
taking place around the world on a regular basis.
218
619949
3212
10:23
This is the sort of thing that AGI could enable.
219
623202
2503
10:25
CA: So within months of AGI being with us, so to speak,
220
625997
4045
10:30
it's quite possible that some of the scientific challenges
221
630084
3087
10:33
that humans have wrestled with for decades, centuries, if you like,
222
633171
3962
10:37
will start to tumble in rapid succession.
223
637133
2961
10:40
SL: Yeah, I think it'll open up all sorts of amazing possibilities.
224
640136
4129
10:44
And it could be really a golden age of humanity
225
644265
4212
10:48
where human intelligence,
226
648519
1835
10:50
which is aided and extended with machine intelligence,
227
650396
4379
10:54
enables us to do all sorts of fantastic things
228
654775
2461
10:57
and solve problems that previously were just intractable.
229
657278
4922
11:02
CA: So let's come back to that.
230
662200
1501
11:03
But I think you also,
231
663743
1585
11:05
you're not like, an irredeemable optimist only,
232
665328
2252
11:07
you see a potential for it to go very badly in a different direction.
233
667622
3920
11:11
Talk about what that pathway could look like.
234
671542
2711
11:14
SL: Well, yeah, I want to explain.
235
674253
1961
11:17
I don't believe the people
236
677048
1585
11:18
who are sure that it's going to go very well,
237
678674
2253
11:20
and I don't believe the people
238
680968
1460
11:22
who are sure that it’s going to go very, very badly.
239
682428
2461
11:24
Because what we’re talking about is an incredibly profound transition.
240
684889
4046
11:29
It's like the arrival of human intelligence in the world.
241
689352
3211
11:32
This is another intelligence arriving in the world.
242
692605
2961
11:35
And so it is an incredibly deep transition,
243
695608
3045
11:38
and we do not fully understand all the implications
244
698653
3128
11:41
and consequences of this.
245
701781
1585
11:43
And so we can't be certain
246
703407
1418
11:44
that it's going to be this, that or the other thing.
247
704867
2461
11:47
So we have to be open-minded about what may happen.
248
707328
3211
11:51
I have some optimism because I think
249
711332
1919
11:53
that if you want to make a system safe,
250
713251
3086
11:56
you need to understand a lot about that system.
251
716379
2669
11:59
You can't make an airplane safe
252
719090
1668
12:00
if you don't know about how airplanes work.
253
720800
2627
12:03
So as we get closer to AGI,
254
723761
2252
12:06
we will understand more and more about these systems,
255
726055
2502
12:08
and we will see more ways to make these systems safe,
256
728557
3003
12:11
make highly ethical AI systems.
257
731560
2712
12:15
But there are many things we don't understand about the future.
258
735064
3587
12:18
So I have to accept that there is a possibility that things may go badly
259
738693
5297
12:23
because I don't know what's going to happen.
260
743990
2085
12:26
I can't know that about the future in such a big change.
261
746117
3336
12:29
And even if the probability of something going bad is quite small,
262
749453
4839
12:34
we should take this extremely seriously.
263
754333
2378
12:36
CA: Paint a scenario of what going bad could look like.
264
756711
2877
12:39
SL: Well, it's hard to do
265
759630
1710
12:41
because you're talking about systems
266
761382
1752
12:43
that potentially have superhuman intelligence, right?
267
763134
3503
12:46
So there are many ways in which things would go bad in the world.
268
766637
4463
12:51
People sometimes point to, I don't know, engineered pathogens, right?
269
771142
3295
12:54
Maybe a superintelligence could design an engineered pathogen.
270
774437
3753
12:58
It could be much more mundane things.
271
778190
2253
13:00
Maybe with AGI, you know,
272
780484
2545
13:03
it gets used to destabilize democracy in the world,
273
783070
4588
13:07
with, you know, propaganda or all sorts of other things like that.
274
787700
3128
13:10
We don't know --
275
790870
1168
13:12
CA: That one might already have happened.
276
792079
2002
13:14
SL: There might be happening a bit already.
277
794081
2002
13:16
But, you know, there may be a lot more of this
278
796125
2169
13:18
if we have more powerful systems.
279
798294
1626
13:19
So there are many ways in which societies can be destabilized.
280
799920
2920
13:22
And you can see that in the history books.
281
802882
2043
13:24
CA: I mean, Shane, if you could have asked all humans,
282
804967
2544
13:27
say, 15 years ago, OK, we can open a door here,
283
807511
4338
13:31
and opening this door could lead to the best-ever outcomes for humanity.
284
811891
3670
13:35
But there's also a meaningful chance,
285
815603
1793
13:37
let's say it's more than five percent,
286
817396
2544
13:39
that we could actually destroy our civilization.
287
819982
2753
13:43
I mean, isn't there a chance that most people would have actually said,
288
823235
3337
13:46
"Don't you dare open that damn door.
289
826614
1752
13:48
Let's wait."
290
828407
1168
13:50
SL: If I had a magic wand and I could slow things down,
291
830326
4087
13:54
I would use that magic wand, but I don't.
292
834455
2336
13:56
There are dozens of companies,
293
836791
1835
13:58
well, there's probably 10 companies in the world now
294
838626
2461
14:01
that can develop the most cutting-edge models, including, I think,
295
841087
4421
14:05
some national intelligence services who have secret projects doing this.
296
845508
4087
14:10
And then there's, I don't know,
297
850262
1502
14:11
dozens of companies that can develop something that's a generation behind.
298
851764
3503
14:15
And remember, intelligence is incredibly valuable.
299
855267
3379
14:18
It's incredibly useful.
300
858687
1168
14:19
We're doing this
301
859855
1168
14:21
because we can see all kinds of value that can be created in this
302
861065
3086
14:24
for all sorts of reasons.
303
864151
1752
14:25
How do you stop this process?
304
865903
2836
14:28
I don't see any realistic plan that I've heard of,
305
868781
2836
14:31
of stopping this process.
306
871659
1209
14:32
Maybe we can --
307
872868
1210
14:34
I think we should think about regulating things.
308
874078
2294
14:36
I think we should do things like this as we do with every powerful technology.
309
876414
3670
14:40
There's nothing special about AI here.
310
880084
1877
14:41
People talk about, oh, you know, how dare you talk about regulating this?
311
881961
3462
14:45
No, we regulate powerful technologies all the time in the interests of society.
312
885464
3837
14:49
And I think this is a very important thing that we should be looking at.
313
889343
3420
14:52
CA: It's kind of the first time we have this superpowerful technology out there
314
892763
3754
14:56
that we literally don't understand in full how it works.
315
896559
3086
14:59
Is the most single, most important thing we must do, in your view,
316
899645
3921
15:03
to understand, to understand better what on Earth is going on,
317
903566
4671
15:08
so that we least have a shot at pointing it in the right direction?
318
908279
3170
15:11
SL: There is a lot of energy behind capabilities at the moment
319
911449
3294
15:14
because there's a lot of value in developing the capabilities.
320
914785
3045
15:17
I think we need to see a lot more energy going into actual science,
321
917830
4504
15:22
understanding how these networks work,
322
922376
3128
15:25
what they're doing, what they're not doing,
323
925504
2169
15:27
where they're likely to fail,
324
927715
1835
15:29
and understanding how we put these things together
325
929592
2419
15:32
so that we're able to find ways
326
932011
2085
15:34
to make these AGI systems profoundly ethical and safe.
327
934138
4129
15:38
I believe it's possible.
328
938267
1793
15:40
But we need to mobilize more people's minds and brains
329
940895
5505
15:46
to finding these solutions
330
946442
2502
15:48
so that we end up in a future
331
948986
1835
15:50
where we have incredibly powerful machine intelligence
332
950863
3962
15:54
that's also profoundly ethical and safe,
333
954867
2836
15:57
and it enhances and extends humanity
334
957745
2753
16:00
into, I think, a new golden period for humanity.
335
960539
4296
16:05
CA: Shane, thank you so much for sharing that vision
336
965920
2460
16:08
and coming to TED.
337
968380
1210
16:09
(Applause)
338
969590
1460
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7