The Dark Side of Competition in AI | Liv Boeree | TED

149,153 views ・ 2023-11-09

TED


Please double-click on the English subtitles below to play the video.

00:04
Competition.
0
4000
1502
00:05
It's a fundamental part of human nature.
1
5502
3003
00:09
I was a professional poker player for 10 years,
2
9047
2628
00:11
so I've very much seen all the good, bad and ugly ways it can manifest.
3
11675
4421
00:16
When it's done right,
4
16763
1502
00:18
it can drive us to incredible feats in sports and innovation,
5
18306
5589
00:23
like when car companies compete over who can build the safest cars
6
23895
4380
00:28
or the most efficient solar panels.
7
28275
2544
00:30
Those are all examples of healthy competition,
8
30860
2169
00:33
because even though individual companies might come and go,
9
33029
3170
00:36
in the long run,
10
36199
1168
00:37
the game between them creates win-win outcomes
11
37409
2419
00:39
where everyone benefits in the end.
12
39828
2043
00:42
But sometimes competition is not so great
13
42581
3003
00:45
and can create lose-lose outcomes where everyone's worse off than before.
14
45625
5464
00:51
Take these AI beauty filters, for example.
15
51715
2878
00:54
As you can see, they're a very impressive technology.
16
54968
2586
00:57
They can salvage almost any picture.
17
57596
2210
01:00
They can even make Angelina and Margot more beautiful.
18
60181
5297
01:05
So they're very handy,
19
65520
1335
01:06
especially for influencers who, now, at the click of a button,
20
66896
3045
01:09
can transform into the most beautiful Hollywood versions of themselves.
21
69983
3795
01:13
But handy doesn't always mean healthy.
22
73778
2419
01:16
And I've personally noticed how quickly these things can train you
23
76573
3879
01:20
to hate your natural face.
24
80493
2378
01:22
And there's growing evidence that they're creating issues
25
82912
2795
01:25
like body dysmorphia, especially in young people.
26
85707
3420
01:29
Nonetheless, these things are now endemic to social media
27
89502
4421
01:33
because the nature of the game demands it.
28
93965
2461
01:37
The platforms are incentivized to provide them
29
97093
2211
01:39
because hotter pictures means more hijacked limbic systems,
30
99346
3253
01:42
which means more scrolling and thus more ad revenue.
31
102641
2502
01:45
And users are incentivized to use them
32
105644
2210
01:47
because hotter pictures get you more followers.
33
107896
2377
01:50
But this is a trap,
34
110815
2461
01:53
because once you start using these things,
35
113318
2044
01:55
it's really hard to go back.
36
115362
2127
01:57
Plus, you don't even get a competitive advantage from them anymore
37
117530
3254
02:00
because everyone else is already using them too.
38
120784
2502
02:03
So influencers are stuck using these things
39
123787
3461
02:07
with all the downsides
40
127290
1460
02:08
and very little upside.
41
128792
1793
02:10
A lose-lose game.
42
130627
1334
02:12
A similar kind of trap is playing out in our news media right now,
43
132462
4504
02:16
but with much worse consequences.
44
136966
2712
02:19
You'd think since the internet came along
45
139719
1961
02:21
that the increased competition between news outlets
46
141680
3169
02:24
would create a sort of positive spiral,
47
144891
2044
02:26
like a race to the top of nuanced, impartial, accurate journalism.
48
146976
4839
02:32
Instead, we're seeing a race to the bottom of clickbait and polarization,
49
152565
6132
02:38
where even respectable papers are increasingly leaning
50
158697
2836
02:41
into these kind of low-brow partisan tactics.
51
161533
3253
02:45
Again, this is due to crappy incentives.
52
165537
3170
02:49
Today, we no longer just read our news.
53
169374
2669
02:52
We interact with it by sharing and commenting.
54
172085
3378
02:56
And headlines that trigger emotions like fear or anger
55
176131
4170
03:00
are far more likely to go viral than neutral or positive ones.
56
180343
4088
03:05
So in many ways, news editors are in a similar kind of trap
57
185098
4087
03:09
as the influencers,
58
189185
1544
03:10
where, the more their competitors lean into clickbaity tactics,
59
190770
4171
03:14
the more they have to as well.
60
194983
1960
03:16
Otherwise, their stories just get lost in the noise.
61
196943
2628
03:20
But this is terrible for everybody,
62
200029
2420
03:22
because now the media get less trust from the public,
63
202490
3295
03:25
but also it becomes harder and harder for anyone to discern truth from fiction,
64
205785
4463
03:30
which is a really big problem for democracy.
65
210248
2127
03:33
Now, this process of competition gone wrong
66
213251
3420
03:36
is actually the driving force behind so many of our biggest issues.
67
216713
4880
03:42
Plastic pollution,
68
222427
1877
03:44
deforestation,
69
224345
2128
03:46
antibiotic overuse in farming,
70
226514
2711
03:49
arms races,
71
229225
1627
03:50
greenhouse gas emissions.
72
230894
1918
03:52
These are all a result of crappy incentives,
73
232854
3754
03:56
of poorly designed games that push their players --
74
236649
3420
04:00
be them people, companies or governments --
75
240069
2753
04:02
into taking strategies and tactics that defer costs and harms to the future.
76
242864
5714
04:09
And what's so ridiculous is that most of the time,
77
249454
2919
04:12
these guys don't even want to be doing this.
78
252415
2878
04:15
You know, it's not like packaging companies
79
255335
2169
04:17
want to fill the oceans with plastic
80
257545
1919
04:19
or farmers want to worsen antibiotic resistance.
81
259506
2836
04:22
But they’re all stuck in the same dilemma of:
82
262759
3420
04:26
"If I don't use this tactic,
83
266221
2294
04:28
I’ll get outcompeted by all the others who do.
84
268515
3211
04:31
So I have to do it, too.”
85
271768
1835
04:34
This is the mechanism we need to fix as a civilization.
86
274604
3545
04:38
And I know what you're probably all thinking, "So it's capitalism."
87
278858
3170
04:42
No, it's not capitalism.
88
282028
1543
04:43
Which, yes, can cause problems,
89
283613
1793
04:45
but it can also solve them and has been fantastic in general.
90
285448
3170
04:48
It's something much deeper.
91
288618
2127
04:51
It's a force of misaligned incentives of game theory itself.
92
291287
5005
04:57
So a few years ago, I retired from poker,
93
297001
2378
04:59
in part because I wanted to understand this mechanism better.
94
299420
3671
05:03
Because it takes many different forms, and it goes by many different names.
95
303633
4213
05:08
These are just some of those names.
96
308346
2085
05:11
You can see they're a little bit abstract and clunky, right?
97
311015
3337
05:14
They don't exactly roll off the tongue.
98
314394
2002
05:16
And given how insidious and connected all of these problems are,
99
316771
5089
05:21
it helps to have a more visceral way of recognizing them.
100
321901
3963
05:27
So this is probably the only time
101
327073
1585
05:28
you're going to hear about the Bible at this conference.
102
328658
2711
05:31
But I want to tell you a quick story from it,
103
331369
2127
05:33
because allegedly, back in the Canaanite days,
104
333496
2544
05:36
there was a cult who wanted money and power so badly,
105
336040
5256
05:41
they were willing to sacrifice their literal children for it.
106
341296
3587
05:45
And they did this by burning them alive in an effigy
107
345592
4212
05:49
of a God that they believed would then reward them
108
349846
2711
05:52
for this ultimate sacrifice.
109
352599
2127
05:55
And the name of this god was Moloch.
110
355643
2962
05:59
Bit of a bummer, as stories go.
111
359772
2461
06:02
But you can see why it's an apt metaphor,
112
362233
2586
06:04
because sometimes we get so lost in winning the game right in front of us,
113
364861
6673
06:11
we lose sight of the bigger picture
114
371534
1960
06:13
and sacrifice too much in our pursuit of victory.
115
373536
3754
06:17
So just like these guys were sacrificing their children for power,
116
377832
4088
06:21
those influencers are sacrificing their happiness for likes.
117
381961
4588
06:27
Those news editors are sacrificing their integrity for clicks,
118
387008
4087
06:31
and polluters are sacrificing the biosphere for profit.
119
391137
3796
06:35
In all these examples,
120
395767
1793
06:37
the short-term incentives of the games themselves are pushing,
121
397602
3962
06:41
they're tempting their players
122
401606
1752
06:43
to sacrifice more and more of their future,
123
403358
3378
06:46
trapping them in a death spiral where they all lose in the end.
124
406778
4129
06:52
That's Moloch's trap.
125
412158
2127
06:54
The mechanism of unhealthy competition.
126
414661
2752
06:58
And the same is now happening in the AI industry.
127
418164
3337
07:02
We're all aware of the race that's heating up
128
422251
2128
07:04
between companies right now
129
424379
1293
07:05
over who can score the most compute,
130
425713
2253
07:08
who can get the biggest funding round or get the top talent.
131
428007
3379
07:11
Well, as more and more companies enter this race,
132
431386
3461
07:14
the greater the pressure for everyone to go as fast as possible
133
434847
3629
07:18
and sacrifice other important stuff like safety testing.
134
438476
4171
07:23
This has all the hallmarks of a Moloch trap.
135
443189
3253
07:27
Because, like, imagine you're a CEO who, you know, in your heart of hearts,
136
447068
3587
07:30
believes that your team is the best
137
450655
2586
07:33
to be able to safely build extremely powerful AI.
138
453241
3462
07:37
Well, if you go too slowly, then you run the risk of other,
139
457120
4170
07:41
much less cautious teams getting there first
140
461332
2628
07:44
and deploying their systems before you can.
141
464002
2168
07:46
So that in turn pushes you to be more reckless yourself.
142
466212
3378
07:49
And given how many different experts and researchers,
143
469966
3628
07:53
both within these companies
144
473594
1585
07:55
but also completely independent ones,
145
475179
2545
07:57
have been warning us about the extreme risks of rushed AI,
146
477765
4296
08:02
this approach is absolutely mad.
147
482103
2711
08:05
Plus, almost all AI companies
148
485398
3170
08:08
are beholden to satisfying their investors,
149
488609
3003
08:11
a short-term incentive which, over time, will inevitably start to conflict
150
491612
4255
08:15
with any benevolent mission.
151
495867
1751
08:18
And this wouldn't be a big deal
152
498202
1502
08:19
if this was really just toasters we're talking about here.
153
499704
3337
08:23
But AI, and especially AGI,
154
503041
2752
08:25
is set to be a bigger paradigm shift
155
505835
1752
08:27
than the agricultural or industrial revolutions.
156
507587
2669
08:30
A moment in time so pivotal,
157
510590
2544
08:33
it's deserving of reverence and reflection,
158
513176
4546
08:37
not something to be reduced to a corporate rat race
159
517764
2419
08:40
of who can score the most daily active users.
160
520183
2544
08:43
I'm not saying I know
161
523186
1209
08:44
what the right trade-off between acceleration and safety is,
162
524437
3420
08:47
but I do know that we'll never find out what that right trade-off is
163
527899
3962
08:51
if we let Moloch dictate it for us.
164
531903
2794
08:56
So what can we do?
165
536699
1794
08:58
Well, the good news is we have managed to coordinate
166
538868
3086
09:01
to escape some of Moloch's traps before.
167
541954
2002
09:04
We managed to save the ozone layer from CFCs
168
544540
2545
09:07
with the help of the Montreal Protocol.
169
547126
2169
09:09
We managed to reduce the number of nuclear weapons on Earth
170
549337
2878
09:12
by 80 percent,
171
552256
1210
09:13
with the help of the Strategic Arms Reduction Treaty in 1991.
172
553466
3378
09:17
So smart regulation may certainly help with AI too,
173
557428
4338
09:21
but ultimately,
174
561766
1335
09:23
it's the players within the game who have the most influence on it.
175
563142
4004
09:27
So we need AI leaders to show us
176
567814
2294
09:30
that they're not only aware of the risks their technologies pose,
177
570108
3586
09:33
but also the destructive nature of the incentives
178
573736
3253
09:37
that they're currently beholden to.
179
577031
1877
09:39
As their technological capabilities reach towards the power of gods,
180
579617
3712
09:43
they're going to need the godlike wisdom to know how to wield them.
181
583329
3712
09:47
So it doesn't fill me with encouragement
182
587917
2211
09:50
when I see a CEO of a very major company saying something like,
183
590169
4338
09:54
"I want people to know we made our competitor dance."
184
594507
2753
09:57
That is not the type of mindset we need here.
185
597677
2669
10:00
We need leaders who are willing to flip Moloch's playbook,
186
600346
3629
10:04
who are willing to sacrifice their own individual chance of winning
187
604016
3671
10:07
for the good of the whole.
188
607687
1626
10:10
Now, fortunately, the three leading labs are showing some signs of doing this.
189
610148
4921
10:15
Anthropic recently announced their responsible scaling policy,
190
615695
3086
10:18
which pledges to only increase capabilities
191
618823
2586
10:21
once certain security criteria have been met.
192
621409
2627
10:24
OpenAI have recently pledged
193
624954
1418
10:26
to dedicate 20 percent of their compute purely to alignment research.
194
626414
3962
10:31
And DeepMind have shown a decade-long focus
195
631627
3087
10:34
of science ahead of commerce,
196
634714
2377
10:37
like their development of AlphaFold,
197
637133
1752
10:38
which they gave away to the science community for free.
198
638885
2627
10:41
These are all steps in the right direction,
199
641846
2044
10:43
but they are still nowhere close to being enough.
200
643931
2836
10:46
I mean, most of these are currently just words,
201
646809
2586
10:49
they're not even proven actions.
202
649437
2127
10:52
So we need a clear way to turn the AI race into a definitive race to the top.
203
652023
6006
10:58
Perhaps companies can start competing over who can be within these metrics,
204
658779
5297
11:04
over who can develop the best security criteria.
205
664118
3128
11:07
A race of who can dedicate the most compute to alignment.
206
667288
4004
11:11
Now that would truly flip the middle finger to Moloch.
207
671292
2669
11:14
Competition can be an amazing tool,
208
674587
3253
11:17
provided we wield it wisely.
209
677882
2085
11:20
And we're going to need to do that
210
680468
1668
11:22
because the stakes we are playing for are astronomical.
211
682136
3170
11:25
If we get AI, and especially AGI, wrong,
212
685806
3421
11:29
it could lead to unimaginable catastrophe.
213
689268
2795
11:32
But if we get it right,
214
692980
1502
11:34
it could be our path out of many of these Moloch traps
215
694523
2837
11:37
that I've mentioned today.
216
697401
1502
11:40
And as things get crazier over the coming years,
217
700446
3045
11:43
which they're probably going to,
218
703532
1836
11:45
it's going to be more important than ever
219
705368
2002
11:47
that we remember that it is the real enemy here, Moloch.
220
707370
4296
11:52
Not any individual CEO or company, and certainly not one another.
221
712208
4379
11:57
So don't hate the players,
222
717255
1751
11:59
change the game.
223
719006
1293
12:01
(Applause)
224
721509
1126
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7