Robert Wright: The logic of non-zero-sum progress

93,052 views ・ 2008-04-15

TED


Please double-click on the English subtitles below to play the video.

00:26
I've got apparently 18 minutes
0
26000
2000
00:28
to convince you that history has a direction, an arrow;
1
28000
3000
00:31
that in some fundamental sense, it's good;
2
31000
3000
00:34
that the arrow points to something positive.
3
34000
3000
00:37
Now, when the TED people first approached me about giving this upbeat talk --
4
37000
4000
00:41
(Laughter)
5
41000
1000
00:42
-- that was before the cartoon of Muhammad had triggered global rioting.
6
42000
6000
00:48
It was before the avian flu had reached Europe.
7
48000
2000
00:50
It was before Hamas had won the Palestinian election,
8
50000
2000
00:52
eliciting various counter-measures by Israel.
9
52000
4000
00:56
And to be honest, if I had known when I was asked to give this upbeat talk
10
56000
4000
01:00
that even as I was giving the upbeat talk,
11
60000
3000
01:03
the apocalypse would be unfolding --
12
63000
3000
01:06
(Laughter)
13
66000
1000
01:07
-- I might have said, "Is it okay if I talk about something else?"
14
67000
4000
01:11
But I didn't, OK. So we're here. I'll do what I can. I'll do what I can.
15
71000
5000
01:17
I've got to warn you:
16
77000
2000
01:19
the sense in which my worldview is upbeat has always been kind of subtle,
17
79000
7000
01:26
sometimes even elusive.
18
86000
2000
01:28
(Laughter)
19
88000
1000
01:29
The sense in which I can be uplifting and inspiring --
20
89000
3000
01:32
I mean, there's always been a kind of a certain grim dimension
21
92000
3000
01:35
to the way I try to uplift, so if grim inspiration --
22
95000
4000
01:39
(Laughter)
23
99000
1000
01:40
-- if grim inspiration is not a contradiction in terms, that is, I'm afraid,
24
100000
4000
01:44
the most you can hope for. OK, today -- that's if I succeed.
25
104000
4000
01:48
I'll see what I can do. OK?
26
108000
2000
01:50
Now, in one sense,
27
110000
2000
01:52
the claim that history has a direction is not that controversial.
28
112000
3000
01:55
If you're just talking about social structure,
29
115000
3000
01:58
OK, clearly that's gotten more complex
30
118000
2000
02:00
a little over the last 10,000 years -- has reached higher and higher levels.
31
120000
3000
02:03
And in fact, that's actually sustaining
32
123000
2000
02:05
a long-standing trend that predates human beings, OK,
33
125000
4000
02:09
that biological evolution was doing for us.
34
129000
3000
02:12
Because what happened in the beginning, this stuff encases itself in a cell,
35
132000
5000
02:17
then cells start hanging out together in societies.
36
137000
3000
02:20
Eventually they get so close, they form multicellular organisms,
37
140000
3000
02:23
then you get complex multicellular organisms; they form societies.
38
143000
5000
02:28
But then at some point, one of these multicellular organisms
39
148000
3000
02:31
does something completely amazing with this stuff, which is
40
151000
3000
02:34
it launches a whole second kind of evolution: cultural evolution.
41
154000
5000
02:39
And amazingly, that evolution sustains the trajectory
42
159000
3000
02:42
that biological evolution had established toward greater complexity.
43
162000
5000
02:47
By cultural evolution we mean the evolution of ideas.
44
167000
3000
02:50
A lot of you have heard the term "memes." The evolution of technology,
45
170000
3000
02:53
I pay a lot of attention to, so, you know,
46
173000
2000
02:55
one of the first things you got was a little hand axe.
47
175000
4000
02:59
Generations go by, somebody says, hey, why don't we put it on a stick?
48
179000
5000
03:04
(Laughter)
49
184000
3000
03:07
Just absolutely delights the little ones.
50
187000
3000
03:10
Next best thing to a video game.
51
190000
2000
03:12
This may not seem to impress,
52
192000
2000
03:14
but technological evolution is progressive, so another 10, 20,000 years,
53
194000
4000
03:18
and armaments technology takes you here.
54
198000
2000
03:20
(Laughter)
55
200000
1000
03:21
Impressive. And the rate of technological evolution speeds up,
56
201000
4000
03:25
so a mere quarter of a century after this, you get this, OK.
57
205000
4000
03:29
(Laughter)
58
209000
2000
03:31
And this.
59
211000
2000
03:33
(Laughter)
60
213000
1000
03:34
I'm sorry -- it was a cheap laugh, but I wanted to find a way
61
214000
3000
03:37
to transition back to this idea of the unfolding apocalypse,
62
217000
3000
03:40
and I thought that might do it.
63
220000
2000
03:42
(Applause)
64
222000
7000
03:49
So, what threatens to happen with this unfolding apocalypse
65
229000
5000
03:54
is the collapse of global social organization.
66
234000
4000
03:58
Now, first let me remind you how much work it took to get us where we are,
67
238000
3000
04:01
to be on the brink of true global social organization.
68
241000
3000
04:04
Originally, you had the most complex societies, the hunter-gatherer village.
69
244000
6000
04:10
Stonehenge is the remnant of a chiefdom,
70
250000
2000
04:12
which is what you get with the invention of agriculture: multi-village polity
71
252000
3000
04:15
with centralized rule.
72
255000
3000
04:18
With the invention of writing, you start getting cities. This is blurry. I kind of like that
73
258000
5000
04:23
because it makes it look like a one-celled organism and reminds you
74
263000
3000
04:26
how many levels organic organization has already moved through
75
266000
4000
04:30
to get to this point. And then you get to, you know, you get empires.
76
270000
6000
04:36
I want to stress, you know, social organization can transcend political bounds.
77
276000
4000
04:40
This is the Silk Road connecting the Chinese Empire and the Roman Empire.
78
280000
4000
04:44
So you had social complexity spanning the whole continent,
79
284000
3000
04:47
even if no polity did similarly. Today, you've got nation states.
80
287000
5000
04:52
Point is: there's obviously collaboration and organization going on
81
292000
3000
04:55
beyond national bounds.
82
295000
2000
04:57
This is actually just a picture of the earth at night,
83
297000
3000
05:00
and I'm just putting it up because I think it's pretty.
84
300000
2000
05:02
Does kind of convey the sense that this is an integrated system.
85
302000
5000
05:07
Now, I explained this growth of complexity by reference to something
86
307000
6000
05:13
called "non-zero sumness."
87
313000
4000
05:17
Assuming that a few of you did not do the assigned reading, very quickly,
88
317000
5000
05:22
the key idea is the distinction between zero-sum games, in which correlations
89
322000
5000
05:27
are inverse: always a winner and a loser.
90
327000
3000
05:30
Non-zero-sum games in which correlations can be positive, OK.
91
330000
4000
05:34
So like in tennis, usually it's win-lose;
92
334000
4000
05:38
it always adds up to zero-zero-sum. But if you're playing doubles,
93
338000
3000
05:41
the person on your side of the net, they're in the same boat as you,
94
341000
2000
05:43
so you're playing a non-zero-sum game with them.
95
343000
2000
05:45
It's either for the better or for the worse, OK.
96
345000
3000
05:48
A lot of forms of non-zero-sum behavior in the realm of economics and so on
97
348000
5000
05:53
in everyday life often leads to cooperation.
98
353000
4000
05:57
The argument I make is basically that, well,
99
357000
2000
05:59
non-zero-sum games have always been part of life.
100
359000
2000
06:01
You have them in hunter-gatherer societies,
101
361000
2000
06:03
but then through technological evolution, new forms of technology arise
102
363000
5000
06:08
that facilitate or encourage the playing of non-zero-sum games,
103
368000
5000
06:13
involving more people over larger territory.
104
373000
3000
06:16
Social structure adapts to accommodate this possibility and
105
376000
4000
06:20
to harness this productive potential, so you get cities, you know,
106
380000
3000
06:23
and you get all the non-zero-sum games you don't think about
107
383000
3000
06:26
that are being played across the world.
108
386000
1000
06:27
Like, have you ever thought when you buy a car,
109
387000
3000
06:30
how many people on how many different continents contributed
110
390000
2000
06:32
to the manufacture of that car? Those are people in effect
111
392000
5000
06:37
you're playing a non-zero-sum game with.
112
397000
2000
06:39
I mean, there are certainly plenty of them around.
113
399000
7000
06:46
Now, this sounds like an intrinsically upbeat worldview in a way,
114
406000
3000
06:49
because when you think of non-zero, you think win-win, you know,
115
409000
3000
06:52
that's good. Well, there are a few reasons
116
412000
2000
06:54
that actually it's not intrinsically upbeat.
117
414000
3000
06:57
First of all, it can accommodate; it doesn't deny the existence
118
417000
4000
07:01
of inequality exploitation war.
119
421000
3000
07:04
But there's a more fundamental reason
120
424000
2000
07:06
that it's not intrinsically upbeat, because a non-zero-sum game,
121
426000
3000
07:09
all it tells you for sure is that the fortunes will be correlated for better or worse.
122
429000
4000
07:13
It doesn't necessarily predict a win-win outcome.
123
433000
6000
07:19
So, in a way, the question is: on what grounds
124
439000
2000
07:21
am I upbeat at all about history? And the answer is,
125
441000
4000
07:25
first of all, on balance I would say people have played their games
126
445000
4000
07:29
to more win-win outcomes than lose-lose outcomes. On balance,
127
449000
4000
07:33
I think history is a net positive in the non-zero-sum game department.
128
453000
6000
07:39
And a testament to this is the thing that most amazes me,
129
459000
5000
07:44
most impresses me, and most uplifts me,
130
464000
2000
07:46
which is that there is a moral dimension to history;
131
466000
5000
07:51
there is a moral arrow. We have seen moral progress over time.
132
471000
3000
07:54
2,500 years ago, members of one Greek city-state
133
474000
4000
07:58
considered members of another Greek city-state subhuman
134
478000
3000
08:01
and treated them that way. And then this moral revolution arrived,
135
481000
5000
08:06
and they decided that actually, no, Greeks are human beings.
136
486000
4000
08:10
It's just the Persians who aren't fully human
137
490000
3000
08:13
and don't deserve to be treated very nicely.
138
493000
2000
08:15
But this was progress -- you know, give them credit. And now today,
139
495000
3000
08:18
we've seen more progress. I think -- I hope -- most people here would say
140
498000
3000
08:21
that all people everywhere are human beings,
141
501000
3000
08:24
deserve to be treated decently,
142
504000
3000
08:27
unless they do something horrendous, regardless of race or religion.
143
507000
4000
08:31
And you have to read your ancient history to realize what a revolution that has been,
144
511000
4000
08:35
OK. This was not a prevalent view,
145
515000
2000
08:37
few thousand years ago, and I attribute it to this non-zero-sum dynamic.
146
517000
5000
08:42
I think that's the reason there is as much tolerance toward nationalities,
147
522000
4000
08:46
ethnicities, religions as there is today. If you asked me,
148
526000
5000
08:51
you know, why am I not in favor of bombing Japan,
149
531000
2000
08:53
well, I'm only half-joking when I say they built my car.
150
533000
3000
08:56
We have this non-zero-sum relationship,
151
536000
2000
08:58
and I think that does lead to a kind of a tolerance to the extent that you realize
152
538000
6000
09:04
that someone else's welfare is positively correlated with yours --
153
544000
4000
09:08
you're more likely to cut them a break.
154
548000
2000
09:10
I kind of think this is a kind of a business-class morality.
155
550000
5000
09:15
Unfortunately, I don't fly trans-Atlantic business class often enough
156
555000
4000
09:19
to know, or any other kind of business class really,
157
559000
3000
09:22
but I assume that in business class, you don't hear many expressions of, you know,
158
562000
4000
09:26
bigotry about racial groups or ethnic groups,
159
566000
3000
09:29
because the people who are flying trans-Atlantic business class
160
569000
3000
09:32
are doing business with all these people; they're making money
161
572000
2000
09:34
off all these people. And I really do think that, in that sense at least,
162
574000
4000
09:38
capitalism has been a constructive force,
163
578000
3000
09:41
and more fundamentally, it's a non-zero-sumness
164
581000
2000
09:43
that has been a constructive force in expanding people's realm
165
583000
4000
09:47
of moral awareness. I think the non-zero-sum dynamic,
166
587000
5000
09:52
which is not only economic by any means -- it's not always commerce --
167
592000
4000
09:56
but it has driven us to the verge of a moral truth,
168
596000
5000
10:01
which is the fundamental equality of everyone. It has done that.
169
601000
3000
10:04
As it has moved global, moved us toward a global level of social organization,
170
604000
5000
10:09
it has driven us toward moral truth.
171
609000
2000
10:11
I think that's wonderful.
172
611000
3000
10:14
Now, back to the unfolding apocalypse.
173
614000
3000
10:17
And you may wonder, OK, that's all fine,
174
617000
2000
10:19
sounds great -- moral direction in history --
175
619000
2000
10:21
but what about this so-called clash of civilizations? Well, first of all,
176
621000
9000
10:30
I would emphasize that it fits into the non-zero-sum framework,
177
630000
3000
10:33
OK. If you look at the relationship
178
633000
2000
10:35
between the so-called Muslim world and Western world --
179
635000
3000
10:38
two terms I don't like, but can't really avoid;
180
638000
3000
10:41
in such a short span of time, they're efficient if nothing else --
181
641000
4000
10:45
it is non-zero-sum. And by that I mean,
182
645000
3000
10:48
if people in the Muslim world get more hateful, more resentful,
183
648000
4000
10:52
less happy with their place in the world,
184
652000
1000
10:53
it'll be bad for the West. If they get more happy, it'll be good for the West.
185
653000
4000
10:57
So that is a non-zero-sum dynamic.
186
657000
5000
11:02
And I would say the non-zero-sum dynamic is only going to grow more intense over time
187
662000
4000
11:06
because of technological trends, but more intense in a kind of negative way.
188
666000
5000
11:11
It's the downside correlation of their fortunes that will become more and more possible.
189
671000
6000
11:17
And one reason is because of something I call the "growing lethality of hatred."
190
677000
5000
11:22
More and more, it's possible for grassroots hatred abroad
191
682000
4000
11:26
to manifest itself in the form of organized violence on American soil.
192
686000
5000
11:31
And that's pretty new, and I think it's probably going to get a lot worse
193
691000
3000
11:34
-- this capacity -- because of
194
694000
3000
11:37
trends in information technology, in technologies that can be used
195
697000
4000
11:41
for purposes of munitions like biotechnology and nanotechnology.
196
701000
6000
11:47
We may be hearing more about that today.
197
707000
2000
11:49
And there's something I worry about especially, which is that
198
709000
3000
11:52
this dynamic will lead to a kind of a feedback cycle that puts us on a slippery slope.
199
712000
6000
11:58
What I have in mind is: terrorism happens here; we overreact to it.
200
718000
3000
12:01
That, you know, we're not sufficiently surgical in our retaliation
201
721000
4000
12:05
leads to more hatred abroad, more terrorism.
202
725000
2000
12:07
We overreact because being human, we feel like retaliating,
203
727000
4000
12:11
and it gets worse and worse and worse.
204
731000
2000
12:13
You could call this the positive feedback of negative vibes,
205
733000
4000
12:17
but I think in something so spooky,
206
737000
2000
12:19
we really shouldn't have the word positive there at all, even in a technical sense.
207
739000
3000
12:22
So let's call it the death spiral of negativity.
208
742000
3000
12:25
(Laughter)
209
745000
2000
12:27
I assure you if it happens, at the end, both the West
210
747000
2000
12:29
and the Muslim world will have suffered.
211
749000
3000
12:32
So, what do we do? Well, first of all, we can do a lot more with arms control,
212
752000
6000
12:38
the international regulation of dangerous technologies.
213
758000
2000
12:40
I have a whole global governance sermon
214
760000
2000
12:42
that I will spare you right now,
215
762000
2000
12:44
because I don't think that's going to be enough anyway, although it's essential.
216
764000
3000
12:47
I think we're going to have to have a major round
217
767000
2000
12:49
of moral progress in the world.
218
769000
2000
12:51
I think you're just going to have to see less hatred among groups,
219
771000
5000
12:56
less bigotry, and, you know, racial groups, religious groups, whatever.
220
776000
6000
13:02
I've got to admit I feel silly saying that.
221
782000
2000
13:04
It sounds so kind of Pollyannaish. I feel like Rodney King, you know,
222
784000
3000
13:07
saying, why can't we all just get along?
223
787000
2000
13:09
But hey, I don't really see any alternative, given the way I read the situation.
224
789000
6000
13:15
There's going to have to be moral progress.
225
795000
3000
13:18
There's going to have to be a lessening of the amount of hatred in the world,
226
798000
4000
13:22
given how dangerous it's becoming.
227
802000
4000
13:26
In my defense, I'd say, as naive as this may sound,
228
806000
3000
13:29
it's ultimately grounded in cynicism.
229
809000
3000
13:32
That is to say --
230
812000
1000
13:33
(Laughter)
231
813000
1000
13:34
-- thank you, thank you. That is to say, remember: my whole view
232
814000
5000
13:39
of morality is that it boils down to self-interest.
233
819000
3000
13:42
It's when people's fortunes are correlated.
234
822000
2000
13:44
It's when your welfare conduces to mine, that I decide, oh yeah,
235
824000
3000
13:47
I'm all in favor of your welfare. That's what's responsible
236
827000
4000
13:51
for this growth of this moral progress so far,
237
831000
3000
13:54
and I'm saying we once again have a correlation of fortunes,
238
834000
3000
13:57
and if people respond to it intelligently, we will see
239
837000
4000
14:01
the development of tolerance and so on --
240
841000
3000
14:04
the norms that we need, you know.
241
844000
3000
14:07
We will see the further evolution of this kind of business-class morality.
242
847000
4000
14:11
So, these two things, you know, if they get people's attention
243
851000
6000
14:17
and drive home the positive correlation and people do what's in their self-interests,
244
857000
3000
14:20
which is further the moral evolution,
245
860000
4000
14:24
then they could actually have a constructive effect.
246
864000
3000
14:27
And that's why I lump growing lethality of hatred
247
867000
3000
14:30
and death spiral of negativity under the general rubric,
248
870000
3000
14:33
reasons to be cheerful.
249
873000
2000
14:35
(Laughter)
250
875000
2000
14:37
Doing the best I can, OK.
251
877000
2000
14:39
(Laughter)
252
879000
1000
14:40
I never called myself Mr. Uplift.
253
880000
2000
14:42
I'm just doing what I can here.
254
882000
3000
14:45
(Laughter)
255
885000
1000
14:46
Now, launching a moral revolution has got to be hard, right?
256
886000
3000
14:49
I mean, what do you do?
257
889000
2000
14:51
And I think the answer is a lot of different people
258
891000
2000
14:53
are going to have to do a lot of different things.
259
893000
3000
14:56
We all start where we are. Speaking as an American
260
896000
5000
15:01
who has children whose security 10, 20, 30 years down the road
261
901000
4000
15:05
I worry about -- what I personally want to start out doing
262
905000
3000
15:08
is figuring out why so many people around the world hate us, OK.
263
908000
4000
15:12
I think that's a worthy research project myself.
264
912000
4000
15:16
I also like it because it's an intrinsically kind of morally redeeming exercise.
265
916000
5000
15:21
Because to understand why somebody
266
921000
2000
15:23
in a very different culture does something --
267
923000
2000
15:25
somebody you're kind of viewing as alien,
268
925000
2000
15:27
who's doing things you consider strange
269
927000
2000
15:29
in a culture you consider strange -- to really understand why they do
270
929000
4000
15:33
the things they do is a morally redeeming accomplishment,
271
933000
4000
15:37
because you've got to relate their experience to yours.
272
937000
2000
15:39
To really understand it, you've got to say, "Oh, I get it.
273
939000
4000
15:43
So when they feel resentful, it's kind of like
274
943000
2000
15:45
the way I feel resentful when this happens,
275
945000
2000
15:47
and for somewhat the same reasons." That's true understanding.
276
947000
4000
15:51
And I think that is an expansion of your moral compass when you manage to do that.
277
951000
6000
15:57
It's especially hard to do when people hate you, OK,
278
957000
3000
16:00
because you don't really, in a sense, want
279
960000
3000
16:03
to completely understand why people hate you.
280
963000
2000
16:05
I mean, you want to hear the reason, but you don't want to be able to relate to it.
281
965000
2000
16:07
You don't want it to make sense, right? (Laughter)
282
967000
2000
16:09
You don't want to say, "Well, yeah, I can kind of understand
283
969000
2000
16:11
how a human being in those circumstances
284
971000
2000
16:13
would hate the country I live in." That's not a pleasant thing,
285
973000
3000
16:16
but I think it's something that we're going to have to get used to and
286
976000
6000
16:22
work on. Now, I want to stress that to understand, you know --
287
982000
11000
16:33
there are people who don't like this whole business of understanding
288
993000
3000
16:36
the grassroots, the root causes of things; they don't want to know
289
996000
4000
16:40
why people hate us. I want to understand it.
290
1000000
3000
16:43
The reason you're trying to understand why they hate us,
291
1003000
2000
16:45
is to get them to quit hating us. The idea
292
1005000
3000
16:48
when you go through this moral exercise of really coming to appreciate
293
1008000
4000
16:52
their humanity and better understand them, is part of an effort
294
1012000
5000
16:57
to get them to appreciate your humanity in the long run.
295
1017000
2000
16:59
I think it's the first step toward that. That's the long-term goal.
296
1019000
4000
17:03
There are people who worry about this, and in fact,
297
1023000
5000
17:08
I, myself, apparently, was denounced on national TV
298
1028000
5000
17:13
a couple of nights ago because of an op-ed I'd written.
299
1033000
4000
17:17
It was kind of along these lines, and the allegation was
300
1037000
2000
17:19
that I have, quote, "affection for terrorists."
301
1039000
4000
17:23
Now, the good news is that the person who said it was Ann Coulter.
302
1043000
3000
17:26
(Laughter)
303
1046000
2000
17:28
(Applause)
304
1048000
2000
17:30
I mean, if you've got to have an enemy, do make it Ann Coulter.
305
1050000
2000
17:32
(Laughter)
306
1052000
1000
17:33
But it's not a crazy concern, OK, because understanding behavior
307
1053000
4000
17:37
can lead to a kind of empathy,
308
1057000
2000
17:39
and it can make it a little harder to deliver tough love, and so on.
309
1059000
3000
17:42
But I think we're a lot closer to erring on the side of not comprehending
310
1062000
7000
17:49
the situation clearly enough, than in comprehending it so clearly
311
1069000
4000
17:53
that we just can't, you know, get the army out to kill terrorists.
312
1073000
3000
17:56
So I'm not really worried about it. So --
313
1076000
3000
17:59
(Laughter)
314
1079000
2000
18:01
-- I mean, we're going to have to work on a lot of fronts,
315
1081000
5000
18:06
but if we succeed -- if we succeed -- then once again,
316
1086000
7000
18:13
non-zero-sumness and the recognition of non-zero-sum dynamics
317
1093000
4000
18:17
will have forced us to a higher moral level.
318
1097000
4000
18:21
And a kind of saving higher moral level,
319
1101000
5000
18:26
something that kind of literally saves the world.
320
1106000
2000
18:28
If you look at the word "salvation" in the Bible --
321
1108000
3000
18:31
the Christian usage that we're familiar with --
322
1111000
3000
18:34
saving souls, that people go to heaven -- that's actually a latecomer.
323
1114000
3000
18:37
The original meaning of the word "salvation" in the Bible is about saving the social system.
324
1117000
6000
18:43
"Yahweh is our Savior" means "He has saved the nation of Israel,"
325
1123000
3000
18:46
which at the time, was a pretty high-level social organization.
326
1126000
3000
18:49
Now, social organization has reached the global level, and I guess,
327
1129000
4000
18:53
if there's good news I can say I'm bringing you, it's just that
328
1133000
4000
18:57
all the salvation of the world requires is the intelligent pursuit
329
1137000
6000
19:03
of self-interests in a disciplined and careful way.
330
1143000
6000
19:09
It's going to be hard. I say we give it a shot anyway
331
1149000
3000
19:12
because we've just come too far to screw it up now.
332
1152000
3000
19:15
Thanks.
333
1155000
2000
19:17
(Applause)
334
1157000
2000
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7