Edward Tenner: Unintended consequences

44,108 views ・ 2011-09-06

TED


Please double-click on the English subtitles below to play the video.

00:15
I didn't always love unintended consequences,
0
15260
3000
00:18
but I've really learned to appreciate them.
1
18260
2000
00:20
I've learned that they're really the essence
2
20260
2000
00:22
of what makes for progress,
3
22260
2000
00:24
even when they seem to be terrible.
4
24260
3000
00:27
And I'd like to review
5
27260
2000
00:29
just how unintended consequences
6
29260
3000
00:32
play the part that they do.
7
32260
3000
00:35
Let's go to 40,000 years before the present,
8
35260
5000
00:40
to the time of the cultural explosion,
9
40260
4000
00:44
when music, art, technology,
10
44260
5000
00:49
so many of the things that we're enjoying today,
11
49260
2000
00:51
so many of the things that are being demonstrated at TED
12
51260
3000
00:54
were born.
13
54260
2000
00:56
And the anthropologist Randall White
14
56260
3000
00:59
has made a very interesting observation:
15
59260
3000
01:02
that if our ancestors
16
62260
2000
01:04
40,000 years ago
17
64260
2000
01:06
had been able to see
18
66260
3000
01:09
what they had done,
19
69260
2000
01:11
they wouldn't have really understood it.
20
71260
2000
01:13
They were responding
21
73260
2000
01:15
to immediate concerns.
22
75260
3000
01:18
They were making it possible for us
23
78260
2000
01:20
to do what they do,
24
80260
2000
01:22
and yet, they didn't really understand
25
82260
2000
01:24
how they did it.
26
84260
2000
01:26
Now let's advance to 10,000 years before the present.
27
86260
5000
01:31
And this is when it really gets interesting.
28
91260
2000
01:33
What about the domestication of grains?
29
93260
3000
01:36
What about the origins of agriculture?
30
96260
3000
01:39
What would our ancestors 10,000 years ago
31
99260
3000
01:42
have said
32
102260
2000
01:44
if they really had technology assessment?
33
104260
2000
01:46
And I could just imagine the committees
34
106260
2000
01:48
reporting back to them
35
108260
2000
01:50
on where agriculture was going to take humanity,
36
110260
3000
01:53
at least in the next few hundred years.
37
113260
3000
01:56
It was really bad news.
38
116260
2000
01:58
First of all, worse nutrition,
39
118260
2000
02:00
maybe shorter life spans.
40
120260
2000
02:02
It was simply awful for women.
41
122260
2000
02:04
The skeletal remains from that period
42
124260
2000
02:06
have shown that they were grinding grain morning, noon and night.
43
126260
5000
02:11
And politically, it was awful.
44
131260
3000
02:14
It was the beginning of a much higher degree
45
134260
3000
02:17
of inequality among people.
46
137260
3000
02:20
If there had been rational technology assessment then,
47
140260
3000
02:23
I think they very well might have said,
48
143260
2000
02:25
"Let's call the whole thing off."
49
145260
3000
02:28
Even now, our choices are having unintended effects.
50
148260
4000
02:32
Historically, for example,
51
152260
2000
02:34
chopsticks -- according to one Japanese anthropologist
52
154260
3000
02:37
who wrote a dissertation about it
53
157260
2000
02:39
at the University of Michigan --
54
159260
2000
02:41
resulted in long-term changes
55
161260
3000
02:44
in the dentition, in the teeth,
56
164260
2000
02:46
of the Japanese public.
57
166260
2000
02:48
And we are also changing our teeth right now.
58
168260
3000
02:51
There is evidence
59
171260
2000
02:53
that the human mouth and teeth
60
173260
2000
02:55
are growing smaller all the time.
61
175260
2000
02:57
That's not necessarily a bad unintended consequence.
62
177260
3000
03:00
But I think from the point of view of a Neanderthal,
63
180260
2000
03:02
there would have been a lot of disapproval
64
182260
2000
03:04
of the wimpish choppers that we now have.
65
184260
3000
03:07
So these things are kind of relative
66
187260
3000
03:10
to where you or your ancestors happen to stand.
67
190260
4000
03:14
In the ancient world
68
194260
2000
03:16
there was a lot of respect for unintended consequences,
69
196260
3000
03:19
and there was a very healthy sense of caution,
70
199260
3000
03:22
reflected in the Tree of Knowledge,
71
202260
2000
03:24
in Pandora's Box,
72
204260
2000
03:26
and especially in the myth of Prometheus
73
206260
2000
03:28
that's been so important
74
208260
2000
03:30
in recent metaphors about technology.
75
210260
2000
03:32
And that's all very true.
76
212260
3000
03:35
The physicians of the ancient world --
77
215260
2000
03:37
especially the Egyptians,
78
217260
2000
03:39
who started medicine as we know it --
79
219260
2000
03:41
were very conscious
80
221260
2000
03:43
of what they could and couldn't treat.
81
223260
2000
03:45
And the translations of the surviving texts say,
82
225260
5000
03:50
"This I will not treat. This I cannot treat."
83
230260
2000
03:52
They were very conscious.
84
232260
2000
03:54
So were the followers of Hippocrates.
85
234260
2000
03:56
The Hippocratic manuscripts also --
86
236260
2000
03:58
repeatedly, according to recent studies --
87
238260
3000
04:01
show how important it is not to do harm.
88
241260
3000
04:04
More recently,
89
244260
2000
04:06
Harvey Cushing,
90
246260
2000
04:08
who really developed neurosurgery as we know it,
91
248260
2000
04:10
who changed it from a field of medicine
92
250260
3000
04:13
that had a majority of deaths resulting from surgery
93
253260
4000
04:17
to one in which there was a hopeful outlook,
94
257260
3000
04:20
he was very conscious
95
260260
2000
04:22
that he was not always going to do the right thing.
96
262260
3000
04:25
But he did his best,
97
265260
2000
04:27
and he kept meticulous records
98
267260
2000
04:29
that let him transform that branch of medicine.
99
269260
3000
04:32
Now if we look forward a bit
100
272260
3000
04:35
to the 19th century,
101
275260
2000
04:37
we find a new style of technology.
102
277260
2000
04:39
What we find is,
103
279260
2000
04:41
no longer simple tools,
104
281260
3000
04:44
but systems.
105
284260
2000
04:46
We find more and more
106
286260
2000
04:48
complex arrangements of machines
107
288260
2000
04:50
that make it harder and harder
108
290260
2000
04:52
to diagnose what's going on.
109
292260
2000
04:54
And the first people who saw that
110
294260
2000
04:56
were the telegraphers of the mid-19th century,
111
296260
3000
04:59
who were the original hackers.
112
299260
2000
05:01
Thomas Edison would have been very, very comfortable
113
301260
3000
05:04
in the atmosphere of a software firm today.
114
304260
3000
05:07
And these hackers had a word
115
307260
3000
05:10
for those mysterious bugs in telegraph systems
116
310260
3000
05:13
that they called bugs.
117
313260
2000
05:15
That was the origin of the word "bug."
118
315260
4000
05:19
This consciousness, though,
119
319260
2000
05:21
was a little slow to seep through the general population,
120
321260
3000
05:24
even people who were very, very well informed.
121
324260
3000
05:27
Samuel Clemens, Mark Twain,
122
327260
2000
05:29
was a big investor
123
329260
2000
05:31
in the most complex machine of all times --
124
331260
3000
05:34
at least until 1918 --
125
334260
2000
05:36
registered with the U.S. Patent Office.
126
336260
2000
05:38
That was the Paige typesetter.
127
338260
2000
05:40
The Paige typesetter
128
340260
2000
05:42
had 18,000 parts.
129
342260
2000
05:44
The patent had 64 pages of text
130
344260
3000
05:47
and 271 figures.
131
347260
4000
05:51
It was such a beautiful machine
132
351260
2000
05:53
because it did everything that a human being did
133
353260
3000
05:56
in setting type --
134
356260
2000
05:58
including returning the type to its place,
135
358260
2000
06:00
which was a very difficult thing.
136
360260
2000
06:02
And Mark Twain, who knew all about typesetting,
137
362260
2000
06:04
really was smitten by this machine.
138
364260
3000
06:07
Unfortunately, he was smitten in more ways than one,
139
367260
3000
06:10
because it made him bankrupt,
140
370260
2000
06:12
and he had to tour the world speaking
141
372260
2000
06:14
to recoup his money.
142
374260
3000
06:17
And this was an important thing
143
377260
2000
06:19
about 19th century technology,
144
379260
2000
06:21
that all these relationships among parts
145
381260
2000
06:23
could make the most brilliant idea fall apart,
146
383260
4000
06:27
even when judged by the most expert people.
147
387260
2000
06:29
Now there is something else, though, in the early 20th century
148
389260
3000
06:32
that made things even more complicated.
149
392260
3000
06:35
And that was that safety technology itself
150
395260
3000
06:38
could be a source of danger.
151
398260
2000
06:40
The lesson of the Titanic, for a lot of the contemporaries,
152
400260
3000
06:43
was that you must have enough lifeboats
153
403260
2000
06:45
for everyone on the ship.
154
405260
2000
06:47
And this was the result
155
407260
3000
06:50
of the tragic loss of lives
156
410260
2000
06:52
of people who could not get into them.
157
412260
2000
06:54
However, there was another case, the Eastland,
158
414260
3000
06:57
a ship that capsized in Chicago Harbor in 1915,
159
417260
4000
07:01
and it killed 841 people --
160
421260
3000
07:04
that was 14 more
161
424260
2000
07:06
than the passenger toll of the Titanic.
162
426260
3000
07:09
The reason for it, in part, was
163
429260
2000
07:11
the extra life boats that were added
164
431260
3000
07:14
that made this already unstable ship
165
434260
3000
07:17
even more unstable.
166
437260
2000
07:19
And that again proves
167
439260
2000
07:21
that when you're talking about unintended consequences,
168
441260
3000
07:24
it's not that easy to know
169
444260
2000
07:26
the right lessons to draw.
170
446260
2000
07:28
It's really a question of the system, how the ship was loaded,
171
448260
3000
07:31
the ballast and many other things.
172
451260
3000
07:35
So the 20th century, then,
173
455260
3000
07:38
saw how much more complex reality was,
174
458260
2000
07:40
but it also saw a positive side.
175
460260
3000
07:43
It saw that invention
176
463260
3000
07:46
could actually benefit from emergencies.
177
466260
2000
07:48
It could benefit
178
468260
2000
07:50
from tragedies.
179
470260
3000
07:53
And my favorite example of that --
180
473260
2000
07:55
which is not really widely known
181
475260
2000
07:57
as a technological miracle,
182
477260
2000
07:59
but it may be one of the greatest of all times,
183
479260
3000
08:02
was the scaling up of penicillin in the Second World War.
184
482260
4000
08:06
Penicillin was discovered in 1928,
185
486260
3000
08:09
but even by 1940,
186
489260
2000
08:11
no commercially and medically useful quantities of it
187
491260
3000
08:14
were being produced.
188
494260
2000
08:16
A number of pharmaceutical companies were working on it.
189
496260
3000
08:19
They were working on it independently,
190
499260
2000
08:21
and they weren't getting anywhere.
191
501260
2000
08:23
And the Government Research Bureau
192
503260
2000
08:25
brought representatives together
193
505260
2000
08:27
and told them that this is something
194
507260
2000
08:29
that has to be done.
195
509260
2000
08:31
And not only did they do it,
196
511260
2000
08:33
but within two years,
197
513260
2000
08:35
they scaled up penicillin
198
515260
2000
08:37
from preparation in one-liter flasks
199
517260
3000
08:40
to 10,000-gallon vats.
200
520260
4000
08:44
That was how quickly penicillin was produced
201
524260
4000
08:48
and became one of the greatest medical advances of all time.
202
528260
4000
08:52
In the Second World War, too,
203
532260
2000
08:54
the existence
204
534260
2000
08:56
of solar radiation
205
536260
2000
08:58
was demonstrated by studies of interference
206
538260
3000
09:01
that was detected by the radar stations of Great Britain.
207
541260
4000
09:05
So there were benefits in calamities --
208
545260
3000
09:08
benefits to pure science,
209
548260
2000
09:10
as well as to applied science
210
550260
2000
09:12
and medicine.
211
552260
3000
09:15
Now when we come to the period after the Second World War,
212
555260
3000
09:18
unintended consequences get even more interesting.
213
558260
4000
09:22
And my favorite example of that
214
562260
2000
09:24
occurred beginning in 1976,
215
564260
3000
09:27
when it was discovered
216
567260
2000
09:29
that the bacteria causing Legionnaires disease
217
569260
3000
09:32
had always been present in natural waters,
218
572260
3000
09:35
but it was the precise temperature of the water
219
575260
4000
09:39
in heating, ventilating and air conditioning systems
220
579260
3000
09:42
that raised the right temperature
221
582260
4000
09:46
for the maximum reproduction
222
586260
3000
09:49
of Legionella bacillus.
223
589260
2000
09:51
Well, technology to the rescue.
224
591260
2000
09:53
So chemists got to work,
225
593260
2000
09:55
and they developed a bactericide
226
595260
2000
09:57
that became widely used in those systems.
227
597260
3000
10:00
But something else happened in the early 1980s,
228
600260
4000
10:04
and that was that there was a mysterious epidemic
229
604260
2000
10:06
of failures of tape drives
230
606260
3000
10:09
all over the United States.
231
609260
2000
10:11
And IBM, which made them,
232
611260
3000
10:14
just didn't know what to do.
233
614260
3000
10:17
They commissioned a group of their best scientists
234
617260
3000
10:20
to investigate,
235
620260
2000
10:22
and what they found was
236
622260
2000
10:24
that all these tape drives
237
624260
2000
10:26
were located near ventilation ducts.
238
626260
3000
10:29
What happened was the bactericide was formulated
239
629260
3000
10:32
with minute traces of tin.
240
632260
2000
10:34
And these tin particles were deposited on the tape heads
241
634260
3000
10:37
and were crashing the tape heads.
242
637260
3000
10:40
So they reformulated the bactericide.
243
640260
3000
10:43
But what's interesting to me
244
643260
2000
10:45
is that this was the first case
245
645260
2000
10:47
of a mechanical device
246
647260
2000
10:49
suffering, at least indirectly, from a human disease.
247
649260
3000
10:52
So it shows that we're really all in this together.
248
652260
3000
10:55
(Laughter)
249
655260
2000
10:57
In fact, it also shows something interesting,
250
657260
3000
11:00
that although our capabilities and technology
251
660260
3000
11:03
have been expanding geometrically,
252
663260
2000
11:05
unfortunately, our ability to model their long-term behavior,
253
665260
3000
11:08
which has also been increasing,
254
668260
2000
11:10
has been increasing only arithmetically.
255
670260
3000
11:13
So one of the characteristic problems of our time
256
673260
3000
11:16
is how to close this gap
257
676260
2000
11:18
between capabilities and foresight.
258
678260
3000
11:21
One other very positive consequence
259
681260
3000
11:24
of 20th century technology, though,
260
684260
3000
11:27
was the way in which other kinds of calamities
261
687260
4000
11:31
could lead to positive advances.
262
691260
3000
11:34
There are two historians of business
263
694260
3000
11:37
at the University of Maryland,
264
697260
2000
11:39
Brent Goldfarb and David Kirsch,
265
699260
2000
11:41
who have done some extremely interesting work,
266
701260
2000
11:43
much of it still unpublished,
267
703260
3000
11:46
on the history of major innovations.
268
706260
2000
11:48
They have combined the list of major innovations,
269
708260
3000
11:51
and they've discovered that the greatest number, the greatest decade,
270
711260
3000
11:54
for fundamental innovations,
271
714260
2000
11:56
as reflected in all of the lists that others have made --
272
716260
4000
12:00
a number of lists that they have merged --
273
720260
2000
12:02
was the Great Depression.
274
722260
3000
12:05
And nobody knows just why this was so,
275
725260
3000
12:08
but one story can reflect something of it.
276
728260
3000
12:11
It was the origin of the Xerox copier,
277
731260
3000
12:14
which celebrated its 50th anniversary
278
734260
3000
12:17
last year.
279
737260
2000
12:19
And Chester Carlson, the inventor,
280
739260
5000
12:24
was a patent attorney.
281
744260
3000
12:27
He really was not intending
282
747260
3000
12:30
to work in patent research,
283
750260
2000
12:32
but he couldn't really find an alternative technical job.
284
752260
4000
12:36
So this was the best job he could get.
285
756260
2000
12:38
He was upset by the low quality and high cost
286
758260
4000
12:42
of existing patent reproductions,
287
762260
3000
12:45
and so he started to develop
288
765260
3000
12:48
a system of dry photocopying,
289
768260
3000
12:51
which he patented in the late 1930s --
290
771260
3000
12:54
and which became the first dry photocopier
291
774260
4000
12:58
that was commercially practical
292
778260
2000
13:00
in 1960.
293
780260
2000
13:02
So we see that sometimes,
294
782260
2000
13:04
as a result of these dislocations,
295
784260
2000
13:06
as a result of people
296
786260
2000
13:08
leaving their original intended career
297
788260
3000
13:11
and going into something else
298
791260
2000
13:13
where their creativity could make a difference,
299
793260
2000
13:15
that depressions
300
795260
2000
13:17
and all kinds of other unfortunate events
301
797260
3000
13:20
can have a paradoxically stimulating effect
302
800260
3000
13:23
on creativity.
303
803260
2000
13:25
What does this mean?
304
805260
2000
13:27
It means, I think,
305
807260
2000
13:29
that we're living in a time of unexpected possibilities.
306
809260
2000
13:31
Think of the financial world, for example.
307
811260
3000
13:34
The mentor of Warren Buffett, Benjamin Graham,
308
814260
3000
13:37
developed his system of value investing
309
817260
5000
13:42
as a result of his own losses
310
822260
2000
13:44
in the 1929 crash.
311
824260
2000
13:46
And he published that book
312
826260
2000
13:48
in the early 1930s,
313
828260
3000
13:51
and the book still exists in further editions
314
831260
2000
13:53
and is still a fundamental textbook.
315
833260
2000
13:55
So many important creative things can happen
316
835260
4000
13:59
when people learn from disasters.
317
839260
3000
14:02
Now think of the large and small plagues that we have now --
318
842260
4000
14:06
bed bugs, killer bees, spam --
319
846260
5000
14:11
and it's very possible that the solutions to those
320
851260
3000
14:14
will really extend well beyond the immediate question.
321
854260
3000
14:17
If we think, for example, of Louis Pasteur,
322
857260
3000
14:20
who in the 1860s
323
860260
2000
14:22
was asked to study
324
862260
2000
14:24
the diseases of silk worms for the silk industry,
325
864260
4000
14:28
and his discoveries were really the beginning
326
868260
3000
14:31
of the germ theory of disease.
327
871260
2000
14:33
So very often, some kind of disaster --
328
873260
3000
14:36
sometimes the consequence, for example,
329
876260
3000
14:39
of over-cultivation of silk worms,
330
879260
3000
14:42
which was a problem in Europe at the time --
331
882260
2000
14:44
can be the key to something much bigger.
332
884260
2000
14:46
So this means
333
886260
2000
14:48
that we need to take a different view
334
888260
2000
14:50
of unintended consequences.
335
890260
2000
14:52
We need to take a really positive view.
336
892260
3000
14:55
We need to see what they can do for us.
337
895260
3000
14:58
We need to learn
338
898260
2000
15:00
from those figures that I mentioned.
339
900260
2000
15:02
We need to learn, for example, from Dr. Cushing,
340
902260
3000
15:05
who killed patients
341
905260
2000
15:07
in the course of his early operations.
342
907260
2000
15:09
He had to have some errors. He had to have some mistakes.
343
909260
3000
15:12
And he learned meticulously from his mistakes.
344
912260
3000
15:15
And as a result,
345
915260
2000
15:17
when we say, "This isn't brain surgery,"
346
917260
3000
15:20
that pays tribute to how difficult it was
347
920260
3000
15:23
for anyone to learn from their mistakes
348
923260
2000
15:25
in a field of medicine
349
925260
2000
15:27
that was considered so discouraging in its prospects.
350
927260
3000
15:30
And we can also remember
351
930260
3000
15:33
how the pharmaceutical companies
352
933260
2000
15:35
were willing to pool their knowledge,
353
935260
2000
15:37
to share their knowledge,
354
937260
2000
15:39
in the face of an emergency,
355
939260
2000
15:41
which they hadn't really been for years and years.
356
941260
3000
15:44
They might have been able to do it earlier.
357
944260
3000
15:47
The message, then, for me,
358
947260
3000
15:50
about unintended consequences
359
950260
2000
15:52
is chaos happens;
360
952260
3000
15:55
let's make better use of it.
361
955260
2000
15:57
Thank you very much.
362
957260
2000
15:59
(Applause)
363
959260
4000
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7