Beware conflicts of interest | Dan Ariely

142,206 views ・ 2011-08-29

TED


Please double-click on the English subtitles below to play the video.

00:16
So, I was in the hospital for a long time.
0
16260
3000
00:19
And a few years after I left, I went back,
1
19260
3000
00:22
and the chairman of the burn department was very excited to see me --
2
22260
3000
00:25
said, "Dan, I have a fantastic new treatment for you."
3
25260
3000
00:28
I was very excited. I walked with him to his office.
4
28260
2000
00:30
And he explained to me that, when I shave,
5
30260
3000
00:33
I have little black dots on the left side of my face where the hair is,
6
33260
3000
00:36
but on the right side of my face
7
36260
2000
00:38
I was badly burned so I have no hair,
8
38260
2000
00:40
and this creates lack of symmetry.
9
40260
2000
00:42
And what's the brilliant idea he had?
10
42260
2000
00:44
He was going to tattoo little black dots
11
44260
2000
00:46
on the right side of my face
12
46260
3000
00:49
and make me look very symmetric.
13
49260
2000
00:51
It sounded interesting. He asked me to go and shave.
14
51260
3000
00:54
Let me tell you, this was a strange way to shave,
15
54260
2000
00:56
because I thought about it
16
56260
2000
00:58
and I realized that the way I was shaving then
17
58260
2000
01:00
would be the way I would shave for the rest of my life --
18
60260
2000
01:02
because I had to keep the width the same.
19
62260
2000
01:04
When I got back to his office,
20
64260
2000
01:06
I wasn't really sure.
21
66260
2000
01:08
I said, "Can I see some evidence for this?"
22
68260
2000
01:10
So he showed me some pictures
23
70260
2000
01:12
of little cheeks with little black dots --
24
72260
2000
01:14
not very informative.
25
74260
2000
01:16
I said, "What happens when I grow older and my hair becomes white?
26
76260
2000
01:18
What would happen then?"
27
78260
2000
01:20
"Oh, don't worry about it," he said.
28
80260
2000
01:22
"We have lasers; we can whiten it out."
29
82260
3000
01:25
But I was still concerned,
30
85260
2000
01:27
so I said, "You know what, I'm not going to do it."
31
87260
3000
01:30
And then came one of the biggest guilt trips of my life.
32
90260
4000
01:34
This is coming from a Jewish guy, all right, so that means a lot.
33
94260
3000
01:37
(Laughter)
34
97260
2000
01:39
And he said, "Dan, what's wrong with you?
35
99260
3000
01:42
Do you enjoy looking non-symmetric?
36
102260
2000
01:44
Do you have some kind of perverted pleasure from this?
37
104260
5000
01:49
Do women feel pity for you
38
109260
2000
01:51
and have sex with you more frequently?"
39
111260
3000
01:54
None of those happened.
40
114260
3000
01:58
And this was very surprising to me,
41
118260
2000
02:00
because I've gone through many treatments --
42
120260
2000
02:02
there were many treatments I decided not to do --
43
122260
2000
02:04
and I never got this guilt trip to this extent.
44
124260
2000
02:06
But I decided not to have this treatment.
45
126260
2000
02:08
And I went to his deputy and asked him, "What was going on?
46
128260
2000
02:10
Where was this guilt trip coming from?"
47
130260
2000
02:12
And he explained that they have done this procedure on two patients already,
48
132260
4000
02:16
and they need the third patient for a paper they were writing.
49
136260
3000
02:19
(Laughter)
50
139260
2000
02:21
Now you probably think that this guy's a schmuck.
51
141260
2000
02:23
Right, that's what he seems like.
52
143260
2000
02:25
But let me give you a different perspective on the same story.
53
145260
3000
02:28
A few years ago, I was running some of my own experiments in the lab.
54
148260
3000
02:31
And when we run experiments,
55
151260
2000
02:33
we usually hope that one group will behave differently than another.
56
153260
3000
02:36
So we had one group that I hoped their performance would be very high,
57
156260
3000
02:39
another group that I thought their performance would be very low,
58
159260
3000
02:42
and when I got the results, that's what we got --
59
162260
2000
02:44
I was very happy -- aside from one person.
60
164260
3000
02:47
There was one person in the group
61
167260
2000
02:49
that was supposed to have very high performance
62
169260
2000
02:51
that was actually performing terribly.
63
171260
2000
02:53
And he pulled the whole mean down,
64
173260
2000
02:55
destroying my statistical significance of the test.
65
175260
3000
02:59
So I looked carefully at this guy.
66
179260
2000
03:01
He was 20-some years older than anybody else in the sample.
67
181260
3000
03:04
And I remembered that the old and drunken guy
68
184260
2000
03:06
came one day to the lab
69
186260
2000
03:08
wanting to make some easy cash
70
188260
2000
03:10
and this was the guy.
71
190260
2000
03:12
"Fantastic!" I thought. "Let's throw him out.
72
192260
2000
03:14
Who would ever include a drunken guy in a sample?"
73
194260
3000
03:17
But a couple of days later,
74
197260
2000
03:19
we thought about it with my students,
75
199260
2000
03:21
and we said, "What would have happened if this drunken guy was not in that condition?
76
201260
3000
03:24
What would have happened if he was in the other group?
77
204260
2000
03:26
Would we have thrown him out then?"
78
206260
2000
03:28
We probably wouldn't have looked at the data at all,
79
208260
2000
03:30
and if we did look at the data,
80
210260
2000
03:32
we'd probably have said, "Fantastic! What a smart guy who is performing this low,"
81
212260
3000
03:35
because he would have pulled the mean of the group lower,
82
215260
2000
03:37
giving us even stronger statistical results than we could.
83
217260
3000
03:41
So we decided not to throw the guy out and to rerun the experiment.
84
221260
3000
03:44
But you know, these stories,
85
224260
3000
03:47
and lots of other experiments that we've done on conflicts of interest,
86
227260
3000
03:50
basically kind of bring two points
87
230260
2000
03:52
to the foreground for me.
88
232260
2000
03:54
The first one is that in life we encounter many people
89
234260
3000
03:57
who, in some way or another,
90
237260
3000
04:00
try to tattoo our faces.
91
240260
2000
04:02
They just have the incentives that get them to be blinded to reality
92
242260
3000
04:05
and give us advice that is inherently biased.
93
245260
3000
04:08
And I'm sure that it's something that we all recognize,
94
248260
2000
04:10
and we see that it happens.
95
250260
2000
04:12
Maybe we don't recognize it every time,
96
252260
2000
04:14
but we understand that it happens.
97
254260
2000
04:16
The most difficult thing, of course, is to recognize
98
256260
2000
04:18
that sometimes we too
99
258260
2000
04:20
are blinded by our own incentives.
100
260260
2000
04:22
And that's a much, much more difficult lesson to take into account.
101
262260
3000
04:25
Because we don't see how conflicts of interest work on us.
102
265260
4000
04:29
When I was doing these experiments,
103
269260
2000
04:31
in my mind, I was helping science.
104
271260
2000
04:33
I was eliminating the data
105
273260
2000
04:35
to get the true pattern of the data to shine through.
106
275260
2000
04:37
I wasn't doing something bad.
107
277260
2000
04:39
In my mind, I was actually a knight
108
279260
2000
04:41
trying to help science move along.
109
281260
2000
04:43
But this was not the case.
110
283260
2000
04:45
I was actually interfering with the process with lots of good intentions.
111
285260
3000
04:48
And I think the real challenge is to figure out
112
288260
2000
04:50
where are the cases in our lives
113
290260
2000
04:52
where conflicts of interest work on us,
114
292260
2000
04:54
and try not to trust our own intuition to overcome it,
115
294260
3000
04:57
but to try to do things
116
297260
2000
04:59
that prevent us from falling prey to these behaviors,
117
299260
2000
05:01
because we can create lots of undesirable circumstances.
118
301260
3000
05:05
I do want to leave you with one positive thought.
119
305260
2000
05:07
I mean, this is all very depressing, right --
120
307260
2000
05:09
people have conflicts of interest, we don't see it, and so on.
121
309260
3000
05:12
The positive perspective, I think, of all of this
122
312260
2000
05:14
is that, if we do understand when we go wrong,
123
314260
3000
05:17
if we understand the deep mechanisms
124
317260
2000
05:19
of why we fail and where we fail,
125
319260
2000
05:21
we can actually hope to fix things.
126
321260
2000
05:23
And that, I think, is the hope. Thank you very much.
127
323260
2000
05:25
(Applause)
128
325260
4000
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7