Paul Bloom: Can prejudice ever be a good thing?

188,139 views ・ 2014-07-03

TED


请双击下面的英文字幕来播放视频。

翻译人员: Zhiting Chen 校对人员: Geoff Chen
00:12
When we think about prejudice and bias,
0
12844
2378
当我们想到偏见和偏爱,
00:15
we tend to think about stupid and evil people
1
15222
2144
我们总会联想到愚蠢又邪恶的人
00:17
doing stupid and evil things.
2
17366
2454
做着愚蠢且邪恶的事。
00:19
And this idea is nicely summarized
3
19820
2070
英国评论家威廉‧哈兹里特
00:21
by the British critic William Hazlitt,
4
21890
2468
非常好地总结了这个想法,
00:24
who wrote, "Prejudice is the child of ignorance."
5
24358
2935
他写道,“偏见是无知的幼子”
00:27
I want to try to convince you here
6
27293
2112
我想要试图游说你
00:29
that this is mistaken.
7
29405
1635
这是错误的。
00:31
I want to try to convince you
8
31040
1732
我想要向你证明
00:32
that prejudice and bias
9
32772
1723
偏见和偏爱
00:34
are natural, they're often rational,
10
34495
3288
是自然而然的, 它们时常是理性的,
00:37
and they're often even moral,
11
37783
1831
时常甚至是道德的,
00:39
and I think that once we understand this,
12
39614
2252
我想当我们理解这,
00:41
we're in a better position to make sense of them
13
41866
2509
当它出现问题的时候
00:44
when they go wrong,
14
44375
1057
当它可能造成严重后果的时候,
00:45
when they have horrible consequences,
15
45432
1768
我们会有更好的应对方式,
00:47
and we're in a better position to know what to do
16
47200
2325
当这一切发生的时候,
00:49
when this happens.
17
49525
1682
我们会知道要如何处理。
00:51
So, start with stereotypes. You look at me,
18
51207
3027
好,让我们从“成见”开始聊起。 你看着我,
00:54
you know my name, you know certain facts about me,
19
54234
2246
你知道我的名字, 以及一些关于我的事情,
00:56
and you could make certain judgments.
20
56480
1829
你可以做出一定的判断。
00:58
You could make guesses about my ethnicity,
21
58309
2853
你可以猜测我的种族,
01:01
my political affiliation, my religious beliefs.
22
61162
3281
我的政治倾向,我的宗教信仰
01:04
And the thing is, these judgments tend to be accurate.
23
64443
2099
这些判断似乎可以是准确的。
01:06
We're very good at this sort of thing.
24
66542
2182
我们对这些事非常擅长。
01:08
And we're very good at this sort of thing
25
68724
1483
我们非常善于这些事的原因是
01:10
because our ability to stereotype people
26
70207
2733
我们“定义”他人的能力
01:12
is not some sort of arbitrary quirk of the mind,
27
72940
3255
不是非常武断的意识行为,
01:16
but rather it's a specific instance
28
76195
2316
而是一个综合过程
01:18
of a more general process,
29
78511
1655
的特定反应,
01:20
which is that we have experience
30
80166
1619
这意谓着,当我们对所经历过的
01:21
with things and people in the world
31
81785
1541
世界上发生的人与事
01:23
that fall into categories,
32
83326
1249
做出分类,
01:24
and we can use our experience to make generalizations
33
84575
2456
我们可以用经验来
01:27
about novel instances of these categories.
34
87031
2359
做出反应,
01:29
So everybody here has a lot of experience
35
89390
2367
这里的每个人都有很多经验
01:31
with chairs and apples and dogs,
36
91757
2253
椅子,苹果,狗
01:34
and based on this, you could see
37
94010
1636
根据这些物件,你可以看到
01:35
unfamiliar examples and you could guess,
38
95646
2352
不熟悉的例子,并且你可以猜测,
01:37
you could sit on the chair,
39
97998
1316
你可以坐在这张椅子上,
01:39
you could eat the apple, the dog will bark.
40
99314
2565
你可以吃这个苹果, 狗会对着你叫。
01:41
Now we might be wrong.
41
101879
1764
我们可能是错的。
01:43
The chair could collapse if you sit on it,
42
103643
1800
当你坐在椅子上的时候, 椅子可能会塌,
01:45
the apple might be poison, the dog might not bark,
43
105443
2222
苹果可能是有毒的, 狗未必会叫,
01:47
and in fact, this is my dog Tessie, who doesn't bark.
44
107665
2870
事实上,这是我的狗泰西, 它不叫。
01:50
But for the most part, we're good at this.
45
110535
2759
但在大多数情况下, 我们对此很擅长。
01:53
For the most part, we make good guesses
46
113294
1916
在大多数情况下, 我们的猜测是合理的
01:55
both in the social domain and the non-social domain,
47
115210
1814
在社会领域或其他领域,
01:57
and if we weren't able to do so,
48
117024
1949
如果我们不具有这样的能力,
01:58
if we weren't able to make guesses about new instances that we encounter,
49
118973
3216
如果我们没有办法对出现的新鲜事物 做出正确的猜测,
02:02
we wouldn't survive.
50
122189
1451
我们将无法生存。
02:03
And in fact, Hazlitt later on in his wonderful essay
51
123640
2869
事实上,哈兹里特后来 在他的佳作中
02:06
concedes this.
52
126509
1485
对此评论做出了让步。
02:07
He writes, "Without the aid of prejudice and custom,
53
127994
2542
他写道,“如果没有偏见和风俗习惯的帮助,
02:10
I should not be able to find my way my across the room;
54
130536
2340
我将无法找到穿越房间的路;
02:12
nor know how to conduct myself in any circumstances,
55
132876
2452
也无法知晓自己在不同条件下 要做出怎样的行为反应,
02:15
nor what to feel in any relation of life."
56
135328
4203
也无法体会人生关系中的任何感觉。”
02:19
Or take bias.
57
139531
1509
现在来讨论偏爱。
02:21
Now sometimes, we break the world up into
58
141040
1708
有时候,我们将世界划分为
02:22
us versus them, into in-group versus out-group,
59
142748
3001
我们对抗他们, 内群体对抗外群体,
02:25
and sometimes when we do this,
60
145749
1161
有时当我们这么做的时候,
02:26
we know we're doing something wrong,
61
146910
1557
我们知道我们正在犯错误,
02:28
and we're kind of ashamed of it.
62
148467
1673
我们甚至会有些惭愧。
02:30
But other times we're proud of it.
63
150140
1483
但其他时间,我们对此很自豪。
02:31
We openly acknowledge it.
64
151623
1813
我们公开承认。
02:33
And my favorite example of this
65
153436
1282
我最喜欢的例子
02:34
is a question that came from the audience
66
154718
2402
是一个来自观众的问题
02:37
in a Republican debate prior to the last election.
67
157120
2717
在一个选前的共和党辩论。
02:39
(Video) Anderson Cooper: Gets to your question,
68
159837
2292
(视频)安德森·库柏:问答时间,
02:42
the question in the hall, on foreign aid? Yes, ma'am.
69
162129
4181
观众提问,有关对外援助? 有请这位女士。
02:46
Woman: The American people are suffering
70
166310
2236
女士:在美国国内,有很多
02:48
in our country right now.
71
168546
2637
美国人民正在经受苦难。
02:51
Why do we continue to send foreign aid
72
171183
3348
为什么我们要持续为其他国家
02:54
to other countries
73
174531
1316
提供援助呢?
02:55
when we need all the help we can get for ourselves?
74
175847
4103
此时我们需要这些援助 使用在本国人身上。
02:59
AC: Governor Perry, what about that?
75
179950
1695
安德森·库柏: 州长佩里,请您解答?
03:01
(Applause)
76
181645
1367
(鼓掌)
03:03
Rick Perry: Absolutely, I think it's—
77
183012
2338
里克·佩里:绝对的,我认为--
03:05
Paul Bloom: Each of the people onstage
78
185350
1660
保罗·布鲁姆:这个台上的每个人
03:07
agreed with the premise of her question,
79
187010
1971
同意她问题的前提,
03:08
which is as Americans, we should care more
80
188981
2119
这个前提就是作为美国人, 我们应该将
03:11
about Americans than about other people.
81
191100
2126
更多的关注给予本国人民 而不是其他人民。
03:13
And in fact, in general, people are often swayed
82
193226
2865
事实上,总的来说, 人们时常容易受到影响
03:16
by feelings of solidarity, loyalty, pride, patriotism,
83
196091
3508
对他们国家以及种族 诸如团结,忠诚,
03:19
towards their country or towards their ethnic group.
84
199599
2716
自豪以及爱国主义。
03:22
Regardless of your politics, many people feel proud to be American,
85
202315
3085
不谈政治倾向,很多人对他们 美国人的身份感到自豪,
03:25
and they favor Americans over other countries.
86
205400
2062
他们偏爱美国多于其他国家。
03:27
Residents of other countries feel the same about their nation,
87
207462
2850
其他国家的人们 也持有这样的态度,
03:30
and we feel the same about our ethnicities.
88
210312
2486
人们对自己的种族 也是如此。
03:32
Now some of you may reject this.
89
212798
1684
一些人会反对这种说法。
03:34
Some of you may be so cosmopolitan
90
214482
1721
你们中的某些人可能 是世界主义者,
03:36
that you think that ethnicity and nationality
91
216213
2334
会认为种族和国籍
03:38
should hold no moral sway.
92
218547
2153
不该影响到人们。
03:40
But even you sophisticates accept
93
220700
2762
但是,即便如此,
03:43
that there should be some pull
94
223462
1834
你仍然会接受
03:45
towards the in-group in the domain of friends and family,
95
225296
2701
群体可以以朋友和家人来做划分,
03:47
of people you're close to,
96
227997
1421
那些与你更亲近的人们
03:49
and so even you make a distinction
97
229418
1561
甚至你也会做一个划分
03:50
between us versus them.
98
230979
1975
区别我们和他们,
03:52
Now, this distinction is natural enough
99
232954
2603
这些区别是自然而然的
03:55
and often moral enough, but it can go awry,
100
235557
2924
时常也是道德的, 但有时也会出错,
03:58
and this was part of the research
101
238481
1729
这是伟大的社会心理学家
04:00
of the great social psychologist Henri Tajfel.
102
240210
2759
亨利·泰吉弗尔研究的一个部分。
04:02
Tajfel was born in Poland in 1919.
103
242969
2605
泰吉弗尔生于1919年的波兰。
04:05
He left to go to university in France,
104
245574
2139
他离开家乡去法国上大学,
04:07
because as a Jew, he couldn't go to university in Poland,
105
247713
2555
因为他是犹太人, 无法在波兰接受大学教育,
04:10
and then he enlisted in the French military
106
250268
2510
随后在第二次世界大战
04:12
in World War II.
107
252778
1283
他应募入伍加入法军。
04:14
He was captured and ended up
108
254061
1769
他被捕了随后被送到
04:15
in a prisoner of war camp,
109
255830
1531
战俘营,
04:17
and it was a terrifying time for him,
110
257361
2267
对他来说这是非常恐怖的经历,
04:19
because if it was discovered that he was a Jew,
111
259628
1688
因为如果一旦发现他是犹太人,
04:21
he could have been moved to a concentration camp,
112
261316
2092
他便会被移送到集中营,
04:23
where he most likely would not have survived.
113
263408
1992
很难活下来。
04:25
And in fact, when the war ended and he was released,
114
265400
2587
事实上,当战争结束的时候, 他被释放了,
04:27
most of his friends and family were dead.
115
267987
2505
绝大多数他的亲友都死亡了。
04:30
He got involved in different pursuits.
116
270492
1837
他参与不同的活动。
04:32
He helped out the war orphans.
117
272329
1531
他帮助战争孤儿。
04:33
But he had a long-lasting interest
118
273860
1731
但他对偏见科学
04:35
in the science of prejudice,
119
275591
1545
有着极高的兴趣
04:37
and so when a prestigious British scholarship
120
277136
2660
因此当一个极有声望的, 有关“刻板印象成见”
04:39
on stereotypes opened up, he applied for it,
121
279796
1845
的英国奖学金机会释出的时候, 他递交了申请,
04:41
and he won it,
122
281641
1357
并拿到了奖学金,
04:42
and then he began this amazing career.
123
282998
2190
这使他开启了精彩的职业生涯。
04:45
And what started his career is an insight
124
285188
2749
他的职业开始于发觉
04:47
that the way most people were thinking
125
287937
1840
当绝大多数人思考大屠杀是错误的
04:49
about the Holocaust was wrong.
126
289777
2116
采取了怎样的方法。
04:51
Many people, most people at the time,
127
291893
2406
很多人,那时候的绝大多数人,
04:54
viewed the Holocaust as sort of representing
128
294299
1901
将大屠杀视为
04:56
some tragic flaw on the part of the Germans,
129
296200
3004
代表某种德国人的悲剧错误,
04:59
some genetic taint, some authoritarian personality.
130
299204
3834
像是基因污点,威权性格。
05:03
And Tajfel rejected this.
131
303038
2058
泰吉弗尔拒绝这样的解释。
05:05
Tajfel said what we see in the Holocaust
132
305096
2543
他说道,大屠杀
05:07
is just an exaggeration
133
307639
2311
只是夸大了
05:09
of normal psychological processes
134
309950
1778
正常的心理状态
05:11
that exist in every one of us.
135
311728
1761
这样的心理状态存在于 我们中的每一个人。
05:13
And to explore this, he did a series of classic studies
136
313489
2685
为了继续研究, 他做了一系列的经典研究
05:16
with British adolescents.
137
316174
1744
有关英国青少年。
05:17
And in one of his studies, what he did was he asked
138
317918
1549
在他的其中一项研究中, 他去询问
05:19
the British adolescents all sorts of questions,
139
319467
2552
英国青少年各种不同的问题,
05:22
and then based on their answers, he said,
140
322019
1884
基于他们的回答,他说,
05:23
"I've looked at your answers, and based on the answers,
141
323903
2357
“我看过你的答案, 基于你的回答,
05:26
I have determined that you are either" —
142
326260
2097
我决定你是”--
05:28
he told half of them —
143
328357
1006
他告诉青少年中一半的人--
05:29
"a Kandinsky lover, you love the work of Kandinsky,
144
329363
2957
“康定斯基迷, 你喜爱康定斯基的作品,
05:32
or a Klee lover, you love the work of Klee."
145
332320
2978
你是克利迷, 你喜爱克利的画作。”
05:35
It was entirely bogus.
146
335298
1816
这完全是胡编的。
05:37
Their answers had nothing to do with Kandinsky or Klee.
147
337114
2290
这些青少年的答案和康定斯基 或者克利一点关系也没有。
05:39
They probably hadn't heard of the artists.
148
339404
2728
他们甚至还未听说过 这两位艺术家的大名。
05:42
He just arbitrarily divided them up.
149
342132
2740
泰吉弗尔只是武断地 把青少年们划分开来。
05:44
But what he found was, these categories mattered,
150
344872
3271
但他发现,这样的类别划分 是有作用的,
05:48
so when he later gave the subjects money,
151
348143
2511
随后,他让这些青少年分配钱,
05:50
they would prefer to give the money
152
350654
1676
他们更愿意将金钱给予
05:52
to members of their own group
153
352330
1800
他们本组的其他人
05:54
than members of the other group.
154
354130
1833
而不是另一个组别的人。
05:55
Worse, they were actually most interested
155
355963
2327
更糟的是,他们真的很乐于
05:58
in establishing a difference
156
358290
2006
建立一个不同来将
06:00
between their group and other groups,
157
360296
2566
自己的组和他组区分开来,
06:02
so they would give up money for their own group
158
362862
1908
为了令别组少拿到些钱
06:04
if by doing so they could give the other group even less.
159
364770
5248
他们甚至愿意放弃自己的钱。
06:10
This bias seems to show up very early.
160
370018
2218
偏爱很快就展现出来。
06:12
So my colleague and wife, Karen Wynn, at Yale
161
372236
2300
我的妻子也是我的同事,凯伦·维恩, 在耶鲁大学
06:14
has done a series of studies with babies
162
374536
1611
做了一系列有关婴儿的研究
06:16
where she exposes babies to puppets,
163
376147
2832
她将幼儿放在玩偶旁边,
06:18
and the puppets have certain food preferences.
164
378979
2265
玩偶有它们各自喜爱的食物。
06:21
So one of the puppets might like green beans.
165
381244
2182
某个玩偶可能喜爱青豆。
06:23
The other puppet might like graham crackers.
166
383426
2575
另个玩偶更爱全麦饼干。
06:26
They test the babies own food preferences,
167
386001
2369
研究人员测试了 幼儿们自身的食物偏好
06:28
and babies typically prefer the graham crackers.
168
388370
2690
幼儿们代表性地更爱全麦饼干。
06:31
But the question is, does this matter to babies
169
391060
2612
问题是,这样的喜好差别
06:33
in how they treat the puppets? And it matters a lot.
170
393672
3116
会影响到幼儿们对待玩偶的态度吗? 确实有很大影响。
06:36
They tend to prefer the puppet
171
396788
1519
幼儿们更喜欢
06:38
who has the same food tastes that they have,
172
398307
3479
和他们有相同口味偏好的玩具,
06:41
and worse, they actually prefer puppets
173
401786
2556
更糟的是,幼儿们喜欢那些惩罚
06:44
who punish the puppet with the different food taste.
174
404342
2985
拥有不同口味同伴的玩偶。
06:47
(Laughter)
175
407327
2277
(笑声)
06:49
We see this sort of in-group, out-group psychology all the time.
176
409604
3632
这样群体内外分别 非常常见。
06:53
We see it in political clashes
177
413236
1664
政治冲突中也会展现
06:54
within groups with different ideologies.
178
414900
2414
在持有不同意识形态的群体。
06:57
We see it in its extreme in cases of war,
179
417314
3626
极端的例子通过战争展现,
07:00
where the out-group isn't merely given less,
180
420940
3217
外群体不是被轻视
07:04
but dehumanized,
181
424157
1588
而是被不当作人类对待,
07:05
as in the Nazi perspective of Jews
182
425745
2240
如同纳粹视犹太人为
07:07
as vermin or lice,
183
427985
2085
害虫或是虱子,
07:10
or the American perspective of Japanese as rats.
184
430070
4236
美国人视日本人为老鼠。
07:14
Stereotypes can also go awry.
185
434306
2214
刻板印象是会歪曲现实的。
07:16
So often they're rational and useful,
186
436520
2261
因此时常他们是理性的,有帮助的,
07:18
but sometimes they're irrational,
187
438781
1574
但时常也会是非理性的,
07:20
they give the wrong answers,
188
440355
1226
会给出错误的答案,
07:21
and other times
189
441581
1217
有时
07:22
they lead to plainly immoral consequences.
190
442798
2175
会导致不道德的后果。
07:24
And the case that's been most studied
191
444973
2808
最常被研究的案例
07:27
is the case of race.
192
447781
1667
是种族。
07:29
There was a fascinating study
193
449448
1407
在 2008 年美国大选前
07:30
prior to the 2008 election
194
450855
2074
有个极好的研究,
07:32
where social psychologists looked at the extent
195
452929
3026
社会心理学家研究
07:35
to which the candidates were associated with America,
196
455955
3442
被测者们是如何通过 对美国国旗不知不觉的联系
07:39
as in an unconscious association with the American flag.
197
459397
3605
和美国联系在一起的。
07:43
And in one of their studies they compared
198
463002
1356
在其中一个研究中, 他们比较了
07:44
Obama and McCain, and they found McCain
199
464358
2014
奥巴马和麦凯恩, 他们发现麦凯恩
07:46
is thought of as more American than Obama,
200
466372
3394
比奥巴马更加“美国”,
07:49
and to some extent, people aren't that surprised by hearing that.
201
469766
2573
某种程度上, 人们甚至并未表示惊讶。
07:52
McCain is a celebrated war hero,
202
472339
1918
麦凯恩是一个著名的 战争英雄,
07:54
and many people would explicitly say
203
474257
1659
很多人明确地说道
07:55
he has more of an American story than Obama.
204
475916
2700
比起奥巴马, 麦凯恩有更多的美国故事。
07:58
But they also compared Obama
205
478616
1937
研究人员也比对了
08:00
to British Prime Minister Tony Blair,
206
480553
2516
奥巴马和英国首相布莱尔,
08:03
and they found that Blair was also thought of
207
483069
2261
他们发现比起奥巴马
08:05
as more American than Obama,
208
485330
2507
人们认为 布莱尔更加“美国”,
08:07
even though subjects explicitly understood
209
487837
2073
即使他们完全知晓
08:09
that he's not American at all.
210
489910
2990
布莱尔根本不是美国人。
08:12
But they were responding, of course,
211
492900
1424
但人们回应,当然,
08:14
to the color of his skin.
212
494324
3051
因为肤色的原因。
08:17
These stereotypes and biases
213
497375
2051
这样的成见和偏好
08:19
have real-world consequences,
214
499426
1450
有着现世的影响,
08:20
both subtle and very important.
215
500876
2872
这有些微妙,也非常重要。
08:23
In one recent study, researchers
216
503748
2662
在最近的一个研究中, 研究人员
08:26
put ads on eBay for the sale of baseball cards.
217
506410
3269
在易趣(eBay)网站上投放广告 销售篮球卡。
08:29
Some of them were held by white hands,
218
509679
2734
有些卖家是白人,
08:32
others by black hands.
219
512413
1218
另一些则是黑人。
08:33
They were the same baseball cards.
220
513631
1579
同样的实验也包括销售棒球卡。
08:35
The ones held by black hands
221
515210
1244
黑人卖家
08:36
got substantially smaller bids
222
516454
2067
得到的来自买家的出价 价位略小于
08:38
than the ones held by white hands.
223
518521
2484
白人卖家。
08:41
In research done at Stanford,
224
521005
2362
在斯坦福的一个研究项目也表明,
08:43
psychologists explored the case of people
225
523367
4230
心理学家研究了
08:47
sentenced for the murder of a white person.
226
527597
3569
因谋杀白人而被判刑的罪犯。
08:51
It turns out, holding everything else constant,
227
531166
2804
结果表明,除去其他因素,
08:53
you are considerably more likely to be executed
228
533970
2370
比起图片左边的人(白人)
08:56
if you look like the man on the right
229
536340
1777
图片右边的人(黑人)
08:58
than the man on the left,
230
538117
1973
更可能被判死刑,
09:00
and this is in large part because
231
540090
2029
这很大程度归结于
09:02
the man on the right looks more prototypically black,
232
542119
2534
图片右边的人是黑人,
09:04
more prototypically African-American,
233
544653
2630
美国黑人,
09:07
and this apparently influences people's decisions
234
547283
2049
很明显这影响到了人们
09:09
over what to do about him.
235
549332
1771
对他所做出的决定。
09:11
So now that we know about this,
236
551103
1547
现在我们知道了 成见和偏爱的存在,
09:12
how do we combat it?
237
552650
1657
我们要怎样对抗这样的想法呢?
09:14
And there are different avenues.
238
554307
1622
有很多不同的方法。
09:15
One avenue is to appeal
239
555929
1434
一个方法是去感化
09:17
to people's emotional responses,
240
557363
2046
人们的情感,
09:19
to appeal to people's empathy,
241
559409
2133
去令人们感同身受,
09:21
and we often do that through stories.
242
561542
1873
通常我们会用故事 来达到这样的效果。
09:23
So if you are a liberal parent
243
563415
2565
如果你是自由的父母
09:25
and you want to encourage your children
244
565980
1872
你想要鼓励你的孩子
09:27
to believe in the merits of nontraditional families,
245
567852
2374
来相信非传统家庭的价值优点,
09:30
you might give them a book like this. ["Heather Has Two Mommies"]
246
570226
2273
你会给他们看这样的书。 [海瑟有两个妈妈]
09:32
If you are conservative and have a different attitude,
247
572499
1726
如果你比较传统 对此持有不同的态度,
09:34
you might give them a book like this.
248
574225
1931
你会给他们看这本书
09:36
(Laughter) ["Help! Mom! There Are Liberals under My Bed!"]
249
576156
1749
(笑声) [“救命呀!妈妈!自由党人藏在我的床下!”]
09:37
But in general, stories can turn
250
577905
3336
总的来说,故事能够
09:41
anonymous strangers into people who matter,
251
581241
2232
让路人从陌生到关注,
09:43
and the idea that we care about people
252
583473
2685
我们在乎人们
09:46
when we focus on them as individuals
253
586158
1702
当我们将他们是做个体
09:47
is an idea which has shown up across history.
254
587860
2279
这样的思想贯穿人类历史。
09:50
So Stalin apocryphally said,
255
590139
2583
因此,斯大林虚情假意地说,
09:52
"A single death is a tragedy,
256
592722
1617
“一个人死亡是悲剧,
09:54
a million deaths is a statistic,"
257
594339
2040
一百万人的死亡则是统计数据,”
09:56
and Mother Teresa said,
258
596379
1451
特蕾莎修女说道,
09:57
"If I look at the mass, I will never act.
259
597830
1541
"假如我看到一群人, 我不会有所行动。
09:59
If I look at the one, I will."
260
599371
2325
假如我看到一个人,我会。"
10:01
Psychologists have explored this.
261
601696
2070
心理学家对此作出研究。
10:03
For instance, in one study,
262
603766
1301
比方说,在一个研究中,
10:05
people were given a list of facts about a crisis,
263
605067
2783
研究人员交给人们一张清单, 上面列举了一些危急的例子,
10:07
and it was seen how much they would donate
264
607850
4256
看人们愿意为了化解危机
10:12
to solve this crisis,
265
612106
1584
捐赠多少,
10:13
and another group was given no facts at all
266
613690
1837
另一个组则未被告知这些事情
10:15
but they were told of an individual
267
615527
2098
但研究人员告诉他们
10:17
and given a name and given a face,
268
617625
2440
个体故事,包括名字,相片,
10:20
and it turns out that they gave far more.
269
620065
3219
结果是,他们比上一组捐赠更多善款。
10:23
None of this I think is a secret
270
623284
1861
上述故事对于从事
10:25
to the people who are engaged in charity work.
271
625145
2111
慈善工作的人来说都不是秘密。
10:27
People don't tend to deluge people
272
627256
2648
慈善工作者不会向人们
10:29
with facts and statistics.
273
629904
1323
展示大量的事实和数据。
10:31
Rather, you show them faces,
274
631227
1022
而是,给人们看相片,
10:32
you show them people.
275
632249
1736
展示灾民的样子。
10:33
It's possible that by extending our sympathies
276
633985
3227
很有可能的是,通过展现我们
10:37
to an individual, they can spread
277
637212
1971
对于个体的同情,
10:39
to the group that the individual belongs to.
278
639183
2878
他们可以进而展示个体从属的群体。
10:42
This is Harriet Beecher Stowe.
279
642061
2466
这是哈里耶持·比彻·斯托。
10:44
The story, perhaps apocryphal,
280
644527
2443
故事,可能是假的,
10:46
is that President Lincoln invited her
281
646970
2074
林肯总统邀请她
10:49
to the White House in the middle of the Civil War
282
649044
1998
在美国内战期间到白宫
10:51
and said to her,
283
651042
1584
对她说,
10:52
"So you're the little lady who started this great war."
284
652626
2664
“你是开始这场战争的女子。”
10:55
And he was talking about "Uncle Tom's Cabin."
285
655290
1885
他谈到“汤姆叔叔的小屋。”
10:57
"Uncle Tom's Cabin" is not a great book of philosophy
286
657175
2531
“汤姆叔叔的小屋”不是 伟大的哲学或宗教故事
10:59
or of theology or perhaps not even literature,
287
659706
3144
甚至可能都不是文学,
11:02
but it does a great job
288
662850
2515
但它起了很大的作用
11:05
of getting people to put themselves in the shoes
289
665365
2498
在人们能够将自己置身于某个故事
11:07
of people they wouldn't otherwise be in the shoes of,
290
667863
2333
那些本不可能属于他们的故事中,
11:10
put themselves in the shoes of slaves.
291
670196
2402
以奴隶的角度来看世界。
11:12
And that could well have been a catalyst
292
672598
1781
这些是催化剂,
11:14
for great social change.
293
674379
1604
催生巨大的社会变革。
11:15
More recently, looking at America
294
675983
2362
近年来,看看美国
11:18
in the last several decades,
295
678345
3069
在过去几十年的表现,
11:21
there's some reason to believe that shows like "The Cosby Show"
296
681414
3149
太多的原因让我们相信 像是“考斯比一家”
11:24
radically changed American attitudes towards African-Americans,
297
684563
2688
从更本上改变了美国人 对美国黑人的看法,
11:27
while shows like "Will and Grace" and "Modern Family"
298
687251
2983
“威尔与格蕾丝”,“摩登家庭”
11:30
changed American attitudes
299
690234
1363
改变了很多美国人
11:31
towards gay men and women.
300
691597
1300
对同性恋男女的态度。
11:32
I don't think it's an exaggeration to say
301
692897
2455
不夸张地说,
11:35
that the major catalyst in America for moral change
302
695352
2661
对美国道德价值改变 做出最大贡献的
11:38
has been a situation comedy.
303
698013
2893
是情景喜剧。
11:40
But it's not all emotions,
304
700906
1416
但这并不全是情感上的,
11:42
and I want to end by appealing
305
702322
1276
最后我想要谈到
11:43
to the power of reason.
306
703598
2235
理性的力量。
11:45
At some point in his wonderful book
307
705833
2156
在他佳作的某些部分
11:47
"The Better Angels of Our Nature,"
308
707989
1223
“唤醒人性中的天使”
11:49
Steven Pinker says,
309
709212
2016
史蒂文·平克说道,
11:51
the Old Testament says love thy neighbor,
310
711228
2582
旧约说到要爱我们的邻居,
11:53
and the New Testament says love thy enemy,
311
713810
2722
新约说到要爱我们的敌人,
11:56
but I don't love either one of them, not really,
312
716532
2686
但我不爱他们中的任何一个,不尽然,
11:59
but I don't want to kill them.
313
719218
1667
但我不想杀了他们。
12:00
I know I have obligations to them,
314
720885
1866
我知道我有义务对他们,
12:02
but my moral feelings to them, my moral beliefs
315
722751
3470
但我对他们的道德感受,
12:06
about how I should behave towards them,
316
726221
1713
我要如何对待他们的道德信念,
12:07
aren't grounded in love.
317
727934
2047
不会是基于爱。
12:09
What they're grounded in is the understanding of human rights,
318
729981
1939
是基于对人权的理解,
12:11
a belief that their life is as valuable to them
319
731920
2223
他们的生命对他们的价值
12:14
as my life is to me,
320
734143
2356
正如我的生命对我的价值,
12:16
and to support this, he tells a story
321
736499
1932
为了支持这个观点, 他讲了一个故事,
12:18
by the great philosopher Adam Smith,
322
738431
1848
关于伟大的哲人亚当·斯密,
12:20
and I want to tell this story too,
323
740279
1686
我现在讲这个故事,
12:21
though I'm going to modify it a little bit
324
741965
1296
我为了使其适应现代
12:23
for modern times.
325
743261
1678
略微做了改动。
12:24
So Adam Smith starts by asking you to imagine
326
744939
1901
亚当斯密让你来想象
12:26
the death of thousands of people,
327
746840
1901
数千人死亡的场景,
12:28
and imagine that the thousands of people
328
748741
2040
想象这数千人
12:30
are in a country you are not familiar with.
329
750781
2239
是在你不熟悉的国家。
12:33
It could be China or India or a country in Africa.
330
753020
3554
可能是中国,或者是印度, 或者是某个非洲国家。
12:36
And Smith says, how would you respond?
331
756574
2484
斯密说到,你会怎样回应?
12:39
And you would say, well that's too bad,
332
759058
2307
你可能会说,这太糟了,
12:41
and you'd go on to the rest of your life.
333
761365
1876
然后继续你的生活。
12:43
If you were to open up The New York Times online or something,
334
763241
2219
如果你打开纽约时报的网站或什么,
12:45
and discover this, and in fact this happens to us all the time,
335
765460
2960
看到这些消息,事实上 这常发生,
12:48
we go about our lives.
336
768420
1521
我们继续我们的生活。
12:49
But imagine instead, Smith says,
337
769941
2194
斯密说,想象另一个画面:
12:52
you were to learn that tomorrow
338
772135
1254
你发现明天
12:53
you were to have your little finger chopped off.
339
773389
2539
你的小手指会被砍掉。
12:55
Smith says, that would matter a lot.
340
775928
2169
斯密说,这太重要了。
12:58
You would not sleep that night
341
778097
1411
你整晚会睡不着觉
12:59
wondering about that.
342
779508
1353
辗转反侧。
13:00
So this raises the question:
343
780861
2019
这就提出了问题:
13:02
Would you sacrifice thousands of lives
344
782880
2466
你会牺牲数千人的生命
13:05
to save your little finger?
345
785346
1969
以求得保全自己小手指吗?
13:07
Now answer this in the privacy of your own head,
346
787315
2318
现在在自己的脑袋里 回答这个问题,
13:09
but Smith says, absolutely not,
347
789633
2919
但是斯密说,绝对不,
13:12
what a horrid thought.
348
792552
1692
这是多么邪恶的想法。
13:14
And so this raises the question,
349
794244
2031
这就提出了问题,
13:16
and so, as Smith puts it,
350
796275
1374
随后,斯密提出这样的疑问,
13:17
"When our passive feelings are almost always
351
797649
2218
“我们的消极情绪总是
13:19
so sordid and so selfish,
352
799867
1448
如此利欲熏心,自私卑鄙,
13:21
how comes it that our active principles
353
801315
1465
我们的行为又怎么可能
13:22
should often be so generous and so noble?"
354
802780
2533
时常很无私和高尚呢?”
13:25
And Smith's answer is, "It is reason,
355
805313
2050
斯密回答道,“因为理性,
13:27
principle, conscience.
356
807363
1775
道德,良知。”
13:29
[This] calls to us,
357
809138
1541
[这] 告诉我们
13:30
with a voice capable of astonishing the most presumptuous of our passions,
358
810679
3425
能够惊人地绝大部分 爆发我们的激情,
13:34
that we are but one of the multitude,
359
814104
1677
但众多思考中,
13:35
in no respect better than any other in it."
360
815781
2441
没有比尊重更重要。“
13:38
And this last part is what is often described
361
818222
2125
最后的这个部分是有关
13:40
as the principle of impartiality.
362
820347
3208
公正的原则。
13:43
And this principle of impartiality manifests itself
363
823555
2629
这样公正的原则
13:46
in all of the world's religions,
364
826184
1747
在全世界宗教中都有所证明,
13:47
in all of the different versions of the golden rule,
365
827951
2258
在各种不同版本的黄金法则,
13:50
and in all of the world's moral philosophies,
366
830209
2454
世界上所有的道德哲学,
13:52
which differ in many ways
367
832663
1307
即使有所不同
13:53
but share the presupposition that we should judge morality
368
833970
2994
但共有的假设是 我们应该从
13:56
from sort of an impartial point of view.
369
836964
2985
公正的角度来评判道德。
13:59
The best articulation of this view
370
839949
1822
这观点最重要的是
14:01
is actually, for me, it's not from a theologian or from a philosopher,
371
841771
3085
事实上,对我来说, 这不是从宗教学家或哲学家听来,
14:04
but from Humphrey Bogart
372
844856
1357
而是从亨弗莱·鲍嘉
14:06
at the end of "Casablanca."
373
846213
1547
在电影“卡萨布兰卡”片尾的表现。
14:07
So, spoiler alert, he's telling his lover
374
847760
3776
警告有剧透,他告诉他的爱人
14:11
that they have to separate
375
851536
1140
为了更伟大的善,
14:12
for the more general good,
376
852676
1593
他们必须要分开,
14:14
and he says to her, and I won't do the accent,
377
854269
1864
他对她说,我不会模仿这口音,
14:16
but he says to her, "It doesn't take much to see
378
856133
1782
他对她说“不用多久就可以看到
14:17
that the problems of three little people
379
857915
1359
这三个小人的问题
14:19
don't amount to a hill of beans in this crazy world."
380
859274
3111
不会使世界瘋狂。”
14:22
Our reason could cause us to override our passions.
381
862385
3280
我们的理性可以驾驭我们的热情。
14:25
Our reason could motivate us
382
865665
1716
我们的理性可以激励我们
14:27
to extend our empathy,
383
867381
1221
扩展我们的同理心,
14:28
could motivate us to write a book like "Uncle Tom's Cabin,"
384
868602
2327
可以激励我们写 “汤姆叔叔的小屋”这样的书
14:30
or read a book like "Uncle Tom's Cabin,"
385
870929
1723
或者看“汤姆叔叔的小屋”,
14:32
and our reason can motivate us to create
386
872652
2694
我们的理性可以促使我们
14:35
customs and taboos and laws
387
875346
1962
创造海关,烟草和法律
14:37
that will constrain us
388
877308
1810
这会限制我们
14:39
from acting upon our impulses
389
879118
1676
冲动的行为,
14:40
when, as rational beings, we feel
390
880794
1589
当理性存在,我们感到
14:42
we should be constrained.
391
882383
1395
我们要被限制。
14:43
This is what a constitution is.
392
883778
2013
这是宪法。
14:45
A constitution is something which was set up in the past
393
885791
2921
宪法是过去撰写的
14:48
that applies now in the present,
394
888712
1307
适用于现在以及未来,
14:50
and what it says is,
395
890019
985
宪法提到的,
14:51
no matter how much we might to reelect
396
891004
2227
无论我们多想
14:53
a popular president for a third term,
397
893231
2603
选举受欢迎的总统 开始第三任期,
14:55
no matter how much white Americans might choose
398
895834
2095
无论美国白人多么想
14:57
to feel that they want to reinstate the institution of slavery, we can't.
399
897929
4068
重新回到奴隶制度,我们不能。
15:01
We have bound ourselves.
400
901997
1676
我们限制自己。
15:03
And we bind ourselves in other ways as well.
401
903673
2417
我们也从别的方式约束自己。
15:06
We know that when it comes to choosing somebody
402
906090
2758
当我们想要选择某人
15:08
for a job, for an award,
403
908848
2951
来从事一项工作,获得一个奖项,
15:11
we are strongly biased by their race,
404
911799
2958
我们很容易受到种族因素的影响,
15:14
we are biased by their gender,
405
914757
2296
我们会因他们的性别产生偏见,
15:17
we are biased by how attractive they are,
406
917053
2215
我们会因为他们的样貌产生偏爱,
15:19
and sometimes we might say, "Well fine, that's the way it should be."
407
919268
2651
有时我们会说, “是的,就是这样。”
15:21
But other times we say, "This is wrong."
408
921919
2307
有时我们会说,“这是错的。”
15:24
And so to combat this,
409
924226
1889
为了对抗这些,
15:26
we don't just try harder,
410
926115
2251
我们不仅更加努力,
15:28
but rather what we do is we set up situations
411
928366
3001
我们建立机构
15:31
where these other sources of information can't bias us,
412
931367
3039
这些信息资源不会有成见,
15:34
which is why many orchestras
413
934406
1315
这就是为什么很多交响乐团
15:35
audition musicians behind screens,
414
935721
2645
面试音乐家时, 让他们站在幕后,
15:38
so the only information they have
415
938366
1244
这样评委唯一的信息来源
15:39
is the information they believe should matter.
416
939610
2693
就是他们认为最重要的。
15:42
I think prejudice and bias
417
942303
2323
我认为偏见和偏爱
15:44
illustrate a fundamental duality of human nature.
418
944626
3094
展示了人性最基础的二元性。
15:47
We have gut feelings, instincts, emotions,
419
947720
3776
我们有胆识,本能,情感,
15:51
and they affect our judgments and our actions
420
951496
2161
这会影响我们
15:53
for good and for evil,
421
953657
2331
对于好与坏的判断和行为,
15:55
but we are also capable of rational deliberation
422
955988
3622
但我们同样有能力做出理性思考
15:59
and intelligent planning,
423
959610
1435
和智能规划,
16:01
and we can use these to, in some cases,
424
961045
2817
我们可以运用这些,在某些情况下,
16:03
accelerate and nourish our emotions,
425
963862
1943
加速和丰富我们的情绪,
16:05
and in other cases staunch them.
426
965805
2768
某些情况下止住它们。
16:08
And it's in this way
427
968573
1234
这样成见和偏爱
16:09
that reason helps us create a better world.
428
969807
2767
就能帮助我们创建更美好的世界。
16:12
Thank you.
429
972574
2344
谢谢。
16:14
(Applause)
430
974918
3705
(鼓掌)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隐私政策

eng.lish.video

Developer's Blog