What are the most important moral problems of our time? | Will MacAskill

377,590 views ・ 2018-10-03

TED


请双击下面的英文字幕来播放视频。

翻译人员: duan JiGang 校对人员: Yiyang Piao
00:12
This is a graph
0
12857
1479
这是一张曲线图
00:14
that represents the economic history of human civilization.
1
14360
3659
它展现了人类文明的经济史.
00:18
[World GDP per capita over the last 200,000 years]
2
18043
2400
(过去20万年的世界人均GDP)
00:23
There's not much going on, is there.
3
23757
2003
没什么在发生,不是吗。
00:26
For the vast majority of human history,
4
26751
2347
在人类历史的长河中,
00:29
pretty much everyone lived on the equivalent of one dollar per day,
5
29835
4048
几乎每人每天用一美元来生活,
00:33
and not much changed.
6
33907
1286
并且没有什么变化。
00:36
But then, something extraordinary happened:
7
36757
2777
但是后来,发生了一个重大的事件:
00:40
the Scientific and Industrial Revolutions.
8
40677
2811
科学和工业革命。
00:43
And the basically flat graph you just saw
9
43512
2785
刚才你们所看到的平缓的曲线图
00:46
transforms into this.
10
46321
2675
变成了这样。
00:50
What this graph means is that, in terms of power to change the world,
11
50612
4635
这幅图表现的是, 就改变世界的能力而言,
00:55
we live in an unprecedented time in human history,
12
55271
3438
我们现在所处的时代是空前的,
00:58
and I believe our ethical understanding hasn't yet caught up with this fact.
13
58733
3944
并且我认为我们伦理上的理解 还没有赶上这一事实。
01:03
The Scientific and Industrial Revolutions
14
63716
1984
科学和工业革命
01:05
transformed both our understanding of the world
15
65724
2909
不光改变了我们对世界的认知
01:08
and our ability to alter it.
16
68657
1669
也提高了我们改变世界的能力。
01:11
What we need is an ethical revolution
17
71505
3667
我们所需要的 是一场伦理道德层面上的革命
01:15
so that we can work out
18
75196
1548
这样我们才能解决
01:16
how do we use this tremendous bounty of resources
19
76768
3152
怎么利用这些丰富的资源
01:19
to improve the world.
20
79944
1395
让世界变得更美好。
01:22
For the last 10 years,
21
82249
1591
在过去的10年中,
01:23
my colleagues and I have developed a philosophy and research program
22
83864
3833
我和同事开展了一项哲学研究计划,
01:27
that we call effective altruism.
23
87721
1835
我们称之为有效的利他主义。
01:30
It tries to respond to these radical changes in our world,
24
90366
3595
这项研究计划尝试去解决 当今世界所发生的巨变,
01:33
uses evidence and careful reasoning to try to answer this question:
25
93985
4476
依据事实证据和缜密的推理 来回答这一问题:
01:40
How can we do the most good?
26
100173
2278
我们怎样才可以 最大效益得做好事呢?
01:44
Now, there are many issues you've got to address
27
104265
3221
现今,有许多待解决的问题
01:47
if you want to tackle this problem:
28
107510
2263
如果你想要解决这个大难题,
01:49
whether to do good through your charity
29
109797
2031
无论你是通过做慈善
01:51
or your career or your political engagement,
30
111852
2152
还是通过你的职业, 或是通过政治活动,
01:54
what programs to focus on, who to work with.
31
114028
2395
你专注的项目, 或你工作的对象来为世界得益。
01:57
But what I want to talk about
32
117624
1476
但是我想讨论的是
01:59
is what I think is the most fundamental problem.
33
119124
2872
我认为在所有问题之中, 最根本的问题
02:02
Of all the many problems that the world faces,
34
122020
2693
这个世界所面临的所有问题中
02:05
which should we be focused on trying to solve first?
35
125962
2659
哪一个应是我们首先解决的问题?
02:10
Now, I'm going to give you a framework for thinking about this question,
36
130668
3468
现在,我将要给出 一个框架来思考这个问题,
02:14
and the framework is very simple.
37
134160
1936
这个框架很简单。
02:16
A problem's higher priority,
38
136842
1699
一个问题的优先级越高,
02:19
the bigger, the more easily solvable and the more neglected it is.
39
139416
4063
这问题就越大,就越容易解决, 也最容易被忽略。
02:24
Bigger is better,
40
144694
1642
越大越好。
02:26
because we've got more to gain if we do solve the problem.
41
146360
2841
因为我们若解决了这个问题, 我们会获利极多。
02:30
More easily solvable is better
42
150221
1569
越容易解决越好,
02:31
because I can solve the problem with less time or money.
43
151814
2824
因为我可以花费更少的钱和 时间来解决这个问题。
02:35
And most subtly,
44
155737
2063
最微妙的是,
02:38
more neglected is better, because of diminishing returns.
45
158681
2849
根据收益递减规律, 问题越容易被忽视越好。
02:42
The more resources that have already been invested into solving a problem,
46
162285
3714
若前期已投入资源过多, 来解决这个问题
02:46
the harder it will be to make additional progress.
47
166023
2905
就越难取得额外的进展。
02:50
Now, the key thing that I want to leave with you is this framework,
48
170560
4059
现在,我想要留给你们最关键的东西 就是这个框架,
02:54
so that you can think for yourself
49
174643
1984
这样你可以自己去思考
02:56
what are the highest global priorities.
50
176651
2321
什么问题具有全球最高的优先级。
02:59
But I and others in the effective altruism community
51
179954
2692
但是在这个高效利他主义群体中, 我和其他的人
03:02
have converged on three moral issues that we believe are unusually important,
52
182670
5879
一致认为有三个的道德问题
03:08
score unusually well in this framework.
53
188573
2182
在这个框架中占据重要的地位。
03:11
First is global health.
54
191151
2813
第一个是全球卫生。
03:13
This is supersolvable.
55
193988
2411
这是最易解决的。
03:16
We have an amazing track record in global health.
56
196423
3397
我们对于全球卫生 有一个惊人的跟踪记录。
03:19
Rates of death from measles, malaria, diarrheal disease
57
199844
5420
麻疹,疟疾和痢疾引起的死亡率
03:25
are down by over 70 percent.
58
205288
2246
已经降低了70%。
03:29
And in 1980, we eradicated smallpox.
59
209534
2789
并且在1980年, 我们根治了天花。
03:33
I estimate we thereby saved over 60 million lives.
60
213815
3667
我估计因此而拯救了 超过6000万条的生命。
03:37
That's more lives saved than if we'd achieved world peace
61
217506
3064
这些生命数量比 若在同时期取得世界和平
03:40
in that same time period.
62
220594
1697
拯救的生命还要多。
03:43
On our current best estimates,
63
223893
2325
根据我们目前的最佳估计,
03:46
we can save a life by distributing long-lasting insecticide-treated bed nets
64
226242
4128
我们可以通过分配长期有效的、 具有杀虫功效的蚊帐来拯救生命,
03:50
for just a few thousand dollars.
65
230394
1903
而这些蚊帐只花费几千元。
03:52
This is an amazing opportunity.
66
232911
1667
这是一个惊人的机会。
03:55
The second big priority is factory farming.
67
235594
2515
第二优先问题就是农业工厂化。
03:58
This is superneglected.
68
238681
1563
这一点被严重忽视了。
04:00
There are 50 billion land animals used every year for food,
69
240768
4143
每年有500亿陆生动物 被作为食物。
04:05
and the vast majority of them are factory farmed,
70
245625
2548
它们中的大部分是工厂化养殖的,
04:08
living in conditions of horrific suffering.
71
248197
2380
这些动物居住在痛苦的生存环境中。
04:10
They're probably among the worst-off creatures on this planet,
72
250601
3151
它们可能是这个地球上最惨的生物,
04:13
and in many cases, we could significantly improve their lives
73
253776
2858
并且在很多情况下, 我们可以大幅度改善它们的生活,
04:16
for just pennies per animal.
74
256658
1602
只需为每一个动物花费几分钱 就可以做到。
04:19
Yet this is hugely neglected.
75
259123
2082
然而这一点也被极度忽略了。
04:21
There are 3,000 times more animals in factory farms
76
261229
3810
工业化农场里的动物数量是
流浪动物的3000倍多,
04:25
than there are stray pets,
77
265063
1561
04:28
but yet, factory farming gets one fiftieth of the philanthropic funding.
78
268600
4373
但是工业化养殖的动物 只获得了1/5的慈善基金。
04:34
That means additional resources in this area
79
274211
2128
这意味着,在这个领域若有更多资源
04:36
could have a truly transformative impact.
80
276363
2150
可以产生一个真正的变革性影响。
04:39
Now the third area is the one that I want to focus on the most,
81
279458
2985
接下来第三个领域是我最想强调的。
04:42
and that's the category of existential risks:
82
282467
2984
这就是存亡风险:
04:45
events like a nuclear war or a global pandemic
83
285475
3873
一些像核战争或者全球性瘟疫的事件
04:50
that could permanently derail civilization
84
290824
2611
这些可能永久地使文明脱节,
04:54
or even lead to the extinction of the human race.
85
294156
2436
或者甚至导致人类种族的灭绝。
04:57
Let me explain why I think this is such a big priority
86
297882
2540
让我来解释下为什么我认为这件事情
05:00
in terms of this framework.
87
300446
1337
是这个框架下最具优先级的事情。
05:02
First, size.
88
302992
1412
第一,规模。
05:05
How bad would it be if there were a truly existential catastrophe?
89
305341
3889
如果出现一个毁灭性的灾难, 情况将会变得多糟糕?
05:10
Well, it would involve the deaths of all seven billion people on this planet
90
310920
6342
它将牵扯到这个星球上 70亿人类的生死,
05:17
and that means you and everyone you know and love.
91
317286
3119
当然包括你自己以及 你认识或者深爱的任何人。
05:21
That's just a tragedy of unimaginable size.
92
321214
2579
这将是一个不可想象规模的悲剧。
05:25
But then, what's more,
93
325684
1976
但是然后,还有呢。
05:27
it would also mean the curtailment of humanity's future potential,
94
327684
3605
它同时预示着人类未来潜力的缩减,
05:31
and I believe that humanity's potential is vast.
95
331313
2952
我相信人类的潜力是巨大的。
05:35
The human race has been around for about 200,000 years,
96
335551
3451
人类已经存在了大约20万年了,
05:39
and if she lives as long as a typical mammalian species,
97
339026
2883
如果她像一个典型的哺乳物种 活得一样长的话,
05:41
she would last for about two million years.
98
341933
2298
她将会持续200万年。
05:46
If the human race were a single individual,
99
346884
2691
如果人类是一个单一的个体,
05:49
she would be just 10 years old today.
100
349599
2419
那么她今天就只有10岁。
05:53
And what's more, the human race isn't a typical mammalian species.
101
353526
4166
更何况,人类不是 一个典型的哺乳动物,
05:58
There's no reason why, if we're careful,
102
358950
1906
没有理由,如果我们细心的话,
06:00
we should die off after only two million years.
103
360880
2205
我们将会在200万年后灭亡。
06:03
The earth will remain habitable for 500 million years to come.
104
363839
4040
而地球在未来的5亿年 仍将保持可居住状态。
06:08
And if someday, we took to the stars,
105
368696
1944
如果有一天,我们登上了星球,
06:11
the civilization could continue for billions more.
106
371640
2516
那么文明将会继续数十亿年。
06:16
So I think the future is going to be really big,
107
376193
2388
所以,我认为未来将会真的很宽广,
06:19
but is it going to be good?
108
379669
1802
但是未来一定会更好吗?
06:21
Is the human race even really worth preserving?
109
381495
2817
人类种族真的还值得保留吗?
06:26
Well, we hear all the time about how things have been getting worse,
110
386540
3929
是的,我们一直都在 听到事情如何变得更糟的消息。
06:31
but I think that when we take the long run,
111
391459
2693
但是我相信当我们做长远打算时,
06:34
things have been getting radically better.
112
394176
2031
情况已经大为好转。
06:37
Here, for example, is life expectancy over time.
113
397453
2294
比如,人类平均寿命在延长。
06:40
Here's the proportion of people not living in extreme poverty.
114
400892
3023
很大比例的人不再生活异常贫困。
06:45
Here's the number of countries over time that have decriminalized homosexuality.
115
405106
4095
将同性恋合法化的国家数量在增加。
06:50
Here's the number of countries over time that have become democratic.
116
410848
3269
民主化的国家数量也在增加。
06:55
Then, when we look to the future, there could be so much more to gain again.
117
415015
4619
所以,当展望未来时, 我们可以获得更多东西。
06:59
We'll be so much richer,
118
419658
1228
我们将变得更富有,
07:00
we can solve so many problems that are intractable today.
119
420910
3595
我们可以解决很多 如今看来很棘手的问题。
07:05
So if this is kind of a graph of how humanity has progressed
120
425389
4445
所以,如果这是一个 人类进步曲线图的话,
07:09
in terms of total human flourishing over time,
121
429858
2890
它基于人类随着时间的推移 如何繁荣昌盛,
07:12
well, this is what we would expect future progress to look like.
122
432772
3355
这将会是未来我们期望看到的进展。
07:16
It's vast.
123
436881
1150
它很广阔。
07:18
Here, for example,
124
438953
1198
比如。
07:20
is where we would expect no one to live in extreme poverty.
125
440175
3746
在这个点,我们期望没人会生活在 极度的贫困中。
07:25
Here is where we would expect everyone to be better off
126
445930
3202
在个点,我们期待每个人都会比
07:29
than the richest person alive today.
127
449156
1853
比当今最富裕的人过的还好。
07:32
Perhaps here is where we would discover the fundamental natural laws
128
452081
3192
或许在这里,我们将会发现统治世界
07:35
that govern our world.
129
455297
1268
的基本自然法则。
07:37
Perhaps here is where we discover an entirely new form of art,
130
457516
3705
或许在这个点,我们将会发现 一个全新的艺术形式,
07:41
a form of music we currently lack the ears to hear.
131
461245
3040
或者是一种当下我们没有能力 去倾听的音乐。
07:45
And this is just the next few thousand years.
132
465072
2222
这仅仅是未来的几千年。
07:47
Once we think past that,
133
467827
2205
每当我们回首过往时,
07:50
well, we can't even imagine the heights that human accomplishment might reach.
134
470056
4167
我们甚至都不敢相信 人类曾经取得的高度。
07:54
So the future could be very big and it could be very good,
135
474247
3040
所以,前景可能会很广阔, 也可能会很好,
07:57
but are there ways we could lose this value?
136
477311
2086
但是,是否会有一些情况, 会使得我们丧失这些价值?
08:00
And sadly, I think there are.
137
480366
1826
悲伤地是,我认为是他们是存在的。
08:02
The last two centuries brought tremendous technological progress,
138
482216
4053
过去两个世纪带来了 巨大的技术进步,
08:06
but they also brought the global risks of nuclear war
139
486293
2622
但是技术进步同时带来了 核战争的全球危机
08:08
and the possibility of extreme climate change.
140
488939
2157
还有极端气候变化的可能性。
08:11
When we look to the coming centuries,
141
491725
1767
当我们观望未来的世纪时,
08:13
we should expect to see the same pattern again.
142
493516
2647
我们应该期望看到同样的样式。
08:16
And we can see some radically powerful technologies on the horizon.
143
496187
3356
我们能够看到未来会有的 一些激进强大的技术。
08:20
Synthetic biology might give us the power to create viruses
144
500132
2849
合成生物学使得我们能够创病毒,
08:23
of unprecedented contagiousness and lethality.
145
503005
3047
而这些病毒具有不可预测的传染性和 杀伤力。
08:27
Geoengineering might give us the power to dramatically alter the earth's climate.
146
507131
4643
地质工程可能给予我们 极大改变气候的力量。
08:31
Artificial intelligence might give us the power to create intelligent agents
147
511798
4199
人工智能可能会给予我们 创造智能终端的力量,
08:36
with abilities greater than our own.
148
516021
2142
而这些智能终端会比人类 具备更强的能力。
08:40
Now, I'm not saying that any of these risks are particularly likely,
149
520222
3888
现在,我并不是说所有这些风险 都特别的有可能实现,
08:44
but when there's so much at stake,
150
524134
1644
但当这一切都岌岌可危时。
08:45
even small probabilities matter a great deal.
151
525802
2967
甚至一个小的可能性 也会造成大的惨剧。
08:49
Imagine if you're getting on a plane and you're kind of nervous,
152
529568
3001
想象你登上一个飞机,你有点紧张,
08:52
and the pilot reassures you by saying,
153
532593
3444
并且飞行员在欢迎你时说到,
08:56
"There's only a one-in-a-thousand chance of crashing. Don't worry."
154
536061
4634
“飞机只有千分之一的几率会坠毁, 别担心。”
09:02
Would you feel reassured?
155
542157
1554
你会感觉安心吗?
09:04
For these reasons, I think that preserving the future of humanity
156
544509
4088
基于这些原因, 我认为保留人类的未来这件事情
09:08
is among the most important problems that we currently face.
157
548621
2984
是我们目前面临的最重要的难题。
09:12
But let's keep using this framework.
158
552546
2150
然而,让我们继续来使用这个框架。
09:14
Is this problem neglected?
159
554720
1310
这个问题被忽略了吗?
09:18
And I think the answer is yes,
160
558085
2282
并且我认为答案是 “是的”,
09:20
and that's because problems that affect future generations
161
560391
3325
这是因为影响人类后代的问题
09:23
are often hugely neglected.
162
563740
1651
通常都被极大地忽略了。
09:26
Why?
163
566930
1406
为什么?
09:28
Because future people don't participate in markets today.
164
568360
3478
因为未来人类没有参与今天的市场。
09:31
They don't have a vote.
165
571862
1522
他们没有任何决策权。
09:33
It's not like there's a lobby representing the interests
166
573931
2673
并没有一个大厅来展示他们的利益,
09:36
of those born in 2300 AD.
167
576628
2023
我指的是出生在公元2300年 这些人的利益。
09:40
They don't get to influence the decisions we make today.
168
580313
3242
他们没法影响我们今天所做的决定。
09:43
They're voiceless.
169
583995
1191
他们是无声的。
09:46
And that means we still spend a paltry amount on these issues:
170
586490
3445
这意味着我们在这些问题上的投入 微乎其微:
09:49
nuclear nonproliferation,
171
589959
1799
防止核扩散,
09:51
geoengineering, biorisk,
172
591782
2330
地质工程,生物风险,
09:55
artificial intelligence safety.
173
595414
1642
还有人工智能的安全性。
09:57
All of these receive only a few tens of millions of dollars
174
597923
2874
所有这些课题, 每年只能接受大约千万美金
10:00
of philanthropic funding every year.
175
600821
1927
的慈善基金。
10:04
That's tiny compared to the 390 billion dollars
176
604044
3929
这些和美国每年花费的
10:08
that's spent on US philanthropy in total.
177
608790
2261
3900亿慈善基金比起来 简直就是九牛一毛。
10:13
The final aspect of our framework then:
178
613885
2484
我们这个框架的最后一个方面:
10:17
Is this solvable?
179
617083
1190
这个问题可以解决吗?
10:19
I believe it is.
180
619289
1289
我相信是可以解决的。
10:21
You can contribute with your money,
181
621014
3047
你可以用你的钱来做贡献,
10:24
your career or your political engagement.
182
624085
2644
或者通过你的职业, 或者你的政治参与来贡献。
10:28
With your money, you can support organizations
183
628225
2175
使用你的钱,你可以支持那些
10:30
that focus on these risks,
184
630424
1302
聚焦在这些风险上的组织机构,
10:31
like the Nuclear Threat Initiative,
185
631750
2555
比如核威胁倡议组织,
10:34
which campaigns to take nuclear weapons off hair-trigger alert,
186
634329
3660
这个组织旨在将核武器 从一触即发的警告中消除,
10:38
or the Blue Ribbon Panel, which develops policy to minimize the damage
187
638013
3571
或者蓝丝带小组, 这个组织旨在开发政策来最小化
10:41
from natural and man-made pandemics,
188
641608
2095
自然或者人为流行病带来的伤害,
10:45
or the Center for Human-Compatible AI, which does technical research
189
645158
3260
或者人类兼容人工智能中心, 这个组织做一些研究
10:48
to ensure that AI systems are safe and reliable.
190
648442
2747
以确保人工智能系统是安全 并且可靠的。
10:52
With your political engagement,
191
652652
1516
通过你的政治参与活动
10:54
you can vote for candidates that care about these risks,
192
654192
3096
你可以投票给 那些关心这些风险的候选人,
10:57
and you can support greater international cooperation.
193
657312
2586
你也可以支持更大的国际化合作。
11:01
And then with your career, there is so much that you can do.
194
661767
3542
然后,通过你的职业, 有很多事情你可以做。
11:05
Of course, we need scientists and policymakers and organization leaders,
195
665333
3672
毫无疑问,我们需要科学家, 需要政策制定者和组织领导者,
11:09
but just as importantly,
196
669865
1152
但是同样重要的是,
11:11
we also need accountants and managers and assistants
197
671041
4117
我们也需要财会,经理人和助理,
11:16
to work in these organizations that are tackling these problems.
198
676691
3754
为解决这些难题的组织工作。
11:20
Now, the research program of effective altruism
199
680469
3492
当前,针对有效利他主义的研究程序
11:25
is still in its infancy,
200
685191
1444
还处于幼儿期。
11:27
and there's still a huge amount that we don't know.
201
687262
2524
我们对很多东西仍不了解。
11:31
But even with what we've learned so far,
202
691173
2343
但是,仅仅就目前我们所学到的东西 而言,
11:34
we can see that by thinking carefully
203
694748
2183
我们可以发现,通过认真思考
11:37
and by focusing on those problems that are big, solvable and neglected,
204
697494
4873
并且聚焦在这些巨大, 可解决并且被忽略的问题上,
11:43
we can make a truly tremendous difference to the world
205
703152
2708
我们可以对世界作出 真正意义上巨大的改变,
11:45
for thousands of years to come.
206
705884
1631
这些改变是为了未来的几千年。
11:47
Thank you.
207
707963
1151
谢谢。
11:49
(Applause)
208
709138
4560
(掌声)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隐私政策

eng.lish.video

Developer's Blog