How you can help transform the internet into a place of trust | Claire Wardle

54,824 views ・ 2019-11-15

TED


请双击下面的英文字幕来播放视频。

翻译人员: Nan Yang 校对人员: Jiasi Hao
00:13
No matter who you are or where you live,
0
13548
2516
无论你是谁或者住在哪,
00:16
I'm guessing that you have at least one relative
1
16088
2356
我猜你们身边至少有一位
00:18
that likes to forward those emails.
2
18468
2334
喜欢转发那些电子邮件的亲戚。
00:21
You know the ones I'm talking about --
3
21206
1900
你们知道我说的那些邮件
是带有可疑声明 或阴谋论录像的邮件。
00:23
the ones with dubious claims or conspiracy videos.
4
23130
2854
00:26
And you've probably already muted them on Facebook
5
26315
2668
而且你们可能已经在 Facebook 上屏蔽了他们,
00:29
for sharing social posts like this one.
6
29007
2348
不想收到这种信息,像这个。
00:31
It's an image of a banana
7
31379
1404
这是一张香蕉的图片,
00:32
with a strange red cross running through the center.
8
32807
2667
香蕉中间有一个奇怪的红色十字。
00:35
And the text around it is warning people
9
35498
2139
旁边的文字在警告人们
00:37
not to eat fruits that look like this,
10
37661
2158
不要吃这样的水果,
00:39
suggesting they've been injected with blood
11
39843
2028
指出了它们已经被注射了
00:41
contaminated with the HIV virus.
12
41895
2130
被 HIV 病毒污染的血液。
00:44
And the social share message above it simply says,
13
44049
2603
而且它上面的分享消息写着:
00:46
"Please forward to save lives."
14
46676
2181
“为了拯救生命,请转发。”
00:49
Now, fact-checkers have been debunking this one for years,
15
49672
3294
事实审查员多年来 一直在揭露这一说法的真相,
00:52
but it's one of those rumors that just won't die.
16
52990
2809
但这只是那些 永远不会消失的传闻之一。
00:55
A zombie rumor.
17
55823
1271
一个僵尸谣言。
00:57
And, of course, it's entirely false.
18
57513
2093
当然它是完全虚假的。
01:00
It might be tempting to laugh at an example like this, to say,
19
60180
2959
去嘲笑这样一个例子可能很诱人,
大家会说: “反正,谁会相信这个呢?”
01:03
"Well, who would believe this, anyway?"
20
63163
1884
01:05
But the reason it's a zombie rumor
21
65419
1626
但说这是僵尸谣言的原因是
01:07
is because it taps into people's deepest fears about their own safety
22
67069
3889
它利用了关于人们对自身安全 和他们所爱之人的
01:10
and that of the people they love.
23
70982
2175
最深层次的恐惧。
01:13
And if you spend as enough time as I have looking at misinformation,
24
73783
3273
而且如果你们花费和我一样多 的时间来看这个误传,
01:17
you know that this is just one example of many
25
77080
2420
你们会知道这只是众多
01:19
that taps into people's deepest fears and vulnerabilities.
26
79524
3047
利用人们最深层的 恐惧和脆弱的例子之一。
01:23
Every day, across the world, we see scores of new memes on Instagram
27
83214
4369
每一天,在世界范围内, 我们都能看到 Instagram 上出现的
01:27
encouraging parents not to vaccinate their children.
28
87607
3039
大量新表情包在鼓励父母 不要给孩子接种疫苗。
01:30
We see new videos on YouTube explaining that climate change is a hoax.
29
90670
4532
我们看到 YoutTube 上的新视频 在解释说气候变化问题是个骗局。
01:35
And across all platforms, we see endless posts designed to demonize others
30
95226
4302
在所有平台上,我们看见了 因为种族、宗教或性取向的不同,
01:39
on the basis of their race, religion or sexuality.
31
99552
3501
而妖魔化他人的无穷无尽的帖子。
01:44
Welcome to one of the central challenges of our time.
32
104314
3030
欢迎来到我们这个时代 的主要挑战之一。
01:47
How can we maintain an internet with freedom of expression at the core,
33
107647
4025
我们如何在维护互联网 言论自由核心的同时,
01:51
while also ensuring that the content that's being disseminated
34
111696
3303
也能确保正在传播的内容
01:55
doesn't cause irreparable harms to our democracies, our communities
35
115023
3886
不会对我们的民主、我们的社区 和我们的身心健康
01:58
and to our physical and mental well-being?
36
118933
2238
造成不可弥补的伤害?
02:01
Because we live in the information age,
37
121998
2087
因为我们生活在信息时代,
02:04
yet the central currency upon which we all depend -- information --
38
124109
3547
但是我们所有人都依赖的 中央货币——信息——
02:07
is no longer deemed entirely trustworthy
39
127680
2357
不再完全值得信赖,
02:10
and, at times, can appear downright dangerous.
40
130061
2328
而且有时会显得非常危险。
02:12
This is thanks in part to the runaway growth of social sharing platforms
41
132811
3937
这部分“归功”于 社交共享媒体的迅猛发展,
02:16
that allow us to scroll through,
42
136772
1642
让我们可以在谎言和真相
02:18
where lies and facts sit side by side,
43
138438
2222
并存的世界里滑屏浏览,
02:20
but with none of the traditional signals of trustworthiness.
44
140684
3071
但是没有任何 传统的可信赖的信号。
02:24
And goodness -- our language around this is horribly muddled.
45
144268
3619
天啊——我们关于这方面 的语言极其混乱。
02:27
People are still obsessed with the phrase "fake news,"
46
147911
3103
人们仍然沉迷于“假新闻”这一短语,
02:31
despite the fact that it's extraordinarily unhelpful
47
151038
2531
尽管事实是,该短语毫无帮助,
02:33
and used to describe a number of things that are actually very different:
48
153593
3460
并且被用来描述一系列实际上 非常不同的东西:
谎言、谣言、恶作剧、 阴谋、宣传鼓吹……
02:37
lies, rumors, hoaxes, conspiracies, propaganda.
49
157077
3386
02:40
And I really wish we could stop using a phrase
50
160911
2912
我非常希望我们可以停止使用
02:43
that's been co-opted by politicians right around the world,
51
163847
2862
这一世界各地的政治家 共同选择的一个短语,
02:46
from the left and the right,
52
166733
1471
从左派到右派,
02:48
used as a weapon to attack a free and independent press.
53
168228
3222
停止将它作为武器来攻击 自由和独立的新闻媒体。
02:52
(Applause)
54
172307
4702
(掌声)
02:57
Because we need our professional news media now more than ever.
55
177033
3462
因为我们比以往任何时候 都更需要我们的专业新闻媒体。
03:00
And besides, most of this content doesn't even masquerade as news.
56
180882
3373
而且除此之外,大多数内容 甚至都没有被伪装成新闻,
03:04
It's memes, videos, social posts.
57
184279
2642
而是以表情包、视频、 社交帖子的形式存在。
03:06
And most of it is not fake; it's misleading.
58
186945
3453
而且大部分内容 不是假的,而是误导。
我们倾向于专注在真假上。
03:10
We tend to fixate on what's true or false.
59
190422
3015
03:13
But the biggest concern is actually the weaponization of context.
60
193461
4032
但是最大的担忧实际上是 语境的武器化。
03:18
Because the most effective disinformation
61
198855
1968
因为最有效的虚假信息
03:20
has always been that which has a kernel of truth to it.
62
200847
3048
一直是具有真实内核的那些。
03:23
Let's take this example from London, from March 2017,
63
203919
3476
举一个 2017 年 3 月伦敦的例子,
03:27
a tweet that circulated widely
64
207419
1540
在威斯敏斯特桥
03:28
in the aftermath of a terrorist incident on Westminster Bridge.
65
208983
3587
恐怖袭击事件发生后 的一个广为流传的推文。
03:32
This is a genuine image, not fake.
66
212594
2428
这是个真实的图片,不是假的。
03:35
The woman who appears in the photograph was interviewed afterwards,
67
215046
3169
照片里的女子在事后接受了采访,
她解释到 她精神上完全受到了打击。
03:38
and she explained that she was utterly traumatized.
68
218239
2409
03:40
She was on the phone to a loved one,
69
220672
1738
事件发生当时 她和爱人在打电话,
03:42
and she wasn't looking at the victim out of respect.
70
222434
2618
出于尊重, 她没有看向身旁的受害者。
03:45
But it still was circulated widely with this Islamophobic framing,
71
225076
3960
但是它仍然以 伊斯兰恐惧症为主题广为流传,
03:49
with multiple hashtags, including: #BanIslam.
72
229060
3046
并带有多个主题标签, 包括:#BanIslam(禁止伊斯兰人)。
03:52
Now, if you worked at Twitter, what would you do?
73
232425
2398
如果你在推特工作, 你会怎么做?
03:54
Would you take that down, or would you leave it up?
74
234847
2562
你会把它撤下平台, 还是让它留在那?
03:58
My gut reaction, my emotional reaction, is to take this down.
75
238553
3429
我的直觉和情感反应 是把它撤下来。
04:02
I hate the framing of this image.
76
242006
2142
我讨厌对这个图片的“包装”。
04:04
But freedom of expression is a human right,
77
244585
2388
但是言论自由是人类的权力,
04:06
and if we start taking down speech that makes us feel uncomfortable,
78
246997
3225
如果我们开始 把让我们感到不舒服的言论撤下,
04:10
we're in trouble.
79
250246
1230
我们就有麻烦了。
04:11
And this might look like a clear-cut case,
80
251500
2294
而且这个可能看起来是 一个清晰明了的案例,
04:13
but, actually, most speech isn't.
81
253818
1698
但实际上,大多数言论不是。
04:15
These lines are incredibly difficult to draw.
82
255540
2436
这些边界很难描绘。
04:18
What's a well-meaning decision by one person
83
258000
2281
一个人的善意决定
04:20
is outright censorship to the next.
84
260305
2077
是对下一个人的公然审查。
04:22
What we now know is that this account, Texas Lone Star,
85
262759
2929
现在我们知道的是, 这个账户,德克萨斯孤星,
04:25
was part of a wider Russian disinformation campaign,
86
265712
3230
是一个更大的 俄罗斯造谣运动的一部分,
04:28
one that has since been taken down.
87
268966
2151
此后已经被扳倒。
04:31
Would that change your view?
88
271141
1563
这会改变你的看法吗?
04:33
It would mine,
89
273322
1159
会改变我的,
04:34
because now it's a case of a coordinated campaign
90
274505
2301
因为现在这个案子的性质 变成了一场播种不和谐种子的
04:36
to sow discord.
91
276830
1215
协作运动。
04:38
And for those of you who'd like to think
92
278069
1961
而对于你们中那些认为
04:40
that artificial intelligence will solve all of our problems,
93
280054
2831
人工智能将会解决我们 所有问题的人来说,
04:42
I think we can agree that we're a long way away
94
282909
2225
我认为我们可以同意
在人工智能能够理解这种帖子前 我们还有很长的一段路要走。
04:45
from AI that's able to make sense of posts like this.
95
285158
2587
04:48
So I'd like to explain three interlocking issues
96
288856
2507
所以我想解释一下 让这事变得如此复杂的
04:51
that make this so complex
97
291387
2373
三个环环相扣的问题,
04:53
and then think about some ways we can consider these challenges.
98
293784
3122
然后琢磨出些方法来思考这些挑战。
04:57
First, we just don't have a rational relationship to information,
99
297348
3890
首先,我们只是没有 与信息建立起理性关系,
05:01
we have an emotional one.
100
301262
1468
我们的关系是感性的。
05:02
It's just not true that more facts will make everything OK,
101
302754
3794
并非更多的事实真相 会让一切顺利,
05:06
because the algorithms that determine what content we see,
102
306572
3100
因为决定我们 能看见什么内容的算法,
05:09
well, they're designed to reward our emotional responses.
103
309696
3127
是被设计来 奖励我们的情感反应。
05:12
And when we're fearful,
104
312847
1381
而当我们恐惧时,
05:14
oversimplified narratives, conspiratorial explanations
105
314252
3174
过于简化的叙述、阴谋论解释,
05:17
and language that demonizes others is far more effective.
106
317450
3418
和妖魔化事物的语言 更加有效。
05:21
And besides, many of these companies,
107
321538
1874
此外,有很多公司,
05:23
their business model is attached to attention,
108
323436
2546
他们的商业模式 与人们的关注度息息相关,
05:26
which means these algorithms will always be skewed towards emotion.
109
326006
3690
这意味着这些算法 总是会偏向情感。
05:30
Second, most of the speech I'm talking about here is legal.
110
330371
4298
第二,我现在谈论的 大多数言论是合法的。
如果我在说的是
05:35
It would be a different matter
111
335081
1446
05:36
if I was talking about child sexual abuse imagery
112
336551
2341
儿童性虐待图片, 或者是煽动暴力的内容,
05:38
or content that incites violence.
113
338916
1927
就是另外一回事。
05:40
It can be perfectly legal to post an outright lie.
114
340867
3270
公开撒谎可以是完全合法的。
05:45
But people keep talking about taking down "problematic" or "harmful" content,
115
345130
4034
但是人们一直在讨论撤下 “有问题的”或“有害的”内容,
05:49
but with no clear definition of what they mean by that,
116
349188
2609
但是并没有对它们是什么 有明确的定义,
05:51
including Mark Zuckerberg,
117
351821
1264
包括马克·扎克伯格,
他最近呼吁 全球监管来缓和言论。
05:53
who recently called for global regulation to moderate speech.
118
353109
3412
05:56
And my concern is that we're seeing governments
119
356870
2215
而我的担心是,我们看见了
05:59
right around the world
120
359109
1292
世界各地的政府
06:00
rolling out hasty policy decisions
121
360425
2676
推出仓促的政策决定,
06:03
that might actually trigger much more serious consequences
122
363125
2746
但是它可能实际上触发了
06:05
when it comes to our speech.
123
365895
1714
对于我们的言论更严重的后果。
而且即使我们可以决定 哪个言论留住或撤下来,
06:08
And even if we could decide which speech to take up or take down,
124
368006
3706
06:11
we've never had so much speech.
125
371736
2174
我们从来没有过 现在这么多的言论。
06:13
Every second, millions of pieces of content
126
373934
2131
每一秒,几百万的内容,
06:16
are uploaded by people right around the world
127
376089
2107
以不同的语言
被世界各地的人上传,
06:18
in different languages,
128
378220
1168
06:19
drawing on thousands of different cultural contexts.
129
379412
2768
吸取了上千种不同的文化背景。
06:22
We've simply never had effective mechanisms
130
382204
2532
我们根本没有有效的机制,
06:24
to moderate speech at this scale,
131
384760
1738
无论是通过人工还是技术手段
06:26
whether powered by humans or by technology.
132
386522
2801
去缓和这种规模的言论内容,
06:30
And third, these companies -- Google, Twitter, Facebook, WhatsApp --
133
390284
3944
然后第三,这些公司—— 谷歌,推特,脸书,WhatsApp ——
06:34
they're part of a wider information ecosystem.
134
394252
2841
它们是广阔的 信息生态系统的一部分。
06:37
We like to lay all the blame at their feet, but the truth is,
135
397117
3352
我们喜欢把所有责任 都推到他们身上,但事实是,
06:40
the mass media and elected officials can also play an equal role
136
400493
3830
大众媒体和民选官员 只要他们想,
06:44
in amplifying rumors and conspiracies when they want to.
137
404347
2913
也可以在扩大谣言和阴谋上 发挥同等作用。
06:47
As can we, when we mindlessly forward divisive or misleading content
138
407800
4944
我们也一样,当我们漫不经心地 转发分裂性或误导性的内容时,
06:52
without trying.
139
412768
1285
甚至没有费力。
06:54
We're adding to the pollution.
140
414077
1800
我们正在加剧这种“污染”。
06:57
I know we're all looking for an easy fix.
141
417236
2618
我知道我们都在 寻找简单的解决方法。
06:59
But there just isn't one.
142
419878
1667
但就是一个都没有。
07:01
Any solution will have to be rolled out at a massive scale, internet scale,
143
421950
4445
任何解决方案都必须 以大规模推出,互联网的规模,
07:06
and yes, the platforms, they're used to operating at that level.
144
426419
3261
而且的确,这些平台 已经习惯在那种级别的规模上运行。
07:09
But can and should we allow them to fix these problems?
145
429704
3472
但是我们可以并应该 允许他们来解决这些问题吗?
07:13
They're certainly trying.
146
433668
1232
他们当然在努力。
07:14
But most of us would agree that, actually, we don't want global corporations
147
434924
4086
但是我们大多数人都会同意, 实际上,我们不希望跨国公司
07:19
to be the guardians of truth and fairness online.
148
439034
2332
成为网上真理与公平的守护者。
07:21
And I also think the platforms would agree with that.
149
441390
2537
我也认为这些平台也会同意。
07:24
And at the moment, they're marking their own homework.
150
444257
2881
此时此刻, 他们正在展现属于自己的成果。
07:27
They like to tell us
151
447162
1198
他们想告诉我们
07:28
that the interventions they're rolling out are working,
152
448384
2579
他们实施的干预措施正在奏效,
07:30
but because they write their own transparency reports,
153
450987
2540
但是因为他们编写的是 他们自己的透明度报告,
07:33
there's no way for us to independently verify what's actually happening.
154
453551
3618
我们无法独立验证实际的情况。
07:38
(Applause)
155
458431
3342
(掌声)
07:41
And let's also be clear that most of the changes we see
156
461797
2952
我们也要清楚, 大多数我们看到的变化
07:44
only happen after journalists undertake an investigation
157
464773
2994
只发生在记者进行了调查
07:47
and find evidence of bias
158
467791
1611
并找到了存在
07:49
or content that breaks their community guidelines.
159
469426
2829
违反社区规则的偏见 和内容的证据之后。
07:52
So yes, these companies have to play a really important role in this process,
160
472815
4595
所以是的,这些公司必须 在这个过程中扮演重要角色,
07:57
but they can't control it.
161
477434
1560
但是他们无法控制它。
07:59
So what about governments?
162
479855
1518
那么政府呢?
08:01
Many people believe that global regulation is our last hope
163
481863
3096
很多人相信全球监管
08:04
in terms of cleaning up our information ecosystem.
164
484983
2880
是清理我们信息生态系统的 最后希望。
08:07
But what I see are lawmakers who are struggling to keep up to date
165
487887
3166
但是我看见的是正在努力跟上
08:11
with the rapid changes in technology.
166
491077
2341
技术迅速变革的立法者。
08:13
And worse, they're working in the dark,
167
493442
1904
更糟的是, 他们在黑暗中摸索工作,
08:15
because they don't have access to data
168
495370
1821
因为他们没有获取数据的权限
08:17
to understand what's happening on these platforms.
169
497215
2650
来了解这些平台上 正在发生些什么。
08:20
And anyway, which governments would we trust to do this?
170
500260
3071
更何况,我们会相信 哪个政府来做这件事呢?
08:23
We need a global response, not a national one.
171
503355
2770
我们需要全球的回应, 不是国家的。
08:27
So the missing link is us.
172
507419
2277
所以缺少的环节是我们。
08:29
It's those people who use these technologies every day.
173
509720
3123
是每天使用这些技术的那些人。
08:33
Can we design a new infrastructure to support quality information?
174
513260
4591
我们能不能设计一个新的基础设施 来支持高质量信息?
08:38
Well, I believe we can,
175
518371
1230
我相信我们可以,
08:39
and I've got a few ideas about what we might be able to actually do.
176
519625
3357
我已经有了一些想法, 关于我们实际上可以做什么。
08:43
So firstly, if we're serious about bringing the public into this,
177
523006
3103
所以首先,如果认真考虑 让公众参与进来,
08:46
can we take some inspiration from Wikipedia?
178
526133
2381
我们可以从维基百科 汲取一些灵感吗?
08:48
They've shown us what's possible.
179
528538
1824
他们已经向我们展示了可能的方法。
08:50
Yes, it's not perfect,
180
530386
1151
是的,它并不完美,
08:51
but they've demonstrated that with the right structures,
181
531561
2634
但是他们已经用正确的结构,
全球的视野 和很高很高的透明度证明了
08:54
with a global outlook and lots and lots of transparency,
182
534219
2635
08:56
you can build something that will earn the trust of most people.
183
536878
3096
你们可以建立一些 将赢得大多数人信任的东西。
08:59
Because we have to find a way to tap into the collective wisdom
184
539998
3162
因为我们必须找到一种方法, 充分利用集体的智慧
09:03
and experience of all users.
185
543184
2309
和所有用户的经验。
09:05
This is particularly the case for women, people of color
186
545517
2646
妇女,有色人种和 未能充分代表大众的群体
09:08
and underrepresented groups.
187
548187
1346
尤其如此。
09:09
Because guess what?
188
549557
1166
猜猜为什么?
09:10
They are experts when it comes to hate and disinformation,
189
550747
2735
他们是仇恨和虚假信息方面的专家,
09:13
because they have been the targets of these campaigns for so long.
190
553506
3516
因为他们很久以来都是 这些信息运动的目标。
09:17
And over the years, they've been raising flags,
191
557046
2350
多年来,他们一直摇旗呐喊,
09:19
and they haven't been listened to.
192
559420
1665
但是从来没有被听见。
09:21
This has got to change.
193
561109
1280
这必须改变。
09:22
So could we build a Wikipedia for trust?
194
562807
4326
所以我们能否为信任 创建一个维基百科?
09:27
Could we find a way that users can actually provide insights?
195
567157
4189
我们能否找到一种 用户可以真正提供见解的方法?
09:31
They could offer insights around difficult content-moderation decisions.
196
571370
3697
他们可以对有难度的 内容审核决定提出见解。
09:35
They could provide feedback
197
575091
1463
当平台决定要推出新变更时,
09:36
when platforms decide they want to roll out new changes.
198
576578
3041
他们可以提出反馈。
09:40
Second, people's experiences with the information is personalized.
199
580241
4162
第二,人们对信息的体验 是个性化的。
09:44
My Facebook news feed is very different to yours.
200
584427
2643
我脸书上的新闻推荐 与你们的非常不同。
09:47
Your YouTube recommendations are very different to mine.
201
587094
2745
你们的 YouTube 推荐 与我的也很不同。
09:49
That makes it impossible for us to actually examine
202
589863
2492
这使得我们无法实际检查
09:52
what information people are seeing.
203
592379
2023
大家看到的是什么信息。
09:54
So could we imagine
204
594815
1389
那么,我们是否可以想象
09:56
developing some kind of centralized open repository for anonymized data,
205
596228
4778
为匿名数据 开发某种集中式开放存储库,
10:01
with privacy and ethical concerns built in?
206
601030
2864
并内置隐私和道德问题?
10:04
Because imagine what we would learn
207
604220
1778
因为想象一下, 如果我们建立一个
10:06
if we built out a global network of concerned citizens
208
606022
3261
由关心且忧虑的 公民组成的全球网络,
10:09
who wanted to donate their social data to science.
209
609307
3294
他们希望将其社交数据捐赠给科学, 那么我们将学到什么?
10:13
Because we actually know very little
210
613141
1722
因为我们实际上对
10:14
about the long-term consequences of hate and disinformation
211
614887
2881
仇恨和虚假信息对 人们态度和行为产生的
10:17
on people's attitudes and behaviors.
212
617792
1975
长期后果知之甚少。
10:20
And what we do know,
213
620236
1167
而我们知道的是,
10:21
most of that has been carried out in the US,
214
621427
2193
其中大部分是在美国进行的,
10:23
despite the fact that this is a global problem.
215
623644
2381
尽管这是一个全球性问题。
10:26
We need to work on that, too.
216
626049
1635
我们也需要为此努力。
10:28
And third,
217
628192
1150
第三点,
我们可以找到方法来连接个体吗?
10:29
can we find a way to connect the dots?
218
629366
2310
10:31
No one sector, let alone nonprofit, start-up or government,
219
631700
3438
没有一个部门可以解决这个问题, 更不用说非营利组织,初创企业
10:35
is going to solve this.
220
635162
1422
或政府部门了。
10:36
But there are very smart people right around the world
221
636608
2564
但是世界各地有 那些非常聪明的人们
10:39
working on these challenges,
222
639196
1381
在应对这些挑战,
10:40
from newsrooms, civil society, academia, activist groups.
223
640601
3576
包括新闻编辑部,民间社会组织, 学术界和维权组织。
10:44
And you can see some of them here.
224
644201
1898
你们在这可以看见其中的一些。
10:46
Some are building out indicators of content credibility.
225
646123
2927
有些正在建立内容可信度的指标。
10:49
Others are fact-checking,
226
649074
1246
其他人在做事实核查,
10:50
so that false claims, videos and images can be down-ranked by the platforms.
227
650344
3591
以至于虚假声明,视频和图像 可以被平台撤下。
10:53
A nonprofit I helped to found, First Draft,
228
653959
2213
我协助建立的一个非营利组织, 名叫“初稿”(First Draft),
10:56
is working with normally competitive newsrooms around the world
229
656196
2968
正在与世界各地通常竞争激烈的 新闻编辑室合作,
10:59
to help them build out investigative, collaborative programs.
230
659188
3503
以帮助他们建立调查性协作项目。
11:03
And Danny Hillis, a software architect,
231
663231
2309
丹尼·希利斯, 一个软件设计师,
11:05
is designing a new system called The Underlay,
232
665564
2381
正在设计一个 叫做 The Underlay 的新系统,
11:07
which will be a record of all public statements of fact
233
667969
2775
它将记录所有与其来源相连接的
11:10
connected to their sources,
234
670768
1329
公开事实陈述,
11:12
so that people and algorithms can better judge what is credible.
235
672121
3654
以便人们和算法可以 更好地判断什么是可信的。
11:16
And educators around the world are testing different techniques
236
676800
3356
而且世界各地的教育者 在测试不同的技术,
11:20
for finding ways to make people critical of the content they consume.
237
680180
3484
以找到能使人们对所看到内容 产生批判的方法。
11:24
All of these efforts are wonderful, but they're working in silos,
238
684633
3141
所有的这些努力都很棒, 但是他们埋头各自为战,
11:27
and many of them are woefully underfunded.
239
687798
2680
而且很多都严重资金不足。
11:30
There are also hundreds of very smart people
240
690502
2053
在这些公司内部
11:32
working inside these companies,
241
692579
1652
也有成百上千的聪明人在努力,
11:34
but again, these efforts can feel disjointed,
242
694255
2325
但是同样, 这些努力让人感到不够连贯,
11:36
because they're actually developing different solutions to the same problems.
243
696604
3937
因为他们正在为同样的问题 建立不同的解决方案。
11:41
How can we find a way to bring people together
244
701205
2269
我们怎么能找到一种方法, 把这些人同时
11:43
in one physical location for days or weeks at a time,
245
703498
3278
聚集在同一个地点几天或几周,
11:46
so they can actually tackle these problems together
246
706800
2396
这样他们可以真正从不同角度
11:49
but from their different perspectives?
247
709220
1818
共同解决这些问题?
11:51
So can we do this?
248
711062
1340
那么我们能做到吗?
11:52
Can we build out a coordinated, ambitious response,
249
712426
3239
我们能否建立一种协调一致, 雄心勃勃的应对措施,
11:55
one that matches the scale and the complexity of the problem?
250
715689
3709
使其与问题的规模和复杂性相匹配?
11:59
I really think we can.
251
719819
1373
我真的认为我们可以。
12:01
Together, let's rebuild our information commons.
252
721216
2959
加入我们,重建我们的信息共享吧。
12:04
Thank you.
253
724819
1190
谢谢。
12:06
(Applause)
254
726033
3728
(掌声)
关于本网站

这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隐私政策

eng.lish.video

Developer's Blog