How technology can fight extremism and online harassment | Yasmin Green

76,668 views ・ 2018-06-27

TED


請雙擊下方英文字幕播放視頻。

譯者: Lilian Chiu 審譯者: Yanyan Hong
00:13
My relationship with the internet reminds me of the setup
0
13131
4411
我與網際網路的關係,
讓我想起老套恐怖片的情境。
00:17
to a clichéd horror movie.
1
17566
1833
00:19
You know, the blissfully happy family moves in to their perfect new home,
2
19867
4386
幸福快樂的家庭, 搬進一間美好的新房子,
00:24
excited about their perfect future,
3
24277
2281
興奮期待完美的未來,
00:26
and it's sunny outside and the birds are chirping ...
4
26582
3521
外頭陽光普照,鳥兒在唱歌……
00:30
And then it gets dark.
5
30857
1839
接著電影就變黑暗了。
00:32
And there are noises from the attic.
6
32720
2348
閣樓傳出噪音。
00:35
And we realize that that perfect new house isn't so perfect.
7
35092
4345
我們發現,那間美好的 新房子並沒有那麼美好。
00:40
When I started working at Google in 2006,
8
40485
3131
2006 年,當我開始 在谷歌(Google)工作時,
00:43
Facebook was just a two-year-old,
9
43640
1767
臉書(Facebook)才剛推出兩年,
00:45
and Twitter hadn't yet been born.
10
45431
2012
推特(Twitter)甚至還沒問世。
00:47
And I was in absolute awe of the internet and all of its promise
11
47848
4410
我對網際網路及它所有的承諾 感到絕對的敬畏,
00:52
to make us closer
12
52282
1437
它承諾要讓我們
00:53
and smarter
13
53743
1296
更靠近且更聰明,
00:55
and more free.
14
55063
1214
還有給予更多自由。
00:57
But as we were doing the inspiring work of building search engines
15
57265
3714
但當我們開始進行 這鼓舞人心的工作,
建立搜尋引擎,
01:01
and video-sharing sites and social networks,
16
61003
2886
建立影片分享網站和社交網路,
01:04
criminals, dictators and terrorists were figuring out
17
64907
4304
罪犯、獨裁者, 及恐怖分子都在想辦法
01:09
how to use those same platforms against us.
18
69235
3202
如何用同樣的平台來對抗我們。
01:13
And we didn't have the foresight to stop them.
19
73417
2455
我們沒有先見之明來阻止他們。
01:16
Over the last few years, geopolitical forces have come online to wreak havoc.
20
76746
5099
在過去幾年,地緣政治學的 勢力也上網展開大破壞。
01:21
And in response,
21
81869
1169
造成的反應是,
01:23
Google supported a few colleagues and me to set up a new group called Jigsaw,
22
83062
4778
谷歌支持我和幾位同事, 成立一個小組,叫做「Jigsaw」,
01:27
with a mandate to make people safer from threats like violent extremism,
23
87864
4596
我們的使命是要讓大家更安全, 避免受到像是極端主義、
01:32
censorship, persecution --
24
92484
2078
審查制度、迫害的威脅——
01:35
threats that feel very personal to me because I was born in Iran,
25
95186
4117
我個人對這些威脅特別有感, 因為我是在伊朗出生的,
01:39
and I left in the aftermath of a violent revolution.
26
99327
2929
在一場暴力革命的餘波中, 我被迫離開了伊朗。
01:43
But I've come to realize that even if we had all of the resources
27
103525
4346
但我漸漸了解到, 即使我們有所有的資源,
01:47
of all of the technology companies in the world,
28
107895
2858
有全世界所有的科技公司,
01:51
we'd still fail
29
111595
1230
如果我們忽略了
01:53
if we overlooked one critical ingredient:
30
113586
2948
一項關鍵因素,我們仍然會失敗:
01:57
the human experiences of the victims and perpetrators of those threats.
31
117653
5789
那些威脅的受害者 與加害者的人類經歷。
02:04
There are many challenges I could talk to you about today.
32
124935
2736
今天我其實可以 與各位談很多的挑戰。
02:07
I'm going to focus on just two.
33
127695
1504
但我只打算著重兩項:
02:09
The first is terrorism.
34
129623
2079
第一是恐怖主義。
02:13
So in order to understand the radicalization process,
35
133563
2557
為了要了解激進化的過程,
02:16
we met with dozens of former members of violent extremist groups.
36
136144
4287
我們和數十名暴力 極端主義團體的前成員見面。
02:21
One was a British schoolgirl,
37
141590
2483
其中一位是英國的女學生,
02:25
who had been taken off of a plane at London Heathrow
38
145049
3699
她曾在倫敦希斯洛機場 被強拉下飛機,
02:28
as she was trying to make her way to Syria to join ISIS.
39
148772
4692
因為當時她打算 去敘利亞加入伊斯蘭國。
02:34
And she was 13 years old.
40
154281
1931
她當時只有十三歲。
02:37
So I sat down with her and her father, and I said, "Why?"
41
157792
4625
我和她及她父親坐下來談, 我說:「為什麼?」
02:42
And she said,
42
162441
1717
她說:
02:44
"I was looking at pictures of what life is like in Syria,
43
164182
3639
「我在看一些敘利亞 生活寫照的圖片,
02:47
and I thought I was going to go and live in the Islamic Disney World."
44
167845
3510
我以為我是要去住到 伊斯蘭的迪士尼樂園。」
02:52
That's what she saw in ISIS.
45
172527
2084
這是她看到的伊斯蘭國。
02:54
She thought she'd meet and marry a jihadi Brad Pitt
46
174635
3492
她以為她要去見一位聖戰士中的 布萊德彼得並嫁給他,
02:58
and go shopping in the mall all day and live happily ever after.
47
178151
3058
整天都能去購物, 從此幸福快樂地生活。
03:02
ISIS understands what drives people,
48
182977
2824
伊斯蘭國知道什麼能驅使人,
03:05
and they carefully craft a message for each audience.
49
185825
3544
他們會為每一位觀眾 精心策劃一則訊息。
光是去看看他們把他們的行銷素材
03:11
Just look at how many languages
50
191122
1511
03:12
they translate their marketing material into.
51
192657
2273
翻譯成多少語言,就能了解了。
03:15
They make pamphlets, radio shows and videos
52
195677
2661
他們會製作小冊子、 廣播節目,和影片,
03:18
in not just English and Arabic,
53
198362
1973
不只用英語和阿拉伯語,
03:20
but German, Russian, French, Turkish, Kurdish,
54
200359
4767
還有德語、俄語、法語、 土耳其語、庫德語、
03:25
Hebrew,
55
205150
1672
希伯來語、
03:26
Mandarin Chinese.
56
206846
1741
華語(中文)。
03:29
I've even seen an ISIS-produced video in sign language.
57
209309
4192
我甚至看過一支伊斯蘭國 製作的影片是用手語的。
03:34
Just think about that for a second:
58
214605
1884
花點時間思考一下:
03:36
ISIS took the time and made the effort
59
216513
2308
伊斯蘭國投入時間和心力,
03:38
to ensure their message is reaching the deaf and hard of hearing.
60
218845
3804
來確保他們的訊息 能夠傳達給聽障人士。
03:45
It's actually not tech-savviness
61
225143
2144
伊斯蘭國能贏得人心和人信,
03:47
that is the reason why ISIS wins hearts and minds.
62
227311
2595
並不是因為他們很精通科技。
03:49
It's their insight into the prejudices, the vulnerabilities, the desires
63
229930
4163
因為他們有洞見,了解 他們試圖接觸的人有什麼
03:54
of the people they're trying to reach
64
234117
1774
偏見、脆弱、慾望,
03:55
that does that.
65
235915
1161
才能做到這一點。
03:57
That's why it's not enough
66
237718
1429
那就說明了為什麼
03:59
for the online platforms to focus on removing recruiting material.
67
239171
4239
線上平台只把焦點放在 移除召募素材是不足的。
04:04
If we want to have a shot at building meaningful technology
68
244518
3581
如果我想要有機會建立 一種有意義的技術,
04:08
that's going to counter radicalization,
69
248123
1874
用它來對抗極端化,
04:10
we have to start with the human journey at its core.
70
250021
2979
我們就得要從它核心的 人類旅程開始著手。
04:13
So we went to Iraq
71
253884
2187
所以,我們去了伊拉克,
04:16
to speak to young men who'd bought into ISIS's promise
72
256095
2831
去和年輕人交談, 我們找的對象曾相信伊斯蘭國
04:18
of heroism and righteousness,
73
258950
3191
所做的關於英雄主義與公正的承諾,
04:22
who'd taken up arms to fight for them
74
262165
1847
曾拿起武器為他們作戰,
04:24
and then who'd defected
75
264036
1338
接著,在目擊了
04:25
after they witnessed the brutality of ISIS's rule.
76
265398
3021
伊斯蘭國統治的殘酷之後選擇變節。
04:28
And I'm sitting there in this makeshift prison in the north of Iraq
77
268880
3192
我坐在北伊拉克的一間臨時監獄裡,
04:32
with this 23-year-old who had actually trained as a suicide bomber
78
272096
4550
會見一位在變節前 受過訓練的自殺炸彈客,
04:36
before defecting.
79
276670
1552
年僅 23 歲。
04:39
And he says,
80
279080
1158
他說:
04:41
"I arrived in Syria full of hope,
81
281119
3220
「我到敘利亞時滿懷著希望,
04:44
and immediately, I had two of my prized possessions confiscated:
82
284363
4365
可一下子我兩項最重要的 東西就被沒收了:
04:48
my passport and my mobile phone."
83
288752
2933
我的護照和我的手機。」
04:52
The symbols of his physical and digital liberty
84
292140
2406
在他抵達時, 這兩樣象徵他實體自由
04:54
were taken away from him on arrival.
85
294570
1760
與數位自由的東西被奪去了。
04:57
And then this is the way he described that moment of loss to me.
86
297248
3510
接著,他這樣向我描述迷失的時刻。
05:01
He said,
87
301356
1586
他說:
05:02
"You know in 'Tom and Jerry,'
88
302966
2329
「在《湯姆貓與傑利鼠》中,
05:06
when Jerry wants to escape, and then Tom locks the door
89
306192
3103
當傑利想要逃脫時,湯姆把門鎖住,
05:09
and swallows the key
90
309319
1156
把鑰匙吞掉,
05:10
and you see it bulging out of his throat as it travels down?"
91
310499
3551
還可以從外表形狀看到鑰匙 延著喉嚨下滑,記得嗎?」
05:14
And of course, I really could see the image that he was describing,
92
314446
3153
當然,我能看見他所描述的畫面,
05:17
and I really did connect with the feeling that he was trying to convey,
93
317623
3661
我真的能和他試圖傳達的 這種感受產生連結,
05:21
which was one of doom,
94
321308
2021
一種在劫難逃的感受,
05:23
when you know there's no way out.
95
323353
1789
你知道沒有路可逃了。
05:26
And I was wondering:
96
326551
1289
而我很納悶:
05:28
What, if anything, could have changed his mind
97
328644
2682
在他離家的那一天,如果有的話,
05:31
the day that he left home?
98
331350
1240
什麼能改變他的心意?
05:32
So I asked,
99
332614
1250
於是,我問:
05:33
"If you knew everything that you know now
100
333888
3178
「如果你當時知道 你現在知道的這些
05:37
about the suffering and the corruption, the brutality --
101
337090
3051
關於苦難、腐敗、殘酷的狀況——
05:40
that day you left home,
102
340165
1415
在離家的那天就知道,
05:41
would you still have gone?"
103
341604
1679
你還會選擇離開嗎?」
05:43
And he said, "Yes."
104
343786
1711
他說:「會。」
05:45
And I thought, "Holy crap, he said 'Yes.'"
105
345846
2282
我心想:「老天爺,他說『會』。」
05:48
And then he said,
106
348694
1219
接著他說:
05:49
"At that point, I was so brainwashed,
107
349937
3001
「在那個時點,我完全被洗腦了,
05:52
I wasn't taking in any contradictory information.
108
352962
3244
我不會接受任何有所矛盾的資訊。
05:56
I couldn't have been swayed."
109
356744
1555
我當時不可能被動搖。」
05:59
"Well, what if you knew everything that you know now
110
359235
2527
「那麼如果你在你離開前 六個月就已經知道
06:01
six months before the day that you left?"
111
361786
2098
你現在知道的這些,結果會如何?」
06:05
"At that point, I think it probably would have changed my mind."
112
365345
3131
「若在那個時點, 我想我可能會改變心意。」
06:10
Radicalization isn't this yes-or-no choice.
113
370138
3397
極端化並不是關於是非的選擇。
06:14
It's a process, during which people have questions --
114
374007
2977
它是一個過程,在這過程中, 人們會有問題——
06:17
about ideology, religion, the living conditions.
115
377008
3776
關於意識型態、宗教、 生活條件的問題。
06:20
And they're coming online for answers,
116
380808
2766
他們會上網尋求答案,
06:23
which is an opportunity to reach them.
117
383598
1917
這就是一個接觸他們的機會。
06:25
And there are videos online from people who have answers --
118
385905
3014
有答案的人會在網路提供影片——
06:28
defectors, for example, telling the story of their journey
119
388943
2876
比如,叛逃者訴說他們
投入和脫離暴力的心路歷程;
06:31
into and out of violence;
120
391843
1583
06:33
stories like the one from that man I met in the Iraqi prison.
121
393450
3487
就像我在伊拉克監獄見到的 那名男子告訴我的故事。
06:37
There are locals who've uploaded cell phone footage
122
397914
2590
有當地人會上傳手機影片,
06:40
of what life is really like in the caliphate under ISIS's rule.
123
400528
3503
呈現在伊斯蘭國統治之下 穆斯林國的真實生活樣貌。
06:44
There are clerics who are sharing peaceful interpretations of Islam.
124
404055
3735
有教會聖職人員分享 關於伊斯蘭的和平詮釋。
06:48
But you know what?
125
408830
1150
但你們知道嗎?
06:50
These people don't generally have the marketing prowess of ISIS.
126
410004
3020
這些人通常都沒有 伊斯蘭國的高超行銷本領。
06:54
They risk their lives to speak up and confront terrorist propaganda,
127
414049
4532
他們冒著生命危險說出來, 和恐怖主義的宣傳對質,
06:58
and then they tragically don't reach the people
128
418605
2211
但很不幸的是,他們無法接觸到
07:00
who most need to hear from them.
129
420840
1682
最需要聽到他們聲音的人。
07:03
And we wanted to see if technology could change that.
130
423173
2612
我們想試看看, 科技是否能改變這一點。
07:06
So in 2016, we partnered with Moonshot CVE
131
426205
4183
2016 年,我們和 Moonshot CVE (信息泄露組織)合作,
07:10
to pilot a new approach to countering radicalization
132
430412
3180
試驗一種對抗極端化的新方法,
07:13
called the "Redirect Method."
133
433616
1780
叫做「重新定向法」。
07:16
It uses the power of online advertising
134
436453
3012
它用線上廣告的力量,
07:19
to bridge the gap between those susceptible to ISIS's messaging
135
439489
4514
在容易受到伊斯蘭國訊息影響的人
與揭發那些訊息真面目的 可靠聲音之間搭起橋樑。
07:24
and those credible voices that are debunking that messaging.
136
444027
3760
07:28
And it works like this:
137
448633
1150
它的運作方式如下:
07:29
someone looking for extremist material --
138
449807
1961
有人在尋找極端主義的素材——
07:31
say they search for "How do I join ISIS?" --
139
451792
2990
比如他們搜尋 「如何加入伊斯蘭國?」——
07:34
will see an ad appear
140
454806
2476
就會有一則廣告出現,
07:37
that invites them to watch a YouTube video of a cleric, of a defector --
141
457306
4882
邀請他們上 YouTube, 看聖職人員、變節者的影片——
07:42
someone who has an authentic answer.
142
462212
2310
有真實答案的人所拍的影片。
07:44
And that targeting is based not on a profile of who they are,
143
464546
3623
這個方法鎖定目標對象的方式 不是依據個人資料,
07:48
but of determining something that's directly relevant
144
468193
3053
而是由與他們的查詢或問題有直接
07:51
to their query or question.
145
471270
1708
相關的東西來決定。
07:54
During our eight-week pilot in English and Arabic,
146
474122
2842
我們用英語和阿拉伯語 做了八週的測試,
07:56
we reached over 300,000 people
147
476988
3279
接觸到了超過三十萬人,
08:00
who had expressed an interest in or sympathy towards a jihadi group.
148
480291
5545
他們都是對聖戰團體 表示感興趣或同情的人。
08:06
These people were now watching videos
149
486626
2264
現在這些人在看的影片
08:08
that could prevent them from making devastating choices.
150
488914
3340
能預防他們做出毀滅性的選擇。
08:13
And because violent extremism isn't confined to any one language,
151
493405
3727
因為暴力極端主義 不侷限於任何一種語言、
08:17
religion or ideology,
152
497156
1804
宗教,或意識形態,
08:18
the Redirect Method is now being deployed globally
153
498984
3501
「重新定向法」現已在全球實施,
08:22
to protect people being courted online by violent ideologues,
154
502509
3804
保護大家上網時不會受到 暴力意識形態的引誘,
08:26
whether they're Islamists, white supremacists
155
506337
2596
不論是伊斯蘭教的、 白人至上主義的,
08:28
or other violent extremists,
156
508957
2103
或其他暴力極端主義的,
08:31
with the goal of giving them the chance to hear from someone
157
511084
2873
我們的目標是要 給他們機會去聽聽看
08:33
on the other side of that journey;
158
513981
2091
在那旅程另一端的人怎麼說;
08:36
to give them the chance to choose a different path.
159
516096
2839
給他們機會去選擇不同的路。
08:40
It turns out that often the bad guys are good at exploiting the internet,
160
520749
5980
事實證明, 壞人通常擅長利用網際網路,
08:46
not because they're some kind of technological geniuses,
161
526753
3744
並不是因為他們是什麼科技天才,
08:50
but because they understand what makes people tick.
162
530521
2985
而是因為他們了解人的癢處。
08:54
I want to give you a second example:
163
534855
2369
我再舉個例子說明:
08:58
online harassment.
164
538019
1391
線上騷擾。
09:00
Online harassers also work to figure out what will resonate
165
540629
3363
線上騷擾者也在致力於 找出什麼能讓
09:04
with another human being.
166
544016
1615
另一個人產生共鳴。
09:05
But not to recruit them like ISIS does,
167
545655
3110
但他們的目的不像 伊斯蘭國是要招募人,
09:08
but to cause them pain.
168
548789
1275
而是造成別人痛苦。
09:11
Imagine this:
169
551259
1342
想像這個狀況:
09:13
you're a woman,
170
553347
1659
你是一名女子,
09:15
you're married,
171
555030
1413
已婚,
09:16
you have a kid.
172
556467
1154
有一個孩子。
09:18
You post something on social media,
173
558834
1784
你在社交媒體上發了一篇文,
09:20
and in a reply, you're told that you'll be raped,
174
560642
2886
你得到一則回應,說你會被強暴,
09:24
that your son will be watching,
175
564577
1560
你的兒子會被監視,
09:26
details of when and where.
176
566825
1856
還有時間和地點的細節資訊。
09:29
In fact, your home address is put online for everyone to see.
177
569148
3143
事實上,在網路上大家 都能看到你家的地址。
09:33
That feels like a pretty real threat.
178
573580
2007
那威脅感覺十分真實。
09:37
Do you think you'd go home?
179
577113
1656
你認為你會回家嗎?
09:39
Do you think you'd continue doing the thing that you were doing?
180
579999
3048
你認為你會繼續做你正在做的事嗎?
09:43
Would you continue doing that thing that's irritating your attacker?
181
583071
3220
你會繼續做那件惹惱了 攻擊你的人的那件事嗎?
09:48
Online abuse has been this perverse art
182
588016
3096
線上虐待一直都是種 刻意作對的藝術,
09:51
of figuring out what makes people angry,
183
591136
3468
找出什麼能讓人生氣,
09:54
what makes people afraid,
184
594628
2132
什麼能讓人害怕,
09:56
what makes people insecure,
185
596784
1641
什麼能讓人沒有安全感,
09:58
and then pushing those pressure points until they're silenced.
186
598449
3067
接著去壓那些對壓力敏感之處, 直到它們被壓制下來。
10:02
When online harassment goes unchecked,
187
602333
2304
當線上騷擾不受管束時,
10:04
free speech is stifled.
188
604661
1667
自由言論就會被扼殺。
10:07
And even the people hosting the conversation
189
607196
2127
即使主持對話的人
10:09
throw up their arms and call it quits,
190
609347
1834
棄械並宣佈到此為止,
10:11
closing their comment sections and their forums altogether.
191
611205
2957
把他們的留言區以及論壇都關閉。
10:14
That means we're actually losing spaces online
192
614186
2849
那意味著,我們其實正在失去線上
10:17
to meet and exchange ideas.
193
617059
1987
可以碰面交換點子的空間。
10:19
And where online spaces remain,
194
619939
2163
在還有線上空間的地方,
10:22
we descend into echo chambers with people who think just like us.
195
622126
4470
我們會陷入到迴音室當中, 只和相同想法的人聚在一起。
10:27
But that enables the spread of disinformation;
196
627688
2499
但那會造成錯誤訊息被散佈出去;
10:30
that facilitates polarization.
197
630211
2184
那會促成兩極化。
10:34
What if technology instead could enable empathy at scale?
198
634508
5269
但如果反之能用科技 來大量產生同理心呢?
10:40
This was the question that motivated our partnership
199
640451
2486
就是這個問題
驅使我們和谷歌的反虐待小組、
10:42
with Google's Counter Abuse team,
200
642961
1819
10:44
Wikipedia
201
644804
1178
維基,
10:46
and newspapers like the New York Times.
202
646006
1934
以及像紐約時報這類報紙合作。
10:47
We wanted to see if we could build machine-learning models
203
647964
2876
我們想要看看我們 是否能建立出能夠了解
10:50
that could understand the emotional impact of language.
204
650864
3606
語言會帶來什麼情緒影響的 機器學習模型,
10:55
Could we predict which comments were likely to make someone else leave
205
655062
3610
我們能否預測什麼樣的意見 有可能會讓另一個人
10:58
the online conversation?
206
658696
1374
離開線上對談?
11:00
And that's no mean feat.
207
660515
3887
那不是容易的事。
11:04
That's no trivial accomplishment
208
664426
1566
人工智慧要做到
11:06
for AI to be able to do something like that.
209
666016
2563
那樣的事,並不是理所當然。
11:08
I mean, just consider these two examples of messages
210
668603
3729
我是指,想想這兩個例子,
都是在上週我有可能 會收到的訊息。
11:12
that could have been sent to me last week.
211
672356
2224
「祝在 TED 順利!」 (直譯:在 TED 斷一條腿。)
11:15
"Break a leg at TED!"
212
675517
1879
11:17
... and
213
677420
1164
以及
11:18
"I'll break your legs at TED."
214
678608
2126
「我會在 TED 打斷你一條腿。」
11:20
(Laughter)
215
680758
1246
(笑聲)
11:22
You are human,
216
682028
1513
你們是人,
11:23
that's why that's an obvious difference to you,
217
683565
2210
那就是為何你們能明顯看出,
11:25
even though the words are pretty much the same.
218
685799
2224
用字幾乎相同的兩個句子有何差別。
11:28
But for AI, it takes some training to teach the models
219
688047
3079
但對人工智慧來說, 要透過訓練來教導模型
11:31
to recognize that difference.
220
691150
1571
去辨視那差別。
11:32
The beauty of building AI that can tell the difference
221
692745
3245
建立能夠分辨出那差別的 人工智慧,有個美好之處,
11:36
is that AI can then scale to the size of the online toxicity phenomenon,
222
696014
5050
就是人工智慧能處理 線上毒素現象的規模,
11:41
and that was our goal in building our technology called Perspective.
223
701088
3287
為此目的,我們建立了 一種出名為「觀點」的技術。
11:45
With the help of Perspective,
224
705056
1427
在「觀點」的協助下,
11:46
the New York Times, for example,
225
706507
1583
以紐約時報為例,
11:48
has increased spaces online for conversation.
226
708114
2487
他們增加了線上交談的空間。
11:51
Before our collaboration,
227
711005
1310
在與我們合作之前,
11:52
they only had comments enabled on just 10 percent of their articles.
228
712339
4305
他們的文章只有 大約 10% 有開放留言。
11:57
With the help of machine learning,
229
717495
1644
在機器學習的協助下,
11:59
they have that number up to 30 percent.
230
719163
1897
這個數字提升到了 30%。
12:01
So they've tripled it,
231
721084
1156
增加了三倍,
12:02
and we're still just getting started.
232
722264
1917
且我們才剛開始合作而已。
12:04
But this is about way more than just making moderators more efficient.
233
724872
3461
這絕對不只是讓版主更有效率。
12:10
Right now I can see you,
234
730076
1850
現在,我可以看見你們,
12:11
and I can gauge how what I'm saying is landing with you.
235
731950
3294
我可以估量我所說的話 會如何對你們產生影響。
12:16
You don't have that opportunity online.
236
736370
1879
在線上沒有這樣的機會。
12:18
Imagine if machine learning could give commenters,
237
738558
3635
想像一下, 當留言者在打字的時候,
如果機器學習能夠
12:22
as they're typing,
238
742217
1162
12:23
real-time feedback about how their words might land,
239
743403
3347
即使給他們回饋, 說明他們的文字會造成什麼影響,
12:27
just like facial expressions do in a face-to-face conversation.
240
747609
3024
就像在面對面交談時, 面部表情的功能。
12:32
Machine learning isn't perfect,
241
752926
1842
機器學習並不完美,
12:34
and it still makes plenty of mistakes.
242
754792
2394
它仍然會犯許多錯誤。
12:37
But if we can build technology
243
757210
1557
但如果我們能建立出
12:38
that understands the emotional impact of language,
244
758791
3293
能了解語言有什麼 情緒影響力的技術,
12:42
we can build empathy.
245
762108
1460
我們就能建立同理心。
12:43
That means that we can have dialogue between people
246
763592
2425
那就表示,我們能讓兩個人對話,
12:46
with different politics,
247
766041
1816
即使他們政治立場不同,
12:47
different worldviews,
248
767881
1216
世界觀不同,
12:49
different values.
249
769121
1246
價值觀不同。
12:51
And we can reinvigorate the spaces online that most of us have given up on.
250
771359
4775
我們能讓大部分人已經放棄的 線上空間再度復甦。
12:57
When people use technology to exploit and harm others,
251
777857
3785
當人們用科技來利用和傷害他人時,
13:01
they're preying on our human fears and vulnerabilities.
252
781666
3642
他們靠的是我們人類的恐懼和脆弱。
13:06
If we ever thought that we could build an internet
253
786461
3508
如果我們認為我們能夠建立一個完全
13:09
insulated from the dark side of humanity,
254
789993
2578
沒有人性黑暗面的網際網路,
13:12
we were wrong.
255
792595
1184
我們就錯了。
13:14
If we want today to build technology
256
794361
2270
如果現今我們想要建立技術
13:16
that can overcome the challenges that we face,
257
796655
3127
來克服我們面臨的挑戰,
13:19
we have to throw our entire selves into understanding the issues
258
799806
4043
我們就得把自己全心全意投入, 並去了解這些議題,
13:23
and into building solutions
259
803873
1893
去建立人類的解決方案,
13:25
that are as human as the problems they aim to solve.
260
805790
3782
來解決人類的問題。
13:30
Let's make that happen.
261
810071
1513
讓我們來實現它吧。
13:31
Thank you.
262
811924
1150
謝謝。
13:33
(Applause)
263
813098
3277
(掌聲)
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隱私政策

eng.lish.video

Developer's Blog