How a handful of tech companies control billions of minds every day | Tristan Harris

950,120 views ・ 2017-07-28

TED


請雙擊下方英文字幕播放視頻。

譯者: 易帆 余 審譯者: Wilde Luo
00:12
I want you to imagine
0
12960
1200
我要各位去想像一個場景,
00:15
walking into a room,
1
15000
1200
你走進了一個房間,
00:17
a control room with a bunch of people,
2
17480
2136
這個控制室裡有一群人,
00:19
a hundred people, hunched over a desk with little dials,
3
19640
2840
一百多人縮在布置着 各種小儀表盤的辦公桌前,
00:23
and that that control room
4
23280
1520
而這間控制室,
00:25
will shape the thoughts and feelings
5
25680
3696
即將影響
十億多人的想法與感受。
00:29
of a billion people.
6
29400
1240
00:32
This might sound like science fiction,
7
32560
1880
這聽起來像是科幻小說,
00:35
but this actually exists
8
35320
2216
但它確實存在,
00:37
right now, today.
9
37560
1200
就在當下,今天。
00:40
I know because I used to be in one of those control rooms.
10
40040
3360
我會知道的原因,是因為 我也曾經是控制室裡的一員。
00:44
I was a design ethicist at Google,
11
44159
2297
我曾經是谷歌的倫理設計學家,
00:46
where I studied how do you ethically steer people's thoughts?
12
46480
3360
我在研究如何在符合道德的前提下, 控制人們的思想。
00:50
Because what we don't talk about is how the handful of people
13
50560
2896
因為我們不會去討論
00:53
working at a handful of technology companies
14
53480
2536
這幾家科技公司裡面的人
00:56
through their choices will steer what a billion people are thinking today.
15
56040
5040
會如何以他們的選擇意志 去控制十幾億人的想法。
01:02
Because when you pull out your phone
16
62400
1736
因為當你拿出手機時,
01:04
and they design how this works or what's on the feed,
17
64160
3096
他們已經設計好如何運作 或者要給你什麼資訊。
01:07
it's scheduling little blocks of time in our minds.
18
67280
3216
它已經在我們的腦中 安排好很多小時段。
01:10
If you see a notification, it schedules you to have thoughts
19
70520
3136
如果你看了通知, 這會促使你產生一個
01:13
that maybe you didn't intend to have.
20
73680
2040
你也許不想要的想法。
01:16
If you swipe over that notification,
21
76400
2616
如果你跳過那個通知,
01:19
it schedules you into spending a little bit of time
22
79040
2381
它就會讓你多花點時間
01:21
getting sucked into something
23
81445
1381
投入到你不想要的東西上,
01:22
that maybe you didn't intend to get sucked into.
24
82850
2955
而你原本也許不想要 花時間在那上面。
01:27
When we talk about technology,
25
87320
1520
當我們在談論科技時,
01:30
we tend to talk about it as this blue sky opportunity.
26
90040
2696
我們傾向於把它當作是 湛藍天空的機會。
01:32
It could go any direction.
27
92760
1480
它可以往任何方向發展。
01:35
And I want to get serious for a moment
28
95400
1856
但我想認真地說,
01:37
and tell you why it's going in a very specific direction.
29
97280
2680
我要告訴各位, 為什麼科技正在往特定的方向發展。
01:40
Because it's not evolving randomly.
30
100840
2200
因為科技的演變不是隨機的。
01:44
There's a hidden goal driving the direction
31
104000
2016
在我們所有創造的科技背後,
01:46
of all of the technology we make,
32
106040
2136
都隱藏著一個特定目標,
01:48
and that goal is the race for our attention.
33
108200
2920
而那個目標就是 競相爭奪我們的注意力。
01:52
Because every news site,
34
112840
2736
因為每一個新網頁──
01:55
TED, elections, politicians,
35
115600
2736
TED 網頁、選舉網頁、政客網頁、
01:58
games, even meditation apps
36
118360
1976
遊戲網頁,甚至是冥想的應用軟體──
02:00
have to compete for one thing,
37
120360
1960
都必須競爭同樣的東西,
02:03
which is our attention,
38
123160
1736
也就是我們的注意力,
02:04
and there's only so much of it.
39
124920
1600
但市場就這麼大。
02:08
And the best way to get people's attention
40
128440
2416
而要獲得人們注意力的最佳方法,
02:10
is to know how someone's mind works.
41
130880
2440
就是去了解使用者的腦袋 是如何運作的。
02:13
And there's a whole bunch of persuasive techniques
42
133800
2336
有很多說服的技巧,
02:16
that I learned in college at a lab called the Persuasive Technology Lab
43
136160
3496
我是從大學的 「說服力技術實驗室」學來的,
02:19
to get people's attention.
44
139680
1600
他們教你如何獲得別人的注意力。
02:21
A simple example is YouTube.
45
141880
1480
一個簡單的例子就是 YouTube。
02:24
YouTube wants to maximize how much time you spend.
46
144000
2936
YouTube 想要最大化 你花費在他們網站的時間。
02:26
And so what do they do?
47
146960
1200
那他們會怎麼做?
02:28
They autoplay the next video.
48
148840
2280
他們會幫你自動撥放下一部片。
02:31
And let's say that works really well.
49
151760
1816
這一招真的很有效,
02:33
They're getting a little bit more of people's time.
50
153600
2416
他們也因此得到 使用者更多的時間。
02:36
Well, if you're Netflix, you look at that and say,
51
156040
2376
但,如果你是 Netflix, 你看到了這樣的狀況,你會說,
02:38
well, that's shrinking my market share,
52
158440
1858
不行,這樣會把我的客戶給搶走,
02:40
so I'm going to autoplay the next episode.
53
160322
2000
所以,我也要自動播放下一集。
02:43
But then if you're Facebook,
54
163320
1376
但如果你是 Facebook,
02:44
you say, that's shrinking all of my market share,
55
164720
2336
你會說,那這樣我的市場 都被你們瓜分掉了,
02:47
so now I have to autoplay all the videos in the newsfeed
56
167080
2656
所以我會在你點擊播放按鍵前
02:49
before waiting for you to click play.
57
169760
1762
自動播放所有的影片給你看。
02:52
So the internet is not evolving at random.
58
172320
3160
所以,網際網路的演化不是隨機的。
02:56
The reason it feels like it's sucking us in the way it is
59
176320
4416
它會讓你感覺欲罷不能的原因,
03:00
is because of this race for attention.
60
180760
2376
就是因為這場注意力的爭奪賽。
03:03
We know where this is going.
61
183160
1416
我們知道這會有什麼後果。
03:04
Technology is not neutral,
62
184600
1520
因為科技不是中立的。
03:07
and it becomes this race to the bottom of the brain stem
63
187320
3416
這個競賽已變成
看誰可以更深地滲入 使用者腦袋的比賽。
03:10
of who can go lower to get it.
64
190760
2200
03:13
Let me give you an example of Snapchat.
65
193920
2336
我跟各位舉個例子, Snapchat。
03:16
If you didn't know, Snapchat is the number one way
66
196280
3696
不知道各位是否了解, Snapchat 目前是
美國年輕人之間, 最熱門的社交軟體。
03:20
that teenagers in the United States communicate.
67
200000
2256
03:22
So if you're like me, and you use text messages to communicate,
68
202280
4176
所以,如果你們和我一樣, 有在用簡訊在與人交流,
03:26
Snapchat is that for teenagers,
69
206480
1776
應該知道 Snapchat 就是專門 設計給年輕人使用的,
03:28
and there's, like, a hundred million of them that use it.
70
208280
2696
差不多將近有一億人在使用它。
03:31
And they invented a feature called Snapstreaks,
71
211000
2216
這家公司發明了一個叫做 Snapstreaks 的特色功能,
03:33
which shows the number of days in a row
72
213240
1896
它會告訴你,
你跟你朋友兩個人之間, 連續不間斷地聊了幾天。
03:35
that two people have communicated with each other.
73
215160
2616
03:37
In other words, what they just did
74
217800
1856
換句話說,它們給予的是一種
03:39
is they gave two people something they don't want to lose.
75
219680
2960
兩人都捨不得放棄的東西。
03:44
Because if you're a teenager, and you have 150 days in a row,
76
224000
3456
因為,如果你是年輕人 而有著連續 150 天的聊天紀錄,
03:47
you don't want that to go away.
77
227480
1976
你不會想讓紀錄就此中斷的。
03:49
And so think of the little blocks of time that that schedules in kids' minds.
78
229480
4160
所以想想孩子們 腦袋裡被設定好的時間模式。
03:54
This isn't theoretical: when kids go on vacation,
79
234160
2336
我沒在騙你:已經有人證實, 當孩子在度假時,
03:56
it's been shown they give their passwords to up to five other friends
80
236520
3256
他們會把密碼給另外五位朋友,
03:59
to keep their Snapstreaks going,
81
239800
2216
請他們幫忙維持 Snapstreaks 的聊天記錄,
04:02
even when they can't do it.
82
242040
2016
就算他們不能用手機。
04:04
And they have, like, 30 of these things,
83
244080
1936
他們差不多有 30 種類似 這樣的東西要維護,
04:06
and so they have to get through taking photos of just pictures or walls
84
246040
3376
所以他們每天要東拍拍、西拍拍
04:09
or ceilings just to get through their day.
85
249440
2480
拍牆壁、拍天花板, 不然當天他們會渾身不舒服。
04:13
So it's not even like they're having real conversations.
86
253200
2696
所以這根本不像是 他們在真正的交流。
04:15
We have a temptation to think about this
87
255920
1936
我們可能會這麽想,
04:17
as, oh, they're just using Snapchat
88
257880
2696
他們用 Snapchat 的方式
04:20
the way we used to gossip on the telephone.
89
260600
2016
就像我們曾經用電話聊八卦一樣。
04:22
It's probably OK.
90
262640
1200
應該還好吧!
04:24
Well, what this misses is that in the 1970s,
91
264480
2256
但,不同於 1970 年代的是:
04:26
when you were just gossiping on the telephone,
92
266760
2615
當你們用電話聊八卦時,
04:29
there wasn't a hundred engineers on the other side of the screen
93
269399
3017
旁邊並沒有數百位工程師在監控你,
04:32
who knew exactly how your psychology worked
94
272440
2056
準確地知道你的心理,
04:34
and orchestrated you into a double bind with each other.
95
274520
2640
並操控著你們倆緊緊地綁在一起。
04:38
Now, if this is making you feel a little bit of outrage,
96
278440
3400
如果現在你有點生氣了,
04:42
notice that that thought just comes over you.
97
282680
2576
有沒有注意到,你生氣了?
04:45
Outrage is a really good way also of getting your attention,
98
285280
3320
因為激怒你也是引起你 注意的方式之一,
04:49
because we don't choose outrage.
99
289880
1576
因為就算我們不想生氣,
04:51
It happens to us.
100
291480
1416
它還是會發生。
04:52
And if you're the Facebook newsfeed,
101
292920
1856
如果你是 Facebook 的新聞推送者,
04:54
whether you'd want to or not,
102
294800
1416
不管你是刻意或是不經意,
04:56
you actually benefit when there's outrage.
103
296240
2736
人們憤怒的時候,實際上你是受益的。
04:59
Because outrage doesn't just schedule a reaction
104
299000
2936
因為憤怒不僅僅在 情感上讓你有個宣洩的出口,
05:01
in emotional time, space, for you.
105
301960
2880
更提供了你一個發洩的空間。
05:05
We want to share that outrage with other people.
106
305440
2416
我們還會想和其他人 分享我們的憤怒。
05:07
So we want to hit share and say,
107
307880
1576
所以我們會按下分享鍵然後說,
05:09
"Can you believe the thing that they said?"
108
309480
2040
「你敢相信他們說的嗎?」
05:12
And so outrage works really well at getting attention,
109
312520
3376
所以讓人發怒,可以吸引到注意力,
05:15
such that if Facebook had a choice between showing you the outrage feed
110
315920
3896
所以,如果 Facebook 可以在
向你展示令人憤怒或者 令人平靜的消息之間進行選擇,
05:19
and a calm newsfeed,
111
319840
1320
05:22
they would want to show you the outrage feed,
112
322120
2136
他們會選擇向你展示令人憤怒的消息,
05:24
not because someone consciously chose that,
113
324280
2056
不是因為有人刻意如此選,
05:26
but because that worked better at getting your attention.
114
326360
2680
只是因為這樣可以讓你 更注意到他們。
05:31
And the newsfeed control room is not accountable to us.
115
331120
5480
新聞控制室不對我們負責。
05:37
It's only accountable to maximizing attention.
116
337040
2296
它只對最大化注意力負責。
05:39
It's also accountable,
117
339360
1216
它也要向──
05:40
because of the business model of advertising,
118
340600
2376
因為商業模式的關係,
05:43
for anybody who can pay the most to actually walk into the control room
119
343000
3336
它也要向走進控制室 付廣告費的人負責,
05:46
and say, "That group over there,
120
346360
1576
他們會說,「那個團體在那邊,
05:47
I want to schedule these thoughts into their minds."
121
347960
2640
我想要灌輸一些想法到他們腦裡。」
05:51
So you can target,
122
351760
1200
所以你可以定位,
05:54
you can precisely target a lie
123
354040
1936
你可以準確直接定位到
05:56
directly to the people who are most susceptible.
124
356000
2920
那些最容易被受到影響的人。
06:00
And because this is profitable, it's only going to get worse.
125
360080
2880
因為這是有利可圖的, 所以狀況只會越來越糟。
06:05
So I'm here today
126
365040
1800
所以我今天來到這裡的原因,
06:08
because the costs are so obvious.
127
368160
2000
是因為這件事的代價太高了。
06:12
I don't know a more urgent problem than this,
128
372280
2136
我認為沒有其它事 比這問題還要緊急,
06:14
because this problem is underneath all other problems.
129
374440
3120
因為其它所有問題的背後, 都與這個問題有關。
06:18
It's not just taking away our agency
130
378720
3176
它不僅剝奪我們的自主權,
06:21
to spend our attention and live the lives that we want,
131
381920
2600
還浪費我們的注意力, 影響我們的生活方式。
06:25
it's changing the way that we have our conversations,
132
385720
3536
也改變了我們的交流方式,
06:29
it's changing our democracy,
133
389280
1736
改變了我們的民主制度,
06:31
and it's changing our ability to have the conversations
134
391040
2616
而且還改變了我們 想要與他人交流、
06:33
and relationships we want with each other.
135
393680
2000
維持關係的能力。
06:37
And it affects everyone,
136
397160
1776
這會影響到每一個人,
06:38
because a billion people have one of these in their pocket.
137
398960
3360
因為十幾億人口的口袋裡 都有一台這個東西。
06:45
So how do we fix this?
138
405360
1400
所以我們要如何修復這個問題?
06:49
We need to make three radical changes
139
409080
2936
我們需要對科技和我們的社會
06:52
to technology and to our society.
140
412040
1800
做三大激進的改變。
06:55
The first is we need to acknowledge that we are persuadable.
141
415720
3800
首先,我們需要了解 我們是會被說服的。
07:00
Once you start understanding
142
420840
1376
一旦你了解
07:02
that your mind can be scheduled into having little thoughts
143
422240
2776
人腦可以被有計劃性地 注入一些思想
07:05
or little blocks of time that you didn't choose,
144
425040
2576
或被占用掉一些時間時,
07:07
wouldn't we want to use that understanding
145
427640
2056
我們難道不會利用這點
07:09
and protect against the way that that happens?
146
429720
2160
來防止這樣的事發生嗎?
07:12
I think we need to see ourselves fundamentally in a new way.
147
432600
3296
我認為我們需要以全新的方式審視自己。
07:15
It's almost like a new period of human history,
148
435920
2216
就像是人類歷史上新的篇章,
07:18
like the Enlightenment,
149
438160
1216
就像啟蒙運動,
07:19
but almost a kind of self-aware Enlightenment,
150
439400
2216
但是是自覺性的啟蒙運動,
07:21
that we can be persuaded,
151
441640
2080
瞭解到我們是會被說服的,
07:24
and there might be something we want to protect.
152
444320
2240
並意識到我們也有想要保護的東西。
07:27
The second is we need new models and accountability systems
153
447400
4576
第二點是我們需要新的 模式和責任系統,
07:32
so that as the world gets better and more and more persuasive over time --
154
452000
3496
如此,當世界變得越來越好、 我們越來越容易被說服時──
07:35
because it's only going to get more persuasive --
155
455520
2336
因為它只會變得更有說服力──
07:37
that the people in those control rooms
156
457880
1856
這樣在控制室裡的那些人
07:39
are accountable and transparent to what we want.
157
459760
2456
才會對我們想要的東西負責 並把它透明化。
07:42
The only form of ethical persuasion that exists
158
462240
2696
道德說服存在的唯一前提就是:
07:44
is when the goals of the persuader
159
464960
1936
只有當說服者的目標
07:46
are aligned with the goals of the persuadee.
160
466920
2200
和被說服者的目標是一致時才存在。
07:49
And that involves questioning big things, like the business model of advertising.
161
469640
3840
而這涉及到對大型事件的質疑, 像是廣告的商業模式。
07:54
Lastly,
162
474720
1576
最後,
07:56
we need a design renaissance,
163
476320
1680
我們需要一個經過設計的 科技文藝復興,
07:59
because once you have this view of human nature,
164
479080
3056
因為一旦你對人性本質 有一定的了解,
08:02
that you can steer the timelines of a billion people --
165
482160
2976
那你就可以控制十億人的時間軸──
08:05
just imagine, there's people who have some desire
166
485160
2736
想像一下,人都會有一些慾望,
08:07
about what they want to do and what they want to be thinking
167
487920
2856
會有想做的事、想思考的事、
08:10
and what they want to be feeling and how they want to be informed,
168
490800
3136
想感受的事物、想了解的事物,
而我們卻只能被引導到其它方向。
08:13
and we're all just tugged into these other directions.
169
493960
2536
十億人只會被引導到這些不同的方向。
08:16
And you have a billion people just tugged into all these different directions.
170
496520
3696
所以,試想一下 一個完整的文藝復興設計,
08:20
Well, imagine an entire design renaissance
171
500240
2056
08:22
that tried to orchestrate the exact and most empowering
172
502320
3096
可以幫助我們引導至 正確的、有自主性的、
08:25
time-well-spent way for those timelines to happen.
173
505440
3136
時間分配良好的路上。
08:28
And that would involve two things:
174
508600
1656
那就得包含兩件事:
08:30
one would be protecting against the timelines
175
510280
2136
一是我們要保護自己的時間軸
08:32
that we don't want to be experiencing,
176
512440
1856
不被支配到不想經歷的事情上、
08:34
the thoughts that we wouldn't want to be happening,
177
514320
2416
不去產生我們不想要的想法,
08:36
so that when that ding happens, not having the ding that sends us away;
178
516760
3336
如此,當簡訊的提醒聲響起時, 我們就不會被牽著鼻子走;
08:40
and the second would be empowering us to live out the timeline that we want.
179
520120
3616
二是要讓我們能活出 自己想要的時光。
08:43
So let me give you a concrete example.
180
523760
1880
我給各位舉個例子。
08:46
Today, let's say your friend cancels dinner on you,
181
526280
2456
比如說,今天你朋友取消了 與你的晚餐約會,
08:48
and you are feeling a little bit lonely.
182
528760
3775
所以你感到有點寂寞。
08:52
And so what do you do in that moment?
183
532559
1817
那當下你會做甚麼?
08:54
You open up Facebook.
184
534400
1279
你會打開 Facebook。
08:56
And in that moment,
185
536960
1696
而就在那一刻,
08:58
the designers in the control room want to schedule exactly one thing,
186
538680
3376
控制室裡的設計者 想要準確地規劃一件事,
09:02
which is to maximize how much time you spend on the screen.
187
542080
3040
那就是最大化你盯著螢幕的時間。
09:06
Now, instead, imagine if those designers created a different timeline
188
546640
3896
現在,反過來,想像一下, 是否這些設計師可以創造出一個
09:10
that was the easiest way, using all of their data,
189
550560
3496
不一樣的時間軸,以最簡單的方法、 利用他們所有的數據,
09:14
to actually help you get out with the people that you care about?
190
554080
3096
幫你準確地約出你關心的人?
09:17
Just think, alleviating all loneliness in society,
191
557200
5416
試想一下,消除社會中所有的寂寞,
09:22
if that was the timeline that Facebook wanted to make possible for people.
192
562640
3496
這樣的時間軸不就是 Facebook 想要為我們實現的嗎?
09:26
Or imagine a different conversation.
193
566160
1715
或試想另一個對話。
09:27
Let's say you wanted to post something supercontroversial on Facebook,
194
567899
3317
比方說你想要在 Facebook 上 發表備受爭議的言論,
09:31
which is a really important thing to be able to do,
195
571240
2416
你覺得這個爭議性話題很重要,
09:33
to talk about controversial topics.
196
573680
1696
需要被拿出來討論。
09:35
And right now, when there's that big comment box,
197
575400
2336
現在,有一個很大的留言區,
09:37
it's almost asking you, what key do you want to type?
198
577760
3376
它就像是在問你, 你想要輸入什麽東西?
09:41
In other words, it's scheduling a little timeline of things
199
581160
2816
換句話說,它正在安排一些時間軸,
09:44
you're going to continue to do on the screen.
200
584000
2136
好讓你可以繼續待在螢幕上。
09:46
And imagine instead that there was another button there saying,
201
586160
2976
試想如果有另一個按鈕跳出來說,
09:49
what would be most time well spent for you?
202
589160
2056
你想怎麼安排你的時間?
09:51
And you click "host a dinner."
203
591240
1576
然後你點選,「舉辦一個晚餐的聚會」。
09:52
And right there underneath the item it said,
204
592840
2096
然後,底下就會跳出一個,
09:54
"Who wants to RSVP for the dinner?"
205
594960
1696
「有誰想要聚餐,請盡速回覆」?
09:56
And so you'd still have a conversation about something controversial,
206
596680
3256
所以,你的爭議性話題 可以被繼續討論,
09:59
but you'd be having it in the most empowering place on your timeline,
207
599960
3736
而且可以放置在你時間軸上 最顯眼的位置,
10:03
which would be at home that night with a bunch of a friends over
208
603720
3016
那天晚上,你就可以邀請到一堆朋友
10:06
to talk about it.
209
606760
1200
來你家裡晚餐並討論這個話題。
10:09
So imagine we're running, like, a find and replace
210
609000
3160
想像我們正在賽跑, 想盡快找到並替換掉
10:13
on all of the timelines that are currently steering us
211
613000
2576
所有那些正在促使我們
10:15
towards more and more screen time persuasively
212
615600
2560
花越來越多的時間在螢幕上的時間軸,
10:19
and replacing all of those timelines
213
619080
2536
並盡快把這些時間軸
10:21
with what do we want in our lives.
214
621640
1640
用我們在生活中想做的事情替代掉。
10:26
It doesn't have to be this way.
215
626960
1480
真的不須要這樣。
10:30
Instead of handicapping our attention,
216
630360
2256
不需要癱瘓我們的注意力,
10:32
imagine if we used all of this data and all of this power
217
632640
2816
試想如果我們利用所有這些數據和能力,
10:35
and this new view of human nature
218
635480
1616
加上對人性本質的全新認識,
10:37
to give us a superhuman ability to focus
219
637120
2856
來讓我們擁有超人般的注意力、
10:40
and a superhuman ability to put our attention to what we cared about
220
640000
4136
讓我們更關心我們在乎的事情、
10:44
and a superhuman ability to have the conversations
221
644160
2616
讓我們擁有超人般的能力,
10:46
that we need to have for democracy.
222
646800
2000
來進行民主所需要的互動交流。
10:51
The most complex challenges in the world
223
651600
2680
世界上最複雜的挑戰,
10:56
require not just us to use our attention individually.
224
656280
3120
不僅需要我們每個人的注意力。
11:00
They require us to use our attention and coordinate it together.
225
660440
3320
也需要我們的同心協力才能克服。
11:04
Climate change is going to require that a lot of people
226
664440
2816
地球暖化議題需要大家
11:07
are being able to coordinate their attention
227
667280
2096
一起使用最有力的方式
11:09
in the most empowering way together.
228
669400
1896
將所有人的注意力整合起來。
11:11
And imagine creating a superhuman ability to do that.
229
671320
3080
試想如果有了這樣的超人能力會怎樣。
11:19
Sometimes the world's most pressing and important problems
230
679000
4160
有時世界上最要緊、最重要的問題
11:24
are not these hypothetical future things that we could create in the future.
231
684040
3840
不是未來我們可以創造的 假設性事物。
11:28
Sometimes the most pressing problems
232
688560
1736
有時最要緊的問題,
11:30
are the ones that are right underneath our noses,
233
690320
2336
就是我們眼前的問題,
11:32
the things that are already directing a billion people's thoughts.
234
692680
3120
已經在影響著十億人思想的事情。
11:36
And maybe instead of getting excited about the new augmented reality
235
696600
3376
與其花時間對擴增實境感到興奮、
11:40
and virtual reality and these cool things that could happen,
236
700000
3296
對虛擬實境這類酷炫產品感到興奮──
11:43
which are going to be susceptible to the same race for attention,
237
703320
3296
對這些注意力競賽的產品感到興奮──
11:46
if we could fix the race for attention
238
706640
2176
不如把時間放在修正注意力競賽上,
11:48
on the thing that's already in a billion people's pockets.
239
708840
2720
修正十億人口袋裡的那台機器上。
11:52
Maybe instead of getting excited
240
712040
1576
與其花時間對刺激、
11:53
about the most exciting new cool fancy education apps,
241
713640
4176
酷炫的教育軟體感到興奮,
11:57
we could fix the way kids' minds are getting manipulated
242
717840
2896
不如花時間找出方法, 來挽救那些已經被操控、
12:00
into sending empty messages back and forth.
243
720760
2480
腦中只想傳些空洞簡訊的孩子們。
12:04
(Applause)
244
724040
4296
(掌聲)
12:08
Maybe instead of worrying
245
728360
1256
與其擔憂那假想的未來:
12:09
about hypothetical future runaway artificial intelligences
246
729640
3776
只為了吸引我們注意力的 人工智慧未來,
12:13
that are maximizing for one goal,
247
733440
1880
12:16
we could solve the runaway artificial intelligence
248
736680
2656
倒不如開始解決我們現有的、
12:19
that already exists right now,
249
739360
2056
已經失去控制的人工智慧,
12:21
which are these newsfeeds maximizing for one thing.
250
741440
2920
也就是這些為了吸引我們注意力 推送新聞的人工智慧機器。
12:26
It's almost like instead of running away to colonize new planets,
251
746080
3816
就像是,與其逃跑到新的殖民星球,
12:29
we could fix the one that we're already on.
252
749920
2056
不如好好地拯救我們的地球。
12:32
(Applause)
253
752000
4120
(掌聲)
12:40
Solving this problem
254
760040
1776
解決這個問題
12:41
is critical infrastructure for solving every other problem.
255
761840
3800
是解決其它問題的關鍵所在。
12:46
There's nothing in your life or in our collective problems
256
766600
4016
在你生命中或是我們彼此的共同問題中,
12:50
that does not require our ability to put our attention where we care about.
257
770640
3560
沒有一件事是不需要我們 花時間關注、花心思在乎的。
12:55
At the end of our lives,
258
775800
1240
畢竟,到了生命的盡頭,
12:58
all we have is our attention and our time.
259
778240
2640
我們最後所擁有的 就是曾經在乎的美好時光。
13:01
What will be time well spent for ours?
260
781800
1896
讓時間為我們所用。
13:03
Thank you.
261
783720
1216
13:04
(Applause)
262
784960
3120
謝謝各位。
(掌聲)
13:17
Chris Anderson: Tristan, thank you. Hey, stay up here a sec.
263
797760
2936
克里斯 · 安德森: 崔斯頓,謝謝你。請留步。
13:20
First of all, thank you.
264
800720
1336
首先,謝謝你。
13:22
I know we asked you to do this talk on pretty short notice,
265
802080
2776
我知道我們很晚才通知你要做這次演講,
13:24
and you've had quite a stressful week
266
804880
2216
你這禮拜壓力很大,
13:27
getting this thing together, so thank you.
267
807120
2440
為了這次的演講,所以謝謝你。
13:30
Some people listening might say, what you complain about is addiction,
268
810680
3976
有些聽眾可能會說, 你抱怨的是沉迷,
13:34
and all these people doing this stuff, for them it's actually interesting.
269
814680
3496
而對於那些已經沉迷的人來說, 他們認為那是興趣。
13:38
All these design decisions
270
818200
1256
所有這些設計過的決策、內容,
13:39
have built user content that is fantastically interesting.
271
819480
3096
對用戶而言是相當有趣的。
13:42
The world's more interesting than it ever has been.
272
822600
2416
這個世界從未如此有趣過。
13:45
What's wrong with that?
273
825040
1256
這有什麽錯嗎?
13:46
Tristan Harris: I think it's really interesting.
274
826320
2256
崔斯頓.哈瑞斯:我認為這確實有趣。
13:48
One way to see this is if you're just YouTube, for example,
275
828600
4016
我們換個角度看這件事, 假設,你在用 Youtube 看影片,
13:52
you want to always show the more interesting next video.
276
832640
2656
你總是希望下一部影片是最有趣的。
13:55
You want to get better and better at suggesting that next video,
277
835320
3016
你希望下一部推薦的影片越來越棒,
但是即使你推薦了一部比一部好看, 所有人都想要看的完美影片,
13:58
but even if you could propose the perfect next video
278
838360
2456
14:00
that everyone would want to watch,
279
840840
1656
這只會讓你一直盯著螢幕看。
14:02
it would just be better and better at keeping you hooked on the screen.
280
842520
3336
而其中遺漏的是要找出 我們的邊界。對吧?
14:05
So what's missing in that equation
281
845880
1656
14:07
is figuring out what our boundaries would be.
282
847560
2136
14:09
You would want YouTube to know something about, say, falling asleep.
283
849720
3216
你會想讓 YouTube 知道, 你甚麼時候會「睡著」嗎?
Netflix 的 CEO 最近說,
14:12
The CEO of Netflix recently said,
284
852960
1616
我們最大的競爭者就是臉書、 YouTube 和「睡著了」,對吧?
14:14
"our biggest competitors are Facebook, YouTube and sleep."
285
854600
2736
14:17
And so what we need to recognize is that the human architecture is limited
286
857360
4456
所以我們需要認識到 人體是有極限的、
14:21
and that we have certain boundaries or dimensions of our lives
287
861840
2976
我們的生活是有某種界線的、
14:24
that we want to be honored and respected,
288
864840
1976
我們都想要被尊重,
14:26
and technology could help do that.
289
866840
1816
而科技可以幫我們做到。
14:28
(Applause)
290
868680
2616
(掌聲)
14:31
CA: I mean, could you make the case
291
871320
1696
克里斯:可否請你說明一下,
14:33
that part of the problem here is that we've got a naïve model of human nature?
292
873040
6056
這裡還有個問題是,會不會 我們把人性想的太天真了?
14:39
So much of this is justified in terms of human preference,
293
879120
2736
如何判別各式各樣 人類偏好的合理性,
14:41
where we've got these algorithms that do an amazing job
294
881880
2616
我們可以用這些演算法 幫我們完成,
14:44
of optimizing for human preference,
295
884520
1696
幫我們優化人類的偏好,
14:46
but which preference?
296
886240
1336
但是,是哪方面的偏好?
14:47
There's the preferences of things that we really care about
297
887600
3496
有我們確實關心的、
14:51
when we think about them
298
891120
1376
在意的事情的偏好,
14:52
versus the preferences of what we just instinctively click on.
299
892520
3056
也有我們只是直覺地點擊的偏好。
14:55
If we could implant that more nuanced view of human nature in every design,
300
895600
4656
如果我們在每個設計中 植入了對人性本質的微妙了解,
15:00
would that be a step forward?
301
900280
1456
這會是一種進步嗎?
15:01
TH: Absolutely. I mean, I think right now
302
901760
1976
崔斯頓:那是肯定的。 我的意思是,我認為現在
15:03
it's as if all of our technology is basically only asking our lizard brain
303
903760
3496
好像我們的科技基本上
只詢問我們本能反應的腦的意見,
15:07
what's the best way to just impulsively get you to do
304
907280
2496
它們最好的方式就是強迫你
15:09
the next tiniest thing with your time,
305
909800
2136
在下一秒、下一刻做出一些小事,
15:11
instead of asking you in your life
306
911960
1656
而不是問你人生中的大事、
15:13
what we would be most time well spent for you?
307
913640
2176
問你花時間在哪方面 才是對你有幫助的?
15:15
What would be the perfect timeline that might include something later,
308
915840
3296
也不是問你接下來 完美的時間安排是怎樣的,
最後一天你要不要 聽一下 TED 演講之類的?
15:19
would be time well spent for you here at TED in your last day here?
309
919160
3176
克里斯: 所以,是不是要臉書、 谷歌在一開始就問我們,
15:22
CA: So if Facebook and Google and everyone said to us first up,
310
922360
2976
15:25
"Hey, would you like us to optimize for your reflective brain
311
925360
2896
嘿,你想要我們優化思考的腦 還是優化本能反應的腦?
15:28
or your lizard brain? You choose."
312
928280
1656
由你來選擇。
15:29
TH: Right. That would be one way. Yes.
313
929960
2080
崔斯頓:是的。這也是個方法。是的。
15:34
CA: You said persuadability, that's an interesting word to me
314
934358
2858
克里斯:你剛提到的「說服能力」, 對我來說這個詞很有趣,
15:37
because to me there's two different types of persuadability.
315
937240
2856
因為在我看來, 有兩種不同的說服能力。
15:40
There's the persuadability that we're trying right now
316
940120
2536
有一種說服能力是我們現在在嘗試的,
15:42
of reason and thinking and making an argument,
317
942680
2176
用前因後果、提出看法的方式說服聽眾,
15:44
but I think you're almost talking about a different kind,
318
944880
2696
但是我覺得你是在談論另一種
15:47
a more visceral type of persuadability,
319
947590
1906
更能引起「情緒上本能反應」 的說服能力,
15:49
of being persuaded without even knowing that you're thinking.
320
949520
2896
那種不經思考就被說服的能力。
沒錯。我會十分關心這個問題的原因是
15:52
TH: Exactly. The reason I care about this problem so much is
321
952440
2856
我在史丹佛的說服力技術實驗室待過,
15:55
I studied at a lab called the Persuasive Technology Lab at Stanford
322
955320
3176
那裡就是在教導人們這些技術。
15:58
that taught [students how to recognize] exactly these techniques.
323
958520
2546
那裡有論壇和討論,教導人們 使用這些偷偷摸摸的方式
16:01
There's conferences and workshops that teach people all these covert ways
324
961106
2990
16:04
of getting people's attention and orchestrating people's lives.
325
964120
2976
來獲得人們的注意力, 指揮人們的生活。
16:07
And it's because most people don't know that that exists
326
967120
2656
正因為大部分人不知道 有這項技術的存在,
16:09
that this conversation is so important.
327
969800
1896
所以我們的討論才會如此重要。
16:11
CA: Tristan, you and I, we both know so many people from all these companies.
328
971720
3776
克里斯:崔斯頓,你和我都認識 許多來自這些公司的人。
16:15
There are actually many here in the room,
329
975520
1976
許多人也在這裡,
16:17
and I don't know about you, but my experience of them
330
977520
2477
我不知道你是怎麽樣的, 但就我的經驗而言,
16:20
is that there is no shortage of good intent.
331
980021
2075
他們也都是善意的。
16:22
People want a better world.
332
982120
2176
大家都想要一個更好的世界。
16:24
They are actually -- they really want it.
333
984320
3520
他們確實──他們真的想要。
16:28
And I don't think anything you're saying is that these are evil people.
334
988320
4176
我不認為你所說的, 是在指這些人是壞人。
16:32
It's a system where there's these unintended consequences
335
992520
3696
你指的是有一個系統, 導致了一些意想不到的後果
16:36
that have really got out of control --
336
996240
1856
且超出了控制範圍──
16:38
TH: Of this race for attention.
337
998120
1496
崔斯頓:關於注意力的競爭。
16:39
It's the classic race to the bottom when you have to get attention,
338
999640
3176
當你想要吸引人家的注意 它就是一場無視規則的精典賽,
16:42
and it's so tense.
339
1002840
1216
而且競爭很激烈。
16:44
The only way to get more is to go lower on the brain stem,
340
1004080
2736
取得更多注意的唯一辦法, 只有更深入腦袋、
16:46
to go lower into outrage, to go lower into emotion,
341
1006840
2416
更引起憤怒、更深入情感、
16:49
to go lower into the lizard brain.
342
1009280
1696
更深入本能反應的腦。
16:51
CA: Well, thank you so much for helping us all get a little bit wiser about this.
343
1011000
3816
克里斯:非常感謝你幫助我們 對這個問題有更進一步的了解。
16:54
Tristan Harris, thank you. TH: Thank you very much.
344
1014840
2416
謝謝你。
崔斯頓:非常感謝。
16:57
(Applause)
345
1017280
2240
(掌聲)
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7