TikTok CEO Shou Chew on Its Future — and What Makes Its Algorithm Different | Live at TED2023

2,036,568 views

2023-04-21 ・ TED


New videos

TikTok CEO Shou Chew on Its Future — and What Makes Its Algorithm Different | Live at TED2023

2,036,568 views ・ 2023-04-21

TED


請雙擊下方英文字幕播放視頻。

00:00
Chris Anderson: It's very nice to have you here.
0
288
2294
譯者: 至磊Zi Le 黃Ng 審譯者: Helen Chang
克里斯.安德森: 很高興你來到這裡。
00:03
Let's see.
1
3166
1126
我們來看看,首先,恭喜你,
00:04
First of all, congratulations.
2
4333
1460
00:05
You really pulled off something remarkable on that grilling,
3
5793
2836
你在那場審問當中 做了一件很了不起的事,
00:08
you achieved something that very few people do,
4
8629
2211
這件事很少人能做到:
00:10
which was, you pulled off a kind of, a bipartisan consensus in US politics.
5
10882
4671
你讓美國政壇出現跨黨派的共識。
00:15
It was great.
6
15553
1168
這是好事。
00:16
(Laughter)
7
16763
1376
(笑聲)
00:18
The bad news was that that consensus largely seemed to be:
8
18723
2753
壞消息是這個共識看起來很像是
00:21
"We must ban TikTok."
9
21517
2294
「我們必須禁用抖音。」
00:23
So we're going to come to that in a bit.
10
23853
1960
我們等一下就會來談這件事。
00:25
And I'm curious, but before we go there, we need to know about you.
11
25813
4588
但在那之前,我對你很好奇, 我們必須了解你。
00:30
You seem to me like a remarkable person.
12
30443
2878
對我而言你看起來很了不起。
00:33
I want to know a bit of your story
13
33362
1627
我想知道你的故事, 還有你起初是如何加入抖音的。
00:35
and how you came to TikTok in the first place.
14
35031
2627
00:37
Shou Chew: Thank you, Chris.
15
37658
1377
周受資:謝謝你,克里斯。
00:39
Before we do that, can I just check, need to know my audience,
16
39035
3003
我可不可以了解一下, 我需要了解我的聽眾,
00:42
how many of you here use TikTok?
17
42080
1876
你們當中有多少人在用抖音?
喔,謝謝你們。
00:45
Oh, thank you.
18
45124
1168
沒在用的人,這裡有免費上網。
00:46
For those who don’t, the Wi-Fi is free.
19
46292
2085
00:48
(Laughter)
20
48419
1919
(笑聲)
00:51
CA: There’s another question, which is,
21
51631
1877
克:還有一個問題是,
00:53
how many of you here have had your lives touched through TikTok,
22
53508
3169
你們有多少人的生命曾被抖音觸動,
00:56
through your kids and other people in your lives?
23
56719
2461
被你的兒女和生命中其他人觸動?
01:00
SC: Oh, that's great to see.
24
60515
1376
周:看起來真好。
01:01
CA: It's basically, if you're alive,
25
61891
1752
克:基本上只要你活著,
01:03
you have had some kind of contact with TikTok at this point.
26
63643
3211
此時此刻你一定曾與抖音 有過某種連結。
01:06
So tell us about you.
27
66854
1418
告訴我們有關你的事。
01:08
SC: So my name is Shou, and I’m from Singapore.
28
68272
2920
周:我叫受資,來自新加坡。
01:11
Roughly 10 years ago,
29
71651
1668
大約十年前,
01:13
I met with two engineers who were building a product.
30
73361
4004
我遇到兩位工程師, 他們正在打造一個專案。
01:17
And the idea behind this was to build a product
31
77365
2961
這個專案背後的想法 是打造一樣產品,
01:20
that recommended content to people not based on who they knew,
32
80368
4796
把內容推薦給眾人, 但不是基於他們認識的人。
01:25
which was, if you think about it, 10 years ago,
33
85206
3003
如果你想一想,十年前,
01:28
the social graph was all in the rage.
34
88209
3045
社交圖譜的想法非常流行。
01:31
And the idea was, you know,
35
91295
1293
社交圖譜的想法是
01:32
your content and the feed that you saw should be based on people that you knew.
36
92588
3921
你所看見的內容 應該基於你認識的人。
01:36
But 10 years ago,
37
96926
1168
但十年前,
01:38
these two engineers thought about something different,
38
98136
2544
這兩位工程師的想法有所不同,
01:40
which is, instead of showing you --
39
100680
3545
與其把你認識的人展示給你,
01:44
instead of showing you people you knew,
40
104225
2461
01:46
why don't we show you content that you liked?
41
106727
2253
何不展示你喜歡的內容?
01:49
And that's sort of the genesis and the birth
42
109021
2211
這某種程度上就是 抖音早期版本的起源。
01:51
of the early iterations of TikTok.
43
111232
2794
01:54
And about five years ago,
44
114443
1585
而在大約五年前,
01:56
with the advent of 4G, short video, mobile phone penetration,
45
116070
5798
隨著 4G 、短影音、 手機普及的來臨,
02:01
TikTok was born.
46
121868
1376
抖音誕生了。
02:03
And a couple of years ago,
47
123286
1293
又在幾年前,
02:04
I had the opportunity to run this company,
48
124620
2419
我有機會經營這家公司,
02:07
and it still excites me every single day.
49
127039
2002
直到現在它依然每天振奮我的心。
02:09
CA: So I want to dig in a little more into this,
50
129750
2253
克:我想要了解得更深入一些,
02:12
about what was it that made this take-off so explosive?
51
132044
2962
是什麼讓抖音的飛躍如此振撼?
02:15
Because the language I hear from people who spent time on it,
52
135047
3963
因為我聽那些抖音使用者所說,
02:19
it's sort of like I mean,
53
139051
2128
我覺得在某種程度上,
02:21
it is a different level of addiction to other media out there.
54
141220
5965
那上癮的程度 跟其他社群媒體截然不同。
02:27
And I don't necessarily mean this in a good way, we'll be coming on to it.
55
147226
3504
我不只是在講好的一面, 我們等一下會聊。
02:30
There’s good and bad things about this type of addiction.
56
150730
2711
這種上癮的影響有好有壞。
我是說一種感覺:
02:33
But it’s the feeling
57
153441
1168
不過用了抖音短短幾天,
02:34
that within a couple of days of experience of TikTok,
58
154650
2503
02:37
it knows you and it surprises you
59
157195
2502
抖音就認識並驚豔了你,
02:39
with things that you didn't know you were going to be interested in,
60
159739
3211
推薦你以前不知道 你會有興趣的事情,
02:42
but you are.
61
162950
1293
但你真的有興趣。
02:44
How?
62
164243
1210
這是怎麼做到的?
02:45
Is it really just, instead of the social graph --
63
165453
2627
不靠社交圖譜,
02:48
What are these algorithms doing?
64
168080
2044
這些演算法在做些什麼?
02:50
SC: I think to describe this, to begin to answer your question,
65
170124
4171
周:我認為要開始回答你的問題,
02:54
we have to talk about the mission of the company.
66
174337
2502
我們必須從這家企業的使命談起。
02:56
Now the mission is to inspire creativity and to bring joy.
67
176839
3504
我們的使命是刺激創意並帶來快樂。
03:00
And I think missions for companies like ours [are] really important.
68
180384
3629
我認為對像我們這種企業而言, 使命是非常重要的。
03:04
Because you have product managers working on the product every single day,
69
184055
3545
因為產品經理每天 都在為這些產品工作,
03:07
and they need to have a North Star, you know,
70
187600
2127
他們需要一顆北極星,
03:09
something to sort of, work towards together.
71
189769
2586
要有某事讓大家一同前行。
03:12
Now, based on this mission,
72
192355
1543
基於這項使命,
03:13
our vision is to provide three things to our users.
73
193898
3378
我們的願景是 提供三樣東西給使用者。
03:17
We want to provide a window to discover,
74
197318
2127
我們想提供發現新事物的窗口,
03:19
and I’ll talk about discovery, you talked about this, in a second.
75
199445
3170
過一會兒我就會談「發現」, 你剛剛提到的。
03:22
We want to give them a canvas to create,
76
202657
2085
我們想提供一塊創作的畫布,
03:24
which is going to be really exciting with new technologies in AI
77
204784
3879
這非常振奮人心,
人工智慧的新科技 會幫助人們創造新事物。
03:28
that are going to help people create new things.
78
208704
3587
03:32
And the final thing is bridges for people to connect.
79
212792
2919
最後一個是人與人建立連結的橋梁。
03:35
So that's sort of the vision of what we're trying to build.
80
215753
2794
這就是我們試著打造的願景。
03:38
Now what really makes TikTok very unique and very different
81
218547
3754
而真正讓抖音獨一無二、與眾不同的
03:42
is the whole discovery engine behind it.
82
222343
3003
是它背後的整個「發現引擎」。
03:45
So there are earlier apps that I have a lot of respect for,
83
225721
2878
我很敬重一些早期的應用程式,
03:48
but they were built for a different purpose.
84
228641
2127
但它們都是為其他目的而打造。
03:50
For example, in the era of search, you know,
85
230768
3128
例如在搜尋的年代,
03:53
there was an app that was built for people who wanted to search things
86
233896
3504
有個應用程式是為了 想要搜尋的人而打造,
03:57
so that is more easily found.
87
237441
3420
讓人更容易找到資訊。
04:00
And then in the era of social graphs,
88
240903
1794
在社交圖譜的年代,
04:02
it was about connecting people and their followers.
89
242697
2586
那是要在人與其追蹤者間建立連結。
04:05
Now what we have done is that ... based on our machine-learning algorithms,
90
245283
4212
而我們做到的事情是, 基於我們的機器學習演算法,
04:09
we're showing people what they liked.
91
249578
2294
我們讓人們看到他們喜歡的東西。
04:11
And what this means is that we have given the everyday person
92
251872
3087
這就代表我們給了平凡人 一個被發現的平臺。
04:15
a platform to be discovered.
93
255001
1918
04:16
If you have talent, it is very, very easy to get discovered on TikTok.
94
256919
3962
如果你有才華, 非常容易能在抖音被發現。
04:20
And I'll just give you one example of this.
95
260923
2544
我將會給你一個實例。
04:23
The biggest creator on TikTok is a guy called Khaby.
96
263467
3629
抖音上最大的創作者 是一個叫卡比的傢伙。
04:27
Khaby was from Senegal,
97
267430
1751
卡比來自塞內加爾,
04:29
he lives in Italy, he was a factory worker.
98
269181
2711
住在義大利,是一位工廠勞工。
04:32
He, for the longest time, didn't even speak in any of his videos.
99
272685
4338
甚至有很長一段時間裡 他沒在任何影片裡講過一句話。
04:37
But what he did was he had talent.
100
277064
2336
但他有才華。
04:39
He was funny, he had a good expression,
101
279442
2252
他很好笑,有很強的表達力,
04:41
he had creativity, so he kept posting.
102
281694
2461
他也有創造力, 所以不停發布新內容。
04:44
And today he has 160 million followers on our platform.
103
284196
3587
今天,他在我們的平臺上 有 1.6 億位追蹤者。
04:47
So every single day we hear stories like that,
104
287825
2169
每一天,我們都聽說這樣的故事,
04:49
businesses, people with talent.
105
289994
2252
聽說才華出眾的事業與人。
04:52
And I think it's very freeing to have a platform
106
292288
3336
這樣的一個平臺讓人擁有自由,
04:55
where, as long as you have talent, you're going to be heard
107
295666
3128
在這裡,只要你有才華, 你就可以被聽見,
04:58
and you have the chance to succeed.
108
298794
1877
你就有機會成功。
05:00
And that's what we're providing to our users.
109
300713
2210
而這就是我們帶給使用者的事物。
05:02
CA: So this is the amazing thing to me.
110
302923
1919
克:這就是讓我驚豔的地方。
05:04
Like, most of us have grown up with, say, network television,
111
304884
4337
我們大多數人成長於 電視聯播網的時代,
05:09
where, for decades you've had thousands of brilliant, creative people
112
309263
3587
數十年來成千上萬聰明、有創意的人
05:12
toiling in the trenches,
113
312892
1168
絞盡腦汁,
05:14
trying to imagine stuff that will be amazing for an audience.
114
314101
4380
試圖發想可以驚豔觀眾的東西。
05:18
And none of them ever remotely came up with anything
115
318522
4255
但沒有任何人想出
05:22
that looked like many of your creators.
116
322818
3921
像你們的創作者一樣的事物。
05:27
So these algorithms,
117
327114
2294
因此這些演算法
05:29
just by observing people's behavior and what they look like,
118
329450
3921
只是藉由觀察人們的行為和外表,
05:33
have discovered things that thousands of brilliant humans never discovered.
119
333412
3587
就發掘出成千上萬聰明人 未曾發現的事物。
05:37
Tell me some of the things that it is looking at.
120
337041
3045
告訴我演算法會觀察哪些資訊。
05:40
So obvious things, like if someone presses like
121
340086
2502
有些比較明顯,像是如果有人按讚,
05:42
or stays on a video for a long time,
122
342630
1877
或是在一部影片上逗留很久,
05:44
that gives you a clue, "more like that."
123
344548
2545
就給了你們線索: 「多點那種東西。」
05:47
But is it subject matter?
124
347134
1210
那題材呢?
05:48
What are the array of things
125
348386
2377
到底有哪些東西
05:50
that you have noticed that you can actually track
126
350805
2419
是你們注意到真的能加以追蹤,
05:53
that provide useful clues?
127
353224
2043
提供有用的線索的?
05:55
SC: I'm going to simplify this a lot,
128
355309
2461
周:我會解釋得非常簡化,
05:57
but the machine learning, the recommendation algorithm
129
357812
2877
但機器學習、提供推薦的演算法
06:00
is really just math.
130
360731
1377
其實就只是數學。
06:02
So, for example, if you liked videos one, two, three and four,
131
362108
4587
舉例來說,如果你 幫影片 1、2、3、4 按讚,
06:06
and I like videos one, two, three and five,
132
366737
2211
而我幫影片 1、2、3、5 按讚,
06:08
maybe he liked videos one, two, three and six.
133
368989
2253
也許他幫影片 1、2、3、6 按讚。
06:11
Now what's going to happen is,
134
371242
1501
那接下來會發生的是,
06:12
because we like one, two, three at the same time,
135
372743
2336
因為我們都對 1、2、3 按了讚,
06:15
he's going to be shown four, five, six, and so are we.
136
375121
2877
他就會看到 4、5、6 ,我們也是。
06:18
And you can think about this repeated at scale in real time
137
378040
3379
你可以想像這件事 即時地大規模重複進行,
06:21
across more than a billion people.
138
381460
1877
橫跨超過十億人。
06:23
That's basically what it is, it's math.
139
383337
2044
基本上它就是這樣,這是數學。
06:25
And of course, you know,
140
385423
1501
當然,你知道,
06:26
AI and machine learning has allowed this to be done
141
386966
2794
人工智慧和機器學習讓這件事
能以極其巨大的規模進行。
06:29
at a very, very big scale.
142
389802
1710
06:31
And what we have seen, the result of this,
143
391554
2085
而我們所看到的結果
06:33
is that it learns the interest signals
144
393681
2669
是它很快就學會 人們流露的興趣訊號,
06:36
that people exhibit very quickly
145
396392
1793
06:38
and shows you content that's really relevant for you
146
398227
3128
用很快的速度把跟你 真正相關的內容推送給你。
06:41
in a very quick way.
147
401355
2461
06:44
CA: So it's a form of collaborative filtering, from what you're saying.
148
404442
3336
克:所以就你所言, 這是協同過濾的一種形式。
06:47
The theory behind it is that these humans are weird people,
149
407820
2794
它的理論是這些人很怪, 我們不知道他們對什麼感興趣,
06:50
we don't really know what they're interested in,
150
410614
2253
但如果我們看到某人跟他人 感興趣的事物有所重疊,
06:52
but if we see that one human is interested,
151
412908
2002
06:54
with an overlap of someone else, chances are, you know,
152
414952
2836
那很有可能,你知道,
06:57
you could make use of the other pieces
153
417830
2002
可以利用其他的東西,
06:59
that are in that overlapped human's repertoire to feed them,
154
419874
5463
把其他人看的東西餵給他們,
07:05
and they'll be surprised.
155
425337
1252
他們會很驚喜。
07:06
But the reason they like it is because their pal also liked it.
156
426589
3670
但他們喜歡的原因是 他們的朋友也喜歡。
07:10
SC: It's pattern recognition based on your interest signals.
157
430301
3253
周:這是基於你的 興趣訊號的模式識別。
07:13
And I think the other thing here
158
433596
1543
而我認為另一件事情是
07:15
is that we don't actually ask you 20 questions
159
435139
2669
我們不會問你 20 道問題,
07:17
on whether you like a piece of content, you know, what are your interests,
160
437808
3504
像你喜不喜歡某個內容、 有什麼興趣,
我們不做那種事。
07:21
we don't do that.
161
441312
1209
我們把那種體驗有機地 打造在應用程式中。
07:22
We built that experience organically into the app experience.
162
442563
2878
你用拇指投票,
07:25
So you are voting with your thumbs by watching a video,
163
445441
3545
不論看影片、滑過去、按讚、分享,
07:29
by swiping it, by liking it, by sharing it,
164
449028
3336
07:32
you are basically exhibiting interest signals.
165
452364
2586
你基本上都是在展現興趣訊號。
07:34
And what it does mathematically is to take those signals,
166
454992
2711
演算法在數學上就是吸收這些訊號,
07:37
put it in a formula and then matches it through pattern recognition.
167
457745
3211
代入一套公式, 然後透過模式識別進行配對。
07:40
That's basically the idea behind it.
168
460998
2211
它背後的點子基本上就是如此。
07:43
CA: I mean, lots of start-ups have tried to use these types of techniques.
169
463626
5422
克:我的意思是, 很多新創企業都用過這類技術。
07:49
I'm wondering what else played a role early on?
170
469089
2211
在草創期還有哪些因素?
我是指以下這件事有多重要,
07:51
I mean, how big a deal was it,
171
471300
1460
07:52
that from the get-go you were optimizing for smartphones
172
472760
3253
從最一開始你們就針對 智慧型手機進行最佳化,
07:56
so that videos were shot in portrait format
173
476013
2711
所以影片是以直立的方向拍攝,
07:58
and they were short.
174
478724
1877
而且很短。
08:00
Was that an early distinguishing thing that mattered?
175
480643
2961
這是早期一個重要的差異嗎?
08:03
SC: I think we were the first to really try this at scale.
176
483646
2752
周:我認為是我們最早開始 大規模測試這套技術。
08:06
You know, the recommendation algorithm is a very important reason
177
486941
3712
你知道,推薦演算法 是一項非常重要的因素,
08:10
as to why the platform is so popular among so many people.
178
490653
5088
可以解釋為何這個平臺 能夠大為流行。
08:15
But beyond that, you know, you mentioned the format itself.
179
495783
4296
但在此之外,你還提到影片的格式。
08:20
So we talked about the vision of the company,
180
500079
2127
我們談到這家企業的願景
08:22
which is to have a window to discover.
181
502206
1835
是提供一扇發現之窗。
08:24
And if you just open the app for the first time,
182
504083
2252
而如果你首次打開這個應用程式,
08:26
you'll see that it takes up your whole screen.
183
506377
2169
會發現它占滿整個螢幕。
這就是我們想要的窗口。
08:28
So that's the window that we want.
184
508546
1626
你可以想見有許多人利用那扇窗
08:30
You can imagine a lot of people using that window
185
510214
2294
08:32
to discover new things in their lives.
186
512508
1835
發現了生命中的新事物。
08:34
Then, you know, through this recommendation algorithm,
187
514385
3545
然後,透過推薦演算法,
08:37
we have found that it connects people together.
188
517972
2210
我們發現它讓人們有所連結。
人們找到社群,
08:40
People find communities,
189
520182
1293
08:41
and I've heard so many stories of people who have found their communities
190
521475
4088
而我聽過很多故事是 人們找到所屬的社群,
08:45
because of the content that they're posting.
191
525604
2086
因為他們發布的內容。
08:47
Now, I'll give you an example.
192
527690
1459
現在,我要給你一個例子。
08:49
I was in DC recently, and I met with a bunch of creators.
193
529567
3670
我最近在華盛頓特區, 遇到了一群創作者。
08:53
CA: I heard.
194
533237
1168
克:我有聽說。
08:54
(Laughter)
195
534446
1669
(笑聲)
08:56
SC: One of them was sitting next to me at a dinner,
196
536156
2420
周:其中一位有次晚餐 坐我旁邊,他叫山繆。
08:58
his name is Samuel.
197
538576
1209
08:59
He runs a restaurant in Phoenix, Arizona, and it's a taco restaurant.
198
539827
4921
他在亞利桑那州的鳳凰城 經營一家墨西哥夾餅餐廳。
09:04
He told me he has never done this before, first venture.
199
544790
2795
他告訴我他從未這麼做, 這是他首度創業。
09:08
He started posting all this content on TikTok,
200
548252
2419
他開始把這些內容發在抖音,
09:10
and I saw his content,
201
550713
1376
我看到了他的內容,
09:12
I was hungry after looking at it, it's great content.
202
552131
3920
看完我就餓了,那是很棒的內容。
09:16
And he's generated so much interest in his business,
203
556051
3254
他從生意中賺取了非常多利潤,
09:19
that last year he made something like a million dollars in revenue
204
559305
3128
去年他的營收就達到一百萬美元,
09:22
just via TikTok.
205
562474
1168
只是透過抖音,就一家餐廳。
09:23
One restaurant.
206
563684
1168
09:24
And again and again, I hear these stories,
207
564893
2211
一次又一次,我聽到這些故事。
09:27
you know, by connecting people together,
208
567104
2419
透過把眾人連結在一起,
09:29
by giving people the window to discover,
209
569565
2210
給予人們一扇發現的窗口,
09:31
we have given many small businesses and many people, your common person,
210
571775
4463
我們賦予許多小企業和平凡的個人
09:36
a voice that they will never otherwise have.
211
576280
2544
一個沒有抖音就不會有的發聲方式。
09:38
And I think that's the power of the platform.
212
578824
2169
我認為這就是這個平臺的力量。
09:41
CA: So you definitely have identified early
213
581327
2043
克:你在更早之前一定說過,
09:43
just how we're social creatures, we need affirmation.
214
583370
3754
我們是社會的動物, 我們需要他人的肯定。
我聽過一個故事,
09:47
I've heard a story,
215
587166
1168
09:48
and you can tell me whether true or not,
216
588375
1961
你可以告訴我它是真是假:
09:50
that one of the keys to your early liftoff
217
590336
2961
你在早期能夠大獲成功的關鍵之一
09:53
was that you wanted to persuade creators who were trying out TikTok
218
593297
4212
是你想要說服正在嘗試抖音的創作者
09:57
that this was a platform where they would get response,
219
597509
4088
這就是他們可以得到回應的平臺。
10:01
early on, when you're trying to grow something,
220
601597
2210
一開始你希望發展這個平臺時,
10:03
the numbers aren't there for response.
221
603849
1835
回應量沒這麼多。
10:05
So you had the brilliant idea of goosing those numbers a bit,
222
605684
3253
於是你想到一個 增加這些數字的主意,
10:08
basically finding ways to give people, you know,
223
608979
2503
基本上就是找到方法給人們
10:11
a bigger sense of like, more likes,
224
611482
1793
比實際情況更多的 按讚次數和參與度,
10:13
more engagement than was actually the case,
225
613317
2044
10:15
by using AI agents somehow in the process.
226
615361
4254
在過程中某種程度上 使用人工智慧代理人。
10:19
Is that a brilliant idea, or is that just a myth?
227
619657
2711
這是一個聰明的點子, 還是只是迷思?
10:22
SC: I would describe it in a different way.
228
622409
4004
周:我會用不同的方式表述它。
10:26
So there are other platforms that exist before TikTok.
229
626413
3504
有一些其他平臺出現於抖音之前。
10:29
And if you think about those platforms,
230
629917
1918
如果你想想這些平臺,
10:31
you sort of have to be famous already in order to get followers.
231
631877
3754
只有已經成名的人才會擁有追蹤者。
10:35
Because the way it’s built is that people come and follow people.
232
635631
3712
因為它們的打造方式 就是人們會來追蹤某些人。
10:39
And if you aren't already famous,
233
639385
2377
而如果你不是已經成名的人,
10:41
the chances that you get discovered are very, very low.
234
641762
3170
你被發現的機會就很小很小。
10:44
Now, what we have done, again,
235
644973
1460
而我們所辦到的,再次強調,
10:46
because of the difference in the way we're recommending content,
236
646433
3420
是因為我們用不同的方式推薦內容,
10:49
is that we have given anyone,
237
649895
1543
因此我們讓任何人、 任何有才華的個體
10:51
any single person with enough talent a stage to be able to be discovered.
238
651480
5339
一個可以被發現的舞臺。
10:56
And I think that actually is the single, probably the most important thing
239
656819
3545
我認為那是促進這個平臺成長
11:00
contributing to the growth of the platform.
240
660364
2085
最重要的單一因素。
11:02
And again and again, you will hear stories from people who use the platform,
241
662491
3587
你會一再聽到 這個平臺的使用者的故事,
這些人定期發布內容,
11:06
who post regularly on it,
242
666120
1626
11:07
that if they have something they want to say,
243
667788
2628
如果他們有話想說,
11:10
the platform gives them the chance and the stage
244
670457
2253
這個平臺給予他們機會和舞臺
11:12
to connect with their audience
245
672751
1460
跟觀眾建立連結,
11:14
in a way that I think no other product in the past has ever offered them.
246
674211
4505
用一種以往的產品 從未提供給他們的方式。
11:19
CA: So I'm just trying to play back what you said there.
247
679341
2628
克:我試著重述你所說的。
11:21
You said you were describing a different way what I said.
248
681969
3128
你說你要用一種不同的方式 描述我所說的事。
11:25
Is it then the case that like, to give someone a decent chance,
249
685139
3712
情況是不是像是 為了給予某人一個好機會,
11:28
someone who's brilliant but doesn't come with any followers initially,
250
688892
3379
這個人很有才華, 但起初沒有追蹤者,
11:32
that you've got some technique to identify talent
251
692312
3629
你們有某些技術去偵測才華,
11:35
and that you will almost encourage them,
252
695983
2085
然後你們會鼓勵這些人,
11:38
you will give them some kind of, you know,
253
698110
2794
給予他們某種,你知道,
11:40
artificially increase the number of followers or likes
254
700946
2544
人為增加的追蹤者人數或按讚數,
11:43
or whatever that they have,
255
703532
1335
或是他們的其他東西,
11:44
so that others are encouraged to go,
256
704867
1751
因此其他人會被激勵:
11:46
"Wow, there's something there."
257
706618
1544
「哇,那裡真有搞頭。」
這是群聚效應的概念,
11:48
Like it's this idea of critical mass that kind of, every entrepreneur,
258
708162
3336
幾乎每個企業家、 活動策劃人都知道的,
11:51
every party planner kind of knows about of
259
711498
2044
11:53
"No, no, this is the hot place in town, everyone come,"
260
713542
2628
「這是最熱門的地方, 所有人都會來,」
11:56
and that that is how you actually gain critical mass?
261
716170
2627
這是你達成群聚效應的實際方式嗎?
11:58
SC: We want to make sure that every person who posts a video
262
718839
3754
周:我們希望確保任何發布影片的人
12:02
is given an equal chance to be able to have some audience to begin with.
263
722634
4713
都能有均等的機會獲得最初的觀眾。
12:07
But this idea that you are maybe alluding to,
264
727389
3379
但你所說的想法可能是在影射
12:10
that we can get people to like something,
265
730809
3003
我們可以讓人們喜歡某些東西,
12:13
it doesn't really work like that.
266
733854
1585
但它並不是這樣運作的。
12:15
CA: Could you get AI agents to like something?
267
735439
2919
克:你能讓人工智慧代理人 喜歡某樣東西嗎?
12:18
Could you seed the network with extra AI agents that could kind of, you know,
268
738400
4046
你可以利用人工智慧代理人 預先設定這個網絡,
12:22
give someone early encouragement?
269
742488
1960
讓某人在早期得到激勵嗎?
12:24
SC: Ultimately, what the machine does is it recognizes people's interests.
270
744490
4546
周:追根究柢,機器做的 就是識別人們的興趣。
12:29
So if you post something that's not interesting to a lot of people,
271
749077
3170
所以如果你發的東西 對多數人來講不有趣,
12:32
even if you gave it a lot of exposure,
272
752289
1835
即便你給它很高的曝光率,
12:34
you're not going to get the virality that you want.
273
754166
2419
你也無法達到你想要的病毒式傳播。
12:36
So it's a lot of ...
274
756585
1793
所以這是很多的……
12:38
There is no push here.
275
758420
2878
其中並沒有人為助力。
12:41
It's not like you can go and push something,
276
761340
2669
你沒有辦法推廣某個事物,
12:44
because I like Chris, I'm going to push your content,
277
764009
2503
因為喜歡克里斯就推廣你的內容,
12:46
it doesn't work like that.
278
766553
1252
不是這麼運作的。
12:47
You've got to have a message that resonates with people,
279
767805
2627
你必須有一個激發共鳴的訊息,
如果有,
12:50
and if it does,
280
770474
1168
它就會自然而然地紅起來。
12:51
then it will automatically just have the virality itself.
281
771683
2711
12:54
That's the beauty of user-generated content.
282
774394
2253
這就是使用者產出內容的美妙之處。
12:56
It's not something that can be engineered or over-thought.
283
776688
4046
它不能被設計或事前規劃。
13:00
It really is something that has to resonate with the audience.
284
780776
3462
它其實就是跟觀眾有共鳴的事物。
13:04
And if it does, then it goes viral.
285
784238
1918
如果它有,就會紅起來。
13:06
CA: Speaking privately with an investor who knows your company quite well,
286
786198
4588
克:我跟一位非常了解你的公司的 投資者私下聊過,
13:10
who said that actually the level of sophistication
287
790786
4463
他說你的演算法老練的程度
13:15
of the algorithms you have going
288
795249
1543
13:16
is just another order of magnitude
289
796834
1918
跟臉書、YouTube 等競爭者比起來 完全是不同的數量級。
13:18
to what competitors like, you know, Facebook or YouTube have going.
290
798794
4379
13:23
Is that just hype or do you really believe you --
291
803924
3253
那只是炒作,還是你真的很有信心?
13:27
like, how complex are these algorithms?
292
807177
3420
像是,這些演算法有多複雜?
13:31
SC: Well, I think in terms of complexity,
293
811014
2002
周:我想,就複雜度而言,
13:33
there are many companies who have a lot of resources
294
813016
2545
很多公司都有大量資源與天才。
13:35
and a lot of talent.
295
815602
1168
13:36
They will figure out even the most complex algorithms.
296
816770
2920
他們會設計出最複雜的演算法。
13:39
I think what is very different is your mission of your company,
297
819982
3795
我認為最為不同的地方 是你的企業的使命,
13:43
how you started the company.
298
823819
1626
你如何開始建立公司。
13:45
Like I said, you know, we started with this idea
299
825445
2294
如我所說,你知道, 我們最初的點子是
13:47
that this was the main use case.
300
827739
2128
這是主要的使用情境:
13:50
The most important use case is you come and you get to see recommended content.
301
830158
4463
最重要的使用情境是 你一來就看到推薦的內容。
13:54
Now for some other apps out there,
302
834663
1835
而其他既有的應用程式
13:56
they are very significant and have a lot of users,
303
836540
2544
影響力很巨大,擁有許多使用者,
13:59
they are built for a different original purpose.
304
839084
2920
它們是為了不同的原始目的而打造。
14:02
And if you are built for something different,
305
842004
2127
而如果你是另有目標,
使用者就會習以為常,
14:04
then your users are used to that
306
844131
1543
14:05
because the community comes in and they expect that sort of experience.
307
845716
3336
因為這個社群加入時 就是期望那一種體驗。
我認為事情的關鍵
14:09
So I think the pivot away from that
308
849052
1710
14:10
is not really just a matter of engineering and algorithms,
309
850762
3379
並不只是工程和演算法,
14:14
it’s a matter of what your company is built to begin with.
310
854141
3837
而是你的企業是為何而打造。
14:18
Which is why I started this by saying you need to have a vision,
311
858020
3003
這就是為什麼我一開始 會說你要有願景,
你必須有使命,那就是北極星。
14:21
you need to have a mission, and that's the North Star.
312
861064
2545
14:23
You can't just shift it halfway.
313
863650
1585
你不能半途而廢。
14:25
CA: Right.
314
865235
1168
克:沒錯。
14:26
And is it fair to say
315
866445
1168
這樣說公不公道:
14:27
that because your start point has been interest algorithms
316
867613
2752
因為你的起始點就是興趣演算法,
而非社交圖譜演算法,
14:30
rather than social graph algorithms,
317
870365
1919
14:32
you've been able to avoid some of the worst of the sort of,
318
872326
2794
所以你們能夠避免 一個非常糟的問題,
也就是其他社交媒體的同溫層,
14:35
the filter bubbles that have happened in other social media
319
875120
2795
14:37
where you have tribes kind of declaring war on each other effectively.
320
877956
3337
你有有效互相對立的各種論點。
14:41
And so much of the noise and energy is around that.
321
881293
4296
而許多議論和活力正是以此為中心。
14:45
Do you believe that you've largely avoided that on TikTok?
322
885631
2752
你是否相信你在抖音 大致避免了同溫層?
14:48
SC: The diversity of content that our users see is very key.
323
888425
4254
周:最重要的關鍵是我們的使用者 看見的內容的多樣性。
14:52
You know, in order for the discovery -- the mission is to discover --
324
892721
4213
你知道,為了促成發現—— 我們的使命就是發現——
14:56
sorry, the vision is to discover.
325
896975
2044
抱歉,我們的願景是發現。
14:59
So in order to facilitate that,
326
899061
1626
為了促成這件事,
15:00
it is very important to us
327
900729
1710
對我們而言很重要的是
15:02
that what the users see is a diversity of content.
328
902481
3128
使用者所見的是多樣的內容。
15:06
Now, generally speaking, you know,
329
906276
2461
大體而言,你知道,
15:08
there are certain issues that you mentioned
330
908779
2002
你提到了一些議題 是這個產業要面對的。
15:10
that the industry faces, you know.
331
910822
1627
15:12
There are some bad actors who come on the internet,
332
912491
2419
網路上有些不好的行為人,
15:14
they post bad content.
333
914952
1126
發布壞內容。
我們的方法是我們有 非常清晰的社群守則。
15:16
Now our approach is that we have very clear community guidelines.
334
916119
4046
15:20
We're very transparent about what is allowed
335
920165
2377
我們對於平臺要容許 和禁止什麼非常透明。
15:22
and what is not allowed on our platform.
336
922542
1961
15:24
No executives make any ad hoc decisions.
337
924544
3003
沒有行政人員會做出臨時的決定。
15:27
And based on that,
338
927589
1126
基於這點,
15:28
we have built a team that is tens of thousands of people plus machines
339
928757
4338
我們成立了一支數萬人 加上機器的小組,
15:33
in order to identify content that is bad
340
933136
2461
以辨別出壞的內容,
15:35
and actively and proactively remove it from the platform.
341
935639
3211
並主動積極地將它從平臺移除。
15:38
CA: Talk about what some of those key guidelines are.
342
938892
2711
克:談談有哪些關鍵守則。
15:41
SC: We have it published on our website.
343
941603
2127
周:我們把它公布在網站上。
15:43
In March, we just iterated a new version to make it more readable.
344
943772
5672
我們三月才修訂了一份新版本, 為了讓它更易讀。
15:49
So there are many things like, for example, no pornography,
345
949444
3712
其中有很多規定,例如禁止色情,
15:53
clearly no child sexual abuse material and other bad things,
346
953156
3546
清楚禁止兒童性虐待題材 和其他不好的事情,
15:56
no violence, for example.
347
956702
2085
像是禁止暴力。
15:58
We also make it clear that it's a differentiated experience
348
958787
3086
我們也清楚表示未滿 18 歲者 會有差異化的體驗。
16:01
if you're below 18 years old.
349
961873
1836
16:03
So if you're below 18 years old, for example,
350
963750
2127
舉例來說,如果你還沒 18 歲,
16:05
your entire app experience is actually more restricted.
351
965919
3253
你整個應用程式體驗 其實是更加受限的。
16:09
We don't allow, as an example,
352
969172
2461
例如我們預設不允許
16:11
users below 16, by default, to go viral.
353
971633
4046
低於 16 歲的使用者爆紅。
16:15
We don't allow that.
354
975679
1209
我們不允許。
16:16
If you're below 16,
355
976930
1168
如果沒有 16 歲,
我們不允許你使用 應用程式的即時訊息功能。
16:18
we don’t allow you to use the instant messaging feature in app.
356
978140
3753
16:22
If you’re below 18, we don’t allow you to use the livestreaming features.
357
982602
3504
如果你未滿 18 歲, 我們不允許你使用直播功能。
16:26
And of course, we give parents a whole set of tools
358
986148
2669
當然,我們也給予家長一整套工具
16:28
to control their teenagers’ experience as well.
359
988817
2211
控制青少年子女的體驗。
克:你們怎麼知道使用者的年紀?
16:31
CA: How do you know the age of your users?
360
991069
2836
16:34
SC: In our industry, we rely mainly on something called age gating,
361
994364
4588
克:在我們這個產業, 我們主要依靠的是年齡限制,
16:38
which is when you sign up for the app for the first time
362
998994
2627
這是說當你首次註冊應用程式,
16:41
and we ask you for the age.
363
1001621
1335
我們會問你的年齡。
16:42
Now, beyond that,
364
1002956
1168
除此之外,
16:44
we also have built tools to go through your public profile for example,
365
1004166
5422
舉例來說,我們也建立了 檢查你的公開資訊的工具,
16:49
when you post a video,
366
1009588
1168
當你發布影片,
16:50
we try to match the age that you said with the video that you just posted.
367
1010797
4255
我們會試著檢查你宣稱的年齡 和你發布的影片是否相符。
16:55
Now, there are questions of can we do more?
368
1015093
2336
問題是我們能不能做得更多?
16:57
And the question always has, for every company, by the way,
369
1017471
3253
這個問題對於我們產業的每家企業, 總是必須跟隱私權取得平衡。
17:00
in our industry, has to be balanced with privacy.
370
1020766
2836
17:03
Now, if, for example, we scan the faces of every single user,
371
1023977
5380
比如說,如果掃描每個使用者的臉,
17:09
then we will significantly increase the ability to tell their age.
372
1029399
3962
我們判斷年齡的能力就會大為提升。
17:13
But we will also significantly increase the amount of data
373
1033403
2711
但我們也會大量增加 從你身上蒐集的資料量。
17:16
that we collect on you.
374
1036156
1293
17:17
Now, we don't want to collect data.
375
1037491
1710
我們不想蒐集資料。
我們不想蒐集你的臉部掃描資料。
17:19
We don't want to scan data on your face to collect that.
376
1039201
2669
17:21
So that balance has to be maintained,
377
1041870
1877
所以必須求取一個平衡,
17:23
and it's a challenge that we are working through
378
1043747
2294
這是我們正在和這個產業 以及監管者努力克服的挑戰。
17:26
together with industry, together with the regulators as well.
379
1046041
3378
17:29
CA: So look, one thing that is unquestionable
380
1049920
2127
克:看起來無庸置疑的是
17:32
is that you have created a platform for literally millions of people
381
1052047
3837
你為數百萬未曾想過 成為內容創作者的人
17:35
who never thought they were going to be a content creator.
382
1055884
2878
建立了一個平臺。
17:38
You've given them an audience.
383
1058804
1626
你給他們觀眾。
17:40
I'd actually like to hear from you one other favorite example
384
1060472
2878
我其實想再聽一個你最喜歡的例子,
17:43
of someone who TikTok has given an audience to
385
1063391
2712
抖音給了某人以前 從未有過的一群觀眾。
17:46
that never had that before.
386
1066144
1460
17:47
SC: So when again,
387
1067604
2252
周:再一次,
17:49
when I travel around the world,
388
1069898
2002
當我前往世界各地時,
17:51
I meet with a whole bunch of creators on our platform.
389
1071942
2961
我遇到一大群我們平臺的創作者。
17:55
I was in South Korea just yesterday, and before that I met with -- yes,
390
1075320
5798
我昨天在南韓,此前我遇到了——
18:01
before that I met with a bunch of --
391
1081118
1751
在我遇到一群人之前。
18:02
People don't expect, for example, teachers.
392
1082869
2044
大家不會想到的人,例如老師。
18:04
There is an English teacher from Arkansas.
393
1084955
3170
有一位來自阿肯色州的英文老師。
18:08
Her name is Claudine, and I met her in person.
394
1088125
3086
她的名字是克勞汀,我遇到她本人。
18:11
She uses our platform to reach out to students.
395
1091253
2669
她用我們的平臺來幫助學生。
18:14
There is another teacher called Chemical Kim.
396
1094381
3128
還有另一位老師叫「化學金」。
18:17
And Chemical Kim teaches chemistry.
397
1097551
2752
化學金教化學。
18:20
What she does is she uses our platform
398
1100595
1877
她使用我們的平臺
18:22
to reach out to a much broader student base
399
1102514
2502
幫助到比在教室裡更多的學生。
18:25
than she has in her classroom.
400
1105058
1543
18:26
And they're both very, very popular.
401
1106643
2085
她們都非常有名。
18:28
You know, in fact,
402
1108770
1168
你知道,事實上,
18:29
what we have realized is that STEM content
403
1109938
4588
我們發現科學、技術、 工程、數學的內容
18:34
has over 116 billion views on our platform globally.
404
1114568
4546
在我們的平臺上 全球超過 1160 億的觀看次數。
18:39
And it's so significant --
405
1119114
1251
這真的很多——
18:40
CA: In a year?
406
1120407
1168
克:一年嗎?
18:41
SC: Cumulatively.
407
1121575
1418
周:總共。
18:43
CA: [116] billion.
408
1123034
1210
克: 1160 億。
18:44
SC: It's so significant, that in the US we have started testing,
409
1124286
3461
周:這真的很多, 於是我們開始在美國測試
18:47
creating a feed just for STEM content.
410
1127789
3087
專門建立科學、技術、 工程、數學內容的頻道。
18:50
Just for STEM content.
411
1130876
1167
專門給 STEM 內容。
18:52
I’ve been using it for a while, and I learned something new.
412
1132085
3128
我已經使用它一陣子了, 我有學到新知識。
18:55
You want to know what it is?
413
1135255
1627
你想知道是什麼嗎?
18:56
Apparently if you flip an egg on your tray,
414
1136923
3629
如果你把雞蛋倒過來放,
19:00
the egg will last longer.
415
1140552
2169
就可以保鮮更久。
19:02
It's science,
416
1142721
1168
這是科學,有一系列影片,
19:03
there’s a whole video on this, I learned this on TikTok.
417
1143930
2628
我在抖音學到的。
你可以搜尋看看。
19:06
You can search for this.
418
1146558
1251
19:07
CA: You want to know something else about an egg?
419
1147809
2294
克:你還想知道雞蛋的知識嗎?
如果你把雞蛋放在一隻手上, 非常用力擠壓它,
19:10
If you put it in just one hand and squeeze it as hard as you can,
420
1150145
3086
它不會破掉。
19:13
it will never break.
421
1153231
1168
周:對,我覺得我也讀到過。
19:14
SC: Yes, I think I read about that, too.
422
1154441
1918
克:那不是真的。
19:16
CA: It's not true.
423
1156359
1168
(笑聲)
19:17
(Laughter)
424
1157569
1126
周:我們可以找找看,搜尋看看。
19:18
SC: We can search for it.
425
1158737
1376
19:21
CA: But look, here's here's the flip side to all this amazingness.
426
1161072
3504
克:但注意看這些驚奇的反面。
19:24
And honestly, this is the key thing,
427
1164618
2210
老實說,這是一個關鍵,
19:26
that I want to have an honest, heart-to-heart conversation with you
428
1166870
4504
因此我想跟你有一段 真誠、推心置腹的對話,
19:31
because it's such an important issue,
429
1171416
1794
因為這是重要議題,
成癮的問題。
19:33
this question of human addiction.
430
1173210
2460
19:35
You know, we are ...
431
1175712
2377
你知道,我們是……
19:38
animals with a prefrontal cortex.
432
1178131
3629
有前額葉皮質的動物。
19:42
That's how I think of us.
433
1182594
1209
我是這麼想的。
19:43
We have these addictive instincts that go back millions of years,
434
1183803
4213
我們這些成癮的本能 可追溯到數百萬年前,
19:48
and we often are in the mode of trying to modulate our own behavior.
435
1188058
6673
我們時常處於試著 調整自己的行為的模式。
19:54
It turns out that the internet is incredibly good
436
1194731
4421
結果發現網際網路極度擅長
19:59
at activating our animal cells
437
1199194
2461
活化我們的動物細胞,
20:01
and getting them so damn excited.
438
1201696
2795
讓它們該死地興奮起來。
20:04
And your company, the company you've built,
439
1204532
2712
而你的公司,你所打造的公司
20:07
is better at it than any other company on the planet, I think.
440
1207244
5338
比這顆星球上任何其他公司 都還要擅長,我覺得。
20:12
So what are the risks of this?
441
1212999
3003
那麼這件事的風險是什麼?
20:16
I mean, how ...
442
1216002
1752
我的意思是,怎麼……
20:17
From a company point of view, for example,
443
1217796
3003
舉例來說,從一家企業的視角,
20:20
it's in your interest to have people on there as long as possible.
444
1220840
3212
讓人們待在上面越久越好 符合你們的利益。
20:24
So some would say, as a first pass,
445
1224094
1960
有些人會說,從一開始,
20:26
you want people to be addicted as long as possible.
446
1226096
2794
你們就希望人們成癮得越久越好。
20:28
That's how advertising money will flow and so forth,
447
1228932
3962
這是廣告費等事物流入的方式,
20:32
and that's how your creators will be delighted.
448
1232936
2461
也是讓你的創作者開心的理由。
20:36
What is too much?
449
1236481
1376
這會不會太超過?
20:37
SC: I don't actually agree with that.
450
1237857
1794
周:我其實並不同意。
20:40
You know, as a company,
451
1240402
1543
你知道,身為一家公司,
20:41
our goal is not to optimize and maximize time spent.
452
1241987
3211
我們的目標並不是極大化 花在我們身上的時間。
20:45
It is not.
453
1245240
1418
並非如此。
20:46
In fact, in order to address people spending too much time on our platform,
454
1246700
4004
事實上,為了處理人們在我們平臺 花費太多時間的問題,
20:50
we have done a number of things.
455
1250704
1585
我們做了幾件事。
20:52
I was just speaking with some of your colleagues backstage.
456
1252289
2836
我剛剛在後臺跟一些你的同事談話。
20:55
One of them told me she has encountered this as well.
457
1255166
3379
其中一位告訴我她也遇到過這件事。
20:58
If you spend too much time on our platform,
458
1258586
2169
如果你在我們的平臺花太多時間,
21:00
we will proactively send you videos to tell you to get off the platform.
459
1260755
4004
我們會主動推送 提醒你離開平臺的影片。
21:05
We will.
460
1265093
1251
我們真的會。
21:06
And depending on the time of the day,
461
1266386
1793
而且取決於時間,
如果夜深了,影片會來得更快。
21:08
if it's late at night, it will come sooner.
462
1268179
2002
21:10
We have also built in tools to limit,
463
1270223
1919
我們也在工具中建立限制,
21:12
if you below 18 years old, by default,
464
1272142
3670
預設是如果你未滿 18 歲,
21:15
we set a 60-minute default time limit.
465
1275812
2377
我們會設定 60 分鐘的限制。
21:18
CA: How many?
466
1278189
1168
克:幾分鐘? 周:60 分鐘。
21:19
SC: Sixty minutes.
467
1279399
1168
我們也給家長和你本人工具,
21:20
And we've given parents tools and yourself tools,
468
1280608
2294
21:22
if you go to settings, you can set your own time limit.
469
1282944
2628
如果你去設定,可以自己限制時間。
我們也給家長工具進行配對,
21:25
We've given parents tools so that you can pair,
470
1285572
2210
如果有家長不知道, 可以去設定,家庭安全配對,
21:27
for the parents who don't know this, go to settings, family pairing,
471
1287824
3212
你可以把你的手機跟孩子的配對起來
21:31
you can pair your phone with your teenager's phone
472
1291036
2335
並設下時間限制。
21:33
and set the time limit.
473
1293413
1168
21:34
And we really encourage parents to have these conversations with their teenagers
474
1294581
3795
而且我們非常鼓勵家長 跟青少年進行討論
21:38
on what is the right amount of screen time.
475
1298418
2044
螢幕使用時間多少才合適。
21:40
I think there’s a healthy relationship that you should have with your screen,
476
1300503
3629
我認為你應該跟螢幕有健康的關係,
而身為一家公司, 我們相信必須取得平衡。
21:44
and as a business, we believe that that balance needs to be met.
477
1304132
3003
所以我們並不是 只想極大化使用時間。
21:47
So it's not true that we just want to maximize time spent.
478
1307135
3754
21:50
CA: If you were advising parents here
479
1310889
2043
克:如果要你建議在座的家長
21:52
what time they should actually recommend to their teenagers,
480
1312974
3003
他們應該跟子女提議多少的時間,
21:56
what do you think is the right setting?
481
1316019
1877
你認為正確的設定是什麼?
21:57
SC: Well, 60 minutes,
482
1317896
1168
周:嗯,60 分鐘, 這不是我們自己想的。
21:59
we did not come up with it ourselves.
483
1319105
1794
22:00
So I went to the Digital Wellness Lab at the Boston Children's Hospital,
484
1320899
3420
我去了波士頓兒童醫院 數位健康實驗室,
22:04
and we had this conversation with them.
485
1324319
1877
跟他們討論了這個問題。
22:06
And 60 minutes was the recommendation that they gave to us,
486
1326237
2795
而 60 分鐘是他們給我們的建議,
所以我們把它預設在應用程式。
22:09
which is why we built this into the app.
487
1329032
1918
22:10
So 60 minutes, take it for what it is,
488
1330950
1877
所以 60 分鐘,就這麼建議吧,
22:12
it’s something that we’ve had some discussions of experts.
489
1332827
3295
這是我們跟專家討論過的。
22:16
But I think for all parents here,
490
1336122
1585
但對於在座的家長,
22:17
it is very important to have these conversations with your teenage children
491
1337749
4212
我認為非常重要的是 跟青少年子女有這些討論,
22:21
and help them develop a healthy relationship with screens.
492
1341961
4588
並幫助他們與螢幕建立健康的關係。
22:26
I think we live in an age where it's completely inevitable
493
1346549
3087
我認為我們生活在一個無可避免
22:29
that we're going to interact with screens and digital content,
494
1349636
4629
要跟螢幕和數位內容互動的時代,
22:34
but I think we should develop healthy habits early on in life,
495
1354307
3504
但我認為我們應該在人生中 盡早發展出健康的習慣,
22:37
and that's something I would encourage.
496
1357811
1918
這是我會鼓勵的事情。
22:39
CA: Curious to ask the audience,
497
1359729
2795
克:好奇問一下各位觀眾,
22:42
which of you who have ever had that video on TikTok appear
498
1362565
3629
你們當中有誰曾經 在抖音看到那個影片,
22:46
saying, “Come off.”
499
1366236
1501
叫你別再用了?
22:48
OK, I mean ...
500
1368655
1626
好,我的意思是……
22:50
So maybe a third of the audience seem to be active TikTok users,
501
1370949
3128
也許有三分之一的聽眾 是抖音的活躍使用者,
22:54
and about 20 people maybe put their hands up there.
502
1374119
4254
但舉手的大概只有 20 人。
22:58
Are you sure that --
503
1378957
1918
你確定——
23:00
like, it feels to me like this is a great thing to have,
504
1380875
4713
對我來說這是一件好事,
23:05
but are you ...
505
1385630
1543
但你會不會……
23:07
isn't there always going to be a temptation
506
1387215
2002
難道不會總有一股誘惑,
23:09
in any given quarter or whatever,
507
1389217
2252
不論是在某一季或其他情況,
23:11
to just push it a bit at the boundary
508
1391469
2169
稍微越過界限,
23:13
and just dial back a bit on that
509
1393680
2210
對於成癮預防做出妥協,
23:15
so that you can hit revenue goals, etc?
510
1395932
3170
以達到營收目標等數據嗎?
23:19
Are you saying that this is used scrupulously?
511
1399727
2878
你是說這會一絲不苟地落實嗎?
23:22
SC: I think, you know, in terms ...
512
1402647
2920
周:我認為,你知道,談到……
23:25
Even if you think about it from a commercial point of view,
513
1405608
2795
就算你在商言商,
23:28
it is always best when your customers have a very healthy relationship
514
1408403
3295
讓顧客跟你的產品 有健康的關係總是最好的。
23:31
with your product.
515
1411698
1168
23:32
It's always best when it's healthy.
516
1412907
1710
它健康永遠是最好的。
23:34
So if you think about very short-term retention, maybe,
517
1414617
2962
如果你想的是短期保有客戶,也許吧,
23:37
but that's not the way we think about it.
518
1417579
2335
但我們不是這麼思考的。
23:40
If you think about it from a longer-term perspective,
519
1420290
2502
如果你從長期的視角來思考,
23:42
what you really want to have is a healthy relationship, you know.
520
1422792
3087
你真正想要有的是 健康的關係,你知道。
23:45
You don’t want people to develop very unhealthy habits,
521
1425879
2752
你不想讓人們發展出 非常不健康的習慣,
23:48
and then at some point they're going to drop it.
522
1428631
2461
然後在某個時間點退出應用程式。
23:51
So I think everything in moderation.
523
1431134
2419
所以我認為一切都要適度。
23:53
CA: There's a claim out there that in China,
524
1433595
3003
周:有一個說法是在中國,
23:56
there's a much more rigorous standards imposed on the amount of time
525
1436598
4462
他們以更嚴格的標準限制使用時間,
24:01
that children, especially, can spend on the TikTok equivalent of that.
526
1441060
5297
尤其是孩童可以花在 抖音及其同類產品的時間。
24:06
SC: That is unfortunately a misconception.
527
1446357
4713
周:很不幸的,這是一個誤解。
24:11
So that experience that is being mentioned for Douyin,
528
1451070
3129
你所述的體驗是中國版抖音, 那是不同的應用程式,
24:14
which is a different app,
529
1454199
1251
24:15
is for an under 14-year-old experience.
530
1455450
2627
對於未滿 14 歲孩童的使用體驗。
24:18
Now, if you compare that in the United States,
531
1458495
2293
如果你拿來跟美國的抖音相比,
24:20
we have an under-13 experience in the US.
532
1460788
2503
我們在美國有未滿 13 歲的體驗。
24:23
It's only available in the US, it's not available here in Canada,
533
1463333
3086
這只有美國可以使用, 在加拿大這裡沒有開放。
24:26
in Canada, we just don't allow it.
534
1466419
1668
在加拿大我們直接禁止。
24:28
If you look at the under-13 experience in the US,
535
1468129
2378
如果你看一下美國 未滿 13 歲的體驗,
24:30
it's much more restricted than the under-14 experience in China.
536
1470548
3754
它比中國未滿 14 歲的 體驗更加受限。
24:34
It's so restrictive,
537
1474344
1334
它的限制程度高到每一個內容
24:35
that every single piece of content is vetted
538
1475720
3003
都要經過我們的 第三方兒童安全專家審查。
24:38
by our third-party child safety expert.
539
1478723
4296
24:43
And we don't allow any under-13s in the US to publish,
540
1483436
4463
而且我們不允許美國 任何未滿 13 歲者發表內容,
24:47
we don’t allow them to post,
541
1487899
1418
我們不准他們發文,
24:49
and we don't allow them to use a lot of features.
542
1489317
2336
我們也不讓他們使用很多功能。
24:51
So I think that that report, I've seen that report too,
543
1491653
2627
所以我認為那篇報導,我也讀過,
它的比較是不公平的。
24:54
it's not doing a fair comparison.
544
1494280
1710
24:56
CA: What do you make of this issue?
545
1496324
1710
克:你對此有何看法?
你知道,你擁有 數百萬名內容創作者,
24:58
You know, you've got these millions of content creators
546
1498034
2878
25:00
and all of them, in a sense, are in a race for attention,
547
1500954
4045
而他們所有人在某種程度上 都在競逐注意力,
25:05
and that race can pull them in certain directions.
548
1505041
2920
這個競賽會把他們帶往特定的方向。
25:07
So, for example, teenage girls on TikTok,
549
1507961
4296
舉例來說,抖音上的青少女,
25:12
sometimes people worry that, to win attention,
550
1512257
3628
有時大家會擔心為了贏得注意力,
25:15
they've discovered that by being more sexual
551
1515885
2086
她們會發現只要更性感一些,
25:18
that they can gain extra viewers.
552
1518012
1669
就可以獲得更多觀看者。
這需要擔心嗎?
25:20
Is this a concern?
553
1520098
1376
25:21
Is there anything you can do about this?
554
1521474
2252
你對這個問題有什麼可以做的嗎?
25:23
SC: We address this in our community guidelines as well.
555
1523768
3295
周:我們也在社群守則 處理了這個問題。
25:28
You know, if you look at sort of the sexualized content on our guidelines,
556
1528147
4130
你知道,如果你看看 我們的守則提到的性感內容,
25:32
if you’re below a certain age,
557
1532318
1543
假如你低於特定年齡,
25:33
you know, for certain themes that are mature,
558
1533903
3253
你知道,那些成年的場景
25:37
we actually remove that from your experience.
559
1537198
2127
其實會被我們從你的體驗移除。
25:39
Again, I come back to this,
560
1539325
1543
我要再次回來這一點,
25:40
you know, we want to have a safe platform.
561
1540910
2127
我們想要有個安全的平臺。
25:43
In fact, at my congressional hearing,
562
1543037
1794
事實上,在國會聽證會,
25:44
I made four commitments to our users and to the politicians in the US.
563
1544872
3754
我對我們的使用者和美國的政治人物 做出了四項承諾。
25:48
And the first one is that we take safety, especially for teenagers,
564
1548626
4338
第一是我們非常嚴肅地看待安全性,
25:53
extremely seriously,
565
1553006
1167
尤其是對青少年,
25:54
and we will continue to prioritize that.
566
1554215
2419
而且我們會繼續將其置於首位。
25:56
You know, I believe that we need to give our teenage users,
567
1556634
3796
你知道,我相信我們必須 給予青少年使用者
26:00
and our users in general,
568
1560471
1460
以及所有使用者
26:01
a very safe experience,
569
1561973
1168
非常安全的體驗。
26:03
because if we don't do that,
570
1563182
1502
因為如果我們不如此,
26:04
then we cannot fulfill --
571
1564684
1293
我們就無法實現 刺激創意與帶來快樂的使命。
26:05
the mission is to inspire creativity and to bring joy.
572
1565977
2919
26:08
If they don't feel safe, I cannot fulfill my mission.
573
1568938
2836
如果他們感覺不安全, 我就無法達成使命。
26:11
So it's all very organic to me as a business
574
1571816
2753
所以對於我的生意而言, 確保我做到這件事
26:14
to make sure I do that.
575
1574569
1293
是非常有機的。
26:15
CA: But in the strange interacting world of human psychology and so forth,
576
1575862
3503
克:但這個奇怪而彼此互動的世界 涉及心理學等事物,
26:19
weird memes can take off.
577
1579365
1418
怪異的迷因可以爆紅。
26:20
I mean, you had this outbreak a couple years back
578
1580825
2461
我是說,你在幾年前 曾爆發過一件事,
26:23
with these devious licks where kids were competing with each other
579
1583328
3420
小孩用這些狡猾的技倆互相競爭,
26:26
to do vandalism in schools and, you know,
580
1586789
2253
故意破壞學校的財產,你知道,
26:29
get lots of followers from it.
581
1589083
1585
為了得到很多追蹤者。
26:30
How on Earth do you battle something like that?
582
1590710
2461
怎麼可能會有人比賽這種事?
26:33
SC: So dangerous challenges are not allowed on our platform.
583
1593963
2878
周:所以我們的平臺 不允許危險的挑戰。
26:36
If you look at our guidelines, it's violative.
584
1596883
2961
如果你看看我們的守則, 它是違規的。
26:39
We proactively invest resources to identify them
585
1599886
3336
我們積極主動地投入資源 以識別這類內容,
26:43
and remove them from our platform.
586
1603222
1794
並把它們從平臺上移除。
26:45
In fact, if you search for dangerous challenges on our platform today,
587
1605058
3336
事實上,現在如果你在我們的平臺 搜尋危險的挑戰,
26:48
we will redirect you to a safety resource page.
588
1608394
2878
我們會把你重新導向到 安全指南的頁面。
26:51
And we actually worked with some creators as well to come up with campaigns.
589
1611314
3587
而且我們其實也跟創作者 構思了一些活動。
26:54
This is another campaign.
590
1614901
1501
這是另一個活動。
26:56
It's the "Stop, Think, Decide Before You Act" campaign
591
1616444
2711
它叫「停一停,想一想, 先決定再行動」,
26:59
where we work with the creators to produce videos,
592
1619155
2419
我們在活動中跟創作者一同製作影片
27:01
to explain to people that some things are dangerous,
593
1621574
2920
來向大家說明有些事物是危險的,
27:04
please don't do it.
594
1624535
1210
請不要去做。
27:05
And we post these videos actively on our platform as well.
595
1625787
3128
而且我們也主動 將這些影片發在平臺上。
27:09
CA: That's cool.
596
1629791
1209
克:這真酷。
27:11
And you've got lots of employees.
597
1631042
1877
你有很多員工。
27:12
I mean, how many employees do you have
598
1632960
1836
我想問,你的員工有多少人
27:14
who are specifically looking at these content moderation things,
599
1634796
3753
專門在處理這些內容審查事務?
27:18
or is that the wrong question?
600
1638549
1502
還是我不該這麼問?
它們是先由人工智慧進行識別,
27:20
Are they mostly identified by AI initially
601
1640093
2377
27:22
and then you have a group who are overseeing
602
1642512
2961
然後你們的團隊才去檢查
27:25
and making the final decision?
603
1645515
1585
並做出最終決定嗎?
27:27
SC: The group is based in Ireland and it's a lot of people,
604
1647100
4045
周:這個團隊設在愛爾蘭, 它的人數眾多,
27:31
it's tens of thousands of people.
605
1651187
1585
有好幾萬人。
27:32
CA: Tens of thousands?
606
1652772
1168
克:好幾萬人?
27:33
SC: It's one of the most important cost items on my PnL,
607
1653981
4004
周:這是我的損益表上 成本最高的項目之一,
而我認為這完全是值得的。
27:38
and I think it's completely worth it.
608
1658027
1794
27:39
Now, most of the moderation has to be done by machines.
609
1659821
2836
多數審查必須由機器做。
27:42
The machines are good, they're quite good,
610
1662657
2627
機器很好,相當好,
27:45
but they're not as good as, you know,
611
1665326
1794
但它們沒有好到,你知道,
27:47
they're not perfect at this point.
612
1667120
1668
它們目前還不完美,
27:48
So you have to complement them with a lot of human beings today.
613
1668788
3211
所以現今你必須以大量真人來輔助。
27:51
And I think, by the way, a lot of the progress in AI in general
614
1671999
4380
而我認為,順帶一提, 人工智慧的許多進展大致上
27:56
is making that kind of content moderation capabilities a lot better.
615
1676421
4004
讓這類內容審查的量能大大改善,
28:00
So we're going to get more precise.
616
1680466
1877
因此我們可以更加準確。
28:02
You know, we’re going to get more specific.
617
1682385
2002
你知道,我們可以更具體。
28:04
And it’s going to be able to handle larger scale.
618
1684429
4296
這也可以用更大的規模去執行。
28:08
And that's something I think that I'm personally looking forward to.
619
1688766
4088
我個人相當期待這件事情。
28:13
CA: What about this perceived huge downside
620
1693813
4296
克:還有一件事被視為很大的問題,
28:18
of use of, certainly Instagram, I think TikTok as well.
621
1698151
3670
一定跟 Instagram 的使用有關, 我覺得抖音也是。
28:21
What people worry that you are amplifying insecurities,
622
1701863
3712
眾人擔心的是你們在增加不安全感,
28:25
especially of teenagers
623
1705575
1626
尤其是對於青少年,
28:27
and perhaps especially of teenage girls.
624
1707243
1919
可能最主要是青少女。
28:29
They see these amazing people on there doing amazing things,
625
1709162
3920
他們看到這些驚奇的人 在平臺上做著驚奇的事情,
28:33
they feel inadequate,
626
1713124
1251
覺得做得不夠,
28:34
there's all these reported cases of depression, insecurity,
627
1714417
3545
所以有一大堆憂鬱、不安全感、 輕生等案例被回報。
28:38
suicide and so forth.
628
1718004
1752
28:39
SC: I take this extremely seriously.
629
1719797
1835
周:我嚴肅看待這個問題。
28:42
So in our guidelines,
630
1722967
2377
所以在我們的守則中,
28:45
for certain themes that we think are mature and not suitable for teenagers,
631
1725386
5089
我們認為特定主題屬於成人, 不適合青少年,
28:50
we actually proactively remove it from their experience.
632
1730516
3254
我們實際上會主動將它 從他們的體驗中移除。
28:54
At the same time, if you search certain terms,
633
1734395
2544
與此同時,如果你搜尋特定關鍵字,
28:56
we will make sure that you get redirected to a resource safety page.
634
1736981
4087
我們會確保你被重新導向 到安全指南的頁面。
29:01
Now we are always working with experts to understand some of these new trends
635
1741068
4338
我們總是跟專家合作,
理解這些可能產生的新趨勢,
29:05
that could emerge
636
1745406
1168
29:06
and proactively try to manage them, if that makes sense.
637
1746616
3795
主動試圖管制它們, 如果有需要的話。
29:10
Now, this is a problem that predates us,
638
1750411
2961
這個問題在抖音出現之前就存在了。
29:13
that predates TikTok.
639
1753372
1168
29:14
It actually predates the internet.
640
1754540
1710
它其實比網路還早。
29:16
But it's our responsibility to make sure
641
1756250
2378
但我們仍然有責任去確保
29:18
that we invest enough to understand and to address the concerns,
642
1758628
3253
我們投入夠多資源 以理解並處理這些擔憂,
29:21
to keep the experience as safe as possible
643
1761923
2043
讓使用越多人的體驗越安全越好。
29:23
for as many people as possible.
644
1763966
1502
29:26
CA: Now, in Congress,
645
1766302
1210
克:在國會,
29:27
the main concern seemed to be not so much what we've talked about,
646
1767512
3461
主要的擔憂看起來不太是 我們先前討論的事情,
29:30
but data, the data of users,
647
1770973
2461
而是資料,使用者的資料,
29:33
the fact that you're owned by ByteDance, Chinese company,
648
1773476
3253
關於你們從屬於 字節跳動這家中國企業,
29:36
and the concern that at any moment
649
1776771
2753
以及擔心在任何時間點
29:39
Chinese government might require or ask for data.
650
1779524
3670
中國政府可能會要求或詢問資料。
29:43
And in fact, there have been instances
651
1783194
1835
而事實上,已經有實例,
29:45
where, I think you've confirmed,
652
1785029
1543
我認為你有承認過,
29:46
that some data of journalists on the platform
653
1786572
3963
有些記者在平臺上的資訊
29:50
was made available to ByteDance's engineers
654
1790576
3003
被設定成字節跳動的工程師可取用,
29:53
and from there, who knows what.
655
1793579
2211
從那裡開始,誰知道會發生什麼?
29:56
Now, your response to this was to have this Project Texas,
656
1796165
4254
你對此的回應是推動 「德克薩斯計畫」,
30:00
where you're moving data to be controlled by Oracle here in the US.
657
1800461
4755
你們會將資料移轉到美國, 由甲骨文公司控制。
30:05
Can you talk about that project and why, if you believe it so,
658
1805258
5130
你可不可以談談這個計畫,
以及你為什麼相信我們 不該這麼擔心這項議題?
30:10
why we should not worry so much about this issue?
659
1810429
3003
30:13
SC: I will say a couple of things about this, if you don't mind.
660
1813432
3045
周:若你不介意, 對此我有幾點要說。
我要說的第一件事是,
30:16
The first thing I would say is that the internet is built
661
1816519
2711
網路建立於全球互操作性之上,
30:19
on global interoperability,
662
1819230
1501
30:20
and we are not the only company that relies on the global talent pool
663
1820773
4171
我們不是唯一一家依靠全球人才庫 來讓產品盡可能變好的公司。
30:24
to make our products as good as possible.
664
1824944
1960
30:26
Technology is a very collaborative effort.
665
1826946
2002
科技非常需要合作。
30:28
I think many people here would say the same thing.
666
1828990
3461
我認為在場很多人也會同意。
30:32
So we are not the first company to have engineers in all countries,
667
1832493
3170
我們不是第一個在全球, 包括中國都有工程師的公司,
30:35
including in China.
668
1835663
1168
30:36
We're not the first one.
669
1836873
1167
我們不是第一家。
我可以理解有些顧慮。
30:38
Now, I understand some of these concerns.
670
1838082
2377
30:40
You know, the data access by employees is not data accessed by government.
671
1840459
4463
你知道,資料可以被員工取用 不代表就可以被政府取用。
30:44
This is very different, and there’s a clear difference in this.
672
1844922
2962
這很不一樣,差異點也十分清楚。
30:47
But we hear the concerns that are raised in the United States.
673
1847925
2962
但我們聽到在美國被提出的擔憂。
30:50
We did not try to avoid discussing.
674
1850928
3420
我們並未試圖迴避討論,
30:54
We did not try to argue our way out of it.
675
1854390
2878
沒有試圖辯駁脫身。
30:57
What we did was we built an unprecedented project
676
1857310
2836
我們所做的是建立 一個前所未有的計畫,
31:00
where we localize American data to be stored on American soil
677
1860187
4171
讓美國人的資料在美國本土儲存,
31:04
by an American company overseen by American personnel.
678
1864400
4004
由美國公司負責,受美國的人士監督。
31:08
So this kind of protection for American data
679
1868404
3212
以這種等級來保護美國人的資料
31:11
is beyond what any other company in our industry has ever done.
680
1871657
4213
遠遠超過我們產業中 任何其他公司所做到的水準。
31:16
Well, money is not the only issue here,
681
1876454
2544
錢不是唯一的障礙,
31:19
but it's very expensive to build something like that.
682
1879040
2502
但要建立這種東西是非常昂貴的。
31:21
And more importantly, you know,
683
1881584
1626
更重要的是,你知道,
我們將資料當地化的方式 基本上沒有企業做到過的。
31:23
we are basically localizing data in a way that no other company has done.
684
1883252
5673
31:28
So we need to be very careful that whilst we are pursuing
685
1888966
3963
因此我們必須非常小心,
當我們在美國追求所謂的數位主權,
31:32
what we call digital sovereignty in the US
686
1892929
2585
31:35
and we are also doing a version of this in Europe,
687
1895556
2503
以及在歐洲做這件事時,
31:38
that we don't balkanize the internet.
688
1898100
1794
我們不能讓網路四分五裂。
31:39
Now we are the first to do it.
689
1899936
1459
我們是首個這麼做的。
31:41
And I expect that, you know,
690
1901437
1460
而我預期,你知道,
31:42
other companies are probably looking at this
691
1902897
2085
其他公司可能也在關注,
31:45
and trying to figure out how you balance between protecting, protected data,
692
1905024
4630
試著找出如何平衡保護資料,
31:49
you know, to make sure that everybody feels secure about it
693
1909654
2836
你知道,確保所有人對此感到安全,
31:52
while at the same time allowing for interoperability
694
1912490
2586
同時允許互操作性持續發生,
31:55
to continue to happen,
695
1915117
1126
因為就是它讓科技和網路這麼棒。
31:56
because that's what makes technology and the internet so great.
696
1916285
3003
31:59
So that's something that we are doing.
697
1919330
1835
這是我們正在做的事情。
32:01
CA: How far are you along that journey with Project Texas?
698
1921207
2794
克:你在德克薩斯計畫的進程如何?
32:04
SC: We are very, very far along today.
699
1924043
1835
周:我們已經做了很多。
32:05
CA: When will there be a clear you know,
700
1925920
3920
克:什麼時候才會清楚地,你知道,
32:09
here it is, it’s done, it’s firewalled, this data is protected?
701
1929882
3087
它已經完成了,設立防火牆, 資料受到保障?
32:12
SC: Today, by default, all new US data
702
1932969
3086
周:今天,美國所有新資料都已經 預設儲存於甲骨文的雲端設備。
32:16
is already stored in the Oracle cloud infrastructure.
703
1936055
2502
32:18
So it's in this protected US environment that we talked about in the United States.
704
1938557
4713
它是在我們談過的 受保護的美國環境中。
32:23
We still have some legacy data to delete in our own servers in Virginia
705
1943270
3963
我們在維吉尼亞和新加坡的伺服器中 還有一些要刪除的舊資料。
32:27
and in Singapore.
706
1947274
1168
32:28
Our data has never been stored in China, by the way.
707
1948484
2461
順帶一提,我們的資料 從未儲存在中國。
32:31
That deletion is a very big engineering effort.
708
1951487
2211
資料的刪除是一項大工程。
32:33
So as we said, as I said at the hearing,
709
1953739
2378
就如我們說的, 就像我在聽證會說的,
32:36
it's going to take us a while to delete them,
710
1956117
2502
我們需要一些時間來將它刪除,
32:38
but I expect it to be done this year.
711
1958619
2294
但我預期在今年就可以完成。
32:43
CA: How much power do you have
712
1963791
3045
克:你個人有多大的權力 去控制特定的事項?
32:46
over your own ability to control certain things?
713
1966877
2837
32:49
So, for example, suppose that, for whatever reason,
714
1969714
2961
舉例而言,假如因為某個原因,
32:52
the Chinese government was to look at an upcoming US election and say,
715
1972717
4421
中國政府要關注 即將來臨的美國大選,說:
32:57
"You know what, we would like this party to win," let's say,
716
1977138
4754
「我們希望這個黨贏」,假設如此,
33:01
or "We would like civil war to break out" or whatever.
717
1981934
2711
或是「我們希望內戰爆發」等等。
33:05
How ...
718
1985104
1376
如何……
33:06
"And we could do this
719
1986522
1418
「我們辦得到,
33:07
by amplifying the content of certain troublemaking, disturbing people,
720
1987982
4212
只要多加推廣特定幾位 製造麻煩的搗亂者的內容,
33:12
causing uncertainty, spreading misinformation," etc.
721
1992236
2795
造成不確定性,傳遞不實訊息」等。
33:16
If you were required via ByteDance to do this,
722
1996032
4963
如果字節跳動要求你做這件事,
33:21
like, first of all, is there a pathway where theoretically that is possible?
723
2001037
4421
首先,有理論上可行的方法嗎?
33:26
What's your personal line in the sand on this?
724
2006542
3462
你個人對此的底線是什麼?
33:30
SC: So during the congressional hearing,
725
2010046
2544
周:在國會聽證會中,
33:32
I made four commitments,
726
2012631
1585
我做了四項承諾,
33:34
we talked about the first one, which is safety.
727
2014216
2211
我們談了第一點:安全。
33:36
The third one is to keep TikTok a place of freedom of expression.
728
2016469
3420
第三點是確保抖音是個 有言論自由的場所。
33:39
By the way, if you go on TikTok today,
729
2019930
1836
如果你今天打開抖音, 你可以搜尋任何東西,
33:41
you can search for anything you want,
730
2021766
1793
只要不要違反我們的社群守則。
33:43
as long as it doesn't violate our community guidelines.
731
2023559
2753
33:46
And to keep it free from any government manipulation.
732
2026312
3003
我們會讓抖音免於受任何政府干預。
33:49
And the fourth one is transparency and third-party monitoring.
733
2029315
4129
第四點是透明度和第三方監督。
33:53
So the way we are trying to address this concern
734
2033444
2294
我們試圖處理這個顧慮的方法是
33:55
is an unprecedented amount of transparency.
735
2035780
2294
提供前所未有的透明度。
33:58
What do I mean by this?
736
2038074
1459
我這麼說是什麼意思?
33:59
We're actually allowing third-party reviewers
737
2039533
3546
我們其實允許第三方的審查者
34:03
to come in and review our source code.
738
2043079
2127
進來審核我們的原始碼。
34:05
I don't know any other company that does this, by the way.
739
2045206
2752
我沒聽過其他公司這麼做。
因為所有事情,如你所知, 都受程式驅動。
34:08
Because everything, as you know, is driven by code.
740
2048417
2670
34:11
So to allow someone else to review the source code
741
2051128
2670
所以允許他人審核原始碼
34:13
is to give this a significant amount of transparency
742
2053798
3003
就是提供非常高的透明度,
34:16
to ensure that the scenarios that you described
743
2056842
2336
以確保你所述的高度假設性情境
34:19
that are highly hypothetical, cannot happen on our platform.
744
2059220
3503
不可能發生在我們平臺。
34:23
Now, at the same time,
745
2063057
1168
與此同時,
34:24
we are releasing more research tools for researchers
746
2064266
3879
我們正在發布更多 給研究者的研究工具,
34:28
so that they can study the output.
747
2068187
1627
讓他們可以研究輸出。
34:29
So the source code is the input.
748
2069814
2210
原始碼是輸入,
34:32
We are also allowing researchers to study the output,
749
2072024
2628
我們也允許研究者研究輸出,
34:34
which is the content on our platform.
750
2074652
1793
也就是我們平臺上的內容。
34:36
I think the easiest way to sort of fend this off is transparency.
751
2076487
4087
我認為最簡單的解決之道 就是透明度。
34:40
You know, we give people access to monitor us,
752
2080574
2753
你知道,我們讓大眾 有管道監督我們,
34:43
and we just make it very, very transparent.
753
2083369
2002
而且我們讓它極為透明。
34:45
And that's our approach to the problem.
754
2085412
1836
這是我們的解決方法。
克:所以你會直接告訴這個小組,
34:47
CA: So you will say directly to this group
755
2087248
2252
34:49
that the scenario I talked about,
756
2089500
2252
我所描述的情境,
34:51
of theoretical Chinese government interference in an American election,
757
2091794
4671
理論上的中國政府干預美國大選,
34:56
you can say that will not happen?
758
2096465
2753
你可以說這不會發生嗎?
34:59
SC: I can say that we are building all the tools
759
2099677
2711
周:我可以說我們建立這些工具
35:02
to prevent any of these actions from happening.
760
2102388
2919
就是為了預防任何這類行動的發生。
35:05
And I'm very confident that with an unprecedented amount of transparency
761
2105933
4338
而且我很有信心,
靠著我們給予這個平臺的 前所未見的透明性,
35:10
that we're giving on the platform,
762
2110271
1626
35:11
we can reduce this risk to as low as zero as possible.
763
2111939
4421
我們可以將風險盡可能降到零。
35:17
CA: To as low as zero as possible.
764
2117403
2753
克:盡可能降到零。
35:20
SC: To as close to zero as possible.
765
2120197
2419
周:盡可能降到零。
35:22
CA: As close to zero as possible.
766
2122616
1877
克:盡可能降到零。
35:25
That's fairly reassuring.
767
2125452
1919
這相當讓人放心。
35:27
Fairly.
768
2127830
1168
相當。
35:28
(Laughter)
769
2128998
1168
(笑聲)
35:33
I mean, how would the world know?
770
2133085
1585
我是指,世界怎麼知道?
35:34
If you discovered this or you thought you had to do it,
771
2134712
2586
如果你發現了, 或是你覺得你必須做,
35:37
is this a line in the sand for you?
772
2137298
3461
這是你的底線嗎?
35:40
Like, are you in a situation you would not let the company that you know now
773
2140801
3587
你是否不會讓你所知道 且正在經營的公司做這件事?
35:44
and that you are running do this?
774
2144430
1585
周:絕對不會。這就是我們 讓第三方來監督的原因,
35:46
SC: Absolutely.
775
2146056
1168
35:47
That's the reason why we're letting third parties monitor,
776
2147266
2711
因為如果他們有所發現, 他們會揭發我們。
35:50
because if they find out, you know, they will disclose this.
777
2150019
2836
35:52
We also have transparency reports, by the way,
778
2152855
2169
我們也有透明度報告,
我們在其中談到一大堆事情:
35:55
where we talk about a whole bunch of things,
779
2155024
2085
35:57
the content that we remove, you know, that violates our guidelines,
780
2157151
3545
我們移除的內容, 你知道,違反守則的內容、
36:00
government requests.
781
2160738
1168
政府的要求。
36:01
You know, it's all published online.
782
2161906
1751
這都公布在網路。
36:03
All you have to do is search for it.
783
2163657
1752
你只需要去搜尋它。
克:我得說,你是非常讓人信服 且喜歡的執行長。
36:05
CA: So you're super compelling
784
2165409
1460
36:06
and likable as a CEO, I have to say.
785
2166911
1751
36:08
And I would like to, as we wrap this up,
786
2168662
2419
我想要,在我們結束前,
36:11
I'd like to give you a chance just to paint, like, what's the vision?
787
2171123
3253
我想給你一個機會 描繪你們的願景是什麼?
36:14
As you look at what TikTok could be,
788
2174418
4087
當你在想抖音可以成為什麼,
36:18
let's move the clock out, say, five years from now.
789
2178547
3295
比如說從現在開始五年之後。
36:21
How should we think about your contribution to our collective future?
790
2181884
4504
我們應該如何看待 你對我們全體的未來的貢獻?
36:26
SC: I think it's still down to the vision that we have.
791
2186430
2586
周:我認為這依然 要回到我們的願景。
36:29
So in terms of the window of discovery,
792
2189058
2794
就成為發現的窗口而言,
36:31
I think there's a huge benefit to the world
793
2191852
2878
我認為當人們可以發現新事物, 這對世界是很大的助益。
36:34
when people can discover new things.
794
2194772
1835
36:36
You know, people think that TikTok is all about dancing and singing,
795
2196607
3211
你知道,大家覺得 抖音都是跳舞唱歌,
36:39
and there’s nothing wrong with that, because it’s super fun.
796
2199860
2836
這也沒有錯,因為這超級有趣。
36:42
There's still a lot of that,
797
2202696
1377
抖音很多這類內容,
但我們看到科學、STEM 的內容。
36:44
but we're seeing science content, STEM content,
798
2204073
2878
36:46
have you about BookTok?
799
2206992
1794
你聽過「BookTok」嗎?
36:48
It's a viral trend that talks about books
800
2208827
2670
這是一個爆紅的潮流,
它談論書籍並鼓勵大眾閱讀。
36:51
and encourages people to read.
801
2211538
1711
36:53
That BookTok has 120 billion views globally,
802
2213290
3963
BookTok 在全球 有 1200 億觀看次數,
36:57
120 billion.
803
2217294
1168
1200 億。
36:58
CA: Billion, with a B.
804
2218462
1460
克:是億。
36:59
SC: People are learning how to cook,
805
2219964
2002
周:人們在學烹飪,人們在學科學,
37:02
people are learning about science,
806
2222007
1627
37:03
people are learning how to golf --
807
2223634
1668
人們在學打高爾夫球——
37:05
well, people are watching videos on golfing, I guess.
808
2225302
2586
嗯,大家只是在看 高爾夫球影片,我猜。
37:07
(Laughter)
809
2227930
1877
(笑聲)
37:10
I haven't gotten better by looking at the videos.
810
2230349
2961
我的球技沒有因為 看這些影片而進步。
37:13
I think there's a huge, huge opportunity here on discovery
811
2233352
3754
我認為這裡有非常大的機會去發現,
37:17
and giving the everyday person a voice.
812
2237147
2044
並讓平凡人也可以發聲。
37:19
If you talk to our creators, you know,
813
2239233
1835
如果你跟我們的創作者談談, 很多人會一再告訴你:
37:21
a lot of people will tell you this again and again, that before TikTok,
814
2241068
3378
在有抖音之前,
37:24
they would never have been discovered.
815
2244446
1836
他們絕對不可能被發現。
37:26
And we have given them the platform to do that.
816
2246323
2211
我們給了他們平臺, 維持它是很重要的。
37:28
And it's important to maintain that.
817
2248534
1752
然後我們談了創作。
37:30
Then we talk about creation.
818
2250286
1376
37:31
You know, there’s all this new technology coming in with AI-generated content
819
2251662
4379
你知道,隨著人工智慧生成內容 新科技的到來,
37:36
that will help people create even more creative content.
820
2256083
3962
它會幫助人們創作出 更有創意的內容。
我認為兩者將會互相合作,
37:40
I think there's going to be a collaboration between,
821
2260045
2461
37:42
and I think there's a speaker who is going to talk about this,
822
2262506
2920
我想有一名講者會談這個主題,
在人類與人工智慧之間,
37:45
between people and AI
823
2265467
1168
37:46
where they can unleash their creativity in a different way.
824
2266635
2795
兩者如何以不同的方式激發創造力。
37:49
You know, like for example, I'm terrible at drawing personally,
825
2269471
3295
你知道,舉個例子, 我個人很不會畫畫,
37:52
but if I had some AI to help me,
826
2272808
2377
但如果我有某種人工智慧的幫忙,
37:55
then maybe I can express myself even better.
827
2275227
2127
也許我可以更好地表達想法。
37:57
Then we talk about bridges to connect
828
2277813
1793
我們談到連結的橋梁,
37:59
and connecting people and the communities together.
829
2279648
2419
以及把眾人和社群連結在一起。
這可以是產品,可以有商機,
38:02
This could be products, this could be commerce,
830
2282067
2253
38:04
five million businesses in the US benefit from TikTok today.
831
2284320
5714
如今有 500 萬家 美國企業從抖音受惠。
38:10
I think we can get that number to a much higher number.
832
2290075
2586
我認為我們可以 讓這個數字高得多。
38:12
And of course, if you look around the world, including in Canada,
833
2292661
3087
當然,如果你看全世界, 包含加拿大,
這個數字將會非常大。
38:15
that number is going to be massive.
834
2295748
1710
所以我認為這些是我們最大的機會,
38:17
So I think these are the biggest opportunities that we have,
835
2297458
2836
而且這讓人相當興奮。
38:20
and it's really very exciting.
836
2300336
1459
38:21
CA: So courtesy of your experience in Congress,
837
2301795
2211
克:由於你在國會的經驗,
38:24
you actually became a bit of a TikTok star yourself, I think.
838
2304048
3586
你自己其實也變得有點像 抖音上的明星,我認為。
38:28
Some of your videos have gone viral.
839
2308010
2210
有些你的影片爆紅。
要不要…… 你有帶手機吧?
38:31
You've got your phone with you.
840
2311055
1501
38:32
Do you want to make a little little TikTok video right now?
841
2312556
2795
你現在想要拍一個 短短的抖音影片嗎?
38:35
Let's do this.
842
2315351
1167
那就做吧!
38:36
SC: If you don't mind ...
843
2316560
1210
周:若你不介意,
38:37
CA: What do you think, should we do this?
844
2317811
1919
克:你覺得要嗎? 周:我們來自拍如何?
38:39
SC: We're just going to do a selfie together, how's that?
845
2319730
2711
克:好的。
38:42
So why don't we just say "Hi."
846
2322816
1669
周:我們說聲「嗨」吧。
38:44
Hi!
847
2324526
1168
嗨!
38:45
Audience: Hi!
848
2325694
1168
克:TED 的哈囉。
38:46
CA: Hello from TED.
849
2326904
1126
克:(笑聲)
38:48
SC: All right, thank you, I hope it goes viral.
850
2328405
2211
周:好,謝謝,我希望它爆紅。
38:50
(Laughter)
851
2330657
2795
(笑聲)
38:53
CA: If that one goes viral, I think I've given up on your algorithm, actually.
852
2333494
3670
克:如果它可以爆紅, 我會放棄你的演算法。
(笑聲)
38:57
(Laughter)
853
2337206
1543
39:00
Shou Chew, you're one of the most influential
854
2340417
2753
周受資,你是世上 最有影響力的人之一,
39:03
and powerful people in the world, whether you know it or not.
855
2343212
2878
無論你知不知道。
我非常感謝你過來分享你的願景。
39:06
And I really appreciate you coming and sharing your vision.
856
2346090
2794
39:08
I really, really hope the upside of what you're talking about comes about.
857
2348884
3545
我真的很希望 你談論的那些好處發生。
39:12
Thank you so much for coming today.
858
2352471
1710
克:非常感謝你今天過來。 周:謝謝,克里斯。
39:14
SC: Thank you, Chris.
859
2354181
1168
39:15
CA: It's really interesting.
860
2355391
1376
克:這真的很有趣。
39:16
(Applause)
861
2356767
1627
(掌聲)
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7