TikTok CEO Shou Chew on Its Future — and What Makes Its Algorithm Different | Live at TED2023

1,560,334 views

2023-04-21 ・ TED


New videos

TikTok CEO Shou Chew on Its Future — and What Makes Its Algorithm Different | Live at TED2023

1,560,334 views ・ 2023-04-21

TED


Please double-click on the English subtitles below to play the video.

00:00
Chris Anderson: It's very nice to have you here.
0
288
2294
00:03
Let's see.
1
3166
1126
00:04
First of all, congratulations.
2
4333
1460
00:05
You really pulled off something remarkable on that grilling,
3
5793
2836
00:08
you achieved something that very few people do,
4
8629
2211
00:10
which was, you pulled off a kind of, a bipartisan consensus in US politics.
5
10882
4671
00:15
It was great.
6
15553
1168
00:16
(Laughter)
7
16763
1376
00:18
The bad news was that that consensus largely seemed to be:
8
18723
2753
00:21
"We must ban TikTok."
9
21517
2294
00:23
So we're going to come to that in a bit.
10
23853
1960
00:25
And I'm curious, but before we go there, we need to know about you.
11
25813
4588
00:30
You seem to me like a remarkable person.
12
30443
2878
00:33
I want to know a bit of your story
13
33362
1627
00:35
and how you came to TikTok in the first place.
14
35031
2627
00:37
Shou Chew: Thank you, Chris.
15
37658
1377
00:39
Before we do that, can I just check, need to know my audience,
16
39035
3003
00:42
how many of you here use TikTok?
17
42080
1876
00:45
Oh, thank you.
18
45124
1168
00:46
For those who don’t, the Wi-Fi is free.
19
46292
2085
00:48
(Laughter)
20
48419
1919
00:51
CA: There’s another question, which is,
21
51631
1877
00:53
how many of you here have had your lives touched through TikTok,
22
53508
3169
00:56
through your kids and other people in your lives?
23
56719
2461
01:00
SC: Oh, that's great to see.
24
60515
1376
01:01
CA: It's basically, if you're alive,
25
61891
1752
01:03
you have had some kind of contact with TikTok at this point.
26
63643
3211
01:06
So tell us about you.
27
66854
1418
01:08
SC: So my name is Shou, and I’m from Singapore.
28
68272
2920
01:11
Roughly 10 years ago,
29
71651
1668
01:13
I met with two engineers who were building a product.
30
73361
4004
01:17
And the idea behind this was to build a product
31
77365
2961
01:20
that recommended content to people not based on who they knew,
32
80368
4796
01:25
which was, if you think about it, 10 years ago,
33
85206
3003
01:28
the social graph was all in the rage.
34
88209
3045
01:31
And the idea was, you know,
35
91295
1293
01:32
your content and the feed that you saw should be based on people that you knew.
36
92588
3921
01:36
But 10 years ago,
37
96926
1168
01:38
these two engineers thought about something different,
38
98136
2544
01:40
which is, instead of showing you --
39
100680
3545
01:44
instead of showing you people you knew,
40
104225
2461
01:46
why don't we show you content that you liked?
41
106727
2253
01:49
And that's sort of the genesis and the birth
42
109021
2211
01:51
of the early iterations of TikTok.
43
111232
2794
01:54
And about five years ago,
44
114443
1585
01:56
with the advent of 4G, short video, mobile phone penetration,
45
116070
5798
02:01
TikTok was born.
46
121868
1376
02:03
And a couple of years ago,
47
123286
1293
02:04
I had the opportunity to run this company,
48
124620
2419
02:07
and it still excites me every single day.
49
127039
2002
02:09
CA: So I want to dig in a little more into this,
50
129750
2253
02:12
about what was it that made this take-off so explosive?
51
132044
2962
02:15
Because the language I hear from people who spent time on it,
52
135047
3963
02:19
it's sort of like I mean,
53
139051
2128
02:21
it is a different level of addiction to other media out there.
54
141220
5965
02:27
And I don't necessarily mean this in a good way, we'll be coming on to it.
55
147226
3504
02:30
There’s good and bad things about this type of addiction.
56
150730
2711
02:33
But it’s the feeling
57
153441
1168
02:34
that within a couple of days of experience of TikTok,
58
154650
2503
02:37
it knows you and it surprises you
59
157195
2502
02:39
with things that you didn't know you were going to be interested in,
60
159739
3211
02:42
but you are.
61
162950
1293
02:44
How?
62
164243
1210
02:45
Is it really just, instead of the social graph --
63
165453
2627
02:48
What are these algorithms doing?
64
168080
2044
02:50
SC: I think to describe this, to begin to answer your question,
65
170124
4171
02:54
we have to talk about the mission of the company.
66
174337
2502
02:56
Now the mission is to inspire creativity and to bring joy.
67
176839
3504
03:00
And I think missions for companies like ours [are] really important.
68
180384
3629
03:04
Because you have product managers working on the product every single day,
69
184055
3545
03:07
and they need to have a North Star, you know,
70
187600
2127
03:09
something to sort of, work towards together.
71
189769
2586
03:12
Now, based on this mission,
72
192355
1543
03:13
our vision is to provide three things to our users.
73
193898
3378
03:17
We want to provide a window to discover,
74
197318
2127
03:19
and I’ll talk about discovery, you talked about this, in a second.
75
199445
3170
03:22
We want to give them a canvas to create,
76
202657
2085
03:24
which is going to be really exciting with new technologies in AI
77
204784
3879
03:28
that are going to help people create new things.
78
208704
3587
03:32
And the final thing is bridges for people to connect.
79
212792
2919
03:35
So that's sort of the vision of what we're trying to build.
80
215753
2794
03:38
Now what really makes TikTok very unique and very different
81
218547
3754
03:42
is the whole discovery engine behind it.
82
222343
3003
03:45
So there are earlier apps that I have a lot of respect for,
83
225721
2878
03:48
but they were built for a different purpose.
84
228641
2127
03:50
For example, in the era of search, you know,
85
230768
3128
03:53
there was an app that was built for people who wanted to search things
86
233896
3504
03:57
so that is more easily found.
87
237441
3420
04:00
And then in the era of social graphs,
88
240903
1794
04:02
it was about connecting people and their followers.
89
242697
2586
04:05
Now what we have done is that ... based on our machine-learning algorithms,
90
245283
4212
04:09
we're showing people what they liked.
91
249578
2294
04:11
And what this means is that we have given the everyday person
92
251872
3087
04:15
a platform to be discovered.
93
255001
1918
04:16
If you have talent, it is very, very easy to get discovered on TikTok.
94
256919
3962
04:20
And I'll just give you one example of this.
95
260923
2544
04:23
The biggest creator on TikTok is a guy called Khaby.
96
263467
3629
04:27
Khaby was from Senegal,
97
267430
1751
04:29
he lives in Italy, he was a factory worker.
98
269181
2711
04:32
He, for the longest time, didn't even speak in any of his videos.
99
272685
4338
04:37
But what he did was he had talent.
100
277064
2336
04:39
He was funny, he had a good expression,
101
279442
2252
04:41
he had creativity, so he kept posting.
102
281694
2461
04:44
And today he has 160 million followers on our platform.
103
284196
3587
04:47
So every single day we hear stories like that,
104
287825
2169
04:49
businesses, people with talent.
105
289994
2252
04:52
And I think it's very freeing to have a platform
106
292288
3336
04:55
where, as long as you have talent, you're going to be heard
107
295666
3128
04:58
and you have the chance to succeed.
108
298794
1877
05:00
And that's what we're providing to our users.
109
300713
2210
05:02
CA: So this is the amazing thing to me.
110
302923
1919
05:04
Like, most of us have grown up with, say, network television,
111
304884
4337
05:09
where, for decades you've had thousands of brilliant, creative people
112
309263
3587
05:12
toiling in the trenches,
113
312892
1168
05:14
trying to imagine stuff that will be amazing for an audience.
114
314101
4380
05:18
And none of them ever remotely came up with anything
115
318522
4255
05:22
that looked like many of your creators.
116
322818
3921
05:27
So these algorithms,
117
327114
2294
05:29
just by observing people's behavior and what they look like,
118
329450
3921
05:33
have discovered things that thousands of brilliant humans never discovered.
119
333412
3587
05:37
Tell me some of the things that it is looking at.
120
337041
3045
05:40
So obvious things, like if someone presses like
121
340086
2502
05:42
or stays on a video for a long time,
122
342630
1877
05:44
that gives you a clue, "more like that."
123
344548
2545
05:47
But is it subject matter?
124
347134
1210
05:48
What are the array of things
125
348386
2377
05:50
that you have noticed that you can actually track
126
350805
2419
05:53
that provide useful clues?
127
353224
2043
05:55
SC: I'm going to simplify this a lot,
128
355309
2461
05:57
but the machine learning, the recommendation algorithm
129
357812
2877
06:00
is really just math.
130
360731
1377
06:02
So, for example, if you liked videos one, two, three and four,
131
362108
4587
06:06
and I like videos one, two, three and five,
132
366737
2211
06:08
maybe he liked videos one, two, three and six.
133
368989
2253
06:11
Now what's going to happen is,
134
371242
1501
06:12
because we like one, two, three at the same time,
135
372743
2336
06:15
he's going to be shown four, five, six, and so are we.
136
375121
2877
06:18
And you can think about this repeated at scale in real time
137
378040
3379
06:21
across more than a billion people.
138
381460
1877
06:23
That's basically what it is, it's math.
139
383337
2044
06:25
And of course, you know,
140
385423
1501
06:26
AI and machine learning has allowed this to be done
141
386966
2794
06:29
at a very, very big scale.
142
389802
1710
06:31
And what we have seen, the result of this,
143
391554
2085
06:33
is that it learns the interest signals
144
393681
2669
06:36
that people exhibit very quickly
145
396392
1793
06:38
and shows you content that's really relevant for you
146
398227
3128
06:41
in a very quick way.
147
401355
2461
06:44
CA: So it's a form of collaborative filtering, from what you're saying.
148
404442
3336
06:47
The theory behind it is that these humans are weird people,
149
407820
2794
06:50
we don't really know what they're interested in,
150
410614
2253
06:52
but if we see that one human is interested,
151
412908
2002
06:54
with an overlap of someone else, chances are, you know,
152
414952
2836
06:57
you could make use of the other pieces
153
417830
2002
06:59
that are in that overlapped human's repertoire to feed them,
154
419874
5463
07:05
and they'll be surprised.
155
425337
1252
07:06
But the reason they like it is because their pal also liked it.
156
426589
3670
07:10
SC: It's pattern recognition based on your interest signals.
157
430301
3253
07:13
And I think the other thing here
158
433596
1543
07:15
is that we don't actually ask you 20 questions
159
435139
2669
07:17
on whether you like a piece of content, you know, what are your interests,
160
437808
3504
07:21
we don't do that.
161
441312
1209
07:22
We built that experience organically into the app experience.
162
442563
2878
07:25
So you are voting with your thumbs by watching a video,
163
445441
3545
07:29
by swiping it, by liking it, by sharing it,
164
449028
3336
07:32
you are basically exhibiting interest signals.
165
452364
2586
07:34
And what it does mathematically is to take those signals,
166
454992
2711
07:37
put it in a formula and then matches it through pattern recognition.
167
457745
3211
07:40
That's basically the idea behind it.
168
460998
2211
07:43
CA: I mean, lots of start-ups have tried to use these types of techniques.
169
463626
5422
07:49
I'm wondering what else played a role early on?
170
469089
2211
07:51
I mean, how big a deal was it,
171
471300
1460
07:52
that from the get-go you were optimizing for smartphones
172
472760
3253
07:56
so that videos were shot in portrait format
173
476013
2711
07:58
and they were short.
174
478724
1877
08:00
Was that an early distinguishing thing that mattered?
175
480643
2961
08:03
SC: I think we were the first to really try this at scale.
176
483646
2752
08:06
You know, the recommendation algorithm is a very important reason
177
486941
3712
08:10
as to why the platform is so popular among so many people.
178
490653
5088
08:15
But beyond that, you know, you mentioned the format itself.
179
495783
4296
08:20
So we talked about the vision of the company,
180
500079
2127
08:22
which is to have a window to discover.
181
502206
1835
08:24
And if you just open the app for the first time,
182
504083
2252
08:26
you'll see that it takes up your whole screen.
183
506377
2169
08:28
So that's the window that we want.
184
508546
1626
08:30
You can imagine a lot of people using that window
185
510214
2294
08:32
to discover new things in their lives.
186
512508
1835
08:34
Then, you know, through this recommendation algorithm,
187
514385
3545
08:37
we have found that it connects people together.
188
517972
2210
08:40
People find communities,
189
520182
1293
08:41
and I've heard so many stories of people who have found their communities
190
521475
4088
08:45
because of the content that they're posting.
191
525604
2086
08:47
Now, I'll give you an example.
192
527690
1459
08:49
I was in DC recently, and I met with a bunch of creators.
193
529567
3670
08:53
CA: I heard.
194
533237
1168
08:54
(Laughter)
195
534446
1669
08:56
SC: One of them was sitting next to me at a dinner,
196
536156
2420
08:58
his name is Samuel.
197
538576
1209
08:59
He runs a restaurant in Phoenix, Arizona, and it's a taco restaurant.
198
539827
4921
09:04
He told me he has never done this before, first venture.
199
544790
2795
09:08
He started posting all this content on TikTok,
200
548252
2419
09:10
and I saw his content,
201
550713
1376
09:12
I was hungry after looking at it, it's great content.
202
552131
3920
09:16
And he's generated so much interest in his business,
203
556051
3254
09:19
that last year he made something like a million dollars in revenue
204
559305
3128
09:22
just via TikTok.
205
562474
1168
09:23
One restaurant.
206
563684
1168
09:24
And again and again, I hear these stories,
207
564893
2211
09:27
you know, by connecting people together,
208
567104
2419
09:29
by giving people the window to discover,
209
569565
2210
09:31
we have given many small businesses and many people, your common person,
210
571775
4463
09:36
a voice that they will never otherwise have.
211
576280
2544
09:38
And I think that's the power of the platform.
212
578824
2169
09:41
CA: So you definitely have identified early
213
581327
2043
09:43
just how we're social creatures, we need affirmation.
214
583370
3754
09:47
I've heard a story,
215
587166
1168
09:48
and you can tell me whether true or not,
216
588375
1961
09:50
that one of the keys to your early liftoff
217
590336
2961
09:53
was that you wanted to persuade creators who were trying out TikTok
218
593297
4212
09:57
that this was a platform where they would get response,
219
597509
4088
10:01
early on, when you're trying to grow something,
220
601597
2210
10:03
the numbers aren't there for response.
221
603849
1835
10:05
So you had the brilliant idea of goosing those numbers a bit,
222
605684
3253
10:08
basically finding ways to give people, you know,
223
608979
2503
10:11
a bigger sense of like, more likes,
224
611482
1793
10:13
more engagement than was actually the case,
225
613317
2044
10:15
by using AI agents somehow in the process.
226
615361
4254
10:19
Is that a brilliant idea, or is that just a myth?
227
619657
2711
10:22
SC: I would describe it in a different way.
228
622409
4004
10:26
So there are other platforms that exist before TikTok.
229
626413
3504
10:29
And if you think about those platforms,
230
629917
1918
10:31
you sort of have to be famous already in order to get followers.
231
631877
3754
10:35
Because the way it’s built is that people come and follow people.
232
635631
3712
10:39
And if you aren't already famous,
233
639385
2377
10:41
the chances that you get discovered are very, very low.
234
641762
3170
10:44
Now, what we have done, again,
235
644973
1460
10:46
because of the difference in the way we're recommending content,
236
646433
3420
10:49
is that we have given anyone,
237
649895
1543
10:51
any single person with enough talent a stage to be able to be discovered.
238
651480
5339
10:56
And I think that actually is the single, probably the most important thing
239
656819
3545
11:00
contributing to the growth of the platform.
240
660364
2085
11:02
And again and again, you will hear stories from people who use the platform,
241
662491
3587
11:06
who post regularly on it,
242
666120
1626
11:07
that if they have something they want to say,
243
667788
2628
11:10
the platform gives them the chance and the stage
244
670457
2253
11:12
to connect with their audience
245
672751
1460
11:14
in a way that I think no other product in the past has ever offered them.
246
674211
4505
11:19
CA: So I'm just trying to play back what you said there.
247
679341
2628
11:21
You said you were describing a different way what I said.
248
681969
3128
11:25
Is it then the case that like, to give someone a decent chance,
249
685139
3712
11:28
someone who's brilliant but doesn't come with any followers initially,
250
688892
3379
11:32
that you've got some technique to identify talent
251
692312
3629
11:35
and that you will almost encourage them,
252
695983
2085
11:38
you will give them some kind of, you know,
253
698110
2794
11:40
artificially increase the number of followers or likes
254
700946
2544
11:43
or whatever that they have,
255
703532
1335
11:44
so that others are encouraged to go,
256
704867
1751
11:46
"Wow, there's something there."
257
706618
1544
11:48
Like it's this idea of critical mass that kind of, every entrepreneur,
258
708162
3336
11:51
every party planner kind of knows about of
259
711498
2044
11:53
"No, no, this is the hot place in town, everyone come,"
260
713542
2628
11:56
and that that is how you actually gain critical mass?
261
716170
2627
11:58
SC: We want to make sure that every person who posts a video
262
718839
3754
12:02
is given an equal chance to be able to have some audience to begin with.
263
722634
4713
12:07
But this idea that you are maybe alluding to,
264
727389
3379
12:10
that we can get people to like something,
265
730809
3003
12:13
it doesn't really work like that.
266
733854
1585
12:15
CA: Could you get AI agents to like something?
267
735439
2919
12:18
Could you seed the network with extra AI agents that could kind of, you know,
268
738400
4046
12:22
give someone early encouragement?
269
742488
1960
12:24
SC: Ultimately, what the machine does is it recognizes people's interests.
270
744490
4546
12:29
So if you post something that's not interesting to a lot of people,
271
749077
3170
12:32
even if you gave it a lot of exposure,
272
752289
1835
12:34
you're not going to get the virality that you want.
273
754166
2419
12:36
So it's a lot of ...
274
756585
1793
12:38
There is no push here.
275
758420
2878
12:41
It's not like you can go and push something,
276
761340
2669
12:44
because I like Chris, I'm going to push your content,
277
764009
2503
12:46
it doesn't work like that.
278
766553
1252
12:47
You've got to have a message that resonates with people,
279
767805
2627
12:50
and if it does,
280
770474
1168
12:51
then it will automatically just have the virality itself.
281
771683
2711
12:54
That's the beauty of user-generated content.
282
774394
2253
12:56
It's not something that can be engineered or over-thought.
283
776688
4046
13:00
It really is something that has to resonate with the audience.
284
780776
3462
13:04
And if it does, then it goes viral.
285
784238
1918
13:06
CA: Speaking privately with an investor who knows your company quite well,
286
786198
4588
13:10
who said that actually the level of sophistication
287
790786
4463
13:15
of the algorithms you have going
288
795249
1543
13:16
is just another order of magnitude
289
796834
1918
13:18
to what competitors like, you know, Facebook or YouTube have going.
290
798794
4379
13:23
Is that just hype or do you really believe you --
291
803924
3253
13:27
like, how complex are these algorithms?
292
807177
3420
13:31
SC: Well, I think in terms of complexity,
293
811014
2002
13:33
there are many companies who have a lot of resources
294
813016
2545
13:35
and a lot of talent.
295
815602
1168
13:36
They will figure out even the most complex algorithms.
296
816770
2920
13:39
I think what is very different is your mission of your company,
297
819982
3795
13:43
how you started the company.
298
823819
1626
13:45
Like I said, you know, we started with this idea
299
825445
2294
13:47
that this was the main use case.
300
827739
2128
13:50
The most important use case is you come and you get to see recommended content.
301
830158
4463
13:54
Now for some other apps out there,
302
834663
1835
13:56
they are very significant and have a lot of users,
303
836540
2544
13:59
they are built for a different original purpose.
304
839084
2920
14:02
And if you are built for something different,
305
842004
2127
14:04
then your users are used to that
306
844131
1543
14:05
because the community comes in and they expect that sort of experience.
307
845716
3336
14:09
So I think the pivot away from that
308
849052
1710
14:10
is not really just a matter of engineering and algorithms,
309
850762
3379
14:14
it’s a matter of what your company is built to begin with.
310
854141
3837
14:18
Which is why I started this by saying you need to have a vision,
311
858020
3003
14:21
you need to have a mission, and that's the North Star.
312
861064
2545
14:23
You can't just shift it halfway.
313
863650
1585
14:25
CA: Right.
314
865235
1168
14:26
And is it fair to say
315
866445
1168
14:27
that because your start point has been interest algorithms
316
867613
2752
14:30
rather than social graph algorithms,
317
870365
1919
14:32
you've been able to avoid some of the worst of the sort of,
318
872326
2794
14:35
the filter bubbles that have happened in other social media
319
875120
2795
14:37
where you have tribes kind of declaring war on each other effectively.
320
877956
3337
14:41
And so much of the noise and energy is around that.
321
881293
4296
14:45
Do you believe that you've largely avoided that on TikTok?
322
885631
2752
14:48
SC: The diversity of content that our users see is very key.
323
888425
4254
14:52
You know, in order for the discovery -- the mission is to discover --
324
892721
4213
14:56
sorry, the vision is to discover.
325
896975
2044
14:59
So in order to facilitate that,
326
899061
1626
15:00
it is very important to us
327
900729
1710
15:02
that what the users see is a diversity of content.
328
902481
3128
15:06
Now, generally speaking, you know,
329
906276
2461
15:08
there are certain issues that you mentioned
330
908779
2002
15:10
that the industry faces, you know.
331
910822
1627
15:12
There are some bad actors who come on the internet,
332
912491
2419
15:14
they post bad content.
333
914952
1126
15:16
Now our approach is that we have very clear community guidelines.
334
916119
4046
15:20
We're very transparent about what is allowed
335
920165
2377
15:22
and what is not allowed on our platform.
336
922542
1961
15:24
No executives make any ad hoc decisions.
337
924544
3003
15:27
And based on that,
338
927589
1126
15:28
we have built a team that is tens of thousands of people plus machines
339
928757
4338
15:33
in order to identify content that is bad
340
933136
2461
15:35
and actively and proactively remove it from the platform.
341
935639
3211
15:38
CA: Talk about what some of those key guidelines are.
342
938892
2711
15:41
SC: We have it published on our website.
343
941603
2127
15:43
In March, we just iterated a new version to make it more readable.
344
943772
5672
15:49
So there are many things like, for example, no pornography,
345
949444
3712
15:53
clearly no child sexual abuse material and other bad things,
346
953156
3546
15:56
no violence, for example.
347
956702
2085
15:58
We also make it clear that it's a differentiated experience
348
958787
3086
16:01
if you're below 18 years old.
349
961873
1836
16:03
So if you're below 18 years old, for example,
350
963750
2127
16:05
your entire app experience is actually more restricted.
351
965919
3253
16:09
We don't allow, as an example,
352
969172
2461
16:11
users below 16, by default, to go viral.
353
971633
4046
16:15
We don't allow that.
354
975679
1209
16:16
If you're below 16,
355
976930
1168
16:18
we don’t allow you to use the instant messaging feature in app.
356
978140
3753
16:22
If you’re below 18, we don’t allow you to use the livestreaming features.
357
982602
3504
16:26
And of course, we give parents a whole set of tools
358
986148
2669
16:28
to control their teenagers’ experience as well.
359
988817
2211
16:31
CA: How do you know the age of your users?
360
991069
2836
16:34
SC: In our industry, we rely mainly on something called age gating,
361
994364
4588
16:38
which is when you sign up for the app for the first time
362
998994
2627
16:41
and we ask you for the age.
363
1001621
1335
16:42
Now, beyond that,
364
1002956
1168
16:44
we also have built tools to go through your public profile for example,
365
1004166
5422
16:49
when you post a video,
366
1009588
1168
16:50
we try to match the age that you said with the video that you just posted.
367
1010797
4255
16:55
Now, there are questions of can we do more?
368
1015093
2336
16:57
And the question always has, for every company, by the way,
369
1017471
3253
17:00
in our industry, has to be balanced with privacy.
370
1020766
2836
17:03
Now, if, for example, we scan the faces of every single user,
371
1023977
5380
17:09
then we will significantly increase the ability to tell their age.
372
1029399
3962
17:13
But we will also significantly increase the amount of data
373
1033403
2711
17:16
that we collect on you.
374
1036156
1293
17:17
Now, we don't want to collect data.
375
1037491
1710
17:19
We don't want to scan data on your face to collect that.
376
1039201
2669
17:21
So that balance has to be maintained,
377
1041870
1877
17:23
and it's a challenge that we are working through
378
1043747
2294
17:26
together with industry, together with the regulators as well.
379
1046041
3378
17:29
CA: So look, one thing that is unquestionable
380
1049920
2127
17:32
is that you have created a platform for literally millions of people
381
1052047
3837
17:35
who never thought they were going to be a content creator.
382
1055884
2878
17:38
You've given them an audience.
383
1058804
1626
17:40
I'd actually like to hear from you one other favorite example
384
1060472
2878
17:43
of someone who TikTok has given an audience to
385
1063391
2712
17:46
that never had that before.
386
1066144
1460
17:47
SC: So when again,
387
1067604
2252
17:49
when I travel around the world,
388
1069898
2002
17:51
I meet with a whole bunch of creators on our platform.
389
1071942
2961
17:55
I was in South Korea just yesterday, and before that I met with -- yes,
390
1075320
5798
18:01
before that I met with a bunch of --
391
1081118
1751
18:02
People don't expect, for example, teachers.
392
1082869
2044
18:04
There is an English teacher from Arkansas.
393
1084955
3170
18:08
Her name is Claudine, and I met her in person.
394
1088125
3086
18:11
She uses our platform to reach out to students.
395
1091253
2669
18:14
There is another teacher called Chemical Kim.
396
1094381
3128
18:17
And Chemical Kim teaches chemistry.
397
1097551
2752
18:20
What she does is she uses our platform
398
1100595
1877
18:22
to reach out to a much broader student base
399
1102514
2502
18:25
than she has in her classroom.
400
1105058
1543
18:26
And they're both very, very popular.
401
1106643
2085
18:28
You know, in fact,
402
1108770
1168
18:29
what we have realized is that STEM content
403
1109938
4588
18:34
has over 116 billion views on our platform globally.
404
1114568
4546
18:39
And it's so significant --
405
1119114
1251
18:40
CA: In a year?
406
1120407
1168
18:41
SC: Cumulatively.
407
1121575
1418
18:43
CA: [116] billion.
408
1123034
1210
18:44
SC: It's so significant, that in the US we have started testing,
409
1124286
3461
18:47
creating a feed just for STEM content.
410
1127789
3087
18:50
Just for STEM content.
411
1130876
1167
18:52
I’ve been using it for a while, and I learned something new.
412
1132085
3128
18:55
You want to know what it is?
413
1135255
1627
18:56
Apparently if you flip an egg on your tray,
414
1136923
3629
19:00
the egg will last longer.
415
1140552
2169
19:02
It's science,
416
1142721
1168
19:03
there’s a whole video on this, I learned this on TikTok.
417
1143930
2628
19:06
You can search for this.
418
1146558
1251
19:07
CA: You want to know something else about an egg?
419
1147809
2294
19:10
If you put it in just one hand and squeeze it as hard as you can,
420
1150145
3086
19:13
it will never break.
421
1153231
1168
19:14
SC: Yes, I think I read about that, too.
422
1154441
1918
19:16
CA: It's not true.
423
1156359
1168
19:17
(Laughter)
424
1157569
1126
19:18
SC: We can search for it.
425
1158737
1376
19:21
CA: But look, here's here's the flip side to all this amazingness.
426
1161072
3504
19:24
And honestly, this is the key thing,
427
1164618
2210
19:26
that I want to have an honest, heart-to-heart conversation with you
428
1166870
4504
19:31
because it's such an important issue,
429
1171416
1794
19:33
this question of human addiction.
430
1173210
2460
19:35
You know, we are ...
431
1175712
2377
19:38
animals with a prefrontal cortex.
432
1178131
3629
19:42
That's how I think of us.
433
1182594
1209
19:43
We have these addictive instincts that go back millions of years,
434
1183803
4213
19:48
and we often are in the mode of trying to modulate our own behavior.
435
1188058
6673
19:54
It turns out that the internet is incredibly good
436
1194731
4421
19:59
at activating our animal cells
437
1199194
2461
20:01
and getting them so damn excited.
438
1201696
2795
20:04
And your company, the company you've built,
439
1204532
2712
20:07
is better at it than any other company on the planet, I think.
440
1207244
5338
20:12
So what are the risks of this?
441
1212999
3003
20:16
I mean, how ...
442
1216002
1752
20:17
From a company point of view, for example,
443
1217796
3003
20:20
it's in your interest to have people on there as long as possible.
444
1220840
3212
20:24
So some would say, as a first pass,
445
1224094
1960
20:26
you want people to be addicted as long as possible.
446
1226096
2794
20:28
That's how advertising money will flow and so forth,
447
1228932
3962
20:32
and that's how your creators will be delighted.
448
1232936
2461
20:36
What is too much?
449
1236481
1376
20:37
SC: I don't actually agree with that.
450
1237857
1794
20:40
You know, as a company,
451
1240402
1543
20:41
our goal is not to optimize and maximize time spent.
452
1241987
3211
20:45
It is not.
453
1245240
1418
20:46
In fact, in order to address people spending too much time on our platform,
454
1246700
4004
20:50
we have done a number of things.
455
1250704
1585
20:52
I was just speaking with some of your colleagues backstage.
456
1252289
2836
20:55
One of them told me she has encountered this as well.
457
1255166
3379
20:58
If you spend too much time on our platform,
458
1258586
2169
21:00
we will proactively send you videos to tell you to get off the platform.
459
1260755
4004
21:05
We will.
460
1265093
1251
21:06
And depending on the time of the day,
461
1266386
1793
21:08
if it's late at night, it will come sooner.
462
1268179
2002
21:10
We have also built in tools to limit,
463
1270223
1919
21:12
if you below 18 years old, by default,
464
1272142
3670
21:15
we set a 60-minute default time limit.
465
1275812
2377
21:18
CA: How many?
466
1278189
1168
21:19
SC: Sixty minutes.
467
1279399
1168
21:20
And we've given parents tools and yourself tools,
468
1280608
2294
21:22
if you go to settings, you can set your own time limit.
469
1282944
2628
21:25
We've given parents tools so that you can pair,
470
1285572
2210
21:27
for the parents who don't know this, go to settings, family pairing,
471
1287824
3212
21:31
you can pair your phone with your teenager's phone
472
1291036
2335
21:33
and set the time limit.
473
1293413
1168
21:34
And we really encourage parents to have these conversations with their teenagers
474
1294581
3795
21:38
on what is the right amount of screen time.
475
1298418
2044
21:40
I think there’s a healthy relationship that you should have with your screen,
476
1300503
3629
21:44
and as a business, we believe that that balance needs to be met.
477
1304132
3003
21:47
So it's not true that we just want to maximize time spent.
478
1307135
3754
21:50
CA: If you were advising parents here
479
1310889
2043
21:52
what time they should actually recommend to their teenagers,
480
1312974
3003
21:56
what do you think is the right setting?
481
1316019
1877
21:57
SC: Well, 60 minutes,
482
1317896
1168
21:59
we did not come up with it ourselves.
483
1319105
1794
22:00
So I went to the Digital Wellness Lab at the Boston Children's Hospital,
484
1320899
3420
22:04
and we had this conversation with them.
485
1324319
1877
22:06
And 60 minutes was the recommendation that they gave to us,
486
1326237
2795
22:09
which is why we built this into the app.
487
1329032
1918
22:10
So 60 minutes, take it for what it is,
488
1330950
1877
22:12
it’s something that we’ve had some discussions of experts.
489
1332827
3295
22:16
But I think for all parents here,
490
1336122
1585
22:17
it is very important to have these conversations with your teenage children
491
1337749
4212
22:21
and help them develop a healthy relationship with screens.
492
1341961
4588
22:26
I think we live in an age where it's completely inevitable
493
1346549
3087
22:29
that we're going to interact with screens and digital content,
494
1349636
4629
22:34
but I think we should develop healthy habits early on in life,
495
1354307
3504
22:37
and that's something I would encourage.
496
1357811
1918
22:39
CA: Curious to ask the audience,
497
1359729
2795
22:42
which of you who have ever had that video on TikTok appear
498
1362565
3629
22:46
saying, “Come off.”
499
1366236
1501
22:48
OK, I mean ...
500
1368655
1626
22:50
So maybe a third of the audience seem to be active TikTok users,
501
1370949
3128
22:54
and about 20 people maybe put their hands up there.
502
1374119
4254
22:58
Are you sure that --
503
1378957
1918
23:00
like, it feels to me like this is a great thing to have,
504
1380875
4713
23:05
but are you ...
505
1385630
1543
23:07
isn't there always going to be a temptation
506
1387215
2002
23:09
in any given quarter or whatever,
507
1389217
2252
23:11
to just push it a bit at the boundary
508
1391469
2169
23:13
and just dial back a bit on that
509
1393680
2210
23:15
so that you can hit revenue goals, etc?
510
1395932
3170
23:19
Are you saying that this is used scrupulously?
511
1399727
2878
23:22
SC: I think, you know, in terms ...
512
1402647
2920
23:25
Even if you think about it from a commercial point of view,
513
1405608
2795
23:28
it is always best when your customers have a very healthy relationship
514
1408403
3295
23:31
with your product.
515
1411698
1168
23:32
It's always best when it's healthy.
516
1412907
1710
23:34
So if you think about very short-term retention, maybe,
517
1414617
2962
23:37
but that's not the way we think about it.
518
1417579
2335
23:40
If you think about it from a longer-term perspective,
519
1420290
2502
23:42
what you really want to have is a healthy relationship, you know.
520
1422792
3087
23:45
You don’t want people to develop very unhealthy habits,
521
1425879
2752
23:48
and then at some point they're going to drop it.
522
1428631
2461
23:51
So I think everything in moderation.
523
1431134
2419
23:53
CA: There's a claim out there that in China,
524
1433595
3003
23:56
there's a much more rigorous standards imposed on the amount of time
525
1436598
4462
24:01
that children, especially, can spend on the TikTok equivalent of that.
526
1441060
5297
24:06
SC: That is unfortunately a misconception.
527
1446357
4713
24:11
So that experience that is being mentioned for Douyin,
528
1451070
3129
24:14
which is a different app,
529
1454199
1251
24:15
is for an under 14-year-old experience.
530
1455450
2627
24:18
Now, if you compare that in the United States,
531
1458495
2293
24:20
we have an under-13 experience in the US.
532
1460788
2503
24:23
It's only available in the US, it's not available here in Canada,
533
1463333
3086
24:26
in Canada, we just don't allow it.
534
1466419
1668
24:28
If you look at the under-13 experience in the US,
535
1468129
2378
24:30
it's much more restricted than the under-14 experience in China.
536
1470548
3754
24:34
It's so restrictive,
537
1474344
1334
24:35
that every single piece of content is vetted
538
1475720
3003
24:38
by our third-party child safety expert.
539
1478723
4296
24:43
And we don't allow any under-13s in the US to publish,
540
1483436
4463
24:47
we don’t allow them to post,
541
1487899
1418
24:49
and we don't allow them to use a lot of features.
542
1489317
2336
24:51
So I think that that report, I've seen that report too,
543
1491653
2627
24:54
it's not doing a fair comparison.
544
1494280
1710
24:56
CA: What do you make of this issue?
545
1496324
1710
24:58
You know, you've got these millions of content creators
546
1498034
2878
25:00
and all of them, in a sense, are in a race for attention,
547
1500954
4045
25:05
and that race can pull them in certain directions.
548
1505041
2920
25:07
So, for example, teenage girls on TikTok,
549
1507961
4296
25:12
sometimes people worry that, to win attention,
550
1512257
3628
25:15
they've discovered that by being more sexual
551
1515885
2086
25:18
that they can gain extra viewers.
552
1518012
1669
25:20
Is this a concern?
553
1520098
1376
25:21
Is there anything you can do about this?
554
1521474
2252
25:23
SC: We address this in our community guidelines as well.
555
1523768
3295
25:28
You know, if you look at sort of the sexualized content on our guidelines,
556
1528147
4130
25:32
if you’re below a certain age,
557
1532318
1543
25:33
you know, for certain themes that are mature,
558
1533903
3253
25:37
we actually remove that from your experience.
559
1537198
2127
25:39
Again, I come back to this,
560
1539325
1543
25:40
you know, we want to have a safe platform.
561
1540910
2127
25:43
In fact, at my congressional hearing,
562
1543037
1794
25:44
I made four commitments to our users and to the politicians in the US.
563
1544872
3754
25:48
And the first one is that we take safety, especially for teenagers,
564
1548626
4338
25:53
extremely seriously,
565
1553006
1167
25:54
and we will continue to prioritize that.
566
1554215
2419
25:56
You know, I believe that we need to give our teenage users,
567
1556634
3796
26:00
and our users in general,
568
1560471
1460
26:01
a very safe experience,
569
1561973
1168
26:03
because if we don't do that,
570
1563182
1502
26:04
then we cannot fulfill --
571
1564684
1293
26:05
the mission is to inspire creativity and to bring joy.
572
1565977
2919
26:08
If they don't feel safe, I cannot fulfill my mission.
573
1568938
2836
26:11
So it's all very organic to me as a business
574
1571816
2753
26:14
to make sure I do that.
575
1574569
1293
26:15
CA: But in the strange interacting world of human psychology and so forth,
576
1575862
3503
26:19
weird memes can take off.
577
1579365
1418
26:20
I mean, you had this outbreak a couple years back
578
1580825
2461
26:23
with these devious licks where kids were competing with each other
579
1583328
3420
26:26
to do vandalism in schools and, you know,
580
1586789
2253
26:29
get lots of followers from it.
581
1589083
1585
26:30
How on Earth do you battle something like that?
582
1590710
2461
26:33
SC: So dangerous challenges are not allowed on our platform.
583
1593963
2878
26:36
If you look at our guidelines, it's violative.
584
1596883
2961
26:39
We proactively invest resources to identify them
585
1599886
3336
26:43
and remove them from our platform.
586
1603222
1794
26:45
In fact, if you search for dangerous challenges on our platform today,
587
1605058
3336
26:48
we will redirect you to a safety resource page.
588
1608394
2878
26:51
And we actually worked with some creators as well to come up with campaigns.
589
1611314
3587
26:54
This is another campaign.
590
1614901
1501
26:56
It's the "Stop, Think, Decide Before You Act" campaign
591
1616444
2711
26:59
where we work with the creators to produce videos,
592
1619155
2419
27:01
to explain to people that some things are dangerous,
593
1621574
2920
27:04
please don't do it.
594
1624535
1210
27:05
And we post these videos actively on our platform as well.
595
1625787
3128
27:09
CA: That's cool.
596
1629791
1209
27:11
And you've got lots of employees.
597
1631042
1877
27:12
I mean, how many employees do you have
598
1632960
1836
27:14
who are specifically looking at these content moderation things,
599
1634796
3753
27:18
or is that the wrong question?
600
1638549
1502
27:20
Are they mostly identified by AI initially
601
1640093
2377
27:22
and then you have a group who are overseeing
602
1642512
2961
27:25
and making the final decision?
603
1645515
1585
27:27
SC: The group is based in Ireland and it's a lot of people,
604
1647100
4045
27:31
it's tens of thousands of people.
605
1651187
1585
27:32
CA: Tens of thousands?
606
1652772
1168
27:33
SC: It's one of the most important cost items on my PnL,
607
1653981
4004
27:38
and I think it's completely worth it.
608
1658027
1794
27:39
Now, most of the moderation has to be done by machines.
609
1659821
2836
27:42
The machines are good, they're quite good,
610
1662657
2627
27:45
but they're not as good as, you know,
611
1665326
1794
27:47
they're not perfect at this point.
612
1667120
1668
27:48
So you have to complement them with a lot of human beings today.
613
1668788
3211
27:51
And I think, by the way, a lot of the progress in AI in general
614
1671999
4380
27:56
is making that kind of content moderation capabilities a lot better.
615
1676421
4004
28:00
So we're going to get more precise.
616
1680466
1877
28:02
You know, we’re going to get more specific.
617
1682385
2002
28:04
And it’s going to be able to handle larger scale.
618
1684429
4296
28:08
And that's something I think that I'm personally looking forward to.
619
1688766
4088
28:13
CA: What about this perceived huge downside
620
1693813
4296
28:18
of use of, certainly Instagram, I think TikTok as well.
621
1698151
3670
28:21
What people worry that you are amplifying insecurities,
622
1701863
3712
28:25
especially of teenagers
623
1705575
1626
28:27
and perhaps especially of teenage girls.
624
1707243
1919
28:29
They see these amazing people on there doing amazing things,
625
1709162
3920
28:33
they feel inadequate,
626
1713124
1251
28:34
there's all these reported cases of depression, insecurity,
627
1714417
3545
28:38
suicide and so forth.
628
1718004
1752
28:39
SC: I take this extremely seriously.
629
1719797
1835
28:42
So in our guidelines,
630
1722967
2377
28:45
for certain themes that we think are mature and not suitable for teenagers,
631
1725386
5089
28:50
we actually proactively remove it from their experience.
632
1730516
3254
28:54
At the same time, if you search certain terms,
633
1734395
2544
28:56
we will make sure that you get redirected to a resource safety page.
634
1736981
4087
29:01
Now we are always working with experts to understand some of these new trends
635
1741068
4338
29:05
that could emerge
636
1745406
1168
29:06
and proactively try to manage them, if that makes sense.
637
1746616
3795
29:10
Now, this is a problem that predates us,
638
1750411
2961
29:13
that predates TikTok.
639
1753372
1168
29:14
It actually predates the internet.
640
1754540
1710
29:16
But it's our responsibility to make sure
641
1756250
2378
29:18
that we invest enough to understand and to address the concerns,
642
1758628
3253
29:21
to keep the experience as safe as possible
643
1761923
2043
29:23
for as many people as possible.
644
1763966
1502
29:26
CA: Now, in Congress,
645
1766302
1210
29:27
the main concern seemed to be not so much what we've talked about,
646
1767512
3461
29:30
but data, the data of users,
647
1770973
2461
29:33
the fact that you're owned by ByteDance, Chinese company,
648
1773476
3253
29:36
and the concern that at any moment
649
1776771
2753
29:39
Chinese government might require or ask for data.
650
1779524
3670
29:43
And in fact, there have been instances
651
1783194
1835
29:45
where, I think you've confirmed,
652
1785029
1543
29:46
that some data of journalists on the platform
653
1786572
3963
29:50
was made available to ByteDance's engineers
654
1790576
3003
29:53
and from there, who knows what.
655
1793579
2211
29:56
Now, your response to this was to have this Project Texas,
656
1796165
4254
30:00
where you're moving data to be controlled by Oracle here in the US.
657
1800461
4755
30:05
Can you talk about that project and why, if you believe it so,
658
1805258
5130
30:10
why we should not worry so much about this issue?
659
1810429
3003
30:13
SC: I will say a couple of things about this, if you don't mind.
660
1813432
3045
30:16
The first thing I would say is that the internet is built
661
1816519
2711
30:19
on global interoperability,
662
1819230
1501
30:20
and we are not the only company that relies on the global talent pool
663
1820773
4171
30:24
to make our products as good as possible.
664
1824944
1960
30:26
Technology is a very collaborative effort.
665
1826946
2002
30:28
I think many people here would say the same thing.
666
1828990
3461
30:32
So we are not the first company to have engineers in all countries,
667
1832493
3170
30:35
including in China.
668
1835663
1168
30:36
We're not the first one.
669
1836873
1167
30:38
Now, I understand some of these concerns.
670
1838082
2377
30:40
You know, the data access by employees is not data accessed by government.
671
1840459
4463
30:44
This is very different, and there’s a clear difference in this.
672
1844922
2962
30:47
But we hear the concerns that are raised in the United States.
673
1847925
2962
30:50
We did not try to avoid discussing.
674
1850928
3420
30:54
We did not try to argue our way out of it.
675
1854390
2878
30:57
What we did was we built an unprecedented project
676
1857310
2836
31:00
where we localize American data to be stored on American soil
677
1860187
4171
31:04
by an American company overseen by American personnel.
678
1864400
4004
31:08
So this kind of protection for American data
679
1868404
3212
31:11
is beyond what any other company in our industry has ever done.
680
1871657
4213
31:16
Well, money is not the only issue here,
681
1876454
2544
31:19
but it's very expensive to build something like that.
682
1879040
2502
31:21
And more importantly, you know,
683
1881584
1626
31:23
we are basically localizing data in a way that no other company has done.
684
1883252
5673
31:28
So we need to be very careful that whilst we are pursuing
685
1888966
3963
31:32
what we call digital sovereignty in the US
686
1892929
2585
31:35
and we are also doing a version of this in Europe,
687
1895556
2503
31:38
that we don't balkanize the internet.
688
1898100
1794
31:39
Now we are the first to do it.
689
1899936
1459
31:41
And I expect that, you know,
690
1901437
1460
31:42
other companies are probably looking at this
691
1902897
2085
31:45
and trying to figure out how you balance between protecting, protected data,
692
1905024
4630
31:49
you know, to make sure that everybody feels secure about it
693
1909654
2836
31:52
while at the same time allowing for interoperability
694
1912490
2586
31:55
to continue to happen,
695
1915117
1126
31:56
because that's what makes technology and the internet so great.
696
1916285
3003
31:59
So that's something that we are doing.
697
1919330
1835
32:01
CA: How far are you along that journey with Project Texas?
698
1921207
2794
32:04
SC: We are very, very far along today.
699
1924043
1835
32:05
CA: When will there be a clear you know,
700
1925920
3920
32:09
here it is, it’s done, it’s firewalled, this data is protected?
701
1929882
3087
32:12
SC: Today, by default, all new US data
702
1932969
3086
32:16
is already stored in the Oracle cloud infrastructure.
703
1936055
2502
32:18
So it's in this protected US environment that we talked about in the United States.
704
1938557
4713
32:23
We still have some legacy data to delete in our own servers in Virginia
705
1943270
3963
32:27
and in Singapore.
706
1947274
1168
32:28
Our data has never been stored in China, by the way.
707
1948484
2461
32:31
That deletion is a very big engineering effort.
708
1951487
2211
32:33
So as we said, as I said at the hearing,
709
1953739
2378
32:36
it's going to take us a while to delete them,
710
1956117
2502
32:38
but I expect it to be done this year.
711
1958619
2294
32:43
CA: How much power do you have
712
1963791
3045
32:46
over your own ability to control certain things?
713
1966877
2837
32:49
So, for example, suppose that, for whatever reason,
714
1969714
2961
32:52
the Chinese government was to look at an upcoming US election and say,
715
1972717
4421
32:57
"You know what, we would like this party to win," let's say,
716
1977138
4754
33:01
or "We would like civil war to break out" or whatever.
717
1981934
2711
33:05
How ...
718
1985104
1376
33:06
"And we could do this
719
1986522
1418
33:07
by amplifying the content of certain troublemaking, disturbing people,
720
1987982
4212
33:12
causing uncertainty, spreading misinformation," etc.
721
1992236
2795
33:16
If you were required via ByteDance to do this,
722
1996032
4963
33:21
like, first of all, is there a pathway where theoretically that is possible?
723
2001037
4421
33:26
What's your personal line in the sand on this?
724
2006542
3462
33:30
SC: So during the congressional hearing,
725
2010046
2544
33:32
I made four commitments,
726
2012631
1585
33:34
we talked about the first one, which is safety.
727
2014216
2211
33:36
The third one is to keep TikTok a place of freedom of expression.
728
2016469
3420
33:39
By the way, if you go on TikTok today,
729
2019930
1836
33:41
you can search for anything you want,
730
2021766
1793
33:43
as long as it doesn't violate our community guidelines.
731
2023559
2753
33:46
And to keep it free from any government manipulation.
732
2026312
3003
33:49
And the fourth one is transparency and third-party monitoring.
733
2029315
4129
33:53
So the way we are trying to address this concern
734
2033444
2294
33:55
is an unprecedented amount of transparency.
735
2035780
2294
33:58
What do I mean by this?
736
2038074
1459
33:59
We're actually allowing third-party reviewers
737
2039533
3546
34:03
to come in and review our source code.
738
2043079
2127
34:05
I don't know any other company that does this, by the way.
739
2045206
2752
34:08
Because everything, as you know, is driven by code.
740
2048417
2670
34:11
So to allow someone else to review the source code
741
2051128
2670
34:13
is to give this a significant amount of transparency
742
2053798
3003
34:16
to ensure that the scenarios that you described
743
2056842
2336
34:19
that are highly hypothetical, cannot happen on our platform.
744
2059220
3503
34:23
Now, at the same time,
745
2063057
1168
34:24
we are releasing more research tools for researchers
746
2064266
3879
34:28
so that they can study the output.
747
2068187
1627
34:29
So the source code is the input.
748
2069814
2210
34:32
We are also allowing researchers to study the output,
749
2072024
2628
34:34
which is the content on our platform.
750
2074652
1793
34:36
I think the easiest way to sort of fend this off is transparency.
751
2076487
4087
34:40
You know, we give people access to monitor us,
752
2080574
2753
34:43
and we just make it very, very transparent.
753
2083369
2002
34:45
And that's our approach to the problem.
754
2085412
1836
34:47
CA: So you will say directly to this group
755
2087248
2252
34:49
that the scenario I talked about,
756
2089500
2252
34:51
of theoretical Chinese government interference in an American election,
757
2091794
4671
34:56
you can say that will not happen?
758
2096465
2753
34:59
SC: I can say that we are building all the tools
759
2099677
2711
35:02
to prevent any of these actions from happening.
760
2102388
2919
35:05
And I'm very confident that with an unprecedented amount of transparency
761
2105933
4338
35:10
that we're giving on the platform,
762
2110271
1626
35:11
we can reduce this risk to as low as zero as possible.
763
2111939
4421
35:17
CA: To as low as zero as possible.
764
2117403
2753
35:20
SC: To as close to zero as possible.
765
2120197
2419
35:22
CA: As close to zero as possible.
766
2122616
1877
35:25
That's fairly reassuring.
767
2125452
1919
35:27
Fairly.
768
2127830
1168
35:28
(Laughter)
769
2128998
1168
35:33
I mean, how would the world know?
770
2133085
1585
35:34
If you discovered this or you thought you had to do it,
771
2134712
2586
35:37
is this a line in the sand for you?
772
2137298
3461
35:40
Like, are you in a situation you would not let the company that you know now
773
2140801
3587
35:44
and that you are running do this?
774
2144430
1585
35:46
SC: Absolutely.
775
2146056
1168
35:47
That's the reason why we're letting third parties monitor,
776
2147266
2711
35:50
because if they find out, you know, they will disclose this.
777
2150019
2836
35:52
We also have transparency reports, by the way,
778
2152855
2169
35:55
where we talk about a whole bunch of things,
779
2155024
2085
35:57
the content that we remove, you know, that violates our guidelines,
780
2157151
3545
36:00
government requests.
781
2160738
1168
36:01
You know, it's all published online.
782
2161906
1751
36:03
All you have to do is search for it.
783
2163657
1752
36:05
CA: So you're super compelling
784
2165409
1460
36:06
and likable as a CEO, I have to say.
785
2166911
1751
36:08
And I would like to, as we wrap this up,
786
2168662
2419
36:11
I'd like to give you a chance just to paint, like, what's the vision?
787
2171123
3253
36:14
As you look at what TikTok could be,
788
2174418
4087
36:18
let's move the clock out, say, five years from now.
789
2178547
3295
36:21
How should we think about your contribution to our collective future?
790
2181884
4504
36:26
SC: I think it's still down to the vision that we have.
791
2186430
2586
36:29
So in terms of the window of discovery,
792
2189058
2794
36:31
I think there's a huge benefit to the world
793
2191852
2878
36:34
when people can discover new things.
794
2194772
1835
36:36
You know, people think that TikTok is all about dancing and singing,
795
2196607
3211
36:39
and there’s nothing wrong with that, because it’s super fun.
796
2199860
2836
36:42
There's still a lot of that,
797
2202696
1377
36:44
but we're seeing science content, STEM content,
798
2204073
2878
36:46
have you about BookTok?
799
2206992
1794
36:48
It's a viral trend that talks about books
800
2208827
2670
36:51
and encourages people to read.
801
2211538
1711
36:53
That BookTok has 120 billion views globally,
802
2213290
3963
36:57
120 billion.
803
2217294
1168
36:58
CA: Billion, with a B.
804
2218462
1460
36:59
SC: People are learning how to cook,
805
2219964
2002
37:02
people are learning about science,
806
2222007
1627
37:03
people are learning how to golf --
807
2223634
1668
37:05
well, people are watching videos on golfing, I guess.
808
2225302
2586
37:07
(Laughter)
809
2227930
1877
37:10
I haven't gotten better by looking at the videos.
810
2230349
2961
37:13
I think there's a huge, huge opportunity here on discovery
811
2233352
3754
37:17
and giving the everyday person a voice.
812
2237147
2044
37:19
If you talk to our creators, you know,
813
2239233
1835
37:21
a lot of people will tell you this again and again, that before TikTok,
814
2241068
3378
37:24
they would never have been discovered.
815
2244446
1836
37:26
And we have given them the platform to do that.
816
2246323
2211
37:28
And it's important to maintain that.
817
2248534
1752
37:30
Then we talk about creation.
818
2250286
1376
37:31
You know, there’s all this new technology coming in with AI-generated content
819
2251662
4379
37:36
that will help people create even more creative content.
820
2256083
3962
37:40
I think there's going to be a collaboration between,
821
2260045
2461
37:42
and I think there's a speaker who is going to talk about this,
822
2262506
2920
37:45
between people and AI
823
2265467
1168
37:46
where they can unleash their creativity in a different way.
824
2266635
2795
37:49
You know, like for example, I'm terrible at drawing personally,
825
2269471
3295
37:52
but if I had some AI to help me,
826
2272808
2377
37:55
then maybe I can express myself even better.
827
2275227
2127
37:57
Then we talk about bridges to connect
828
2277813
1793
37:59
and connecting people and the communities together.
829
2279648
2419
38:02
This could be products, this could be commerce,
830
2282067
2253
38:04
five million businesses in the US benefit from TikTok today.
831
2284320
5714
38:10
I think we can get that number to a much higher number.
832
2290075
2586
38:12
And of course, if you look around the world, including in Canada,
833
2292661
3087
38:15
that number is going to be massive.
834
2295748
1710
38:17
So I think these are the biggest opportunities that we have,
835
2297458
2836
38:20
and it's really very exciting.
836
2300336
1459
38:21
CA: So courtesy of your experience in Congress,
837
2301795
2211
38:24
you actually became a bit of a TikTok star yourself, I think.
838
2304048
3586
38:28
Some of your videos have gone viral.
839
2308010
2210
38:31
You've got your phone with you.
840
2311055
1501
38:32
Do you want to make a little little TikTok video right now?
841
2312556
2795
38:35
Let's do this.
842
2315351
1167
38:36
SC: If you don't mind ...
843
2316560
1210
38:37
CA: What do you think, should we do this?
844
2317811
1919
38:39
SC: We're just going to do a selfie together, how's that?
845
2319730
2711
38:42
So why don't we just say "Hi."
846
2322816
1669
38:44
Hi!
847
2324526
1168
38:45
Audience: Hi!
848
2325694
1168
38:46
CA: Hello from TED.
849
2326904
1126
38:48
SC: All right, thank you, I hope it goes viral.
850
2328405
2211
38:50
(Laughter)
851
2330657
2795
38:53
CA: If that one goes viral, I think I've given up on your algorithm, actually.
852
2333494
3670
38:57
(Laughter)
853
2337206
1543
39:00
Shou Chew, you're one of the most influential
854
2340417
2753
39:03
and powerful people in the world, whether you know it or not.
855
2343212
2878
39:06
And I really appreciate you coming and sharing your vision.
856
2346090
2794
39:08
I really, really hope the upside of what you're talking about comes about.
857
2348884
3545
39:12
Thank you so much for coming today.
858
2352471
1710
39:14
SC: Thank you, Chris.
859
2354181
1168
39:15
CA: It's really interesting.
860
2355391
1376
39:16
(Applause)
861
2356767
1627
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7