How to seek truth in the era of fake news | Christiane Amanpour

143,721 views

2017-10-30 ・ TED


New videos

How to seek truth in the era of fake news | Christiane Amanpour

143,721 views ・ 2017-10-30

TED


Please double-click on the English subtitles below to play the video.

00:12
Chris Anderson: Christiane, great to have you here.
0
12931
2337
00:15
So you've had this amazing viewpoint,
1
15292
1843
00:17
and perhaps it's fair to say that in the last few years,
2
17159
3060
00:20
there have been some alarming developments that you're seeing.
3
20243
3753
00:24
What's alarmed you most?
4
24020
1564
00:25
Christiane Amanpour: Well, just listening to the earlier speakers,
5
25608
3192
00:28
I can frame it in what they've been saying:
6
28824
2472
00:31
climate change, for instance -- cities, the threat to our environment
7
31320
3422
00:34
and our lives.
8
34766
1260
00:36
It basically also boils down to understanding the truth
9
36440
3894
00:40
and to be able to get to the truth of what we're talking about
10
40358
3011
00:43
in order to really be able to solve it.
11
43393
2092
00:45
So if 99.9 percent of the science on climate
12
45509
3927
00:49
is empirical, scientific evidence,
13
49460
3057
00:52
but it's competing almost equally with a handful of deniers,
14
52541
4895
00:57
that is not the truth;
15
57460
1227
00:58
that is the epitome of fake news.
16
58711
2498
01:01
And so for me, the last few years -- certainly this last year --
17
61233
5102
01:06
has crystallized the notion of fake news in a way that's truly alarming
18
66359
4260
01:10
and not just some slogan to be thrown around.
19
70643
2659
01:13
Because when you can't distinguish between the truth and fake news,
20
73326
3811
01:17
you have a very much more difficult time trying to solve
21
77161
3891
01:21
some of the great issues that we face.
22
81076
2451
01:24
CA: Well, you've been involved in this question of,
23
84512
3421
01:27
what is balance, what is truth, what is impartiality,
24
87957
2912
01:30
for a long time.
25
90893
1255
01:32
You were on the front lines reporting the Balkan Wars 25 years ago.
26
92172
5870
01:38
And back then, you famously said,
27
98066
3412
01:41
by calling out human right abuses,
28
101502
2621
01:44
you said, "Look, there are some situations one simply cannot be neutral about,
29
104147
4329
01:48
because when you're neutral,
30
108500
1380
01:49
you are an accomplice."
31
109904
1887
01:53
So, do you feel that today's journalists aren't heeding that advice
32
113243
4897
01:58
about balance?
33
118164
1472
01:59
CA: Well, look, I think for journalists, objectivity is the golden rule.
34
119660
4106
02:03
But I think sometimes we don't understand what objectivity means.
35
123790
4416
02:08
And I actually learned this very, very young in my career,
36
128230
2994
02:11
which was during the Balkan Wars.
37
131248
1572
02:12
I was young then.
38
132844
1216
02:14
It was about 25 years ago.
39
134084
2539
02:16
And what we faced was the wholesale violation, not just of human rights,
40
136647
5781
02:22
but all the way to ethnic cleansing and genocide,
41
142452
2979
02:25
and that has been adjudicated in the highest war crimes court
42
145455
4006
02:29
in the world.
43
149485
1164
02:30
So, we know what we were seeing.
44
150673
1653
02:32
Trying to tell the world what we were seeing
45
152350
2537
02:34
brought us accusations of bias,
46
154911
2775
02:37
of siding with one side,
47
157710
1889
02:39
of not seeing the whole side,
48
159623
1862
02:41
and just, you know, trying to tell one story.
49
161509
2297
02:43
I particularly and personally was accused of siding with,
50
163830
4307
02:48
for instance, the citizens of Sarajevo --
51
168161
1982
02:50
"siding with the Muslims,"
52
170167
1427
02:51
because they were the minority who were being attacked
53
171618
3052
02:54
by Christians on the Serb side
54
174694
3738
02:58
in this area.
55
178456
1719
03:00
And it worried me.
56
180199
1342
03:01
It worried me that I was being accused of this.
57
181565
2191
03:03
I thought maybe I was wrong,
58
183780
1342
03:05
maybe I'd forgotten what objectivity was.
59
185146
2348
03:07
But then I started to understand that what people wanted
60
187518
3007
03:10
was actually not to do anything --
61
190549
1798
03:12
not to step in,
62
192371
1417
03:13
not to change the situation,
63
193812
1570
03:15
not to find a solution.
64
195406
1449
03:16
And so, their fake news at that time,
65
196879
2353
03:19
their lie at that time --
66
199256
1382
03:20
including our government's, our democratically elected government's,
67
200662
3530
03:24
with values and principles of human rights --
68
204216
2272
03:26
their lie was to say that all sides are equally guilty,
69
206512
3507
03:30
that this has been centuries of ethnic hatred,
70
210043
2793
03:32
whereas we knew that wasn't true,
71
212860
1882
03:34
that one side had decided to kill, slaughter and ethnically cleanse
72
214766
3647
03:38
another side.
73
218437
1157
03:39
So that is where, for me,
74
219618
1496
03:41
I understood that objectivity means giving all sides an equal hearing
75
221138
5306
03:46
and talking to all sides,
76
226468
2105
03:48
but not treating all sides equally,
77
228597
3622
03:52
not creating a forced moral equivalence or a factual equivalence.
78
232243
4788
03:57
And when you come up against that crisis point
79
237055
4479
04:01
in situations of grave violations of international and humanitarian law,
80
241558
5671
04:07
if you don't understand what you're seeing,
81
247253
2342
04:09
if you don't understand the truth
82
249619
2160
04:11
and if you get trapped in the fake news paradigm,
83
251803
3513
04:15
then you are an accomplice.
84
255340
1590
04:17
And I refuse to be an accomplice to genocide.
85
257658
2997
04:20
(Applause)
86
260679
3283
04:26
CH: So there have always been these propaganda battles,
87
266402
2778
04:29
and you were courageous in taking the stand you took back then.
88
269204
4026
04:33
Today, there's a whole new way, though,
89
273652
3727
04:37
in which news seems to be becoming fake.
90
277403
2204
04:39
How would you characterize that?
91
279631
1634
04:41
CA: Well, look -- I am really alarmed.
92
281289
2084
04:43
And everywhere I look,
93
283397
2202
04:45
you know, we're buffeted by it.
94
285623
1837
04:47
Obviously, when the leader of the free world,
95
287484
2202
04:49
when the most powerful person in the entire world,
96
289710
2473
04:52
which is the president of the United States --
97
292207
2246
04:54
this is the most important, most powerful country in the whole world,
98
294477
4819
04:59
economically, militarily, politically in every which way --
99
299320
4240
05:04
and it seeks to, obviously, promote its values and power around the world.
100
304415
5017
05:09
So we journalists, who only seek the truth --
101
309456
3976
05:13
I mean, that is our mission --
102
313456
1521
05:15
we go around the world looking for the truth
103
315001
2117
05:17
in order to be everybody's eyes and ears,
104
317142
1973
05:19
people who can't go out in various parts of the world
105
319139
2519
05:21
to figure out what's going on about things that are vitally important
106
321682
3369
05:25
to everybody's health and security.
107
325075
1956
05:27
So when you have a major world leader accusing you of fake news,
108
327055
6686
05:33
it has an exponential ripple effect.
109
333765
3843
05:37
And what it does is, it starts to chip away
110
337632
4272
05:42
at not just our credibility,
111
342472
2888
05:45
but at people's minds --
112
345384
2029
05:48
people who look at us, and maybe they're thinking,
113
348372
2364
05:50
"Well, if the president of the United States says that,
114
350760
2669
05:53
maybe somewhere there's a truth in there."
115
353453
2135
05:56
CH: Presidents have always been critical of the media --
116
356148
4184
06:00
CA: Not in this way.
117
360356
1601
06:01
CH: So, to what extent --
118
361981
1505
06:03
(Laughter)
119
363510
1064
06:04
(Applause)
120
364598
3120
06:07
CH: I mean, someone a couple years ago looking at the avalanche of information
121
367742
6896
06:14
pouring through Twitter and Facebook and so forth,
122
374662
3236
06:17
might have said,
123
377922
1158
06:19
"Look, our democracies are healthier than they've ever been.
124
379104
2841
06:21
There's more news than ever.
125
381969
1521
06:23
Of course presidents will say what they'll say,
126
383514
2211
06:25
but everyone else can say what they will say.
127
385749
2233
06:28
What's not to like? How is there an extra danger?"
128
388006
4155
06:32
CA: So, I wish that was true.
129
392185
1542
06:34
I wish that the proliferation of platforms upon which we get our information
130
394992
6093
06:41
meant that there was a proliferation of truth and transparency
131
401109
3878
06:45
and depth and accuracy.
132
405011
1868
06:46
But I think the opposite has happened.
133
406903
2455
06:49
You know, I'm a little bit of a Luddite,
134
409382
2090
06:51
I will confess.
135
411496
1196
06:53
Even when we started to talk about the information superhighway,
136
413147
3384
06:56
which was a long time ago,
137
416555
1628
06:58
before social media, Twitter and all the rest of it,
138
418207
2651
07:00
I was actually really afraid
139
420882
1824
07:02
that that would put people into certain lanes and tunnels
140
422730
4021
07:06
and have them just focusing on areas of their own interest
141
426775
4342
07:11
instead of seeing the broad picture.
142
431141
2333
07:13
And I'm afraid to say that with algorithms, with logarithms,
143
433498
4586
07:18
with whatever the "-ithms" are
144
438108
1648
07:19
that direct us into all these particular channels of information,
145
439780
4266
07:24
that seems to be happening right now.
146
444070
1870
07:25
I mean, people have written about this phenomenon.
147
445964
2544
07:28
People have said that yes, the internet came,
148
448532
2198
07:30
its promise was to exponentially explode our access to more democracy,
149
450754
5743
07:36
more information,
150
456521
1714
07:38
less bias,
151
458259
1892
07:40
more varied information.
152
460175
2389
07:42
And, in fact, the opposite has happened.
153
462588
2325
07:44
And so that, for me, is incredibly dangerous.
154
464937
4018
07:48
And again, when you are the president of this country and you say things,
155
468979
4515
07:53
it also gives leaders in other undemocratic countries the cover
156
473518
5425
08:00
to affront us even worse,
157
480009
2306
08:02
and to really whack us -- and their own journalists --
158
482339
2860
08:05
with this bludgeon of fake news.
159
485223
1823
08:08
CH: To what extent is what happened, though,
160
488000
2184
08:10
in part, just an unintended consequence,
161
490208
2066
08:12
that the traditional media that you worked in
162
492298
2802
08:15
had this curation-mediation role,
163
495124
2080
08:17
where certain norms were observed,
164
497228
2026
08:19
certain stories would be rejected because they weren't credible,
165
499278
3153
08:22
but now that the standard for publication and for amplification
166
502455
6499
08:28
is just interest, attention, excitement, click,
167
508978
3328
08:32
"Did it get clicked on?"
168
512330
1163
08:33
"Send it out there!"
169
513517
1155
08:34
and that's what's -- is that part of what's caused the problem?
170
514696
3504
08:38
CA: I think it's a big problem, and we saw this in the election of 2016,
171
518224
3595
08:41
where the idea of "clickbait" was very sexy and very attractive,
172
521843
5107
08:46
and so all these fake news sites and fake news items
173
526974
4306
08:51
were not just haphazardly and by happenstance being put out there,
174
531304
4122
08:55
there's been a whole industry in the creation of fake news
175
535450
4451
08:59
in parts of Eastern Europe, wherever,
176
539925
2990
09:02
and you know, it's planted in real space and in cyberspace.
177
542939
3260
09:06
So I think that, also,
178
546223
2359
09:08
the ability of our technology to proliferate this stuff
179
548606
5121
09:13
at the speed of sound or light, just about --
180
553751
3511
09:17
we've never faced that before.
181
557286
1983
09:19
And we've never faced such a massive amount of information
182
559293
4867
09:24
which is not curated
183
564184
1565
09:25
by those whose profession leads them to abide by the truth,
184
565773
5296
09:31
to fact-check
185
571093
1202
09:32
and to maintain a code of conduct and a code of professional ethics.
186
572319
4834
09:37
CH: Many people here may know people who work at Facebook
187
577177
3343
09:40
or Twitter and Google and so on.
188
580544
2324
09:42
They all seem like great people with good intention --
189
582892
3132
09:46
let's assume that.
190
586048
1380
09:47
If you could speak with the leaders of those companies,
191
587452
3675
09:51
what would you say to them?
192
591151
1291
09:52
CA: Well, you know what --
193
592466
1769
09:54
I'm sure they are incredibly well-intentioned,
194
594259
2344
09:56
and they certainly developed an unbelievable, game-changing system,
195
596627
5218
10:01
where everybody's connected on this thing called Facebook.
196
601869
3211
10:05
And they've created a massive economy for themselves
197
605104
3801
10:08
and an amazing amount of income.
198
608929
2680
10:11
I would just say,
199
611633
1180
10:12
"Guys, you know, it's time to wake up and smell the coffee
200
612837
4234
10:17
and look at what's happening to us right now."
201
617095
2702
10:19
Mark Zuckerberg wants to create a global community.
202
619821
2932
10:22
I want to know: What is that global community going to look like?
203
622777
3219
10:26
I want to know where the codes of conduct actually are.
204
626020
4067
10:30
Mark Zuckerberg said --
205
630111
1825
10:31
and I don't blame him, he probably believed this --
206
631960
2718
10:34
that it was crazy to think
207
634702
2356
10:37
that the Russians or anybody else could be tinkering and messing around
208
637082
4109
10:41
with this avenue.
209
641215
1243
10:42
And what have we just learned in the last few weeks?
210
642482
2482
10:44
That, actually, there has been a major problem in that regard,
211
644988
2958
10:47
and now they're having to investigate it and figure it out.
212
647970
3118
10:51
Yes, they're trying to do what they can now
213
651112
3279
10:54
to prevent the rise of fake news,
214
654415
2158
10:56
but, you know,
215
656597
1383
10:58
it went pretty unrestricted for a long, long time.
216
658004
5091
11:03
So I guess I would say, you know,
217
663119
1900
11:05
you guys are brilliant at technology;
218
665043
2099
11:07
let's figure out another algorithm.
219
667166
1891
11:09
Can we not?
220
669081
1171
11:10
CH: An algorithm that includes journalistic investigation --
221
670276
2887
11:13
CA: I don't really know how they do it, but somehow, you know --
222
673187
3356
11:16
filter out the crap!
223
676567
1819
11:18
(Laughter)
224
678410
1150
11:19
And not just the unintentional --
225
679584
2002
11:21
(Applause)
226
681610
3254
11:24
but the deliberate lies that are planted
227
684888
2206
11:27
by people who've been doing this as a matter of warfare
228
687118
4325
11:31
for decades.
229
691467
1302
11:32
The Soviets, the Russians --
230
692793
1933
11:34
they are the masters of war by other means, of hybrid warfare.
231
694750
5244
11:40
And this is a --
232
700618
1444
11:42
this is what they've decided to do.
233
702689
2984
11:45
It worked in the United States,
234
705697
1605
11:47
it didn't work in France,
235
707326
1321
11:48
it hasn't worked in Germany.
236
708671
1673
11:50
During the elections there, where they've tried to interfere,
237
710368
2941
11:53
the president of France right now, Emmanuel Macron,
238
713333
2602
11:55
took a very tough stand and confronted it head on,
239
715959
2523
11:58
as did Angela Merkel.
240
718506
1158
11:59
CH: There's some hope to be had from some of this, isn't there?
241
719688
2985
12:02
That the world learns.
242
722697
1151
12:03
We get fooled once,
243
723872
1318
12:05
maybe we get fooled again,
244
725214
1332
12:06
but maybe not the third time.
245
726570
1455
12:08
Is that true?
246
728049
1168
12:09
CA: I mean, let's hope.
247
729241
1156
12:10
But I think in this regard that so much of it is also about technology,
248
730421
3387
12:13
that the technology has to also be given some kind of moral compass.
249
733832
3445
12:17
I know I'm talking nonsense, but you know what I mean.
250
737301
2816
12:20
CH: We need a filter-the-crap algorithm with a moral compass --
251
740141
3708
12:23
CA: There you go.
252
743873
1157
12:25
CH: I think that's good.
253
745054
1152
12:26
CA: No -- "moral technology."
254
746230
1671
12:27
We all have moral compasses -- moral technology.
255
747925
3106
12:31
CH: I think that's a great challenge. CA: You know what I mean.
256
751055
2979
12:34
CH: Talk just a minute about leadership.
257
754058
1944
12:36
You've had a chance to speak with so many people across the world.
258
756026
3136
12:39
I think for some of us --
259
759186
1239
12:40
I speak for myself, I don't know if others feel this --
260
760449
2692
12:43
there's kind of been a disappointment of:
261
763165
1996
12:45
Where are the leaders?
262
765185
1859
12:47
So many of us have been disappointed --
263
767068
2314
12:49
Aung San Suu Kyi, what's happened recently,
264
769406
2016
12:51
it's like, "No! Another one bites the dust."
265
771446
2085
12:53
You know, it's heartbreaking.
266
773555
1599
12:55
(Laughter)
267
775178
1235
12:56
Who have you met
268
776437
2021
12:58
who you have been impressed by, inspired by?
269
778482
2870
13:01
CA: Well, you talk about the world in crisis,
270
781376
2504
13:03
which is absolutely true,
271
783904
1354
13:05
and those of us who spend our whole lives immersed in this crisis --
272
785282
4487
13:09
I mean, we're all on the verge of a nervous breakdown.
273
789793
2993
13:12
So it's pretty stressful right now.
274
792810
2676
13:15
And you're right --
275
795510
1159
13:16
there is this perceived and actual vacuum of leadership,
276
796693
3110
13:19
and it's not me saying it, I ask all these --
277
799827
2850
13:22
whoever I'm talking to, I ask about leadership.
278
802701
2453
13:25
I was speaking to the outgoing president of Liberia today,
279
805178
4510
13:29
[Ellen Johnson Sirleaf,]
280
809712
1810
13:31
who --
281
811546
1154
13:32
(Applause)
282
812724
2215
13:34
in three weeks' time,
283
814963
1542
13:36
will be one of the very rare heads of an African country
284
816529
3944
13:40
who actually abides by the constitution
285
820497
2178
13:42
and gives up power after her prescribed term.
286
822699
3612
13:46
She has said she wants to do that as a lesson.
287
826335
3857
13:50
But when I asked her about leadership,
288
830216
2032
13:52
and I gave a quick-fire round of certain names,
289
832272
2683
13:54
I presented her with the name of the new French president,
290
834979
2977
13:57
Emmanuel Macron.
291
837980
1433
13:59
And she said --
292
839437
1336
14:00
I said, "So what do you think when I say his name?"
293
840797
2506
14:03
And she said,
294
843327
1273
14:05
"Shaping up potentially to be
295
845578
2325
14:07
a leader to fill our current leadership vacuum."
296
847927
4066
14:12
I thought that was really interesting.
297
852017
1833
14:13
Yesterday, I happened to have an interview with him.
298
853874
2456
14:16
I'm very proud to say,
299
856354
1158
14:17
I got his first international interview. It was great. It was yesterday.
300
857536
3419
14:20
And I was really impressed.
301
860979
1292
14:22
I don't know whether I should be saying that in an open forum,
302
862295
2928
14:25
but I was really impressed.
303
865247
1455
14:26
(Laughter)
304
866726
1218
14:28
And it could be just because it was his first interview,
305
868867
2675
14:31
but -- I asked questions, and you know what?
306
871566
2095
14:33
He answered them!
307
873685
1208
14:34
(Laughter)
308
874917
1933
14:36
(Applause)
309
876874
3269
14:40
There was no spin,
310
880167
1593
14:41
there was no wiggle and waggle,
311
881784
2391
14:44
there was no spend-five-minutes- to-come-back-to-the-point.
312
884199
2829
14:47
I didn't have to keep interrupting,
313
887052
1668
14:48
which I've become rather renowned for doing,
314
888744
2083
14:50
because I want people to answer the question.
315
890851
2532
14:53
And he answered me,
316
893407
2051
14:55
and it was pretty interesting.
317
895482
2614
14:58
And he said --
318
898120
1431
14:59
CH: Tell me what he said.
319
899575
1778
15:01
CA: No, no, you go ahead.
320
901377
1220
15:02
CH: You're the interrupter, I'm the listener.
321
902621
2228
15:04
CA: No, no, go ahead.
322
904873
1158
15:06
CH: What'd he say?
323
906055
1155
15:07
CA: OK. You've talked about nationalism and tribalism here today.
324
907234
3078
15:10
I asked him, "How did you have the guts to confront the prevailing winds
325
910336
3762
15:14
of anti-globalization, nationalism, populism
326
914122
4535
15:18
when you can see what happened in Brexit,
327
918681
1962
15:20
when you could see what happened in the United States
328
920667
2555
15:23
and what might have happened in many European elections
329
923246
2595
15:25
at the beginning of 2017?"
330
925865
1717
15:27
And he said,
331
927606
1319
15:29
"For me, nationalism means war.
332
929597
3274
15:33
We have seen it before,
333
933486
1673
15:35
we have lived through it before on my continent,
334
935183
2258
15:37
and I am very clear about that."
335
937465
2686
15:40
So he was not going to, just for political expediency,
336
940175
3961
15:44
embrace the, kind of, lowest common denominator
337
944160
3442
15:47
that had been embraced in other political elections.
338
947626
4005
15:51
And he stood against Marine Le Pen, who is a very dangerous woman.
339
951655
4441
15:56
CH: Last question for you, Christiane.
340
956928
2032
16:00
TED is about ideas worth spreading.
341
960093
1998
16:02
If you could plant one idea into the minds of everyone here,
342
962115
4647
16:06
what would that be?
343
966786
1197
16:08
CA: I would say really be careful where you get your information from;
344
968007
5114
16:13
really take responsibility for what you read, listen to and watch;
345
973145
5322
16:18
make sure that you go to the trusted brands to get your main information,
346
978491
4887
16:23
no matter whether you have a wide, eclectic intake,
347
983402
4689
16:28
really stick with the brand names that you know,
348
988115
2995
16:31
because in this world right now, at this moment right now,
349
991134
3592
16:34
our crises, our challenges, our problems are so severe,
350
994750
4339
16:39
that unless we are all engaged as global citizens
351
999113
3551
16:42
who appreciate the truth,
352
1002688
1903
16:44
who understand science, empirical evidence and facts,
353
1004615
4345
16:48
then we are just simply going to be wandering along
354
1008984
3499
16:52
to a potential catastrophe.
355
1012507
1961
16:54
So I would say, the truth,
356
1014492
1364
16:55
and then I would come back to Emmanuel Macron
357
1015880
2256
16:58
and talk about love.
358
1018160
1300
17:00
I would say that there's not enough love going around.
359
1020022
4469
17:04
And I asked him to tell me about love.
360
1024515
2692
17:07
I said, "You know, your marriage is the subject of global obsession."
361
1027231
3592
17:10
(Laughter)
362
1030847
1635
17:12
"Can you tell me about love?
363
1032506
1413
17:13
What does it mean to you?"
364
1033943
1314
17:15
I've never asked a president or an elected leader about love.
365
1035281
2941
17:18
I thought I'd try it.
366
1038246
1158
17:19
And he said -- you know, he actually answered it.
367
1039428
3915
17:23
And he said, "I love my wife, she is part of me,
368
1043367
4161
17:27
we've been together for decades."
369
1047552
1627
17:29
But here's where it really counted,
370
1049203
1685
17:30
what really stuck with me.
371
1050912
1503
17:32
He said,
372
1052439
1241
17:33
"It is so important for me to have somebody at home
373
1053704
3520
17:37
who tells me the truth."
374
1057248
1899
17:40
So you see, I brought it home. It's all about the truth.
375
1060618
2712
17:43
(Laughter)
376
1063354
1006
17:44
CH: So there you go. Truth and love. Ideas worth spreading.
377
1064384
2807
17:47
Christiane Amanpour, thank you so much. That was great.
378
1067215
2663
17:49
(Applause)
379
1069902
1068
17:50
CA: Thank you. CH: That was really lovely.
380
1070994
2334
17:53
(Applause)
381
1073352
1215
17:54
CA: Thank you.
382
1074591
1165
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7