Fake News: Fact & Fiction - Episode 5: Why do people share fake news?

38,640 views ・ 2023-10-10

BBC Learning English


Please double-click on the English subtitles below to play the video.

00:06
Hello and welcome to Fake News: Fact and Fiction, from BBC Learning English.
0
6809
3540
00:10
I'm Hugo.
1
10439
791
00:11
And I'm Sam.
2
11339
781
00:12
In the programme today,
3
12619
791
00:13
we'll look at some of the reasons people share fake news.
4
13490
2659
00:16
We'll hear from disinformation researcher Samantha Bradshaw.
5
16229
2901
00:20
One reason why people share fake news that they know is fake,
6
20039
3781
00:24
has to do with this thing that we call identity signalling.
7
24179
4410
00:28
Identity signalling. Have you heard of that Sam?
8
28949
2520
00:31
Not until I saw the script for this programme actually.
9
31890
3780
00:35
Ah, so make sure you keep watching to find out what identity signalling is
10
35750
3530
00:40
and how it's a part of the spread of fake news.
11
40530
2939
00:43
Before we get to that, we have some key vocabulary which you're going to
12
43979
3271
00:47
tell us about Sam.
13
47329
820
00:48
What do you have for us today?
14
48299
1091
00:49
Yes, so today I am talking about: agenda, actors and motives.
15
49649
5610
00:59
If you have to go to meetings and who doesn't love a good meeting?
16
59189
4841
01:04
I hate them.
17
64700
1080
01:05
They're such a waste of time.
18
65859
2261
01:08
You will probably be familiar with agendas.
19
68239
3410
01:12
An agenda is a list of things that are going to be discussed
20
72120
4040
01:16
at the meeting.
21
76370
1040
01:17
But there is another meaning of the word 'agenda' that is important
22
77659
4291
01:22
in the area of fake news.
23
82040
2450
01:24
If an individual or organisation or a government 'has an agenda',
24
84590
5599
01:30
it means that there is something they want to achieve.
25
90530
3179
01:33
This could be clearly stated objectives.
26
93799
3151
01:37
But having an agenda can also suggest a secret or hidden aim.
27
97030
5559
01:42
Let's park the word agenda for the moment because there is another word
28
102669
4031
01:46
I want to talk about that has a very well-known meaning but also
29
106780
4120
01:51
another meaning which is perhaps not so well-known - actor.
30
111230
4340
01:56
It's probably a very familiar word.
31
116780
3030
01:59
Somebody who performs on stage or in films is an actor.
32
119890
3869
02:03
But the same word is used to describe a person, group or even
33
123879
4511
02:08
a state that has an agenda that's either politically or economically motivated
34
128470
6099
02:15
and they have influence or power that they use
35
135020
2430
02:17
to promote that agenda.
36
137530
2659
02:20
In this context you can hear the phrases:
37
140389
2761
02:23
state actors, political actors, and even - bad actors.
38
143449
6180
02:30
State actors are governments,
39
150360
3709
02:34
political actors are those with, you guessed it, political power and agendas
40
154150
4359
02:38
and bad actors - well that doesn't mean people who aren't very good at acting.
41
158990
5250
02:44
A bad actor uses power or influence in a dishonest way
42
164840
4949
02:50
to push a hidden agenda.
43
170210
2500
02:52
They're trying to achieve something but it's not clear what it is.
44
172909
3680
02:56
And it's not clear who is doing it.
45
176719
2871
02:59
For example, if a government from one country wants to influence an
46
179670
4160
03:03
election in another country, they may spread fake news stories
47
183910
4180
03:08
but hide the fact that they're doing it.
48
188180
2910
03:11
This could be described as the actions of a bad actor or
49
191170
4060
03:15
a bad state actor.
50
195620
1530
03:18
People who spread disinformation online may not be doing it because
51
198169
4050
03:22
they believe it but because it suits their agenda.
52
202300
4810
03:27
This spreading of disinformation can be very sophisticated and the
53
207190
4539
03:31
motives of those doing it are often not obvious.
54
211810
3490
03:36
Motive.
55
216259
1231
03:37
That's another useful word in this context.
56
217570
2770
03:41
If you like TV crime dramas, and who doesn't love a good crime drama?,
57
221240
4889
03:46
- me, I never really understand them -
58
226590
3934
03:50
you will often hear the word 'motive'.
59
230604
3835
03:54
"He can't be guilty. He didn't have a motive."
60
234520
3749
03:59
A motive is a reason for doing something and not everyone's motives
61
239060
4979
04:04
for what they post online are genuine.
62
244120
2519
04:07
But how can you tell?
63
247340
1169
04:09
Let's go back to the studio to see if there are any more useful tips.
64
249349
4111
04:16
Thanks for that Sam.
65
256449
881
04:17
Really interesting, I was really impressed with your acting there.
66
257410
3040
04:20
Well, I may be a bad actor but I'm not a bad actor.
67
260990
5169
04:26
My motives are purely genuine.
68
266239
2290
04:28
Well that's an interesting point about fake news I think.
69
268609
2100
04:31
We've heard that fake news is produced by people who may want to
70
271160
3210
04:34
make money or have a political or social agenda.
71
274449
3391
04:37
Those are their motives,
72
277920
1190
04:39
but they can only be successful if we share those stories.
73
279189
3801
04:44
Absolutely. But why do people share news, stories, images that are fake?
74
284120
5310
04:49
Well researcher Samantha Bradshaw from the Oxford Internet Institute
75
289720
3820
04:53
at Oxford University shared her observations with us earlier.
76
293620
3490
04:58
People share fake news for a variety of reasons.
77
298910
3380
05:02
In many cases sometimes people don't even realise the stories
78
302449
3871
05:06
that they're sharing are fake.
79
306400
1819
05:08
They might find the content funny or humorous or they might think that
80
308329
4110
05:12
people in their network such as their friends or family or broader
81
312519
3731
05:16
acquaintances might enjoy this content.
82
316329
3760
05:20
Another reason people might share fake news has to do with
83
320169
3520
05:24
the emotional appeal of fake news because it tends to be very negative.
84
324050
5549
05:29
It tends to also elicit a reaction from people and they might be angry
85
329679
4541
05:34
and they might type and share and discuss what is being talked about
86
334300
5649
05:40
in this article because they find it so outrageous that they need
87
340029
4151
05:44
to share it.
88
344259
1620
05:45
The third reason why people might share fake news has to do with what
89
345959
3650
05:49
we call 'identity signalling' where people know that the information
90
349689
4571
05:54
being shared in this story is false.
91
354340
2530
05:56
But the purpose of sharing it is to signal to a broader community that
92
356949
4210
06:01
they share certain values that are reflected in that news story.
93
361239
4910
06:06
And so it gives people a sense of belonging and
94
366680
5130
06:12
fitting into a broader community of people.
95
372139
2671
06:14
Has much more to do with identity than the actual facts of the story.
96
374889
4241
06:20
So what do you make of that Sam?
97
380449
1330
06:22
OK, well, I can definitely see how we might accidentally share
98
382790
4039
06:26
things that aren't true for, for example I sometimes see some
99
386909
3800
06:30
amazing photos that I find out later are actually photoshopped.
100
390789
3311
06:34
and things that might make me angry can make me reach for the share button.
101
394939
4230
06:39
But what stood out for me was this idea of identity signalling.
102
399249
4361
06:44
We might want to show that we belong to a particular group.
103
404060
3040
06:47
We share some beliefs and opinions and that it really doesn't matter
104
407180
3930
06:51
whether the content is true or not.
105
411350
1729
06:53
And this is something that the people who produce fake news know well
106
413399
3501
06:57
and they know how to make us share those stories.
107
417050
2669
06:59
Here's Samantha Bradshaw again.
108
419799
2090
07:02
So the purveyors of fake news will use a variety of strategies to
109
422329
5731
07:08
get people clicking and sharing their content.
110
428139
4000
07:12
One of these strategies has to do with just the style that fake news
111
432219
5410
07:17
stories are often written in.
112
437709
1781
07:19
They'll tend to use highly emotional language and content also that gets
113
439759
6750
07:26
people's emotions high.
114
446589
2090
07:28
And this is because psychologically we also like to share
115
448759
4861
07:33
this kind of information.
116
453699
1490
07:35
And this kind of content, especially highly emotive content,
117
455360
4109
07:39
will go much further because of the algorithms of the platforms.
118
459549
4561
07:44
And we also know that emotions like anger, outrage, fear are
119
464189
5690
07:49
much stronger emotions than things like happiness and joy.
120
469959
3340
07:53
And so that's why we see disinformation and fake news stories
121
473379
4031
07:57
go viral compared to facts and the boring truth and the mundane truth
122
477490
6039
08:03
behind our factual daily news briefs.
123
483609
5261
08:08
We of course still see our cat memes and cute animal videos going viral too
124
488949
5171
08:14
because joy and happiness is a strong emotion but it's not quite as strong
125
494199
5020
08:19
as things like anger and fear.
126
499299
2751
08:22
Now we know that many fake news stories are very emotive.
127
502129
2880
08:25
They generate strong emotions; the sense that something's not fair
128
505089
4000
08:29
or people or organisations are doing things they shouldn't and that can
129
509269
3390
08:32
make us look for other content to support that feeling.
130
512740
2990
08:36
Yeah but good to know the cute cat videos are being shared too.
131
516620
4430
08:41
And dogs don't forget dogs!
132
521279
2951
08:44
But she also talked about algorithms,
133
524899
2070
08:47
the algorithms that social media companies use.
134
527090
2969
08:50
So Sam what is an algorithm?
135
530289
2291
08:53
OK, thanks for that Hugo!
136
533090
1449
08:55
I might get into trouble here from computer experts but let's have a go.
137
535190
5490
09:00
Have you ever lost a huge amount of time because, for example,
138
540760
4390
09:05
you've gone on YouTube to watch a video and then you see a list of
139
545230
4180
09:09
suggested related videos which start playing automatically?
140
549490
4059
09:14
All the time, all the time. I think we've all done that.
141
554000
3009
09:17
Absolutely. So you click on one video and then another
142
557940
2420
09:20
and another and another and then it's been hours and hours and hours.
143
560629
4121
09:25
Exactly. Yeah. OK. So what's this got to do with algorithms?
144
565169
3050
09:28
Well an algorithm is a set of instructions a computer system
145
568299
4811
09:33
follows to complete a particular task.
146
573190
3329
09:36
Social media companies use algorithms to make that list of
147
576600
4130
09:40
related videos just for you.
148
580810
3370
09:44
These are personalised based on your behaviour online and
149
584260
4179
09:48
they get us clicking.
150
588519
1500
09:50
This is because the companies want to keep you on their platforms for
151
590100
3830
09:54
as long as possible. That's how they make their money.
152
594010
3389
09:57
The longer you're on their site the more links you click the more
153
597529
3331
10:00
advertisements you see and the more cash they collect.
154
600940
2889
10:04
And the more we click on a particular kind of story, the more
155
604720
3600
10:08
the algorithms give us similar content.
156
608400
2790
10:11
Now there's nothing necessarily wrong with that because I guess
157
611270
3540
10:14
after all we're getting a free service which has to be paid for somehow.
158
614890
4840
10:20
But it does mean that we usually don't see things that
159
620090
3150
10:23
we don't like, that we don't agree with.
160
623320
1859
10:25
And this can make us feel that our view, our opinion
161
625259
3680
10:29
is the right one which everyone shares.
162
629029
2120
10:31
Which is why I think it's a good idea to try and look for different opinions.
163
631460
4189
10:35
So we don't end up thinking that everyone thinks the way we do.
164
635730
4069
10:39
So does that mean after watching all your cat videos you search for dog videos,
165
639879
4821
10:44
for a bit of balance?
166
644780
1409
10:46
I love cats and dogs equally actually but that's not exactly what I was thinking.
167
646429
4551
10:51
I'm just putting your leg there.
168
651060
979
10:52
Sam please remind us of today's vocabulary.
169
652549
2921
10:55
Certainly Hugo.
170
655860
600
10:56
So we started with 'agenda' which is a plan of things to be done.
171
656540
3659
11:00
And it's a hidden, and if it's a hidden agenda, it means the reason
172
660679
3841
11:04
for someone's actions might be secret or hidden.
173
664600
3779
11:08
Then there was the word 'actor'.
174
668480
1659
11:10
In the context of fake news an actor is someone with an agenda that uses
175
670549
4680
11:15
power and influence to try and promote that agenda.
176
675309
3761
11:20
A 'motive' is a reason for doing something.
177
680029
2770
11:22
And like hidden agendas sometimes the motives for a particular action
178
682990
4419
11:27
on the Internet are not always clear and transparent.
179
687490
3979
11:31
Samantha Bradshaw talked about identity signalling.
180
691549
3641
11:35
This is the description of behaviour online when people share things
181
695299
3691
11:39
without being interested in whether they're true or not.
182
699320
2889
11:42
What's important is not how true it is but that it demonstrates that
183
702289
4110
11:46
they think in the same way or share the same beliefs as the group.
184
706480
5149
11:51
And finally 'algorithm'.
185
711710
1380
11:53
An algorithm is a set of software instructions that for example can
186
713600
3990
11:57
produce a personalised list of recommended videos based on what
187
717670
4089
12:01
an individual has searched for or watched in the past.
188
721840
3160
12:06
Thank you Sam. That was very comprehensive.
189
726250
2200
12:08
That's all from us today. We look forward to your company again soon.
190
728530
3269
12:11
Goodbye. And stay safe.
191
731879
2260
12:14
Bye bye!
192
734259
500
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7