BOX SET: 6 Minute English - 'Technology and lifestyle' English mega-class! 30 minutes of vocabulary!

278,468 views

2023-05-21 ・ BBC Learning English


New videos

BOX SET: 6 Minute English - 'Technology and lifestyle' English mega-class! 30 minutes of vocabulary!

278,468 views ・ 2023-05-21

BBC Learning English


Please double-click on the English subtitles below to play the video.

00:05
Hello and welcome to
0
5920
1040
00:06
6 Minute English. I'm Dan.
1
6960
1760
00:08
And I'm Rob. So, Dan, what's that…
2
8720
2720
00:11
Oh, sorry. Oh, it's my wife.
3
11440
4960
00:16
Err… hang on…
4
16400
1920
00:18
You didn't answer!
5
18320
1276
00:19
Well, don't take this personally, Dan,
6
19596
2404
00:22
but I'm not exactly crazy about
7
22000
2080
00:24
someone eavesdropping on my
8
24080
1200
00:25
phone call. If you eavesdrop on
9
25280
1920
00:27
something, you secretly listen to
10
27200
1680
00:28
someone's conversation.
11
28880
1360
00:30
Some things are private, you know?
12
30240
1280
00:31
Oh! Of course! I totally understand.
13
31520
2882
00:34
One quick question for you though…
14
34402
2478
00:36
do you have a smart speaker?
15
36880
1680
00:38
You know, like the Google Assistant,
16
38560
1840
00:40
Amazon's Alexa or Apple's Siri.
17
40400
2800
00:43
Oh sure, yes, I've got one! It's great!
18
43200
2320
00:45
I can ask it all sorts of questions, it
19
45520
2320
00:47
tells me about the news and weather,
20
47840
1600
00:49
it plays music when I want… it does
21
49440
2080
00:51
all sorts! You just give it a voice
22
51520
1840
00:53
command and it does what you want!
23
53360
2240
00:55
So it can hear you, can it?
24
55600
1520
00:57
Of course! How else can you give
25
57120
1840
00:58
it a voice command?
26
58960
1200
01:00
All the time?
27
60160
880
01:01
Well, I assume so.
28
61040
1280
01:02
So how do you know it's not
29
62320
1200
01:03
eavesdropping on you?
30
63520
1760
01:05
Well, I… oh… I see. I hadn't thought
31
65280
3280
01:08
of that.
32
68560
821
01:09
That's our topic for this
33
69381
1419
01:10
6 Minute English.
34
70800
1200
01:12
How safe is your smart speaker?
35
72000
2400
01:14
However, before that, here is our
36
74400
2080
01:16
quiz question. By what percentage
37
76480
1880
01:18
has the number of smart speakers
38
78360
1560
01:19
used in US households increased
39
79920
2320
01:22
from December 2017 to December
40
82240
2320
01:24
2018? Is it…
41
84560
1360
01:25
a) around 40%
42
85920
1680
01:27
b) around 60%, or
43
87600
1920
01:29
c) around 80%?
44
89520
2080
01:31
Oh, well, I know they are very
45
91600
2000
01:33
popular even in my household.
46
93600
1760
01:35
So I'm going to go for c) around 80%.
47
95360
3017
01:38
Well, we'll find out if you're right later
48
98377
2183
01:40
in the programme. So, smart
49
100560
2320
01:42
speakers and privacy! Florian Schaub
50
102880
2720
01:45
is an assistant professor at the
51
105600
1440
01:47
University of Michigan School
52
107040
1600
01:48
of Information. Here he is speaking
53
108640
2400
01:51
on the BBC World Service programme
54
111040
1840
01:52
The Why Factor about smart
55
112880
1760
01:54
speakers. What does he say people
56
114640
2320
01:56
are introducing into their homes?
57
116960
2828
02:00
You're basically introducing...
58
120480
1600
02:02
listening bug in your home, in your
59
122080
1920
02:04
most intimate space. While the
60
124000
1760
02:05
companies say they are only actively
61
125760
2400
02:08
listening to what's going on in your
62
128160
1120
02:09
home when they hear the keyword,
63
129280
1920
02:11
the microphone is still on the whole
64
131200
1440
02:12
time in order to be able to detect
65
132640
1760
02:14
that keyword. We don't know to
66
134400
1360
02:15
what extent companies are
67
135760
1360
02:17
co-operating with the government
68
137120
1440
02:18
or to what extent the government
69
138560
1280
02:19
might try to circumvent company
70
139840
2480
02:22
security mechanisms in order to
71
142320
2160
02:24
then be able to listen to what
72
144480
1360
02:25
you're doing.
73
145840
1298
02:28
So, what did he say people are
74
148064
1536
02:29
introducing, Rob?
75
149600
1280
02:30
He basically said we're introducing
76
150880
2400
02:33
a listening bug. Now, a bug is a
77
153280
2640
02:35
small electronic device used for
78
155920
1600
02:37
secretly listening to conversations.
79
157520
2000
02:39
Much like a spy would use.
80
159520
1840
02:41
Yes, and he mentioned it was in
81
161360
1920
02:43
our most intimate space! 'Intimate'
82
163280
2800
02:46
means 'private and personal'.
83
166080
2187
02:48
Well, I can't think of anywhere
84
168267
1333
02:49
more intimate than my home.
85
169600
2160
02:51
Indeed! He also said that the
86
171760
2400
02:54
smart speaker's microphone is
87
174160
1680
02:55
on the whole time – even though
88
175840
1760
02:57
the companies insist that they're
89
177600
1520
02:59
only actively listening when the
90
179120
1760
03:00
keyword is said.
91
180880
1280
03:02
Yes, he suggested that we can't
92
182160
1440
03:03
know how far a company might
93
183600
1440
03:05
be co-operating with a government
94
185040
1680
03:06
to eavesdrop on people.
95
186720
1520
03:08
Or whether a government might
96
188240
1200
03:09
be circumventing a smart speaker's
97
189440
2080
03:11
security and listening in anyway
98
191520
2800
03:14
without the company's or owner's
99
194320
1760
03:16
permission!
100
196080
1120
03:17
'Circumvent' means 'cleverly bypass
101
197200
2560
03:19
or go around'. So, if all this
102
199760
2400
03:22
eavesdropping is possible, why are
103
202160
2320
03:24
smart speakers so popular?
104
204480
2240
03:26
Good question! And here's Florian
105
206720
2160
03:28
Schaub again with an answer. He
106
208880
2000
03:30
conducted a study on people's
107
210880
1920
03:32
attitudes to privacy when it comes
108
212800
1760
03:34
to smart speakers. How do people
109
214560
2320
03:36
feel about having a smart speaker
110
216880
1760
03:38
that could eavesdrop on them?
111
218640
1865
03:40
What we often saw is people just
112
220880
2080
03:42
being resigned to 'this is the
113
222960
2000
03:44
trade-off they have to make' if
114
224960
1360
03:46
they want to enjoy the convenience
115
226320
1920
03:48
that a smart speaker provides to them.
116
228240
2192
03:50
He said that people are resigned
117
230960
2320
03:53
to the privacy trade-off. If you are
118
233280
2480
03:55
resigned to something, you
119
235760
1360
03:57
accept something unpleasant that
120
237120
1760
03:58
can't be changed.
121
238880
1120
04:00
Yes and a trade-off is a compromise.
122
240000
2560
04:02
You accept something bad to also
123
242560
2080
04:04
receive something good.
124
244640
1360
04:06
So people accept that a smart
125
246000
1600
04:07
speaker gives them advantages,
126
247600
2240
04:09
even though there could be
127
249840
1440
04:11
downsides?
128
251280
800
04:12
Yes. In the grand scheme of things,
129
252080
2080
04:14
the data that these devices hear
130
254160
2000
04:16
is probably not that significant
131
256160
2080
04:18
considering all the data companies
132
258240
1760
04:20
have about us already anyway!
133
260000
2279
04:22
So, can I have the answer to the
134
262857
1543
04:24
quiz then?
135
264400
640
04:25
Of course! Earlier I asked by what
136
265040
2880
04:27
percentage the number of smart
137
267920
1440
04:29
speakers used in US households
138
269360
2080
04:31
increased from December 2017
139
271440
2240
04:33
to December 2018? Was it…
140
273680
2080
04:35
a) around 40%
141
275760
1680
04:37
b) around 60%, or
142
277440
1760
04:39
c) around 80%?
143
279200
1760
04:40
What did you say, Rob?
144
280960
1040
04:42
I said c) around 80%.
145
282000
2080
04:44
And you are right. The answer
146
284080
2720
04:46
is around 80% - from 66 million
147
286800
2720
04:49
in December 2017 to 118 million
148
289520
3760
04:53
in December 2018, and around
149
293280
2800
04:56
ten million people in the UK now
150
296080
2080
04:58
use one too! I guess they're
151
298160
2080
05:00
really not worried about
152
300240
720
05:00
eavesdropping.
153
300960
896
05:01
Nice slide into the vocabulary
154
301856
2464
05:04
there, Dan. If someone eavesdrops
155
304320
2320
05:06
on you, it means they secretly
156
306640
1600
05:08
listen to your conversation.
157
308240
1440
05:09
They could be eavesdropping on
158
309680
1360
05:11
you through a bug, which is a small
159
311040
2240
05:13
electronic device used to secretly
160
313280
2080
05:15
listen to conversations.
161
315360
1840
05:17
Yes, they may have bugged your
162
317200
1520
05:18
most intimate, or private and
163
318720
1840
05:20
personal, spaces.
164
320560
1440
05:22
Next we had circumvent. If you
165
322000
2160
05:24
circumvent something, such as
166
324160
1440
05:25
security, you cleverly or bypass
167
325600
2240
05:27
it or go around it.
168
327840
1920
05:29
Then we had resigned. If you are
169
329760
2080
05:31
resigned to something, it means
170
331840
1840
05:33
you accept something unpleasant
171
333680
2000
05:35
that can't be changed.
172
335680
1440
05:37
And lastly, we had trade-off.
173
337120
2320
05:39
A trade-off is a compromise.
174
339440
1760
05:41
You get something good, but you
175
341200
1840
05:43
also get something bad.
176
343040
1920
05:44
Right - like 6 Minute English!
177
344960
1920
05:46
A great discussion and vocabulary,
178
346880
2320
05:49
but the trade-off is it only lasts
179
349200
2240
05:51
six minutes!
180
351440
1120
05:52
Which is just about now,
181
352560
1440
05:54
actually - time to go. So until next
182
354000
2240
05:56
time, find us all over the place
183
356240
1583
05:57
online and on social media.
184
357823
2402
06:00
Just search for BBC Learning English.
185
360225
2335
06:02
Bye for now.
186
362560
720
06:03
Goodbye!
187
363280
841
06:09
Hello. This is 6 Minute English
188
369680
1600
06:11
from BBC Learning English.
189
371280
1440
06:12
I’m Neil.
190
372720
880
06:13
And I’m Sam.
191
373600
960
06:14
People collect all kinds of things
192
374560
1760
06:16
for a hobby, from stamps and coins
193
376320
1840
06:18
to comics and football stickers.
194
378160
2000
06:20
Do you collect anything, Sam?
195
380160
1920
06:22
I used to have a big collection
196
382080
1680
06:23
of Pokémon cards but I have no
197
383760
2080
06:25
idea where they are now.
198
385840
1600
06:27
Well, maybe you should start
199
387440
1280
06:28
looking for them because all
200
388720
1360
06:30
kinds of collectables – that’s
201
390080
2000
06:32
objects that people want to
202
392080
1280
06:33
collect – are selling for big
203
393360
1680
06:35
money on the internet. It’s
204
395040
1680
06:36
all part of a new tech craze
205
396720
1680
06:38
called non-fungible tokens,
206
398400
2000
06:40
or NFTs for short.
207
400400
1760
06:42
Non-fungible tokens?
208
402160
1760
06:43
Basically, NTFs are unique,
209
403920
2160
06:46
one-of-kind items that can
210
406080
1600
06:47
be bought and sold like any
211
407680
1600
06:49
physical object but only exist
212
409280
2320
06:51
in the digital world. ‘Tokens’
213
411600
2320
06:53
can be thought of as certificates
214
413920
1680
06:55
of ownership for these virtual
215
415600
1840
06:57
possessions.
216
417440
600
06:58
Hmmm, OK. I understand the
217
418040
2920
07:00
‘token’ part but what does
218
420960
2000
07:02
‘non-fungible’ mean?
219
422960
1600
07:04
If something is ‘fungible’ it can
220
424560
1760
07:06
be interchanged, like money,
221
426320
2000
07:08
for example. With money you
222
428320
1680
07:10
can swap a £10 note for two
223
430000
1920
07:11
£5 notes and it will have the
224
431920
1600
07:13
same value.
225
433520
1086
07:14
So, something non-fungible
226
434606
2194
07:16
cannot be interchanged with
227
436800
1520
07:18
something else. Is that because
228
438320
1680
07:20
it has special features that
229
440000
1760
07:21
make it unique?
230
441760
1360
07:23
Exactly. Imagine something
231
443120
1680
07:24
totally unique like the Mona Lisa.
232
444800
2080
07:26
You can take a photo of the
233
446880
1280
07:28
painting or buy a copy of it, but
234
448160
2080
07:30
there will only ever be the one
235
450240
1840
07:32
original painting.
236
452080
1200
07:33
I can see that the Mona Lisa is
237
453280
2320
07:35
one-of-a-kind and extremely
238
455600
1920
07:37
valuable, but it’s not for sale
239
457520
2000
07:39
on the internet!
240
459520
1040
07:40
True, but lots of other things
241
460560
1520
07:42
are, from signed celebrity
242
462080
1440
07:43
artwork to virtual football
243
463520
1760
07:45
cards. NFTs are like autographed
244
465280
2640
07:47
photos – collectors want something
245
467920
2080
07:50
no-one else has, even though
246
470000
1600
07:51
there’s nothing physical they
247
471600
1440
07:53
actually own and keep.
248
473040
1440
07:54
And the value of NFTs is going
249
474480
2480
07:56
up and up?
250
476960
1040
07:58
Massively. A digital sticker of
251
478000
2240
08:00
French footballer, Kylian Mbappé,
252
480240
1760
08:02
recently sold for £25,000! Which
253
482000
3680
08:05
reminds me of my quiz
254
485680
1040
08:06
question, Sam.
255
486720
1200
08:07
In June 2021, Sir Tim Berners-Lee
256
487920
2960
08:10
sold the original source code he
257
490880
1760
08:12
used to invent the World Wide
258
492640
1520
08:14
Web as an NFT at a charity auction.
259
494160
3040
08:17
The sale started at $1000 but
260
497200
2080
08:19
how much did the source code
261
499280
1600
08:20
eventually sell for? Was it:
262
500880
1760
08:22
a) 5.4 thousand dollars?
263
502640
2480
08:25
b) 5.4 million dollars? or
264
505120
2400
08:27
c) 5.4 billion dollars?
265
507520
3065
08:30
If you’re asking me how much
266
510585
2055
08:32
the internet sold for, I’d say
267
512640
2640
08:35
c) 5.4 billion dollars.
268
515280
3120
08:38
OK, Sam, we’ll find out later
269
518400
1600
08:40
if that’s right.
270
520000
960
08:40
Sir Tim Berners-Lee famously
271
520960
1760
08:42
never made any money from the
272
522720
1840
08:44
World Wide Web, insisting that
273
524560
2000
08:46
his invention should be free
274
526560
1680
08:48
for everyone. But the world of
275
528240
1600
08:49
NFTs is controversial and not
276
529840
2560
08:52
everyone thinks Sir Tim should
277
532400
1680
08:54
be getting involved.
278
534080
1360
08:55
That’s certainly the view of
279
535440
1280
08:56
Shona Ghosh, technology
280
536720
1520
08:58
editor at the ‘The Insider’ website,
281
538240
2000
09:00
as she told BBC World Service
282
540240
1840
09:02
programme, Tech Tent:
283
542080
1280
09:03
I think there’s a balance to be
284
543360
1120
09:04
struck between exploring new
285
544480
1840
09:06
technologies and
286
546320
1760
09:08
Sir Tim Berners-Lee is an
287
548080
2000
09:10
amazing figure but NFTs are
288
550080
2480
09:12
a Wild West. Not everything
289
552560
2960
09:15
associated with NFTs are
290
555520
1840
09:17
rogue, but these so-called digital
291
557360
2320
09:19
collectables are going for
292
559680
1040
09:20
lots of money.
293
560720
1360
09:22
Internet NFTs are very new and
294
562080
2480
09:24
there are no rules controlling what
295
564560
2080
09:26
can be sold and for how much, so
296
566640
2640
09:29
Shona describes them as a 'Wild West'.
297
569280
3360
09:32
The Wild West means a situation
298
572640
2080
09:34
where people can do whatever
299
574720
1280
09:36
they want because there are no
300
576000
1440
09:37
laws or controls, like the early
301
577440
1760
09:39
history of the western part of
302
579200
1600
09:40
the United States.
303
580800
1120
09:41
She also calls some things about
304
581920
1600
09:43
NFTs rogue - behaving differently
305
583520
3440
09:46
from what’s normal or expected,
306
586960
2320
09:49
often in a way that causes damage.
307
589280
2880
09:52
Love them or hate them, there are
308
592160
1440
09:53
strong opinions on both sides of
309
593600
1760
09:55
the NFT debate, as
310
595360
1440
09:56
Rory Cellan-Jones, presenter
311
596800
1920
09:58
of BBC World Service’s
312
598720
1600
10:00
Tech Tent, explains:
313
600320
1440
10:01
To some NFTs are a brilliant
314
601760
2480
10:04
innovation which has promised
315
604240
1200
10:05
to put a value on digital artefacts.
316
605440
2880
10:08
To others they’re little more than
317
608320
1600
10:09
a dubious pyramid scheme with a
318
609920
2080
10:12
damaging impact on the environment
319
612000
1920
10:13
because of the way the tokens
320
613920
1440
10:15
are created.
321
615360
1120
10:16
Some people are suspicious of
322
616480
1680
10:18
the large amounts of money
323
618160
1280
10:19
collectors are willing to pay,
324
619440
2000
10:21
comparing NFTs to pyramid
325
621440
2160
10:23
schemes - business tricks or
326
623600
2320
10:25
scams where money is
327
625920
1520
10:27
obtained dishonestly.
328
627440
1840
10:29
But for others NFTs are a
329
629280
1760
10:31
legitimate and useful way
330
631040
1520
10:32
to put a price on rare digital
331
632560
1840
10:34
artefacts – items, such as
332
634400
1840
10:36
images, videos and music,
333
636240
1920
10:38
that are produced and stored
334
638160
1360
10:39
as electronic versions.
335
639520
1680
10:41
Items like, for example, the
336
641200
2080
10:43
original source code for the
337
643280
1440
10:44
internet. So, how much money
338
644720
1520
10:46
did Sir Tim Berners-Lee raise for
339
646240
1840
10:48
charity when he sold it off, Neil?
340
648080
2160
10:50
In my quiz question, I asked
341
650240
1440
10:51
Sam how much the NFT for
342
651680
1920
10:53
the original internet source
343
653600
1360
10:54
code sold for at auction.
344
654960
2240
10:57
I said c) 5.4 billion dollars.
345
657200
3120
11:00
You were… wrong! In fact, it
346
660320
2160
11:02
was less - 5.4 million dollars –
347
662480
2320
11:04
but still far too expensive
348
664800
1360
11:06
for me to collect.
349
666160
800
11:06
Hmmm, now I’m wondering
350
666960
960
11:07
how much my Pokémon card
351
667920
1440
11:09
collection would sell for. If only
352
669360
1680
11:11
I could find it!
353
671040
1360
11:12
Well, while you look for your
354
672400
1120
11:13
collection, Sam, let’s recap
355
673520
1520
11:15
the vocabulary from this
356
675040
1200
11:16
programme all about NFTs
357
676240
2000
11:18
which are digital artefacts or
358
678240
2080
11:20
items that are produced and
359
680320
1440
11:21
stored as electronic versions.
360
681760
2080
11:23
NFTs are virtual collectables –
361
683840
2640
11:26
desirable objects that people
362
686480
1680
11:28
search for and collect.
363
688160
1840
11:30
And they are non-fungible
364
690000
1680
11:31
meaning they have special and
365
691680
1360
11:33
unique characteristics that
366
693040
1360
11:34
cannot be interchanged
367
694400
1280
11:35
with anything else.
368
695680
1040
11:36
If something is rogue it behaves
369
696720
2000
11:38
in a different way from what’s
370
698720
1600
11:40
normal or expected, often
371
700320
2000
11:42
causing damage.
372
702320
1440
11:43
The Wild West describes a
373
703760
1440
11:45
chaotic situation without
374
705200
1600
11:46
laws or controls.
375
706800
1360
11:48
And finally, a pyramid scheme
376
708160
2000
11:50
is a business scam where
377
710160
1600
11:51
money is gained dishonestly.
378
711760
2160
11:53
If you’d like to know more
379
713920
1280
11:55
about non-fungible tokens, bitcoin
380
715200
1920
11:57
and other trending internet topics
381
717120
1680
11:58
there’s plenty to find on the
382
718800
1360
12:00
BBC website.
383
720160
1040
12:01
And for more interesting
384
721200
960
12:02
conversation and useful
385
722160
1360
12:03
vocabulary, remember to join us
386
723520
1920
12:05
again soon here at 6 Minute
387
725440
1920
12:07
English. Bye for now!
388
727360
1600
12:08
Bye!
389
728960
880
12:15
Welcome to 6 Minute English,
390
735520
1760
12:17
the programme where we
391
737280
1200
12:18
explore an interesting topic
392
738480
1680
12:20
and bring you six items of
393
740160
1680
12:21
useful vocabulary. I'm Catherine.
394
741840
2400
12:24
And I'm Rob.
395
744240
960
12:25
I have a question for you, Rob:
396
745200
2000
12:27
how would you feel about
397
747200
1600
12:28
having therapy from a robot?
398
748800
3120
12:31
I'm not too sure about that – you'll
399
751920
1600
12:33
need to tell me more! But first
400
753520
1680
12:35
things first, the word 'therapy'
401
755200
2480
12:37
refers to a kind of treatment
402
757680
1920
12:39
that helps someone feel better –
403
759600
1920
12:41
including treatment for mental
404
761520
1360
12:42
health issues. Someone who
405
762880
1520
12:44
delivers therapy is called a therapist.
406
764400
2640
12:47
We'll find out more about this
407
767040
1760
12:48
robot therapist in just a moment,
408
768800
2320
12:51
but first, Rob, I've got a question for
409
771120
2480
12:53
you about the scale of mental
410
773600
2080
12:55
health issues globally. So, roughly
411
775680
3280
12:58
how many people do you think
412
778960
1760
13:00
experience mental health issues
413
780720
2320
13:03
at some point during their lifetime?
414
783040
2880
13:05
Is it…
415
785920
800
13:06
a) one in ten people,
416
786720
2400
13:09
b) one in four, or
417
789120
2400
13:11
c) one in three?
418
791520
1600
13:13
I'll go for one in four, but I know
419
793120
1840
13:14
whichever answer is right – it's a
420
794960
2160
13:17
big issue. How might a
421
797120
1360
13:18
robot therapist help?
422
798480
1440
13:19
We're not talking about a robot
423
799920
1920
13:21
in the Star Wars sense – so, there's
424
801840
2240
13:24
no flashing lights and mechanical
425
804080
1680
13:25
arms, Rob! It's actually an app in
426
805760
2400
13:28
your smartphone that talks to you –
427
808160
2880
13:31
and it's called Woebot.
428
811040
3120
13:34
So – it has a sense of humour.
429
814160
2000
13:36
'Woe' means 'sadness'; so this
430
816160
2160
13:38
is a 'woe' bot, not a robot.
431
818320
2640
13:40
And it was developed by
432
820960
1520
13:42
psychologist Dr Alison Darcy
433
822480
2480
13:44
from Stanford University in the
434
824960
1840
13:46
US. Here she is, talking to the BBC
435
826800
2880
13:49
radio programme All in the Mind.
436
829680
2480
13:52
Well, after you start an initial
437
832160
2240
13:54
conversation with the Woebot, and
438
834400
960
13:55
he'll take you through sort of what
439
835360
1120
13:56
he can do and what he can't do, he'll
440
836480
1680
13:58
just essentially check in with you
441
838160
1120
13:59
every day and just give you a sort
442
839280
1840
14:01
of figurative tap on the shoulder
443
841120
1440
14:02
and say: "Hey Claudia, how are
444
842560
1600
14:04
you doing? What's going on in your
445
844160
1600
14:05
day? How do you feel?" So if you
446
845760
1520
14:07
say, like "I'm really, really stressed out",
447
847280
2320
14:09
Woebot might offer to help talk
448
849600
2640
14:12
you through something.
449
852240
1488
14:14
Woebot checks in with you every
450
854101
2139
14:16
day and asks you how you are.
451
856240
2080
14:18
So here, to check in with someone
452
858320
2240
14:20
doesn't mean to register at a hotel
453
860560
1760
14:22
with that person! It's an informal
454
862320
2000
14:24
way of saying you talk to someone
455
864320
1760
14:26
in order to report or find out information.
456
866080
2640
14:28
And this usage is more common
457
868720
1760
14:30
in the United States. So, for
458
870480
1920
14:32
example: "I can't meet you today, Rob,
459
872400
2240
14:34
but I'll check in with you tomorrow
460
874640
1520
14:36
to see how the project is getting on."
461
876160
2160
14:38
So, this robot checks in with you
462
878320
2160
14:40
every day. It tracks your mood and
463
880480
2160
14:42
talks to you about your emotions, using
464
882640
2080
14:44
a technique called cognitive
465
884720
1760
14:46
behavioural therapy.
466
886480
1520
14:48
Cognitive behavioural therapy is a
467
888000
2800
14:50
common therapeutic technique that
468
890800
2160
14:52
helps people deal with problems by
469
892960
2320
14:55
changing the way they think.
470
895280
1680
14:56
That all sounds great, but does
471
896960
1600
14:58
Woebot actually work?
472
898560
1920
15:00
They've done some trials which
473
900480
1360
15:01
show it can be more effective than
474
901840
2160
15:04
simply reading information about
475
904000
1840
15:05
mental health. But they haven't
476
905840
2240
15:08
compared Woebot to a real therapist
477
908080
2880
15:10
due to ethical concerns.
478
910960
2080
15:13
Yes, it could be unethical to deny
479
913040
2160
15:15
a real patient access to a human
480
915200
2000
15:17
therapist for the sake of a trial.
481
917200
2400
15:19
'Ethical' basically means 'morally right'.
482
919600
2720
15:22
And another concern is privacy.
483
922320
2720
15:25
People who use apps like this are
484
925040
2240
15:27
not protected by strong privacy laws.
485
927280
2880
15:30
Despite these fears, digital therapy is
486
930160
2560
15:32
booming – and Woebot is just one of
487
932720
2160
15:34
an increasing number of electronic
488
934880
1840
15:36
services. One reason for this could
489
936720
2160
15:38
be using an app carries less stigma
490
938880
2240
15:41
than maybe seeing a human therapist.
491
941120
2240
15:43
And stigma refers to the negative
492
943360
2080
15:45
associations that people have
493
945440
1760
15:47
about something, especially when
494
947200
1760
15:48
these associations are not fair. Even
495
948960
3120
15:52
though mental health is now being
496
952080
1680
15:53
talked about more openly than before,
497
953760
2800
15:56
some people do still see mental
498
956560
2400
15:58
health issues and therapy negatively.
499
958960
2800
16:01
Whatever you think of robot therapy,
500
961760
2400
16:04
Dr Darcy believes that in the modern
501
964160
1920
16:06
world people need to self-reflect more –
502
966080
2720
16:08
which means thinking deeply about
503
968800
1760
16:10
yourself, in order to understand the
504
970560
1840
16:12
reasons behind your feelings.
505
972400
1920
16:14
The world that we live in right now
506
974320
1280
16:15
is very noisy. Particularly digitally. You
507
975600
2960
16:18
know, since we've had these little
508
978560
2080
16:20
computers in our pockets with us
509
980640
1840
16:22
everywhere we go, there aren't that
510
982480
1840
16:24
many opportunities for real silence
511
984320
2400
16:26
or self-reflection. You know, even a
512
986720
2080
16:28
commute on the tube might have
513
988800
1360
16:30
been a moment to just take a second
514
990160
2640
16:32
to yourself, but now that void can be
515
992800
2240
16:35
filled always with super-engaging
516
995040
2160
16:37
content by looking at your phone.
517
997200
2240
16:39
Darcy believes that we don't have
518
999440
1760
16:41
much time for self-reflection because
519
1001200
2080
16:43
there are so many distractions in
520
1003280
1680
16:44
life – especially smartphones!
521
1004960
2560
16:47
After discussing all this – would
522
1007520
1440
16:48
you actually try a therapy app
523
1008960
1920
16:50
like this?
524
1010880
960
16:51
Yes I would, actually – I think it
525
1011840
1280
16:53
might be quite helpful.
526
1013120
1360
16:54
And how about the question
527
1014480
960
16:55
you asked me at the beginning
528
1015440
880
16:56
of the programme: how many
529
1016320
1440
16:57
people experience mental
530
1017760
1680
16:59
health issues?
531
1019440
1040
17:00
The answer was: one in four,
532
1020480
2240
17:02
according the World Health
533
1022720
1520
17:04
Organisation and the World
534
1024240
1680
17:05
Federation for Mental Health.
535
1025920
2000
17:07
But the WHO say that as many
536
1027920
2400
17:10
as two-thirds of people never seek
537
1030320
2320
17:12
help from a health professional –
538
1032640
1760
17:14
with stigma being one of the
539
1034400
1920
17:16
main reasons.
540
1036320
1280
17:17
And just there we had stigma
541
1037600
1920
17:19
again; let's now run through the
542
1039520
1680
17:21
other words we learned today.
543
1041200
1520
17:22
So we had 'woe' - meaning 'sadness'.
544
1042720
3120
17:25
I'm full of woe. Woe is me!
545
1045840
3440
17:29
Maybe you need some therapy –
546
1049280
1600
17:30
that's the process of receiving
547
1050880
1520
17:32
treatment for a particular health
548
1052400
1760
17:34
issue, especially mental
549
1054160
1440
17:35
health illness.
550
1055600
880
17:36
And we had – to check in
551
1056480
1600
17:38
with someone. After we finish
552
1058080
1840
17:39
this programme, I need to check
553
1059920
1680
17:41
in with the boss about my
554
1061600
1440
17:43
new project.
555
1063040
880
17:43
We also had self-reflection – that's
556
1063920
2160
17:46
the process of thinking deeply
557
1066080
1520
17:47
about yourself.
558
1067600
1600
17:49
And finally, we had 'ethical'. If you
559
1069200
2880
17:52
describe something as ethical, you
560
1072080
2080
17:54
mean it's morally right.
561
1074160
1840
17:56
So woe, stigma, therapy, check
562
1076000
3040
17:59
in with, self-reflection and ethical.
563
1079040
2960
18:02
That's it for this edition of
564
1082000
1120
18:03
6 Minute English. We'll leave you to
565
1083120
2000
18:05
self-reflect – and after you've done
566
1085120
1760
18:06
that, do visit our Facebook, Twitter,
567
1086880
2160
18:09
Instagram and YouTube pages, and
568
1089040
2080
18:11
of course our website!
569
1091120
1120
18:12
Bye for now.
570
1092240
800
18:13
Bye bye!
571
1093040
800
18:19
Hello. Welcome to 6 Minute
572
1099440
1680
18:21
English. I'm Rob.
573
1101120
1040
18:22
And I'm Neil.
574
1102160
1166
18:23
Now, Neil, can you remember the first time
575
1103326
2274
18:25
you ever used the World Wide Web
576
1105600
2400
18:28
or as we often call it, the internet, and
577
1108000
2480
18:30
what you used it for?
578
1110480
1280
18:31
Oh, that's a good question. I do
579
1111760
2000
18:33
remember. And... well, nothing really
580
1113760
3360
18:37
changes does it? Because I looked up
581
1117120
1920
18:39
pictures of cats!
582
1119040
1422
18:40
Cats! Very useful, anyway do you
583
1120462
3298
18:43
think the internet has generally
584
1123760
1200
18:44
been positive or negative for
585
1124960
1760
18:46
the world?
586
1126720
800
18:47
Wow! Now, that's a big question. A huge
587
1127520
3200
18:50
question. I don't know if I can
588
1130720
1789
18:52
answer that.
589
1132509
560
18:53
Well, one person who perhaps
590
1133069
2211
18:55
can answer it, is the man who
591
1135280
1760
18:57
invented it: British computer scientist
592
1137040
2560
18:59
Tim Berners-Lee. We'll find out what
593
1139600
2240
19:01
he thinks has become of his 'child'
594
1141840
1840
19:03
shortly but before that, a question
595
1143680
2240
19:05
for you all. When did Berners-Lee
596
1145920
2320
19:08
first suggest the idea for what
597
1148240
1760
19:10
would become the World Wide Web?
598
1150000
2560
19:12
Was it in...
599
1152560
560
19:13
a) 1985,
600
1153120
1680
19:14
b) 1989,
601
1154800
1494
19:16
or, c) 1991?
602
1156294
2585
19:18
Tricky, but I think it's earlier than
603
1158879
4081
19:22
people think so I'm going to go
604
1162960
2240
19:25
for 1985.
605
1165200
2160
19:27
Well, that was a long time ago but
606
1167360
1520
19:28
we'll reveal the answer a little later
607
1168880
1840
19:30
in the programme. I think it's true to
608
1170720
2320
19:33
say that the internet has been one of,
609
1173040
2240
19:35
if not the most important
610
1175280
1520
19:36
technological developments perhaps
611
1176800
2080
19:38
of all time. Would you agree, Neil?
612
1178880
1840
19:40
Well, it's hard to imagine living
613
1180720
1520
19:42
without it. Not impossible, but not
614
1182240
2560
19:44
nearly as convenient.
615
1184800
1520
19:46
These days we take the internet
616
1186320
1680
19:48
for granted. We share our lives on
617
1188000
2000
19:50
social media and not just with
618
1190000
1920
19:51
friends and family. And that isn't
619
1191920
2080
19:54
always a positive thing, according
620
1194000
2080
19:56
to the father of the internet,
621
1196080
1680
19:57
Tim Berners-Lee. In a recent
622
1197760
2080
19:59
BBC Tech Tent programme, he
623
1199840
2240
20:02
talked about his concerns with
624
1202080
1440
20:03
the internet and particularly the
625
1203520
1840
20:05
companies that control its information.
626
1205360
2480
20:07
Companies which he calls
627
1207840
1760
20:09
'internet giants'. What does he say
628
1209600
2240
20:11
he thought these companies
629
1211840
1280
20:13
had to do?
630
1213120
1280
20:14
Initially, I felt the main thing an
631
1214400
2720
20:17
internet giant had to do was
632
1217120
1520
20:18
just to be neutral, just be a platform
633
1218640
2320
20:20
and humanity, once connected by
634
1220960
1920
20:22
technology, will do wonderful
635
1222880
1200
20:24
things. And, clearly, it doesn't
636
1224080
2400
20:26
work like that. If you connect
637
1226480
1520
20:28
humanity via Wikipedia then
638
1228000
3040
20:31
they do produce, in general,
639
1231040
1360
20:32
wonderful things. If you connect
640
1232400
1920
20:34
people by social network where
641
1234320
3040
20:37
they have anonymity, then it can
642
1237360
1840
20:39
bring out the very nastiest of people.
643
1239200
1840
20:41
So, what did he say he thought these
644
1241040
1680
20:42
internet giants had to do?
645
1242720
1840
20:44
He said that he thought initially, that
646
1244560
2080
20:46
they just had to be neutral. 'Initially'
647
1246640
2560
20:49
means 'at first', 'in the beginning' and
648
1249200
2400
20:51
it also suggests that later he
649
1251600
1840
20:53
changed his mind. Anyway, he said
650
1253440
2800
20:56
that he thought they just had to
651
1256240
1440
20:57
be neutral. 'Neutral' here means
652
1257680
2160
20:59
that they didn't need to do anything,
653
1259840
1600
21:01
they didn't need to control the
654
1261440
1200
21:02
internet or information. He thought
655
1262640
1920
21:04
it would be a tool to connect
656
1264560
1440
21:06
people and ideas and information
657
1266000
1840
21:07
and it would be wonderful.
658
1267840
1486
21:09
But it's not all good, is it?
659
1269326
1874
21:11
No. He does say that giving
660
1271200
1840
21:13
people access to sources of
661
1273040
1440
21:14
information is generally a good
662
1274480
1680
21:16
thing but that when it comes to
663
1276160
1680
21:17
social networks, social media,
664
1277840
2160
21:20
people have anonymity.
665
1280000
2000
21:22
Anonymity?
666
1282000
960
21:22
Yes. It means that on the internet
667
1282960
2000
21:24
people can hide their true identity
668
1284960
2000
21:26
or personality. Some people write
669
1286960
2320
21:29
things that they would never say to
670
1289280
1520
21:30
someone in person because they
671
1290800
1520
21:32
think there will be no consequences.
672
1292320
2119
21:34
Berners-Lee says anonymity can
673
1294439
2121
21:36
bring out the nastiest side of people.
674
1296560
2640
21:39
People saying horrible and terrible
675
1299200
1760
21:40
things to each other.
676
1300960
1200
21:42
Berners-Lee does have some
677
1302160
1600
21:43
suggestions for how this could
678
1303760
1600
21:45
be changed. And it's based on the
679
1305360
1920
21:47
idea of likes and shares, which he
680
1307280
2480
21:49
calls kudos. What's his suggestion?
681
1309760
2880
21:52
The different social networks and
682
1312640
1040
21:53
different platforms are in different
683
1313680
1840
21:55
situations and, in some cases, they
684
1315520
2320
21:57
have acknowledged there is an issue.
685
1317840
2080
21:59
I think they realise that the issue
686
1319920
2400
22:02
could be hugely ameliorated by
687
1322320
2320
22:04
tweaking the way the thing works
688
1324640
2080
22:06
by changing the way retweets are
689
1326720
2400
22:09
propagated or changing the way
690
1329120
2400
22:11
people get kudos - give them more
691
1331520
2000
22:13
kudos for being constructive,
692
1333520
1680
22:15
for example.
693
1335200
800
22:16
So, how does he think companies
694
1336000
1600
22:17
could address the problem?
695
1337600
1600
22:19
Well, he says that some of the
696
1339200
1440
22:20
social networks have agreed
697
1340640
1280
22:21
that there is a problem and they
698
1341920
1600
22:23
know what could improve it.
699
1343520
1760
22:25
He didn't use the word 'improve'
700
1345280
1520
22:26
though, did he?
701
1346800
960
22:27
No, he actually used the rather
702
1347760
1840
22:29
formal verb 'ameliorate', which
703
1349600
2000
22:31
means 'to improve or make
704
1351600
1280
22:32
something better'.
705
1352880
1280
22:34
So, how does he suggest the
706
1354160
1360
22:35
problem could be ameliorated?
707
1355520
2240
22:37
By tweaking the way in which
708
1357760
1520
22:39
people give or receive kudos.
709
1359280
2320
22:41
'Tweaking' means 'making a small
710
1361600
1760
22:43
change to the way something
711
1363360
1200
22:44
works'. Much of what happens on
712
1364560
1920
22:46
the internet is driven by our desire
713
1366480
2000
22:48
to get likes and shares – this is
714
1368480
2240
22:50
the kudos that Berners-Lee talks
715
1370720
1920
22:52
about. He feels that tweaking this
716
1372640
2160
22:54
could lead to a better experience.
717
1374800
2080
22:56
For example, getting more kudos
718
1376880
1760
22:58
for constructive or positive actions.
719
1378640
1920
23:00
Mmm, interesting – but I wonder
720
1380560
2480
23:03
who would decide if something
721
1383040
1280
23:04
is constructive?
722
1384320
1280
23:05
Well, that's another big question
723
1385600
1840
23:07
for another day, I guess.
724
1387440
1520
23:08
For now though, let's have the
725
1388960
1440
23:10
answer to our small question. In
726
1390400
2080
23:12
what year did Berners-Lee present
727
1392480
1840
23:14
the idea for what would become the
728
1394320
2240
23:16
World Wide Web? The options were
729
1396560
2560
23:19
a) 1985, b)1989, or c) 1991.
730
1399120
4960
23:24
It was, in fact, 1989.
731
1404080
2713
23:26
Now, before we go,
732
1406793
1207
23:28
let's have a quick recap of
733
1408000
1766
23:29
today's vocabulary.
734
1409766
1354
23:31
'Initially' – means 'at first - in
735
1411120
2240
23:33
the beginning'. Then we
736
1413360
1307
23:34
had 'neutral'.
737
1414667
1514
23:36
In this case, it meant 'not controlling'
738
1416181
2219
23:38
or 'not taking any action to control'.
739
1418400
2880
23:41
Then, there was the noun 'anonymity'
740
1421280
2320
23:43
which is the state of having a
741
1423600
1520
23:45
hidden identity or personality.
742
1425120
2160
23:47
Next, to 'ameliorate' a situation is
743
1427280
2640
23:49
to make it better.
744
1429920
1200
23:51
To 'tweak' something is to 'make
745
1431120
1680
23:52
a small change to the way
746
1432800
1680
23:54
something works'.
747
1434480
1120
23:55
And then we had, of course, 'kudos'.
748
1435600
1680
23:57
'Kudos' is praise and appreciation
749
1437280
1920
23:59
for something you've done.
750
1439200
1120
24:00
Well, kudos to you, Rob, for
751
1440320
1520
24:01
today's programme.
752
1441840
800
24:02
Thank you very much.
753
1442640
1040
24:03
Well, thank you, Neil, and
754
1443680
1200
24:04
thank you everyone for listening.
755
1444880
1280
24:06
That's all we have today, but you
756
1446160
2400
24:08
can, find us on Facebook, Twitter,
757
1448560
1680
24:10
Instagram and YouTube and, of
758
1450240
1440
24:11
course, our website
759
1451680
1280
24:12
bbclearningenglish.com!
760
1452960
2400
24:15
Bye for now.
761
1455360
720
24:16
Thanks for joining us
762
1456080
800
24:16
and goodbye.
763
1456880
960
24:23
Hello. This is 6 Minute English
764
1463440
2160
24:25
and I'm Rob.
765
1465600
800
24:26
And I'm Neil.
766
1466400
1040
24:27
Today we’re talking about buttons.
767
1467440
2720
24:30
Yes, buttons. Buttons are what
768
1470160
2320
24:32
we have on our clothes to fasten
769
1472480
1680
24:34
them but the word is also used
770
1474160
1840
24:36
for things that we push to make
771
1476000
1760
24:37
things happen.
772
1477760
960
24:38
Things like your bedside alarm,
773
1478720
2160
24:40
radio, toaster, kettle. We press
774
1480880
2560
24:43
hundreds of buttons every week
775
1483440
1680
24:45
without thinking about it. Not everyone
776
1485120
2400
24:47
likes buttons though, particularly
777
1487520
2080
24:49
the ones we have on our clothes.
778
1489600
2480
24:52
It’s a recognised phobia.
779
1492080
1760
24:53
What is this fear called? Is it:
780
1493840
2640
24:56
A: buttonophobia,
781
1496480
2278
24:58
B: koumpounophobia,
782
1498758
3036
25:01
or, C: coulrophobia?
783
1501794
2366
25:04
Any ideas Neil?
784
1504160
1280
25:05
Er – I think I’ve got a fear of
785
1505440
1440
25:06
pronouncing these words!
786
1506880
1622
25:08
I've got no idea what the answer is, I
787
1508502
2378
25:10
think buttonophobia is much
788
1510880
1520
25:12
too obvious – so it’s one of the
789
1512400
1760
25:14
others – or is it? It’s a hard one.
790
1514160
3093
25:17
Well, I'll have the answer later in
791
1517253
2027
25:19
the programme. Thinking Allowed
792
1519280
2320
25:21
is a BBC Radio 4 programme
793
1521600
1680
25:23
which covers a range of interesting
794
1523280
1760
25:25
topics. Recently, they featured a
795
1525040
2080
25:27
discussion about buttons and
796
1527120
1840
25:28
how important they are to
797
1528960
1280
25:30
everyday life. Steven Connor,
798
1530240
2320
25:32
Professor of English at the
799
1532560
1440
25:34
University of Cambridge, was on
800
1534000
1600
25:35
the programme and he talked
801
1535600
1440
25:37
about appeal of buttons. What does
802
1537040
2960
25:40
say about children and buttons?
803
1540000
2160
25:42
We do love buttons and I think the
804
1542160
2080
25:44
pleasure and the temptation of
805
1544240
2240
25:46
buttons… this temptation and
806
1546480
1920
25:48
everyone feels it, which is why
807
1548400
2320
25:50
buttons have to be very convenient.
808
1550720
1920
25:52
On the other hand, they have to
809
1552640
1040
25:53
be kept away from children, so
810
1553680
1280
25:54
they've got to be put high up
811
1554960
1760
25:56
on the wall, and buttons that
812
1556720
2320
25:59
really matter have to be made
813
1559040
1600
26:00
quite hard to push – like put
814
1560640
1520
26:02
behind glass or something.
815
1562160
1909
26:04
OK. What does he say about
816
1564069
1771
26:05
children and buttons?
817
1565840
1360
26:07
That they have to be kept away
818
1567200
1680
26:08
from them! As I said before, I
819
1568880
2080
26:10
loved pressing buttons as a child.
820
1570960
1840
26:12
I would press any that I saw, so
821
1572800
1920
26:14
important ones did have to be
822
1574720
1760
26:16
out of my reach.
823
1576480
960
26:17
Yes, he said buttons are a
824
1577440
2240
26:19
temptation. A temptation is
825
1579680
2160
26:21
something that makes you want
826
1581840
1280
26:23
to do something and it’s often
827
1583120
1840
26:24
used when it’s something you
828
1584960
1120
26:26
shouldn’t really do. So, buttons
829
1586080
2160
26:28
that lead to potentially dangerous
830
1588240
2000
26:30
or serious consequences, like a
831
1590240
2240
26:32
fire alarm, need to be protected,
832
1592480
2320
26:34
maybe behind glass, so temptation
833
1594800
2400
26:37
doesn’t get the better of us.
834
1597200
1600
26:38
But many buttons have a useful,
835
1598800
1840
26:40
practical purpose in everyday life,
836
1600640
1840
26:42
like calling a lift - so, these buttons
837
1602480
2080
26:44
have to be easy to use without
838
1604560
1680
26:46
difficulty. The adjective for this
839
1606240
2160
26:48
is 'convenient'. These everyday
840
1608400
2320
26:50
buttons have to be convenient.
841
1610720
2000
26:52
Professor Connor goes on to
842
1612720
1600
26:54
say a bit more about why
843
1614320
1680
26:56
buttons are so appealing.
844
1616000
2240
26:58
What’s his opinion?
845
1618240
1440
26:59
You know what I think? I think
846
1619680
1040
27:00
it's down to the fidgeting
847
1620720
2240
27:02
instinct of very digital or
848
1622960
3280
27:06
manual creatures. I think we
849
1626240
1680
27:07
want to fidget with things and
850
1627920
2080
27:10
adjust them, we want to make
851
1630000
1280
27:11
them slightly better… I guess it's
852
1631280
2320
27:13
the grooming instinct in apes.
853
1633600
2000
27:15
So then, why can’t we resist buttons?
854
1635600
2320
27:17
Because as humans we have a
855
1637920
2240
27:20
fidgeting instinct. We can’t stay still
856
1640160
2800
27:22
for very long, we need to move
857
1642960
1440
27:24
around a lot because we are
858
1644400
1520
27:25
very digital creatures.
859
1645920
1840
27:27
The use of digital though, is nothing
860
1647760
1920
27:29
to do with modern online
861
1649680
1600
27:31
technology, is it?
862
1651280
1200
27:32
No, a 'digit' is a finger or toe. So, we
863
1652480
2720
27:35
are digital creatures – we have
864
1655200
1840
27:37
fingers and we like to use them.
865
1657040
2160
27:39
And one thing other digital
866
1659200
1520
27:40
creatures do, creatures like apes,
867
1660720
1840
27:42
is grooming. That is they use their
868
1662560
2480
27:45
hands to clean the body hair of
869
1665040
2080
27:47
other apes. They look through
870
1667120
1520
27:48
the hair for insects and bugs and
871
1668640
1840
27:50
pull them out and eat them.
872
1670480
1680
27:52
But we can also use the word
873
1672160
1360
27:53
grooming for humans, someone
874
1673520
2160
27:55
who is well-groomed for example
875
1675680
1760
27:57
is neat and tidy, clean and well
876
1677440
2240
27:59
presented. Here’s Professor
877
1679680
1360
28:01
Connor again.
878
1681040
880
28:01
You know what I think? I think it's
879
1681920
1440
28:03
down to the fidgeting instinct of
880
1683360
3040
28:06
very digital or manual creatures.
881
1686400
3040
28:09
I think we want to fidget with things
882
1689440
2560
28:12
and adjust them, we want to make
883
1692000
1440
28:13
them slightly better… I guess it's
884
1693440
2400
28:15
the grooming instinct in apes.
885
1695840
2240
28:18
Before we wrap up, time to get the
886
1698080
1920
28:20
answer to this week’s question.
887
1700000
2160
28:22
Some people have a fear of
888
1702160
1360
28:23
buttons, it’s a recognised phobia, but
889
1703520
2560
28:26
what’s it called? Is it
890
1706080
1920
28:28
A: buttonophobia
891
1708000
2000
28:30
B: koumpounophobia, or
892
1710000
2400
28:32
C: coulrophobia?
893
1712400
1680
28:34
And, Neil, you said?
894
1714080
1399
28:35
Well, I didn’t, but I don’t think it
895
1715479
2041
28:37
can be ‘buttonophobia’, that would be
896
1717520
1840
28:39
too easy. And I think coulrophobia
897
1719360
2960
28:42
is a fear of clowns, so I’m going
898
1722320
4000
28:46
for the other one -
899
1726320
2800
28:49
koumpounophobia.
900
1729120
1600
28:50
That it right. Buttonophobia is a
901
1730720
2320
28:53
made up word, and as you said,
902
1733040
1920
28:54
coulrophobia is a fear of clowns.
903
1734960
2800
28:57
Right, now let’s review
904
1737760
1280
28:59
today’s vocabulary.
905
1739040
1120
29:00
We’ve been talking about buttons.
906
1740160
1920
29:02
These can be small round things
907
1742080
1520
29:03
we use to fasten our clothes, or
908
1743600
2000
29:05
the things that we push to make
909
1745600
1840
29:07
something happen.
910
1747440
1120
29:08
Buttons can be a temptation.
911
1748560
2080
29:10
We see one, we want to push it.
912
1750640
1920
29:12
So, a 'temptation' is something
913
1752560
1680
29:14
that makes us want to do
914
1754240
1120
29:15
something we know we shouldn’t.
915
1755360
2000
29:17
And then we had the adjective
916
1757360
1840
29:19
'convenient'. Something that is
917
1759200
1440
29:20
convenient is easy to use without
918
1760640
1920
29:22
difficulty. For example, the buttons
919
1762560
2080
29:24
to call a lift are at a very convenient
920
1764640
2240
29:26
height, they can be reached easily.
921
1766880
2480
29:29
Professor Connor went on to talk
922
1769360
1680
29:31
about our fidgeting instinct.
923
1771040
2640
29:33
As humans we love to fidget, we like
924
1773680
2320
29:36
to keep moving around, we can’t
925
1776000
1520
29:37
stay still for very long and we love
926
1777520
1840
29:39
to do stuff with our hands.
927
1779360
1600
29:40
The professor talked about us being
928
1780960
2000
29:42
digital creatures, which means
929
1782960
1520
29:44
creatures with fingers – a digit is
930
1784480
2240
29:46
another word for a finger or toe.
931
1786720
2240
29:48
And finally, we had grooming.
932
1788960
2240
29:51
This is the habit of making ourselves
933
1791200
1840
29:53
look nice by cleaning, washing and
934
1793040
2000
29:55
doing our hair. It’s something some
935
1795040
1840
29:56
animals, such as apes, do for
936
1796880
1600
29:58
each other.
937
1798480
960
29:59
Well, the button here in the studio
938
1799440
1760
30:01
is flashing, which tells me it’s time
939
1801200
1680
30:02
to wrap up for today. Do join us
940
1802880
2000
30:04
next time and if you can’t wait, you
941
1804880
1680
30:06
can always find us on Instagram,
942
1806560
1840
30:08
Facebook, Twitter, YouTube our app
943
1808400
2400
30:10
and, of course, on our website
944
1810800
1760
30:12
bbclearningenglish.com.
945
1812560
1840
30:14
Bye bye for now.

946
1814400
1120
30:15
Bye!
947
1815520
937
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7