Are computers always right? 6 Minute English

255,743 views ・ 2018-07-19

BBC Learning English


Please double-click on the English subtitles below to play the video.

00:06
Catherine: Hello and welcome to 6 Minute
0
6480
2280
00:08
English. I'm Catherine.
1
8760
1420
00:10
Rob: And hello, I'm Rob.
2
10180
1400
00:11
Catherine: Today we have another
3
11580
1389
00:12
technology topic.
4
12969
761
00:13
Rob: Oh good! I love technology. It makes
5
13730
2470
00:16
things easier, it's fast and means I can
6
16240
2160
00:18
have gadgets.
7
18400
1400
00:19
Catherine: Do you think that technology
8
19800
1760
00:21
can actually do things better than humans?
9
21560
2200
00:23
Rob: For some things, yes. I think cars
10
23760
2640
00:26
that drive themselves will be safer than
11
26400
1600
00:28
humans but that will take away some of
12
28000
3180
00:31
the pleasure of driving. So I guess it
13
31180
2180
00:33
depends on what you mean by better.
14
33360
1940
00:35
Catherine: Good point, Rob. And that
15
35300
1680
00:36
actually ties in very closely with today's
16
36980
2420
00:39
topic which is technochauvinism.
17
39400
3200
00:42
Rob: What's that?
18
42600
1480
00:44
Catherine: We'll find out shortly, Rob, but
19
44080
1520
00:45
before we do, today's quiz question.
20
45600
3200
00:48
Artificial Intelligence, or A.I., is an area of
21
48800
3260
00:52
computer science that develops the
22
52060
2040
00:54
ability of computers to learn to do things
23
54100
2540
00:56
like solve problems or drive cars without
24
56640
2495
00:59
crashing. But in what decade was the
25
59140
2800
01:01
term 'Artificial Intelligence'
26
61940
2640
01:04
coined? Was it: a) the 1940s, b) the 1950s
27
64589
4901
01:09
or c) the 1960s?
28
69490
2070
01:11
Rob: I think it's quite a new expression so
29
71560
3000
01:14
I'll go for c) the 1960s.
30
74560
2380
01:16
Catherine: Good luck with that, Rob, and
31
76940
1420
01:18
we'll give you the answer later in the
32
78360
1300
01:19
programme. Now, let's get back to our
33
79660
2940
01:22
topic of technochauvinism.
34
82600
2040
01:24
Rob: I know what a chauvinist is. It's
35
84640
1920
01:26
someone who thinks that their country or
36
86580
2380
01:28
race or sex is better than others. But how
37
88960
3140
01:32
does this relate to technology?
38
92100
1340
01:33
Catherine: We're about to find out.
39
93440
1891
01:35
Meredith Broussard is Professor of
40
95340
1984
01:37
Journalism at New York University and
41
97324
2236
01:39
she's written a book called
42
99560
1660
01:41
Artificial Unintelligence. She appeared on
43
101220
3420
01:44
the BBC Radio 4 programme More or Less
44
104640
2580
01:47
to talk about it. Listen carefully and find
45
107220
2620
01:49
out her definition of technochauvinism.
46
109840
3320
01:53
Meredith Broussard: Technochauvinism is
47
113160
1811
01:54
the idea that technology is always the
48
114980
2540
01:57
highest and best solution. So somehow
49
117520
3120
02:00
over the past couple of decades we got
50
120640
2260
02:02
into the habit of
51
122900
1360
02:04
thinking that doing something with a
52
124270
3410
02:07
computer is always the best and most
53
127680
3480
02:11
objective way to do something and
54
131180
2060
02:13
that's simply not true.
55
133240
1540
02:14
Computers are not objective, they are
56
134780
3000
02:17
proxies for the people who make them.
57
137780
2300
02:20
Catherine: What is Meredith Broussard's
58
140080
2200
02:22
definition of technochauvinism?
59
142300
2280
02:24
Rob: It's this idea that using technology
60
144580
2250
02:26
is better than not using technology.
61
146830
2450
02:29
Catherine: She says that we have this idea
62
149280
2360
02:31
that a computer is objective. Something
63
151640
2800
02:34
that is objective is neutral, it doesn't have
64
154440
2500
02:36
an opinion, it's fair and it's unbiased - so
65
156940
3160
02:40
it's the opposite of being a chauvinist. But
66
160100
3080
02:43
Meredith Broussard says this is not true.
67
163180
2860
02:46
Rob: She argues that computers are not
68
166040
2635
02:48
objective. They are proxies for the people
69
168680
2880
02:51
that make them. You might know the
70
171560
1800
02:53
word proxy when you are using your
71
173360
1735
02:55
computer in one country and want to look
72
175095
2235
02:57
at something that is only available
73
177330
1871
02:59
in a different country. You can use a piece
74
179201
2298
03:01
of software called a proxy to do that.
75
181500
2640
03:04
Catherine: But a proxy is also a person or
76
184140
2840
03:06
a thing that carries out your wishes and
77
186980
2660
03:09
your instructions for you.
78
189640
2000
03:11
So computers are only as smart or as
79
191640
2860
03:14
objective as the people that
80
194500
2560
03:17
programme them. Computers are proxies
81
197060
3060
03:20
for their programmers. Broussard says
82
200120
2620
03:22
that believing too much in Artificial
83
202740
2400
03:25
Intelligence can make the world worse.
84
205140
2820
03:27
Let's hear a bit more. This time find out
85
207960
2920
03:30
what serious problems in society
86
210880
2820
03:33
does she think may be reflected in AI?
87
213700
3900
03:37
Meredith Broussard: It's a nuanced
88
217600
1400
03:39
problem. What we have is data on the
89
219000
2700
03:41
world as it is and we have serious
90
221700
3100
03:44
problems with racism, sexism, classism,
91
224800
2920
03:47
ageism, in the world right now so there is
92
227720
2300
03:50
no such thing as perfect data. We also
93
230020
3760
03:53
have a problem inside the tech world
94
233780
2379
03:56
where the creators of algorithms do not
95
236160
3080
03:59
have sufficient awareness of social
96
239240
3960
04:03
issues such that they can make good
97
243200
2680
04:05
technology that gets us closer to a world
98
245880
3800
04:09
as it should be.
99
249680
1660
04:11
Rob: She said that society has problems
100
251340
2640
04:13
with racism, sexism, classism and ageism.
101
253984
3296
04:17
Catherine: And she says it's a nuanced
102
257280
2531
04:19
problem. A nuanced problem is not
103
259820
2231
04:22
simple, but it does have small and
104
262060
1960
04:24
important areas which may be
105
264020
2200
04:26
hard to spot, but they need to be considered.
106
266220
3020
04:29
Rob: And she also talked about
107
269240
1800
04:31
algorithms used to program these
108
271053
2220
04:33
technological systems.
109
273273
1527
04:34
An algorithm is a set of instructions that
110
274800
2497
04:37
computers use to perform their tasks.
111
277300
2240
04:39
Essentially it's the rules that they use
112
279540
1680
04:41
to come up with their answers and
113
281220
2220
04:43
Broussard believes that technology will
114
283440
2560
04:46
reflect the views of those who create the algorithms.
115
286000
2680
04:48
Catherine: Next time you're using a piece
116
288680
1640
04:50
of software or your favourite app, you
117
290320
2120
04:52
might find yourself wondering if it's a
118
292440
2040
04:54
useful tool or does it contain these little
119
294480
3020
04:57
nuances that
120
297500
1100
04:58
reflect the views of the developer.
121
298600
2220
05:00
Rob: Right, Catherine. How about the
122
300830
1735
05:02
answer to this week's question then?
123
302565
1735
05:04
Catherine: I asked in which decade was
124
304300
2160
05:06
the term 'Artificial Intelligence' coined.
125
306460
2300
05:08
Was it the 40s, the 50s or the 60s?
126
308760
2740
05:11
Rob: And I said the 60s.
127
311500
1060
05:12
Catherine: But it was actually the 1950s.
128
312560
2500
05:15
Never mind, Rob. Let's review today's vocabulary.
129
315060
3520
05:18
Rob: Well, we had a chauvinist - that's
130
318580
2100
05:20
someone who believes their country, race
131
320700
2460
05:23
or sex is better than any others.
132
323160
2000
05:25
Catherine: And this gives us
133
325160
1540
05:26
technochauvinism,
134
326700
1380
05:28
the belief that a technological solution is
135
328080
2809
05:30
always a better solution to a problem.
136
330889
2180
05:33
Rob: Next - someone or something that is
137
333069
2776
05:35
objective is neutral, fair and balanced.
138
335845
2775
05:38
Catherine: A proxy is a piece of software
139
338620
2760
05:41
but also someone who does something
140
341380
1660
05:43
for you, on your behalf.
141
343040
2340
05:45
A nuanced problem is a subtle
142
345380
2220
05:47
one, it's not a simple case of right or
143
347600
3000
05:50
wrong, in a nuanced problem there are
144
350600
2740
05:53
small but important things that you need
145
353340
2140
05:55
to consider.
146
355480
800
05:56
Rob: And an algorithm is a set of software
147
356280
2560
05:58
instructions for a computer system.
148
358842
2128
06:00
Catherine: Well, that's all we have time for
149
360970
1430
06:02
today. Goodbye for now.
150
362400
1780
06:04
Rob: Bye bye!
151
364180
700
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7