6 big ethical questions about the future of AI | Genevieve Bell

93,420 views ・ 2021-01-14

TED


Please double-click on the English subtitles below to play the video.

00:13
Let me tell you a story about artificial intelligence.
0
13286
2999
00:16
There's a building in Sydney at 1 Bligh Street.
1
16309
3207
00:19
It houses lots of government apartments
2
19540
2230
00:21
and busy people.
3
21794
1449
00:23
From the outside, it looks like something out of American science fiction:
4
23267
3524
00:26
all gleaming glass and curved lines,
5
26815
2714
00:29
and a piece of orange sculpture.
6
29553
2167
00:31
On the inside, it has excellent coffee on the ground floor
7
31744
2928
00:34
and my favorite lifts in Sydney.
8
34696
1651
00:36
They're beautiful;
9
36371
1301
00:37
they look almost alive.
10
37696
2120
00:40
And it turns out I'm fascinated with lifts.
11
40220
2088
00:42
For lots of reasons.
12
42332
1364
00:43
But because lifts are one of the places you can see the future.
13
43720
3143
00:46
In the 21st century, lifts are interesting
14
46887
2492
00:49
because they're one of the first places that AI will touch you
15
49403
2912
00:52
without you even knowing it happened.
16
52339
2039
00:54
In many buildings all around the world,
17
54919
2565
00:57
the lifts are running a set of algorithms.
18
57508
2491
01:00
A form of protoartificial intelligence.
19
60023
2667
01:03
That means before you even walk up to the lift to press the button,
20
63071
4284
01:07
it's anticipated you being there.
21
67379
1775
01:09
It's already rearranging all the carriages.
22
69490
2936
01:12
Always going down, to save energy,
23
72450
1627
01:14
and to know where the traffic is going to be.
24
74101
2143
01:16
By the time you've actually pressed the button,
25
76268
2262
01:18
you're already part of an entire system
26
78554
1872
01:20
that's making sense of people and the environment
27
80450
2310
01:22
and the building and the built world.
28
82784
1991
01:25
I know when we talk about AI, we often talk about a world of robots.
29
85481
3810
01:29
It's easy for our imaginations to be occupied with science fiction,
30
89720
3468
01:33
well, over the last 100 years.
31
93212
1833
01:35
I say AI and you think "The Terminator."
32
95069
2534
01:38
Somewhere, for us, making the connection between AI and the built world,
33
98728
4532
01:43
that's a harder story to tell.
34
103284
1690
01:45
But the reality is AI is already everywhere around us.
35
105395
3727
01:49
And in many places.
36
109434
1215
01:50
It's in buildings and in systems.
37
110673
2587
01:53
More than 200 years of industrialization
38
113284
2063
01:55
suggest that AI will find its way to systems-level scale relatively easily.
39
115371
4303
02:00
After all, one telling of that history
40
120173
2054
02:02
suggests that all you have to do is find a technology,
41
122251
2540
02:04
achieve scale and revolution will follow.
42
124815
2426
02:07
The story of mechanization, automation and digitization
43
127919
4277
02:12
all point to the role of technology and its importance.
44
132220
3182
02:16
Those stories of technological transformation
45
136274
2461
02:18
make scale seem, well, normal.
46
138759
2714
02:21
Or expected.
47
141497
1150
02:23
And stable.
48
143052
1341
02:24
And sometimes even predictable.
49
144417
2066
02:27
But it also puts the focus squarely on technology and technology change.
50
147012
4023
02:31
But I believe that scaling a technology and building a system
51
151964
3483
02:35
requires something more.
52
155471
1680
02:38
We founded the 3Ai Institute at the Australian National University
53
158260
3762
02:42
in September 2017.
54
162046
1729
02:44
It has one deceptively simple mission:
55
164101
2794
02:46
to establish a new branch of engineering
56
166919
1911
02:48
to take AI safely, sustainably and responsibly to scale.
57
168854
4198
02:53
But how do you build a new branch of engineering in the 21st century?
58
173076
3253
02:56
Well, we're teaching it into existence
59
176353
2389
02:58
through an experimental education program.
60
178766
2658
03:01
We're researching it into existence
61
181876
2019
03:03
with locations as diverse as Shakespeare's birthplace,
62
183919
2769
03:06
the Great Barrier Reef,
63
186712
1802
03:08
not to mention one of Australia's largest autonomous mines.
64
188538
2888
03:11
And we're theorizing it into existence,
65
191892
2509
03:14
paying attention to the complexities of cybernetic systems.
66
194425
3023
03:18
We're working to build something new and something useful.
67
198250
2954
03:21
Something to create the next generation of critical thinkers and critical doers.
68
201228
3776
03:25
And we're doing all of that
69
205736
1461
03:27
through a richer understanding of AI's many pasts and many stories.
70
207221
3973
03:31
And by working collaboratively and collectively
71
211964
3024
03:35
through teaching and research and engagement,
72
215012
3540
03:38
and by focusing as much on the framing of the questions
73
218576
2684
03:41
as the solving of the problems.
74
221284
1761
03:43
We're not making a single AI,
75
223505
1929
03:45
we're making the possibilities for many.
76
225458
2278
03:48
And we're actively working to decolonize our imaginations
77
228280
3079
03:51
and to build a curriculum and a pedagogy
78
231383
2373
03:53
that leaves room for a range of different conversations and possibilities.
79
233780
3745
03:58
We are making and remaking.
80
238514
2277
04:00
And I know we're always a work in progress.
81
240815
2838
04:04
But here's a little glimpse
82
244593
1302
04:05
into how we're approaching that problem of scaling a future.
83
245919
3282
04:09
We start by making sure we're grounded in our own history.
84
249982
2834
04:13
In December of 2018,
85
253537
1365
04:14
I took myself up to the town of Brewarrina
86
254926
2056
04:17
on the New South Wales-Queensland border.
87
257006
2587
04:19
This place was a meeting place for Aboriginal people,
88
259617
2515
04:22
for different groups,
89
262156
1159
04:23
to gather, have ceremonies, meet, to be together.
90
263339
3256
04:26
There, on the Barwon River, there's a set of fish weirs
91
266969
2600
04:29
that are one of the oldest and largest systems
92
269593
2158
04:31
of Aboriginal fish traps in Australia.
93
271775
2098
04:34
This system is comprised of 1.8 kilometers of stone walls
94
274214
3283
04:37
shaped like a series of fishnets
95
277521
1691
04:39
with the "Us" pointing down the river,
96
279236
2341
04:41
allowing fish to be trapped at different heights of the water.
97
281601
2937
04:44
They're also fish holding pens with different-height walls for storage,
98
284562
3349
04:47
designed to change the way the water moves
99
287935
2150
04:50
and to be able to store big fish and little fish
100
290109
2253
04:52
and to keep those fish in cool, clear running water.
101
292386
2673
04:56
This fish-trap system was a way to ensure that you could feed people
102
296060
3263
04:59
as they gathered there in a place that was both a meeting of rivers
103
299347
3237
05:02
and a meeting of cultures.
104
302608
1309
05:04
It isn't about the rocks or even the traps per se.
105
304441
3562
05:08
It is about the system that those traps created.
106
308027
3063
05:11
One that involves technical knowledge,
107
311114
2023
05:13
cultural knowledge
108
313161
1485
05:14
and ecological knowledge.
109
314670
1680
05:16
This system is old.
110
316792
1730
05:18
Some archaeologists think it's as old as 40,000 years.
111
318546
2880
05:21
The last time we have its recorded uses is in the nineteen teens.
112
321760
3870
05:26
It's had remarkable longevity and incredible scale.
113
326164
3245
05:29
And it's an inspiration to me.
114
329871
2000
05:32
And a photo of the weir is on our walls here at the Institute,
115
332263
3149
05:35
to remind us of the promise and the challenge
116
335436
2110
05:37
of building something meaningful.
117
337570
1873
05:39
And to remind us that we're building systems
118
339467
2055
05:41
in a place where people have built systems
119
341546
2015
05:43
and sustained those same systems for generations.
120
343585
2476
05:46
It isn't just our history,
121
346514
1666
05:48
it's our legacy as we seek to establish a new branch of engineering.
122
348204
3611
05:52
To build on that legacy and our sense of purpose,
123
352284
2872
05:55
I think we need a clear framework for asking questions about the future.
124
355180
3929
05:59
Questions for which there aren't ready or easy answers.
125
359744
3065
06:03
Here, the point is the asking of the questions.
126
363262
2944
06:06
We believe you need to go beyond the traditional approach
127
366659
2690
06:09
of problem-solving,
128
369373
1405
06:10
to the more complicated one of question asking
129
370802
3085
06:13
and question framing.
130
373911
1680
06:15
Because in so doing, you open up all kinds of new possibilities
131
375950
2969
06:18
and new challenges.
132
378943
1475
06:20
For me, right now,
133
380442
1978
06:22
there are six big questions that frame our approach
134
382444
2794
06:25
for taking AI safely, sustainably and responsibly to scale.
135
385262
3056
06:28
Questions about autonomy,
136
388730
1486
06:30
agency, assurance,
137
390240
1837
06:32
indicators, interfaces and intentionality.
138
392101
2777
06:36
The first question we ask is a simple one.
139
396284
2406
06:39
Is the system autonomous?
140
399038
1680
06:41
Think back to that lift on Bligh Street.
141
401284
2118
06:43
The reality is, one day, that lift may be autonomous.
142
403879
2627
06:46
Which is to say it will be able to act without being told to act.
143
406530
3134
06:50
But it isn't fully autonomous, right?
144
410219
1937
06:52
It can't leave that Bligh Street building
145
412180
2580
06:54
and wonder down to Circular Quay for a beer.
146
414784
2800
06:58
It goes up and down, that's all.
147
418093
2230
07:00
But it does it by itself.
148
420347
2341
07:02
It's autonomous in that sense.
149
422712
2000
07:05
The second question we ask:
150
425744
2381
07:08
does this system have agency?
151
428149
1866
07:10
Does this system have controls and limits that live somewhere
152
430910
3778
07:14
that prevent it from doing certain kinds of things under certain conditions.
153
434712
3793
07:19
The reality with lifts, that's absolutely the case.
154
439061
3270
07:22
Think of any lift you've been in.
155
442355
1904
07:24
There's a red keyslot in the elevator carriage
156
444283
2620
07:26
that an emergency services person can stick a key into
157
446927
2587
07:29
and override the whole system.
158
449538
1719
07:31
But what happens when that system is AI-driven?
159
451641
2500
07:34
Where does the key live?
160
454165
1206
07:35
Is it a physical key, is it a digital key?
161
455395
2230
07:37
Who gets to use it?
162
457649
1349
07:39
Is that the emergency services people?
163
459022
1928
07:40
And how would you know if that was happening?
164
460974
2189
07:43
How would all of that be manifested to you in the lift?
165
463187
3287
07:47
The third question we ask is how do we think about assurance.
166
467752
3817
07:51
How do we think about all of its pieces:
167
471593
1912
07:53
safety, security, trust, risk, liability, manageability,
168
473529
3557
07:57
explicability, ethics, public policy, law, regulation?
169
477110
3657
08:01
And how would we tell you that the system was safe and functioning?
170
481545
3778
08:06
The fourth question we ask
171
486514
1699
08:08
is what would be our interfaces with these AI-driven systems.
172
488237
2874
08:11
Will we talk to them?
173
491135
1174
08:12
Will they talk to us, will they talk to each other?
174
492333
2421
08:14
And what will it mean to have a series of technologies we've known,
175
494778
3174
08:17
for some of us, all our lives,
176
497976
1475
08:19
now suddenly behave in entirely different ways?
177
499475
2237
08:21
Lifts, cars, the electrical grid, traffic lights, things in your home.
178
501736
3938
08:27
The fifth question for these AI-driven systems:
179
507038
3301
08:30
What will the indicators be to show that they're working well?
180
510363
3055
08:33
Two hundred years of the industrial revolution
181
513442
2159
08:35
tells us that the two most important ways to think about a good system
182
515625
3286
08:38
are productivity and efficiency.
183
518935
2133
08:41
In the 21st century,
184
521712
1175
08:42
you might want to expand that just a little bit.
185
522911
2262
08:45
Is the system sustainable,
186
525197
1245
08:46
is it safe, is it responsible?
187
526466
2000
08:48
Who gets to judge those things for us?
188
528920
2428
08:51
Users of the systems would want to understand
189
531372
2120
08:53
how these things are regulated, managed and built.
190
533516
2581
08:57
And then there's the final, perhaps most critical question
191
537737
2769
09:00
that you need to ask of these new AI systems.
192
540530
2817
09:03
What's its intent?
193
543989
1440
09:05
What's the system designed to do
194
545822
1533
09:07
and who said that was a good idea?
195
547379
1792
09:09
Or put another way,
196
549195
1667
09:10
what is the world that this system is building,
197
550886
2675
09:13
how is that world imagined,
198
553585
1643
09:15
and what is its relationship to the world we live in today?
199
555252
2864
09:18
Who gets to be part of that conversation?
200
558546
2166
09:20
Who gets to articulate it?
201
560736
1730
09:22
How does it get framed and imagined?
202
562490
2400
09:26
There are no simple answers to these questions.
203
566149
2916
09:29
Instead, they frame what's possible
204
569089
2071
09:31
and what we need to imagine,
205
571184
2055
09:33
design, build, regulate and even decommission.
206
573263
3317
09:37
They point us in the right directions
207
577278
1855
09:39
and help us on a path to establish a new branch of engineering.
208
579157
3317
09:42
But critical questions aren't enough.
209
582895
3016
09:46
You also need a way of holding all those questions together.
210
586880
3293
09:50
For us at the Institute,
211
590809
1166
09:51
we're also really interested in how to think about AI as a system,
212
591999
4578
09:56
and where and how to draw the boundaries of that system.
213
596601
2992
09:59
And those feel like especially important things right now.
214
599617
2857
10:03
Here, we're influenced by the work that was started way back in the 1940s.
215
603061
3968
10:07
In 1944, along with anthropologists Gregory Bateson and Margaret Mead,
216
607529
4142
10:11
mathematician Norbert Wiener convened a series of conversations
217
611695
3485
10:15
that would become known as the Macy Conferences on Cybernetics.
218
615204
3365
10:18
Ultimately, between 1946 and 1953,
219
618593
3373
10:21
ten conferences were held under the banner of cybernetics.
220
621990
2809
10:25
As defined by Norbert Wiener,
221
625553
1438
10:27
cybernetics sought to "develop a language and techniques
222
627015
3803
10:30
that will enable us to indeed attack the problem of control and communication
223
630842
5124
10:35
in advanced computing technologies."
224
635990
2400
10:38
Cybernetics argued persuasively
225
638847
1825
10:40
that one had to think about the relationship
226
640696
2064
10:42
between humans, computers
227
642784
1962
10:44
and the broader ecological world.
228
644770
1778
10:46
You had to think about them as a holistic system.
229
646572
2332
10:49
Participants in the Macy Conferences were concerned with how the mind worked,
230
649293
3715
10:53
with ideas about intelligence and learning,
231
653032
2029
10:55
and about the role of technology in our future.
232
655085
2317
10:57
Sadly, the conversations that started with the Macy Conference
233
657847
3040
11:00
are often forgotten when the talk is about AI.
234
660911
2483
11:03
But for me, there's something really important to reclaim here
235
663815
3349
11:07
about the idea of a system that has to accommodate culture,
236
667188
3326
11:10
technology and the environment.
237
670538
2261
11:13
At the Institute, that sort of systems thinking is core to our work.
238
673649
3523
11:17
Over the last three years,
239
677950
1302
11:19
a whole collection of amazing people have joined me here
240
679276
2642
11:21
on this crazy journey to do this work.
241
681942
2373
11:24
Our staff includes anthropologists,
242
684871
2357
11:27
systems and environmental engineers, and computer scientists
243
687252
3365
11:30
as well as a nuclear physicist,
244
690641
2166
11:32
an award-winning photo journalist,
245
692831
1643
11:34
and at least one policy and standards expert.
246
694498
2507
11:37
It's a heady mix.
247
697506
1484
11:39
And the range of experience and expertise is powerful,
248
699014
3387
11:42
as are the conflicts and the challenges.
249
702425
2303
11:45
Being diverse requires a constant willingness
250
705085
2262
11:47
to find ways to hold people in conversation.
251
707371
2627
11:50
And to dwell just a little bit with the conflict.
252
710323
2865
11:53
We also worked out early
253
713911
1880
11:55
that the way to build a new way of doing things
254
715815
2198
11:58
would require a commitment to bringing others along on that same journey with us.
255
718037
4143
12:02
So we opened our doors to an education program very quickly,
256
722204
3135
12:05
and we launched our first master's program in 2018.
257
725363
2944
12:08
Since then, we've had two cohorts of master's students
258
728704
2593
12:11
and one cohort of PhD students.
259
731321
1944
12:13
Our students come from all over the world
260
733683
1960
12:15
and all over life.
261
735667
1222
12:16
Australia, New Zealand, Nigeria, Nepal,
262
736913
2595
12:19
Mexico, India, the United States.
263
739532
1959
12:22
And they range in age from 23 to 60.
264
742286
2400
12:24
They variously had backgrounds in maths and music,
265
744997
2984
12:28
policy and performance,
266
748005
1745
12:29
systems and standards,
267
749774
1334
12:31
architecture and arts.
268
751132
1459
12:33
Before they joined us at the Institute,
269
753385
1881
12:35
they ran companies, they worked for government,
270
755290
2238
12:37
served in the army, taught high school,
271
757552
2197
12:39
and managed arts organizations.
272
759773
2067
12:42
They were adventurers
273
762329
1564
12:43
and committed to each other,
274
763917
1414
12:45
and to building something new.
275
765355
1475
12:47
And really, what more could you ask for?
276
767630
2513
12:50
Because although I've spent 20 years in Silicon Valley
277
770922
2532
12:53
and I know the stories about the lone inventor
278
773478
2175
12:55
and the hero's journey,
279
775677
1436
12:57
I also know the reality.
280
777137
1603
12:58
That it's never just a hero's journey.
281
778764
2186
13:00
It's always a collection of people who have a shared sense of purpose
282
780974
3643
13:04
who can change the world.
283
784641
1680
13:06
So where do you start?
284
786807
1680
13:09
Well, I think you start where you stand.
285
789299
2412
13:12
And for me, that means I want to acknowledge
286
792077
2063
13:14
the traditional owners of the land upon which I'm standing.
287
794164
2777
13:16
The Ngunnawal and Ngambri people,
288
796965
1587
13:18
this is their land,
289
798576
1167
13:19
never ceded, always sacred.
290
799767
1365
13:21
And I pay my respects to the elders, past and present, of this place.
291
801156
3373
13:25
I also acknowledge that we're gathering today
292
805315
2111
13:27
in many other places,
293
807450
1445
13:28
and I pay my respects to the traditional owners and elders
294
808919
2718
13:31
of all those places too.
295
811661
1510
13:33
It means a lot to me to get to say those words
296
813617
2207
13:35
and to dwell on what they mean and signal.
297
815848
2531
13:38
And to remember that we live in a country
298
818855
1961
13:40
that has been continuously occupied for at least 60,000 years.
299
820840
3261
13:44
Aboriginal people built worlds here,
300
824979
1905
13:46
they built social systems, they built technologies.
301
826908
2499
13:49
They built ways to manage this place
302
829837
1729
13:51
and to manage it remarkably over a protracted period of time.
303
831590
3254
13:55
And every moment any one of us stands on a stage as Australians,
304
835935
3000
13:58
here or abroad,
305
838959
1284
14:00
we carry with us a privilege and a responsibility
306
840267
2287
14:02
because of that history.
307
842578
1361
14:04
And it's not just a history.
308
844241
1706
14:05
It's also an incredibly rich set of resources,
309
845971
2517
14:08
worldviews and knowledge.
310
848512
1944
14:10
And it should run through all of our bones
311
850480
2444
14:12
and it should be the story we always tell.
312
852948
2415
14:15
Ultimately, it's about thinking differently,
313
855778
2182
14:17
asking different kinds of questions,
314
857984
2556
14:20
looking holistically at the world and the systems,
315
860564
2737
14:23
and finding other people who want to be on that journey with you.
316
863325
3249
14:26
Because for me,
317
866598
1201
14:27
the only way to actually think about the future and scale
318
867823
3306
14:31
is to always be doing it collectively.
319
871153
2437
14:33
And because for me,
320
873614
1349
14:34
the notion of humans in it together
321
874987
2537
14:37
is one of the ways we get to think about things
322
877548
2214
14:39
that are responsible, safe
323
879786
2357
14:42
and ultimately, sustainable.
324
882167
1872
14:45
Thank you.
325
885143
1150
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7