Why we have an emotional connection to robots | Kate Darling

137,191 views ・ 2018-11-06

TED


Please double-click on the English subtitles below to play the video.

00:00
Translator: Joseph Geni Reviewer: Krystian Aparta
0
0
7000
Prevodilac: Dragana Savanovic Lektor: Ivana Krivokuća
00:13
There was a day, about 10 years ago,
1
13047
2508
Jednog dana, pre oko 10 godina,
00:15
when I asked a friend to hold a baby dinosaur robot upside down.
2
15579
3944
zamolila sam prijatelja da drži robota bebu dinosaurusa naopačke.
00:21
It was this toy called a Pleo that I had ordered,
3
21889
3446
Bila je to igračka po imenu Plio koju sam naručila,
00:25
and I was really excited about it because I've always loved robots.
4
25359
4401
i zbog nje sam bila jako uzbuđena jer sam oduvek volela robote.
00:29
And this one has really cool technical features.
5
29784
2279
A ovaj ima stvarno kul tehnička svojstva.
00:32
It had motors and touch sensors
6
32087
2119
Imao je motor, senzore dodira
00:34
and it had an infrared camera.
7
34230
2244
i infracrvenu kameru.
00:36
And one of the things it had was a tilt sensor,
8
36498
2763
Još jedna od stvari koju je imao bio je senzor naginjanja,
00:39
so it knew what direction it was facing.
9
39285
2318
pa je znao u kom smeru je okrenut.
00:42
And when you held it upside down,
10
42095
2134
Ako biste ga držali naopačke,
00:44
it would start to cry.
11
44253
1572
počeo bi da plače.
00:46
And I thought this was super cool, so I was showing it off to my friend,
12
46527
3496
Mislila sam da je ovo jako zabavno, pa sam ga pokazala prijatelju,
00:50
and I said, "Oh, hold it up by the tail. See what it does."
13
50047
2805
i rekla sam: „Drži ga za rep. Videćeš šta radi.“
00:55
So we're watching the theatrics of this robot
14
55268
3625
I tako smo gledali performans ovog robota
00:58
struggle and cry out.
15
58917
2199
koji se borio i plakao.
01:02
And after a few seconds,
16
62767
2047
Nakon nekoliko sekundi,
01:04
it starts to bother me a little,
17
64838
1972
počelo je malo da mi smeta,
01:07
and I said, "OK, that's enough now.
18
67744
3424
pa sam rekla: „Okej, sad je dosta.
01:11
Let's put him back down."
19
71930
2305
Hajde da ga spustimo dole.“
01:14
And then I pet the robot to make it stop crying.
20
74259
2555
Zatim sam mazila robota da bi prestao da plače.
01:18
And that was kind of a weird experience for me.
21
78973
2452
I to mi je bilo prilično čudno iskustvo.
01:22
For one thing, I wasn't the most maternal person at the time.
22
82084
4569
Prvo, tada baš i nisam imala majčinski instinkt.
01:26
Although since then I've become a mother, nine months ago,
23
86677
2731
Iako sam od tada postala majka, pre devet meseci,
01:29
and I've learned that babies also squirm when you hold them upside down.
24
89432
3433
i shvatila da se i bebe vrpolje kada ih držite naopačke.
01:32
(Laughter)
25
92889
1563
(Smeh)
01:35
But my response to this robot was also interesting
26
95023
2358
Moja reakcija na ovog robota je bila zanimljiva
01:37
because I knew exactly how this machine worked,
27
97405
4101
i zato što sam tačno znala kako ova mašina radi,
01:41
and yet I still felt compelled to be kind to it.
28
101530
3262
a ipak sam se osećala primorano da budem dobra prema njoj.
01:46
And that observation sparked a curiosity
29
106450
2707
To opažanje je izazvalo radoznalost
01:49
that I've spent the past decade pursuing.
30
109181
2832
koju sam sledila proteklih deset godina.
01:52
Why did I comfort this robot?
31
112911
1793
Zašto sam tešila ovog robota?
01:56
And one of the things I discovered was that my treatment of this machine
32
116228
3579
Jedna od stvari koju sam otkrila je da je moje ponašanje prema ovoj mašini
01:59
was more than just an awkward moment in my living room,
33
119831
3701
bilo više od neprijatnog trenutka u mojoj dnevnoj sobi,
02:03
that in a world where we're increasingly integrating robots into our lives,
34
123556
5420
da u svetu u kome robote sve više uključujemo u svoje živote
02:09
an instinct like that might actually have consequences,
35
129000
3126
takav instinkt u stvari može imati posledice,
02:13
because the first thing that I discovered is that it's not just me.
36
133452
3749
jer sam prvo otkrila da nisam samo ja u pitanju.
02:19
In 2007, the Washington Post reported that the United States military
37
139249
4802
List „Vašington post“ je 2007. godine izvestio da je vojska SAD-a
02:24
was testing this robot that defused land mines.
38
144075
3230
testirala nekog robota koji je uklanjao mine.
02:27
And the way it worked was it was shaped like a stick insect
39
147329
2912
Radio je tako što je imao oblik izduženog insekta
02:30
and it would walk around a minefield on its legs,
40
150265
2651
koji bi šetao po minskom polju na svojim nogama,
02:32
and every time it stepped on a mine, one of the legs would blow up,
41
152940
3206
a svaki put kad bi stao na minu, jedna noga bi mu eksplodirala,
02:36
and it would continue on the other legs to blow up more mines.
42
156170
3057
i nastavio bi na ostalim nogama da razminira još mina.
02:39
And the colonel who was in charge of this testing exercise
43
159251
3786
Pukovnik koji je bio zadužen za ovu vežbu
02:43
ends up calling it off,
44
163061
2118
ju je na kraju otkazao,
02:45
because, he says, it's too inhumane
45
165203
2435
jer, kako kaže, previše je nehumano
02:47
to watch this damaged robot drag itself along the minefield.
46
167662
4516
gledati oštećenog robota kako se vuče minskim poljem.
02:54
Now, what would cause a hardened military officer
47
174978
3897
E sad, šta bi nateralo očeličenog vojnika
02:58
and someone like myself
48
178899
2043
i nekog kao što sam ja
03:00
to have this response to robots?
49
180966
1857
na ovakvu reakciju na robote?
03:03
Well, of course, we're primed by science fiction and pop culture
50
183537
3310
Pa, naravno, naveli su nas naučna fantastika i masovna kultura
03:06
to really want to personify these things,
51
186871
2579
da zaista želimo da personifikujemo ove stvari,
03:09
but it goes a little bit deeper than that.
52
189474
2789
ali istina ide malo dublje od toga.
03:12
It turns out that we're biologically hardwired to project intent and life
53
192287
5309
Ispostavlja se da nam je biološki urođeno da projektujemo smisao i život
03:17
onto any movement in our physical space that seems autonomous to us.
54
197620
4766
na bilo šta što se kreće u fizičkom svetu što nam se čini samostalno.
03:23
So people will treat all sorts of robots like they're alive.
55
203214
3465
Tako da će ljudi tretirati sve vrste robota kao da su živi.
03:26
These bomb-disposal units get names.
56
206703
2683
Ovi koji uklanjaju bombe dobijaju imena.
03:29
They get medals of honor.
57
209410
1682
Dobijaju medalje časti.
03:31
They've had funerals for them with gun salutes.
58
211116
2325
Organizovane su im sahrane uz počasnu paljbu.
03:34
And research shows that we do this even with very simple household robots,
59
214380
3833
Istraživanje pokazuje da radimo isto to čak i sa prostim kućnim robotima,
03:38
like the Roomba vacuum cleaner.
60
218237
2135
kao što je usisivač Rumba.
03:40
(Laughter)
61
220396
1291
(Smeh)
03:41
It's just a disc that roams around your floor to clean it,
62
221711
3089
To je samo disk koji tumara vašim podom i čisti ga,
03:44
but just the fact it's moving around on its own
63
224824
2306
ali sama činjenica da se sam kreće okolo
03:47
will cause people to name the Roomba
64
227154
2167
navešće ljude da mu daju ime
03:49
and feel bad for the Roomba when it gets stuck under the couch.
65
229345
3182
i da ga sažaljevaju kad se zaglavi ispod kauča.
03:52
(Laughter)
66
232551
1865
(Smeh)
03:54
And we can design robots specifically to evoke this response,
67
234440
3340
Možemo posebno dizajnirati robote da izazovu ovu reakciju,
03:57
using eyes and faces or movements
68
237804
3461
koristeći oči, lice ili pokrete
04:01
that people automatically, subconsciously associate
69
241289
3259
koje ljudi automatski, podsvesno povezuju
04:04
with states of mind.
70
244572
2020
sa stanjem svesti.
04:06
And there's an entire body of research called human-robot interaction
71
246616
3293
Postoji čitava oblast istraživanja zvana interakcija čoveka i robota
04:09
that really shows how well this works.
72
249933
1826
koja pokazuje kako ovo funkcioniše.
04:11
So for example, researchers at Stanford University found out
73
251783
3126
Na primer, istraživači sa Stenforda su otkrili
04:14
that it makes people really uncomfortable
74
254933
2001
da je ljudima jako neprijatno
04:16
when you ask them to touch a robot's private parts.
75
256958
2472
ako im tražite da dotaknu intimne delove robota.
04:19
(Laughter)
76
259454
2120
(Smeh)
04:21
So from this, but from many other studies,
77
261598
2023
Iz ovog, ali i mnogih drugih istraživanja,
04:23
we know, we know that people respond to the cues given to them
78
263645
4223
znamo da ljudi reaguju na date signale
04:27
by these lifelike machines,
79
267892
2022
ovih čovekolikih mašina,
04:29
even if they know that they're not real.
80
269938
2017
iako znaju da nisu prave.
04:33
Now, we're headed towards a world where robots are everywhere.
81
273654
4056
Uskoro ćemo živeti u svetu u kome su roboti svuda.
04:37
Robotic technology is moving out from behind factory walls.
82
277734
3065
Robotska tehnologija se širi van fabričkih zidova.
04:40
It's entering workplaces, households.
83
280823
3013
Ulazi u poslovni svet, domaćinstva.
04:43
And as these machines that can sense and make autonomous decisions and learn
84
283860
6209
Pošto ove mašine koje mogu da osećaju, samostalno donose odluke i uče
04:50
enter into these shared spaces,
85
290093
2552
stupaju u zajednički prostor,
04:52
I think that maybe the best analogy we have for this
86
292669
2496
mislim da je najbolja analogija za ovo
04:55
is our relationship with animals.
87
295189
1935
naša veza sa životinjama.
04:57
Thousands of years ago, we started to domesticate animals,
88
297523
3888
Pre više hiljada godina, počeli smo da pripitomljavamo životinje,
05:01
and we trained them for work and weaponry and companionship.
89
301435
4045
da ih dresiramo za rad, oružje i drugarstvo.
05:05
And throughout history, we've treated some animals like tools or like products,
90
305504
4985
Kroz istoriju, neke životinje smo tretirali kao alate ili proizvode,
05:10
and other animals, we've treated with kindness
91
310513
2174
a druge sa ljubaznošću
05:12
and we've given a place in society as our companions.
92
312711
3078
i pružili im mesto u društvu kao našim drugarima.
05:15
I think it's plausible we might start to integrate robots in similar ways.
93
315813
3849
Moguće je da ćemo verovatno početi da integrišemo robote na sličan način.
05:21
And sure, animals are alive.
94
321484
3096
Naravno, životinje su žive.
05:24
Robots are not.
95
324604
1150
Roboti nisu.
05:27
And I can tell you, from working with roboticists,
96
327626
2580
Mogu vam reći, pošto sam radila sa robotičarima,
05:30
that we're pretty far away from developing robots that can feel anything.
97
330230
3522
da smo prilično daleko od stvaranja robota koji mogu bilo šta da osete.
05:35
But we feel for them,
98
335072
1460
Ali mi saosećamo sa njima,
05:37
and that matters,
99
337835
1207
i to je važno,
05:39
because if we're trying to integrate robots into these shared spaces,
100
339066
3627
jer ako pokušavamo da integrišemo robote u ove zajedničke prostore,
05:42
we need to understand that people will treat them differently than other devices,
101
342717
4628
moramo razumeti da će ih ljudi tretirati drugačije od ostalih uređaja,
05:47
and that in some cases,
102
347369
1844
a to u nekim slučajevima,
05:49
for example, the case of a soldier who becomes emotionally attached
103
349237
3172
na primer, u slučaju vojnika koji se emocionalno veže
05:52
to the robot that they work with,
104
352433
2047
za robota sa kojim radi,
05:54
that can be anything from inefficient to dangerous.
105
354504
2504
samo može da bude neefikasno i opasno.
05:58
But in other cases, it can actually be useful
106
358551
2138
U drugim slučajevima, može čak biti i korisno
06:00
to foster this emotional connection to robots.
107
360713
2623
negovanje ove emotivne veze sa robotima.
06:04
We're already seeing some great use cases,
108
364184
2134
Već imamo sjajne načine korišćenja robota,
06:06
for example, robots working with autistic children
109
366342
2604
na primer, roboti koji rade sa autističnom decom
06:08
to engage them in ways that we haven't seen previously,
110
368970
3634
kako bi ih aktivirali na nove načine,
06:12
or robots working with teachers to engage kids in learning with new results.
111
372628
4000
ili oni koji pomažu nastavnicima
da uključe decu u proces učenja sa novim rezultatima.
06:17
And it's not just for kids.
112
377433
1381
I to nije samo za decu.
06:19
Early studies show that robots can help doctors and patients
113
379750
3223
Rane studije pokazuju da roboti mogu pomoći doktorima i pacijentima
06:22
in health care settings.
114
382997
1427
u zdravstvenoj zaštiti.
06:25
This is the PARO baby seal robot.
115
385535
1810
Ovo je robotska beba foke PARO.
06:27
It's used in nursing homes and with dementia patients.
116
387369
3285
Koristi se u staračkim domovima i sa dementnim pacijentima.
06:30
It's been around for a while.
117
390678
1570
Koristi se već neko vreme.
06:32
And I remember, years ago, being at a party
118
392272
3325
Sećam se da sam, pre dosta godina, bila na nekoj žurci
06:35
and telling someone about this robot,
119
395621
2571
i pričala nekome o ovom robotu,
06:38
and her response was,
120
398216
2126
a njen odgovor je bio:
06:40
"Oh my gosh.
121
400366
1262
„Zaboga.
06:42
That's horrible.
122
402508
1188
To je strašno,
06:45
I can't believe we're giving people robots instead of human care."
123
405056
3397
Ne mogu da verujem da ljudima umesto ljudske nege dajemo robote.“
06:50
And this is a really common response,
124
410540
1875
Ovo je veoma česta reakcija,
06:52
and I think it's absolutely correct,
125
412439
2499
i mislim da je potpuno ispravna,
06:54
because that would be terrible.
126
414962
2040
jer bi to bilo strašno.
06:57
But in this case, it's not what this robot replaces.
127
417795
2484
Ali u ovom slučaju, nije do toga šta robot zamenjuje.
07:00
What this robot replaces is animal therapy
128
420858
3120
Robot zamenjuje terapiju životinja
07:04
in contexts where we can't use real animals
129
424002
3198
u kontekstima u kojima ne možemo koristiti prave životinje
07:07
but we can use robots,
130
427224
1168
ali možemo robote,
07:08
because people will consistently treat them more like an animal than a device.
131
428416
5230
jer će ih ljudi dosledno tretirati više kao životinju nego kao uređaj.
07:15
Acknowledging this emotional connection to robots
132
435502
2380
Prepozavanje ove emotivne veze sa robotima
07:17
can also help us anticipate challenges
133
437906
1969
može nam pomoći i da predvidimo izazove
07:19
as these devices move into more intimate areas of people's lives.
134
439899
3451
kako ovi uređaji ulaze u intimnije sfere ljudskih života.
07:24
For example, is it OK if your child's teddy bear robot
135
444111
3404
Na primer, je li u redu ako robot meda vašeg deteta
07:27
records private conversations?
136
447539
2237
snima privatne razgovore?
07:29
Is it OK if your sex robot has compelling in-app purchases?
137
449800
4063
Je li u redu ako vaš robot za seks ima
primamljive ponude za kupovinu u sastavu aplikacije?
07:33
(Laughter)
138
453887
1396
(Smeh)
07:35
Because robots plus capitalism
139
455307
2501
Jer roboti plus kapitalizam
07:37
equals questions around consumer protection and privacy.
140
457832
3705
daju pitanja vezana za zaštitu potrošača i privatnost.
07:42
And those aren't the only reasons
141
462549
1612
I to nisu jedini razlozi
07:44
that our behavior around these machines could matter.
142
464185
2570
zbog kojih je naše ponašanje sa ovim mašinama važno.
07:48
A few years after that first initial experience I had
143
468747
3270
Par godina nakon svog prvog iskustva
07:52
with this baby dinosaur robot,
144
472041
2311
sa robotom bebe dinosaurusa,
07:54
I did a workshop with my friend Hannes Gassert.
145
474376
2501
imala sam radionicu sa prijateljem Hanesom Gasertom.
07:56
And we took five of these baby dinosaur robots
146
476901
2897
Uzeli smo pet ovih robotskih beba dinosaurusa
07:59
and we gave them to five teams of people.
147
479822
2453
i podelili ih ljudima u pet timova.
08:02
And we had them name them
148
482299
1697
Tražili smo da im daju imena,
08:04
and play with them and interact with them for about an hour.
149
484020
3809
da se igraju i druže s njima oko sat vremena.
08:08
And then we unveiled a hammer and a hatchet
150
488707
2206
Onda smo im dali čekić i sekiru
08:10
and we told them to torture and kill the robots.
151
490937
2278
i rekli im da muče i ubiju robote.
08:13
(Laughter)
152
493239
3007
(Smeh)
08:16
And this turned out to be a little more dramatic
153
496857
2294
Ovo se ispostavilo malo dramatičnijim
nego što smo očekivali,
08:19
than we expected it to be,
154
499175
1278
08:20
because none of the participants would even so much as strike
155
500477
3072
jer niko od učesnika nije hteo čak ni da udari
08:23
these baby dinosaur robots,
156
503573
1307
robote beba dinosaurusa,
08:24
so we had to improvise a little, and at some point, we said,
157
504904
5150
pa smo morali malo da improvizujemo, i u jednom momentu smo rekli:
08:30
"OK, you can save your team's robot if you destroy another team's robot."
158
510078
4437
„Okej, možete spasiti robota svog tima ako uništite robota drugog tima.“
08:34
(Laughter)
159
514539
1855
(Smeh)
08:36
And even that didn't work. They couldn't do it.
160
516839
2195
Čak ni to nije uspelo. Nisu mogli.
Na kraju smo rekli:
08:39
So finally, we said,
161
519058
1151
08:40
"We're going to destroy all of the robots
162
520233
2032
„Mi ćemo uništiti sve robote
08:42
unless someone takes a hatchet to one of them."
163
522289
2285
ukoliko neko ne podigne sekiru na jednog.“
08:45
And this guy stood up, and he took the hatchet,
164
525586
3579
Jedan tip je ustao i podigao sekiru,
08:49
and the whole room winced as he brought the hatchet down
165
529189
2706
i cela prostorija je drhtala dok je spuštao sekiru
08:51
on the robot's neck,
166
531919
1780
na robotov vrat,
08:53
and there was this half-joking, half-serious moment of silence in the room
167
533723
6338
i onda se javio polusmešan, poluozbiljan momenat tišine u prostoriji
09:00
for this fallen robot.
168
540085
1698
za palog robota.
09:01
(Laughter)
169
541807
1406
(Smeh)
09:03
So that was a really interesting experience.
170
543237
3694
To je bilo baš zanimljivo iskustvo.
09:06
Now, it wasn't a controlled study, obviously,
171
546955
2459
To očigledno nije bila kontrolisana studija,
09:09
but it did lead to some later research that I did at MIT
172
549438
2850
ali je kasnije dovela do mog istraživanja
09:12
with Palash Nandy and Cynthia Breazeal,
173
552312
2228
na MIT-u sa Palašem Nandijem i Sintijom Brezil,
09:14
where we had people come into the lab and smash these HEXBUGs
174
554564
3627
kada su ljudi dolazili u laboratoriju i uništavali robote HEXBUG
09:18
that move around in a really lifelike way, like insects.
175
558215
3087
koje se kreću kao da su živi, kao insekti.
09:21
So instead of choosing something cute that people are drawn to,
176
561326
3134
Umesto da izaberemo nešto slatko što privlači ljude,
09:24
we chose something more basic,
177
564484
2093
izabrali smo nešto prostije,
09:26
and what we found was that high-empathy people
178
566601
3480
i saznali smo da ljudi sa razvijenom empatijom
09:30
would hesitate more to hit the HEXBUGS.
179
570105
2143
više oklevaju da udare robote HEXBUG.
09:33
Now this is just a little study,
180
573575
1564
To je samo mala studija,
09:35
but it's part of a larger body of research
181
575163
2389
ali je deo većeg istraživanja
09:37
that is starting to indicate that there may be a connection
182
577576
2944
koje počinje da pokazuje da može postojati veza
09:40
between people's tendencies for empathy
183
580544
2373
između sklonosti ljudi ka empatiji
09:42
and their behavior around robots.
184
582941
1976
i njihovog ponašanja prema robotima.
09:45
But my question for the coming era of human-robot interaction
185
585721
3627
Ali moje pitanje za buduću eru interakcije između čoveka i robota
09:49
is not: "Do we empathize with robots?"
186
589372
3055
nije: „Da li saosećamo sa robotima?“
09:53
It's: "Can robots change people's empathy?"
187
593211
2920
Nego: „Da li roboti mogu promeniti ljudsku empatiju?“
09:57
Is there reason to, for example,
188
597489
2287
Da li postoji razlog da, na primer,
09:59
prevent your child from kicking a robotic dog,
189
599800
2333
branimo detetu da šutne psa robota,
10:03
not just out of respect for property,
190
603228
2914
ne samo iz poštovanja prema vlasništvu,
10:06
but because the child might be more likely to kick a real dog?
191
606166
2953
već zato što je veća šansa da će šutnuti pravog psa?
10:10
And again, it's not just kids.
192
610507
1883
I opet, nisu u pitanju samo deca.
10:13
This is the violent video games question, but it's on a completely new level
193
613564
4056
Ovo je pitanje nasilnih video igrica, ali na potpuno novom nivou
10:17
because of this visceral physicality that we respond more intensely to
194
617644
4760
zbog te primarne fizikalnosti na koju intenzivnije reagujemo
10:22
than to images on a screen.
195
622428
1547
nego na slike na ekranu.
10:25
When we behave violently towards robots,
196
625674
2578
Kada se ponašamo nasilno prema robotima,
10:28
specifically robots that are designed to mimic life,
197
628276
3120
posebno onima dizajniranim da imitiraju živa bića,
10:31
is that a healthy outlet for violent behavior
198
631420
3892
da li je to zdrav izduvni ventil za nasilno ponašanje
10:35
or is that training our cruelty muscles?
199
635336
2544
ili tako treniramo mišiće za okrutnost?
10:39
We don't know ...
200
639511
1150
Ne znamo...
10:42
But the answer to this question has the potential to impact human behavior,
201
642622
3945
Odgovor na ovo pitanje može uticati na ljudsko ponašanje,
10:46
it has the potential to impact social norms,
202
646591
2768
kao i na društvene norme,
10:49
it has the potential to inspire rules around what we can and can't do
203
649383
3849
može uticati na pravila o tome šta možemo a šta ne
10:53
with certain robots,
204
653256
1151
sa određenim robotima,
10:54
similar to our animal cruelty laws.
205
654431
1848
kao u zakonima o zlostavljanju životinja.
10:57
Because even if robots can't feel,
206
657228
2864
Jer iako roboti ne mogu da osećaju,
11:00
our behavior towards them might matter for us.
207
660116
3080
naše ponašanje prema njima može biti važno za nas.
11:04
And regardless of whether we end up changing our rules,
208
664889
2855
Bez obzira na to da li ćemo na kraju promeniti pravila,
11:08
robots might be able to help us come to a new understanding of ourselves.
209
668926
3556
roboti mogu da nam pomognu da bolje razumemo sami sebe.
11:14
Most of what I've learned over the past 10 years
210
674276
2316
Većina onoga što sam naučila u poslednjih 10 godina
11:16
has not been about technology at all.
211
676616
2238
uopšte nema veze sa tehnologijom.
11:18
It's been about human psychology
212
678878
2503
Nego o ljudskoj psihologiji, o empatiji
11:21
and empathy and how we relate to others.
213
681405
2603
i o tome kako se odnosimo prema drugima.
11:25
Because when a child is kind to a Roomba,
214
685524
2365
Jer kada je dete ljubazno prema Rumbi,
11:29
when a soldier tries to save a robot on the battlefield,
215
689262
4015
kada vojnik pokuša da spase robota na bojnom polju,
11:33
or when a group of people refuses to harm a robotic baby dinosaur,
216
693301
3638
ili kada grupa ljudi odbije da povredi robotsku bebu dinosaurusa,
11:38
those robots aren't just motors and gears and algorithms.
217
698248
3191
ti roboti nisu samo motori, zupčanici i algoritmi.
11:42
They're reflections of our own humanity.
218
702501
1905
Oni su odraz naše ljudskosti.
11:45
Thank you.
219
705523
1151
Hvala vam.
11:46
(Applause)
220
706698
3397
(Aplauz)
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7