Rodney Brooks: How robots will invade our lives

60,837 views ใƒป 2008-10-10

TED


ืื ื ืœื—ืฅ ืคืขืžื™ื™ื ืขืœ ื”ื›ืชื•ื‘ื™ื•ืช ื‘ืื ื’ืœื™ืช ืœืžื˜ื” ื›ื“ื™ ืœื”ืคืขื™ืœ ืืช ื”ืกืจื˜ื•ืŸ.

ืžืชืจื’ื: Ido Dekkers ืžื‘ืงืจ: Sigal Tifferet
00:18
What I want to tell you about today is how I see robots invading our lives
0
18330
5000
ืžื” ืฉืื ื™ ืจื•ืฆื” ืœืกืคืจ ืœื›ื ื”ื™ื•ื ื”ื•ื ืื™ืš ืื ื™ ืจื•ืื” ืืช ื”ืจื•ื‘ื•ื˜ื™ื ืคื•ืœืฉื™ื ืœื—ื™ื™ื ื•
00:23
at multiple levels, over multiple timescales.
1
23330
3000
ื‘ืžื’ื•ื•ืŸ ืจืžื•ืช, ื•ื‘ื–ืžื ื™ื ืฉื•ื ื™ื.
00:26
And when I look out in the future, I can't imagine a world, 500 years from now,
2
26330
4000
ื•ื›ืฉืื ื™ ืžื‘ื™ื˜ ืœืขืชื™ื“, ืื ื™ ืœื ื™ื›ื•ืœ ืœื“ืžื™ื™ืŸ ืขื•ืœื, 500 ืฉื ื” ืžื”ื™ื•ื,
00:30
where we don't have robots everywhere.
3
30330
2000
ื‘ื• ืื™ืŸ ืœื ื• ืจื•ื‘ื•ื˜ื™ื ื‘ื›ืœ ืžืงื•ื.
00:32
Assuming -- despite all the dire predictions from many people about our future --
4
32330
5000
ื‘ื”ื ื—ื” -- ืœืžืจื•ืช ื›ืœ ื”ืชื—ื–ื™ื•ืช ื”ืงื•ื“ืจื•ืช ืฉืœ ื”ืจื‘ื” ืื ืฉื™ื ืขืœ ื”ืขืชื™ื“ ืฉืœื ื• --
00:37
assuming we're still around, I can't imagine the world not being populated with robots.
5
37330
4000
ื‘ื”ื ื—ื” ื•ืขื“ื™ื™ืŸ ื ื”ื™ื™ื” ื›ืืŸ, ืื ื™ ืœื ื™ื›ื•ืœ ืœื“ืžื™ื™ืŸ ืืช ื”ืขื•ืœื ืœื ืžืื•ื›ืœืก ื‘ืจื•ื‘ื•ื˜ื™ื.
00:41
And then the question is, well, if they're going to be here in 500 years,
6
41330
3000
ื•ืื– ื”ืฉืืœื” ื”ื™ื, ื•ื‘ื›ืŸ, ืื ื”ื ืขื•ืžื“ื™ื ืœื”ื™ื•ืช ืคื” ืขื•ื“ 500 ืฉื ื”,
00:44
are they going to be everywhere sooner than that?
7
44330
2000
ื”ืื ื”ื ืขื•ืžื“ื™ื ืœื”ื™ื•ืช ื‘ื›ืœ ืžืงื•ื ืžื•ืงื“ื ื™ื•ืชืจ?
00:46
Are they going to be around in 50 years?
8
46330
2000
ื”ืื ื”ื ืขื•ืžื“ื™ื ืœื”ื™ื•ืช ื‘ืกื‘ื™ื‘ื” ื‘ืขื•ื“ 50 ืฉื ื”?
00:48
Yeah, I think that's pretty likely -- there's going to be lots of robots everywhere.
9
48330
3000
ื›ืŸ, ืื ื™ ื—ื•ืฉื‘ ืฉื–ื” ืžืื•ื“ ืืคืฉืจื™ -- ืขื•ืžื“ื™ื ืœื”ื™ื•ืช ื”ืžื•ืŸ ืจื•ื‘ื•ื˜ื™ื ื‘ื›ืœ ืžืงื•ื.
00:51
And in fact I think that's going to be a lot sooner than that.
10
51330
3000
ื•ืœืžืขืฉื” ืื ื™ ื—ื•ืฉื‘ ืฉื–ื” ืขื•ืžื“ ืœื”ื™ื•ืช ื”ืจื‘ื” ื™ื•ืชืจ ืงืจื•ื‘ ืžื–ื”.
00:54
I think we're sort of on the cusp of robots becoming common,
11
54330
4000
ืื ื™ ื—ื•ืฉื‘ ืฉืื ื—ื ื• ืžืžืฉ ืขืœ ื”ืกืฃ ืฉืœ ื”ื–ืžืŸ ื‘ื• ื”ืจื•ื‘ื•ื˜ื™ื ื™ื”ื™ื• ื ืคื•ืฆื™ื,
00:58
and I think we're sort of around 1978 or 1980 in personal computer years,
12
58330
6000
ื•ืื ื™ ื—ื•ืฉื‘ ืฉืื ื—ื ื• ื‘ืกื‘ื™ื‘ื•ืช 1978 ืื• 1980 ื‘ืฉื ื•ืช ืžื—ืฉื‘ื™ื ืื™ืฉื™ื™ื,
01:04
where the first few robots are starting to appear.
13
64330
3000
ื‘ื”ื ื”ืจื•ื‘ื•ื˜ื™ื ื”ืจืืฉื•ื ื™ื ืžืชื—ื™ืœื™ื ืœื”ื•ืคื™ืข.
01:07
Computers sort of came around through games and toys.
14
67330
4000
ืžื—ืฉื‘ื™ื ื”ื’ื™ืขื• ื“ืจืš ืžืฉื—ืงื™ื ื•ืฆืขืฆื•ืขื™ื.
01:11
And you know, the first computer most people had in the house
15
71330
3000
ื•ืืชื ื™ื•ื“ืขื™ื, ื”ืžื—ืฉื‘ ื”ืจืืฉื•ืŸ ืฉืœืจื•ื‘ ื”ืื ืฉื™ื ื”ื™ื” ื‘ื‘ื™ืช
01:14
may have been a computer to play Pong,
16
74330
2000
ืื•ืœื™ ื”ื™ื” ืžื—ืฉื‘ ื›ื“ื™ ืœืฉื—ืง ืคื•ื ื’,
01:16
a little microprocessor embedded,
17
76330
2000
ืžื™ืงืจื• ืžืขื‘ื“ ืงื˜ืŸ ืžื•ื˜ืžืข,
01:18
and then other games that came after that.
18
78330
3000
ื•ืื– ืžืฉื—ืงื™ื ืื—ืจื™ื ืฉื”ื’ื™ืขื• ืื—ืจื™ื•.
01:21
And we're starting to see that same sort of thing with robots:
19
81330
3000
ื•ืื ื—ื ื• ืžืชื—ื™ืœื™ื ืœืจืื•ืช ืืช ืื•ืชื• ื”ื“ื‘ืจ ืขื ืจื•ื‘ื•ื˜ื™ื:
01:24
LEGO Mindstorms, Furbies -- who here -- did anyone here have a Furby?
20
84330
4000
ืœื’ื• ืžื™ื™ื ื“ืกื˜ื•ืจืžืก, ืคืจื‘ื™ -- ืžื™ ืคื” -- ืœืžื™ืฉื”ื• ื”ื™ื” ืคืจื‘ื™?
01:28
Yeah, there's 38 million of them sold worldwide.
21
88330
3000
ื›ืŸ, ื ืžื›ืจื• 38 ืžื™ืœื™ื•ื ื™ื ื›ืืœื” ื‘ืขื•ืœื.
01:31
They are pretty common. And they're a little tiny robot,
22
91330
2000
ื”ื ืžืื•ื“ ื ืคื•ืฆื™ื. ื•ื”ื ืจื•ื‘ื•ื˜ ืงื˜ื ื˜ืŸ,
01:33
a simple robot with some sensors,
23
93330
2000
ืจื•ื‘ื•ื˜ ืคืฉื•ื˜ ืขื ื›ืžื” ืกื ืกื•ืจื™ื,
01:35
a little bit of processing actuation.
24
95330
2000
ื•ืžืขื˜ ื™ื›ื•ืœืช ืžื—ืฉื•ื‘.
01:37
On the right there is another robot doll, who you could get a couple of years ago.
25
97330
3000
ืžืฆื“ ื™ืžื™ืŸ ืฉื ื™ืฉ ืขื•ื“ ื‘ื•ื‘ืช ืจื•ื‘ื•ื˜, ืฉื™ื›ื•ืœืชื ืœื”ืฉื™ื’ ืœืคื ื™ ื›ืžื” ืฉื ื™ื.
01:40
And just as in the early days,
26
100330
2000
ื•ื›ืžื• ื‘ื™ืžื™ื ื”ืžื•ืงื“ืžื™ื,
01:42
when there was a lot of sort of amateur interaction over computers,
27
102330
5000
ื›ืฉื”ื™ืชื” ืคืขื™ืœื•ืช ื—ื•ื‘ื‘ื ื™ืช ื‘ื”ืงืฉืจ ืœืžื—ืฉื‘ื™ื,
01:47
you can now get various hacking kits, how-to-hack books.
28
107330
4000
ืืชื ื™ื›ื•ืœื™ื ืœืงื‘ืœ ืขื›ืฉื™ื• ื›ืžื” ืขืจื›ื•ืช ืœืคืจื™ืฆื”, ื•ืกืคืจื™ื ืขืœ ืื™ืš ืœืคืจื•ืฅ.
01:51
And on the left there is a platform from Evolution Robotics,
29
111330
4000
ื•ืžืฉืžืืœ ื™ืฉ ืคืœื˜ืคื•ืจืžื” ืžืื‘ื•ืœื•ืฉื™ื•ืŸ ืจื•ื‘ื•ื˜ื™ืงื”,
01:55
where you put a PC on, and you program this thing with a GUI
30
115330
3000
ื‘ื• ืืชื ืฉืžื™ื ืžื—ืฉื‘, ื•ืืชื ืžืชื›ื ื ื™ื ืืช ื”ื“ื‘ืจ ื”ื–ื” ืขื ืžืžืฉืง ืžืฉืชืžืฉ
01:58
to wander around your house and do various stuff.
31
118330
3000
ื›ื“ื™ ืœื˜ื™ื™ืœ ื‘ื‘ื™ืช ื•ืขื•ืฉื™ื ื“ื‘ืจื™ื ืฉื•ื ื™ื.
02:01
And then there's a higher price point sort of robot toys --
32
121330
3000
ื•ืื– ื™ืฉ ืืช ื ืงื•ื“ืช ื”ืžื—ื™ืจ ื”ื’ื‘ื•ื”ื” ื™ื•ืชืจ ืฉืœ ืฆืขืฆื•ืขื™ ืจื•ื‘ื•ื˜ื™ื --
02:04
the Sony Aibo. And on the right there, is one that the NEC developed,
33
124330
4000
ื”ืื™ื™ื‘ื• ืฉืœ ืกื•ื ื™. ื•ืžื™ืžื™ืŸ ืคื”, ื–ื” ืื—ื“ ืฉNEC ืคื™ืชื—ื•,
02:08
the PaPeRo, which I don't think they're going to release.
34
128330
3000
ื” PaPeRo, ืฉืื ื™ ืœื ื—ื•ืฉื‘ ืฉื”ื ื™ืฉื—ืจืจื•.
02:11
But nevertheless, those sorts of things are out there.
35
131330
3000
ืื‘ืœ ืขื“ื™ื™ืŸ, ื”ื“ื‘ืจื™ื ื”ืืœื” ื ืžืฆืื™ื ืฉื ื‘ื—ื•ืฅ.
02:14
And we've seen, over the last two or three years, lawn-mowing robots,
36
134330
4000
ื•ืจืื™ื ื•, ื‘ืžื”ืœืš ื”ืฉื ืชื™ื™ื ืื• ืฉืœื•ืฉ ื”ืื—ืจื•ื ื•ืช, ืจื•ื‘ื•ื˜ื™ื ืžื›ืกื—ื™ ื“ืฉื,
02:18
Husqvarna on the bottom, Friendly Robotics on top there, an Israeli company.
37
138330
6000
ื”ื•ืกืงื•ื•ืจื ื” ืžืœืžื˜ื”, ืคืจื ื“ืœื™ ืจื•ื‘ื•ื˜ื™ืงืก ืžืœืžืขืœื” ืฉื, ื—ื‘ืจื” ื™ืฉืจืืœื™ืช.
02:24
And then in the last 12 months or so
38
144330
2000
ื•ืื– ื‘12 ื”ื—ื•ื“ืฉื™ื ื”ืื—ืจื•ื ื™ื ื‘ืขืจืš
02:26
we've started to see a bunch of home-cleaning robots appear.
39
146330
4000
ื”ืชื—ืœื ื• ืœืจืื•ืช ื—ื‘ื•ืจื” ืฉืœ ืจื•ื‘ื•ื˜ื™ื ืžื ืงื™ ื‘ืชื™ื ืžืชื—ื™ืœื” ืœื”ื•ืคื™ืข.
02:30
The top left one is a very nice home-cleaning robot
40
150330
3000
ื”ืฉืžืืœื™ ื”ืขืœื™ื•ืŸ ื”ื•ื ืจื•ื‘ื•ื˜ ื ื™ืงื•ื™ ื‘ืชื™ื ื ื—ืžื“
02:33
from a company called Dyson, in the U.K. Except it was so expensive --
41
153330
4000
ืžื—ื‘ืจื” ืฉื ืงืจืืช ื“ื™ื™ืกื•ืŸ, ืžืื ื’ืœื™ื”. ืจืง ืฉื”ื•ื ื”ื™ื” ื›ืœ ื›ืš ื™ืงืจ --
02:37
3,500 dollars -- they didn't release it.
42
157330
2000
3,500 ื“ื•ืœืจ -- ื”ื ืœื ืฉื™ื—ืจืจื• ืื•ืชื•.
02:39
But at the bottom left, you see Electrolux, which is on sale.
43
159330
3000
ืื‘ืœ ืœืžื˜ื” ืžืฉืžืืœ, ืืชื ืจื•ืื™ื ืืช ืืœืงื˜ืจื•ืœื•ืงืก, ืฉื ืžื›ืจ ืขื›ืฉื™ื•.
02:42
Another one from Karcher.
44
162330
2000
ืขื•ื“ ืื—ื“ ืžืงืจืฆ'ืจ.
02:44
At the bottom right is one that I built in my lab
45
164330
2000
ืžื™ืžื™ืŸ ืœืžื˜ื” ื™ืฉ ืื—ื“ ืฉื ื‘ื ื” ื‘ืžืขื‘ื“ื” ืฉืœื™
02:46
about 10 years ago, and we finally turned that into a product.
46
166330
3000
ื‘ืขืจืš ืœืคื ื™ 10 ืฉื ื™ื, ื•ืœื‘ืกื•ืฃ ื”ืคื›ื ื• ืื•ืชื• ืœืžื•ืฆืจ.
02:49
And let me just show you that.
47
169330
2000
ื•ืชื ื• ืœื™ ืจืง ืœื”ืจืื•ืช ืœื›ื ืืช ื–ื”.
02:51
We're going to give this away I think, Chris said, after the talk.
48
171330
4000
ืื ื—ื ื• ืขื•ืžื“ื™ื ืœืชืช ืืช ื–ื” ืœื“ืขืชื™, ื›ืจื™ืก ืืžืจ, ืื—ืจื™ ื”ื”ืจืฆืื”.
02:55
This is a robot that you can go out and buy, and that will clean up your floor.
49
175330
6000
ื–ื” ืจื•ื‘ื•ื˜ ืฉืืชื ื™ื›ื•ืœื™ื ืœืฆืืช ื•ืœืงื ื•ืช, ื•ื”ื•ื ื™ื ืงื” ืœื›ื ืืช ื”ืจืฆืคื”.
03:05
And it starts off sort of just going around in ever-increasing circles.
50
185330
5000
ื•ื–ื” ืžืชื—ื™ืœ ืคืฉื•ื˜ ื‘ืชื ื•ืขื” ื‘ืกื™ื‘ื•ื‘ื™ื ื”ื•ืœื›ื™ื ื•ืžืชืจื—ื‘ื™ื.
03:10
If it hits something -- you people see that?
51
190330
4000
ืื ื”ื•ื ืคื•ื’ืข ื‘ืžืฉื”ื• -- ืืชื ืจื•ืื™ื ืืช ื–ื”?
03:14
Now it's doing wall-following, it's following around my feet
52
194330
3000
ืขื›ืฉื™ื• ื”ื•ื ืขื•ืฉื” ืขืงื™ื‘ืช ืงื™ืจื•ืช, ื”ื•ื ืขื•ืงื‘ ืื—ืจ ื”ืจื’ืœื™ื™ื ืฉืœื™
03:17
to clean up around me. Let's see, let's --
53
197330
4000
ื›ื“ื™ ืœื ืงื•ืช ืžืกื‘ื™ื‘ื™. ื‘ื•ืื• ื ืจืื”, ื‘ื•ืื• --
03:21
oh, who stole my Rice Krispies? They stole my Rice Krispies!
54
201330
5000
ืื•, ืžื™ ื’ื ื‘ ืœื™ ืืช ืคืฆืคื•ืฆื™ ื”ืื•ืจื–? ื”ื ื’ื ื‘ื• ืœื™ ืืช ืคืฆืคื•ืฆื™ ื”ืื•ืจื–!
03:26
(Laughter)
55
206330
6000
(ืฆื—ื•ืง)
03:32
Don't worry, relax, no, relax, it's a robot, it's smart!
56
212330
3000
ืืœ ืชื“ืื’ื•, ืชืจื’ืขื•, ืœื, ืชืจื’ืขื•, ื–ื” ืจื•ื‘ื•ื˜, ื”ื•ื ื—ื›ื!
03:35
(Laughter)
57
215330
3000
(ืฆื—ื•ืง)
03:38
See, the three-year-old kids, they don't worry about it.
58
218330
4000
ืชืจืื•, ื™ืœื“ื™ื ื‘ื ื™ 3, ื”ื ืœื ื“ื•ืื’ื™ื ืžื–ื”.
03:42
It's grown-ups that get really upset.
59
222330
2000
ื–ื” ื”ืžื‘ื•ื’ืจื™ื ืฉืžืคื—ื“ื™ื.
03:44
(Laughter)
60
224330
1000
(ืฆื—ื•ืง)
03:45
We'll just put some crap here.
61
225330
2000
ืื– ื‘ื•ืื• ื ืœื›ืœืš ื›ืืŸ.
03:47
(Laughter)
62
227330
4000
(ืฆื—ื•ืง)
03:51
Okay.
63
231330
2000
ืื•ืงื™ื™,
03:53
(Laughter)
64
233330
4000
(ืฆื—ื•ืง)
03:57
I don't know if you see -- so, I put a bunch of Rice Krispies there,
65
237330
3000
ืื ื™ ืœื ื™ื•ื“ืข ืื ืืชื ืจื•ืื™ื -- ืื–, ืื ื™ ืฉื ืงืฆืช ืคืฆืคื•ืฆื™ ืื•ืจื– ืคื”,
04:00
I put some pennies, let's just shoot it at that, see if it cleans up.
66
240330
7000
ืื ื™ ืฉื ื›ืžื” ืžื˜ื‘ืขื•ืช, ื‘ื•ืื• ืคืฉื•ื˜ ื ื–ืจื•ืง ืืช ื–ื” ืขืœื™ื•, ื ืจืื” ืื ื”ื•ื ืžื ืงื”.
04:10
Yeah, OK. So --
67
250330
2000
ื™ื”, ืื•ืงื™ื™. ืื– --
04:12
we'll leave that for later.
68
252330
4000
ืื ื™ ืืฉืื™ืจ ืืช ื–ื” ืœืื—ืจ ื›ืš.
04:16
(Applause)
69
256330
5000
(ืžื—ื™ืื•ืช ื›ืคื™ื™ื)
04:22
Part of the trick was building a better cleaning mechanism, actually;
70
262330
4000
ื—ืœืง ืžื”ื˜ืจื™ืง ื”ื™ื” ืœื‘ื ื•ืช ืžื ื’ื ื•ืŸ ืžื ืงื” ื˜ื•ื‘ ื™ื•ืชืจ, ืœืžืขืฉื”;
04:26
the intelligence on board was fairly simple.
71
266330
4000
ื”ืื™ื ื˜ืœื™ื’ื ืฆื™ื” ืขืœ ื”ืœื•ื— ื”ื™ืชื” ืžืื•ื“ ืคืฉื•ื˜ื”.
04:30
And that's true with a lot of robots.
72
270330
2000
ื•ื–ื” ื ื›ื•ืŸ ืขื ื”ืจื‘ื” ืจื•ื‘ื•ื˜ื™ื.
04:32
We've all, I think, become, sort of computational chauvinists,
73
272330
4000
ื›ื•ืœื ื•, ืื ื™ ื—ื•ืฉื‘, ื”ืคื›ื ื•, ืœืžืขื™ืŸ ืฉื•ื‘ื™ื ื™ืกื˜ื™ื™ื ื—ื™ืฉื•ื‘ื™ื™ื,
04:36
and think that computation is everything,
74
276330
2000
ื•ื—ื•ืฉื‘ื™ื ืฉืžื—ืฉื•ื‘ ื–ื” ื”ื›ืœ,
04:38
but the mechanics still matter.
75
278330
2000
ืื‘ืœ ื”ืžื›ืื ื™ืงื” ืขื“ื™ื™ืŸ ื—ืฉื•ื‘ื”.
04:40
Here's another robot, the PackBot,
76
280330
3000
ื”ื ื” ืขื•ื“ ืจื•ื‘ื•ื˜, ื”ืคืืงื‘ื•ื˜,
04:43
that we've been building for a bunch of years.
77
283330
2000
ืฉื‘ื ื™ื ื• ื‘ืžืฉืš ื›ืžื” ืฉื ื™ื.
04:45
It's a military surveillance robot, to go in ahead of troops --
78
285330
6000
ื–ื” ืจื•ื‘ื•ื˜ ืชืฆืคื™ืช ืฆื‘ืื™, ืฉื™ืœืš ืœืคื ื™ ื”ื›ื•ื—ื•ืช --
04:51
looking at caves, for instance.
79
291330
3000
ื•ื™ื‘ื™ื˜ ืœืชื•ืš ืžืขืจื•ืช, ืœื“ื•ื’ืžื”.
04:54
But we had to make it fairly robust,
80
294330
2000
ืื‘ืœ ื”ื™ื™ื ื• ืฆืจื™ื›ื™ื ืœืขืฉื•ืช ืื•ืชื• ื“ื™ ื—ื–ืง,
04:56
much more robust than the robots we build in our labs.
81
296330
7000
ื”ืจื‘ื” ื™ื•ืชืจ ืขืžื™ื“ ืžืจื•ื‘ื•ื˜ื™ื ืฉืื ื—ื ื• ื‘ื•ื ื™ื ื‘ืžืขื‘ื“ื•ืช ืฉืœื ื•.
05:03
(Laughter)
82
303330
3000
(ืฆื—ื•ืง)
05:12
On board that robot is a PC running Linux.
83
312330
4000
ืขืœ ื”ืจื•ื‘ื•ื˜ ื™ืฉ ืžื—ืฉื‘ ืฉืžืจื™ืฅ ืœื™ื ื•ืงืก.
05:16
It can withstand a 400G shock. The robot has local intelligence:
84
316330
6000
ื”ื•ื ื™ื›ื•ืœ ืœืขืžื•ื“ ื‘ืžื›ื•ืช ืฉืœ 400G. ืœืจื•ื‘ื•ื˜ ื™ืฉ ืื™ื ื˜ืœื™ื’ื ืฆื™ื” ืžืงื•ืžื™ืช:
05:22
it can flip itself over, can get itself into communication range,
85
322330
6000
ื”ื•ื ื™ื›ื•ืœ ืœื”ืคื•ืš ืืช ืขืฆืžื•, ื™ื›ื•ืœ ืœื”ื›ื ื™ืก ืืช ืขืฆืžื• ืœื˜ื•ื•ื— ื”ืชืงืฉื•ืจืช,
05:28
can go upstairs by itself, et cetera.
86
328330
3000
ื™ื›ื•ืœ ืœืขืœื•ืช ื‘ืžื“ืจื’ื•ืช ืœื‘ื“, ื•ืขื•ื“.
05:38
Okay, so it's doing local navigation there.
87
338330
4000
ืื•ืงื™ื™, ืื– ื”ื•ื ืขื•ืฉื” ื ื™ื•ื•ื˜ ืžืงื•ืžื™ ืคื”.
05:42
A soldier gives it a command to go upstairs, and it does.
88
342330
6000
ื—ื™ื™ืœ ื ื•ืชืŸ ืœื• ืคืงื•ื“ื” ืœืขืœื•ืช ื‘ืžื“ืจื’ื•ืช, ื•ื”ื•ื ืขื•ืœื”.
05:49
That was not a controlled descent.
89
349330
3000
ื–ื• ืœื ื™ืจื™ื“ื” ื ืฉืœื˜ืช.
05:52
(Laughter)
90
352330
2000
(ืฆื—ื•ืง)
05:54
Now it's going to head off.
91
354330
2000
ืขื›ืฉื™ื• ื”ื•ื ืขื•ืžื“ ืœื”ืžืฉื™ืš ื”ืœืื”.
05:56
And the big breakthrough for these robots, really, was September 11th.
92
356330
5000
ื•ื”ืคืจื™ืฆื” ื”ื’ื“ื•ืœื” ืœืจื•ื‘ื•ื˜ื™ื ื”ืืœื”, ื‘ืืžืช, ื”ื™ืชื” ื‘ืกืคื˜ืžื‘ืจ 11.
06:01
We had the robots down at the World Trade Center late that evening.
93
361330
4000
ื”ื™ื• ืœื ื• ืจื•ื‘ื•ื˜ื™ื ื‘ืžืจื›ื– ื”ืกื—ืจ ื”ืขื•ืœืžื™ ืžืื•ื—ืจ ื‘ืื•ืชื• ืขืจื‘.
06:06
Couldn't do a lot in the main rubble pile,
94
366330
2000
ืœื ื™ื›ื•ืœื ื• ืœืขืฉื•ืช ื”ืจื‘ื” ื‘ื”ืจื™ืกื•ืช ื”ืขื™ืงืจื™ื•ืช,
06:08
things were just too -- there was nothing left to do.
95
368330
3000
ื”ื“ื‘ืจื™ื ื”ื™ื• ืคืฉื•ื˜ ื™ื•ืชืจ ืžื“ื™ -- ืœื ื ืฉืืจ ืžื” ืœืขืฉื•ืช.
06:11
But we did go into all the surrounding buildings that had been evacuated,
96
371330
5000
ืื‘ืœ ื›ืŸ ื ื›ื ืกื ื• ืœืชื•ืš ื›ืœ ื”ื‘ื ื™ื™ื ื™ื ืžืกื‘ื™ื‘ ืฉืคื•ื ื•,
06:16
and searched for possible survivors in the buildings
97
376330
3000
ื•ื—ื™ืคืฉื ื• ื ื™ืฆื•ืœื™ื ื‘ื‘ื ื™ื™ื ื™ื
06:19
that were too dangerous to go into.
98
379330
2000
ืฉื”ื™ื• ืžืกื•ื›ื ื™ื ืžื›ื“ื™ ืœื”ื›ื ืก ืืœื™ื”ื.
06:21
Let's run this video.
99
381330
2000
ื‘ื•ืื• ื ืจื™ืฅ ืืช ื”ื•ื™ื“ืื•.
06:23
Reporter: ...battlefield companions are helping to reduce the combat risks.
100
383330
3000
ื›ืชื‘: ืขื•ื–ืจื™ื ืœืฉื“ื” ื”ืงืจื‘ ืขื•ื–ืจื™ื ืœื”ืคื—ื™ืช ืืช ื”ืกื™ื›ื•ื ื™ื ื‘ืงืจื‘.
06:26
Nick Robertson has that story.
101
386330
3000
ื ื™ืง ืจื•ื‘ืจื˜ืกื•ืŸ ืžืกืคืจ ืืช ื”ืกื™ืคื•ืจ.
06:31
Rodney Brooks: Can we have another one of these?
102
391330
2000
ืจื•ื“ื ื™ ื‘ืจื•ืงืก: ืื ื—ื ื• ื™ื›ื•ืœื™ื ืœื”ื‘ื™ื ืขื•ื“ ืื—ื“ ื›ื–ื”?
06:38
Okay, good.
103
398330
2000
ืื•ืงื™ื™, ื˜ื•ื‘.
06:43
So, this is a corporal who had seen a robot two weeks previously.
104
403330
3000
ืื–, ื–ื” ืจื‘"ื˜ ืฉืœื ืจืื” ืจื•ื‘ื•ื˜ ืฉื‘ื•ืขื™ื™ื ืงื•ื“ื.
06:48
He's sending robots into caves, looking at what's going on.
105
408330
4000
ื”ื•ื ืฉื•ืœื— ืจื•ื‘ื•ื˜ื™ื ืœืชื•ืš ืžืขืจื•ืช, ืจื•ืื” ืžื” ืงื•ืจื” ืฉื.
06:52
The robot's being totally autonomous.
106
412330
2000
ื”ืจื•ื‘ื•ื˜ื™ื ืขืฆืžืื™ื™ื ืœื—ืœื•ื˜ื™ืŸ.
06:54
The worst thing that's happened in the cave so far
107
414330
4000
ื”ื“ื‘ืจ ื”ื’ืจื•ืข ื‘ื™ื•ืชืจ ืฉืงืจื” ื‘ืžืขืจื” ืขื“ ืขื›ืฉื™ื•
06:58
was one of the robots fell down ten meters.
108
418330
3000
ื”ื™ื” ืฉืื—ื“ ื”ืจื•ื‘ื•ื˜ื™ื ื ืคืœ ืžื’ื•ื‘ื” ืขืฉืจื” ืžื˜ืจื™ื.
07:08
So one year ago, the US military didn't have these robots.
109
428330
3000
ืื– ืœืคื ื™ ืฉื ื”, ืœืฆื‘ื ื”ืืžืจื™ืงืื™ ืœื ื”ื™ื• ืืช ื”ืจื•ื‘ื•ื˜ื™ื ื”ืืœื”.
07:11
Now they're on active duty in Afghanistan every day.
110
431330
2000
ืขื›ืฉื™ื• ื”ื ื‘ืฉืจื•ืช ืคืขื™ืœ ื‘ืืคื’ื ื™ืกื˜ืืŸ ื›ืœ ื™ื•ื.
07:13
And that's one of the reasons they say a robot invasion is happening.
111
433330
3000
ื•ื–ื• ืื—ืช ื”ืกื™ื‘ื•ืช ืฉื”ื ืื•ืžืจื™ื ืฉื™ืฉ ืคืœื™ืฉืช ืจื•ื‘ื•ื˜ื™ื.
07:16
There's a sea change happening in how -- where technology's going.
112
436330
4000
ื™ืฉ ืฉื™ื ื•ื™ ืขืฆื•ื ืฉืงื•ืจื” ืขื›ืฉื™ื• -- ื‘ืžืงื•ื ืืœื™ื• ื˜ื›ื ื•ืœื•ื’ื™ื” ื”ื•ืœื›ืช.
07:20
Thanks.
113
440330
2000
ืชื•ื“ื”.
07:23
And over the next couple of months,
114
443330
2000
ื•ื‘ื—ื•ื“ืฉื™ื ื”ืงืจื•ื‘ื™ื,
07:25
we're going to be sending robots in production
115
445330
3000
ืื ื—ื ื• ืขื•ืžื“ื™ื ืœืฉืœื•ื— ืจื•ื‘ื•ื˜ื™ื ื‘ื™ืฆื•ืจ
07:28
down producing oil wells to get that last few years of oil out of the ground.
116
448330
4000
ืœื‘ืืจื•ืช ื ืคื˜ ืคืขื™ืœื•ืช ื›ื“ื™ ืœื”ื•ืฆื™ื ืืช ืฉืืจื™ื•ืช ื”ื ืคื˜ ื”ืื—ืจื•ื ื•ืช ืžื”ืื“ืžื”.
07:32
Very hostile environments, 150หš C, 10,000 PSI.
117
452330
4000
ืกื‘ื™ื‘ื•ืช ืžืื•ื“ ืขื•ื™ื ื•ืช, 150 ืžืขืœื•ืช ืฆืœื–ื™ื•ืก, 10,000 PSI ืœื—ืฅ.
07:36
Autonomous robots going down, doing this sort of work.
118
456330
4000
ืจื•ื‘ื•ื˜ื™ื ืขืฆืžืื™ื™ื ื™ื•ืจื“ื™ื, ื•ืขื•ืฉื™ื ืืช ื”ืขื‘ื•ื“ื” ื”ื–ื•.
07:40
But robots like this, they're a little hard to program.
119
460330
3000
ืื‘ืœ ืจื•ื‘ื•ื˜ื™ื ื›ืžื• ืืœื”, ื”ื ืงืฆืช ืงืฉื™ื ืœืชื›ื ื•ืช.
07:43
How, in the future, are we going to program our robots
120
463330
2000
ืื™ืš, ื‘ืขืชื™ื“, ืื ื—ื ื• ื ืชื›ื ืช ืจื•ื‘ื•ื˜ื™ื
07:45
and make them easier to use?
121
465330
2000
ื•ื ื”ืคื•ืš ืื•ืชื ืœืงืœื™ื ื™ื•ืชืจ ืœืฉื™ืžื•ืฉ?
07:47
And I want to actually use a robot here --
122
467330
3000
ื•ืื ื™ ืจื•ืฆื” ืœื”ืฉืชืžืฉ ื‘ืจื•ื‘ื•ื˜ ืœืžืขืฉื” --
07:50
a robot named Chris -- stand up. Yeah. Okay.
123
470330
5000
ืจื•ื‘ื•ื˜ ืฉื ืงืจื ื›ืจื™ืก -- ืชืขืžื•ื“. ื›ืŸ. ืื•ืงื™ื™.
07:57
Come over here. Now notice, he thinks robots have to be a bit stiff.
124
477330
4000
ื‘ื•ื ื”ื ื”. ืขื›ืฉื™ื• ืฉื™ืžื• ืœื‘, ื”ื•ื ื—ื•ืฉื‘ ืฉืจื•ื‘ื•ื˜ื™ื ืฆืจื™ื›ื™ื ืœื”ื™ื•ืช ืงืฆืช ืงืฉื™ื—ื™ื.
08:01
He sort of does that. But I'm going to --
125
481330
3000
ื”ื•ื ืขื•ืฉื” ื›ืš. ืื‘ืœ ืื ื™ ืขื•ืžื“ --
08:04
Chris Anderson: I'm just British. RB: Oh.
126
484330
2000
ื›ืจื™ืก ืื ื“ืจืกื•ืŸ: ืื ื™ ืจืง ื‘ืจื™ื˜ื™. ืจ.ื‘: ืื•ื”.
08:06
(Laughter)
127
486330
2000
(ืฆื—ื•ืง)
08:08
(Applause)
128
488330
2000
(ืžื—ื™ืื•ืช ื›ืคื™ื™ื)
08:10
I'm going to show this robot a task. It's a very complex task.
129
490330
3000
ืื ื™ ืขื•ืžื“ ืœื”ืจืื•ืช ืœืจื•ื‘ื•ื˜ ื”ื–ื” ืžืฉื™ืžื”. ื–ื• ืžืฉื™ืžื” ืžืื•ื“ ืžืกื•ื‘ื›ืช.
08:13
Now notice, he nodded there, he was giving me some indication
130
493330
3000
ืฉื™ืžื• ืœื‘ ืขื›ืฉื™ื•, ื”ื•ื ื”ื ื”ืŸ ืคื”, ื”ื•ื ื ืชืŸ ืœื™ ื—ื™ื•ื•ื™
08:16
he was understanding the flow of communication.
131
496330
3000
ื”ื•ื ื”ื‘ื™ืŸ ืืช ืชื”ืœื™ืš ื”ืชืงืฉื•ืจืช.
08:19
And if I'd said something completely bizarre
132
499330
2000
ื•ืื ื”ื™ื™ืชื™ ืื•ืžืจ ืžืฉื”ื• ืžืžืฉ ืžื•ื–ืจ
08:21
he would have looked askance at me, and regulated the conversation.
133
501330
3000
ื”ื•ื ื”ื™ื” ืžื‘ื™ื˜ ื‘ื™ ื‘ื—ืฉื“ื ื•ืช, ื•ืžืกื“ื™ืจ ืืช ื”ืฉื™ื—ื”.
08:24
So now I brought this up in front of him.
134
504330
3000
ืื– ืขื›ืฉื™ื• ื”ื‘ืืชื™ ืืช ื–ื” ืœืคื ื™ื•.
08:27
I'd looked at his eyes, and I saw his eyes looked at this bottle top.
135
507330
4000
ื”ื‘ื˜ืชื™ ื‘ืขื™ื ื™ื•, ื•ืจืื™ืชื™ ืืช ืขื™ื ื™ื• ืžื‘ื™ื˜ื•ืช ื‘ืคืงืง ื”ื‘ืงื‘ื•ืง ื”ื–ื”.
08:31
And I'm doing this task here, and he's checking up.
136
511330
2000
ื•ืื ื™ ืขื•ืžื“ ืœืขืฉื•ืช ืืช ื”ืžืฉื™ืžื” ื”ื–ื• ืคื”, ื•ื”ื•ื ื‘ื•ื“ืง.
08:33
His eyes are going back and forth up to me, to see what I'm looking at --
137
513330
3000
ื”ืขื™ื™ื ื™ื™ื ืฉืœื• ื–ื–ื•ืช ืื—ื•ืจื” ื•ืงื“ื™ืžื” ื‘ื—ื–ืจื” ืืœื™, ื›ื“ื™ ืœืจืื•ืช ืขืœ ืžื” ืื ื™ ืžืกืชื›ืœ --
08:36
so we've got shared attention.
138
516330
2000
ืื– ื™ืฉ ืœื ื• ืชืฉื•ืžืช ืœื‘ ืžืฉื•ืชืคืช.
08:38
And so I do this task, and he looks, and he looks to me
139
518330
3000
ืื– ืื ื™ ืขื•ืฉื” ืืช ื”ืžืฉื™ืžื” ื”ื–ื•, ื•ื”ื•ื ืžื‘ื™ื˜ ื‘ื™
08:41
to see what's happening next. And now I'll give him the bottle,
140
521330
4000
ืœืจืื•ืช ืžื” ื™ืงืจื” ืขื›ืฉื™ื•. ื•ืขื›ืฉื™ื• ืื ื™ ืืชืŸ ืœื• ืืช ื”ื‘ืงื‘ื•ืง,
08:45
and we'll see if he can do the task. Can you do that?
141
525330
2000
ื•ื ืจืื” ืื ื”ื•ื ื™ื›ื•ืœ ืœืขืฉื•ืช ืืช ื”ืžืฉื™ืžื” ื”ื–ืืช. ืืชื” ื™ื›ื•ืœ ืœืขืฉื•ืช ืืช ื–ื”?
08:47
(Laughter)
142
527330
3000
(ืฆื—ื•ืง)
08:50
Okay. He's pretty good. Yeah. Good, good, good.
143
530330
4000
ืื•ืงื™ื™, ื”ื•ื ื“ื™ ื˜ื•ื‘. ื›ืŸ. ื˜ื•ื‘, ื˜ื•ื‘ , ื˜ื•ื‘.
08:54
I didn't show you how to do that.
144
534330
2000
ืœื ื”ืจืื™ืชื™ ืœืš ืื™ืš ืœืขืฉื•ืช ืืช ื–ื”.
08:56
Now see if you can put it back together.
145
536330
2000
ืขื›ืฉื™ื• ืชืจืื” ืื ืชื•ื›ืœ ืœื”ืจื›ื™ื‘ ืื•ืชื• ื‘ื—ื–ืจื”.
08:58
(Laughter)
146
538330
2000
(ืฆื—ื•ืง)
09:00
And he thinks a robot has to be really slow.
147
540330
1000
ื•ื”ื•ื ื—ื•ืฉื‘ ืฉืจื•ื‘ื•ื˜ ืฆืจื™ืš ืœื”ื™ื•ืช ืžืžืฉ ืื™ื˜ื™.
09:01
Good robot, that's good.
148
541330
2000
ืจื•ื‘ื•ื˜, ื˜ื•ื‘, ื–ื” ื˜ื•ื‘.
09:03
So we saw a bunch of things there.
149
543330
2000
ืื– ืจืื™ื ื• ื›ืžื” ื“ื‘ืจื™ื ืฉื.
09:06
We saw when we're interacting,
150
546330
3000
ืจืื™ื ื• ืฉื›ืฉืื ื—ื ื• ืžืชืงืฉืจื™ื,
09:09
we're trying to show someone how to do something, we direct their visual attention.
151
549330
4000
ืื ื—ื ื• ืžื ืกื™ื ืœื”ืจืื•ืช ืœืžื™ืฉื”ื• ืื™ืš ืœืขืฉื•ืช ืžืฉื”ื•, ืื ื—ื ื• ืžืคื ื™ื ืืช ื”ืงืฉื‘ ื”ื•ื™ื–ื•ืืœื™ ืฉืœื”ื.
09:13
The other thing communicates their internal state to us,
152
553330
4000
ื”ื“ื‘ืจ ื”ืฉื ื™ ืžืจืื” ืœื ื• ืืช ื”ืžืฆื‘ ื”ืคื ื™ืžื™ ืฉืœื”ื.
09:17
whether he's understanding or not, regulates a social interaction.
153
557330
3000
ื‘ื™ืŸ ืื ื”ื•ื ืžื‘ื™ืŸ ืื•ืชื• ืœื, ืžื ื˜ืจ ืืช ื”ืชืงืฉื•ืจืช ื”ื—ื‘ืจืชื™ืช.
09:20
There was shared attention looking at the same sort of thing,
154
560330
2000
ื”ื™ืชื” ืชืฉื•ืžืช ืœื‘ ืžืฉื•ืชืคืช ื‘ืœื”ื‘ื™ื˜ ื‘ืื•ืชื• ืกื•ื’ ืฉืœ ื“ื‘ืจ,
09:22
and recognizing socially communicated reinforcement at the end.
155
562330
4000
ื•ืœื–ื”ื•ืช ื—ื™ื–ื•ืง ื—ื‘ืจืชื™ ืžืชื•ืงืฉืจ ื‘ืกื•ืฃ.
09:26
And we've been trying to put that into our lab robots
156
566330
3000
ื•ื ื™ืกื™ื ื• ืœืฉืœื‘ ืืช ื–ื” ื‘ืจื•ื‘ื•ื˜ื™ ื”ืžืขื‘ื“ื” ืฉืœื ื•
09:29
because we think this is how you're going to want to interact with robots in the future.
157
569330
4000
ืžืคื ื™ ืฉืื ื—ื ื• ื—ื•ืฉื‘ื™ื ืฉื–ื” ืื™ืš ืฉืชืจืฆื• ืœืชืงืฉืจ ืขื ืจื•ื‘ื•ื˜ื™ื ื‘ืขืชื™ื“.
09:33
I just want to show you one technical diagram here.
158
573330
2000
ืื ื™ ืจืง ืจื•ืฆื” ืœื”ืจืื•ืช ืœื›ื ื“ื™ืื’ืจืžื” ื˜ื›ื ื™ืช ืื—ืช ืคื”.
09:35
The most important thing for building a robot that you can interact with socially
159
575330
4000
ื”ื“ื‘ืจ ื”ื—ืฉื•ื‘ ื‘ื™ื•ืชืจ ื‘ื‘ื ื™ื™ืช ืจื•ื‘ื•ื˜ ื‘ื™ื›ื•ืœืช ืœืชืงืฉืจ ืื™ืชื• ื—ื‘ืจืชื™ืช
09:39
is its visual attention system.
160
579330
2000
ื”ื•ื ื”ืžืขืจื›ืช ืœืงืฉื‘ ื•ื™ื–ื•ืืœื™.
09:41
Because what it pays attention to is what it's seeing
161
581330
3000
ืžืคื ื™ ืฉืžื” ืฉื”ื•ื ืงืฉื•ื‘ ืืœื™ื• ื”ื•ื ืžื” ืฉื”ื•ื ืจื•ืื”
09:44
and interacting with, and what you're understanding what it's doing.
162
584330
3000
ื•ืžืชืงืฉืจ ืื™ืชื•, ื•ืžื” ืฉืืชื ืžื‘ื™ื ื™ื ืžืžื” ืฉื”ื•ื ืขื•ืฉื”.
09:47
So in the videos I'm about to show you,
163
587330
3000
ืื– ื‘ืกืจื˜ื•ื ื™ื ืฉืื ื™ ืขื•ืžื“ ืœื”ืจืื•ืช ืœื›ื,
09:50
you're going to see a visual attention system on a robot
164
590330
4000
ืื ื—ื ื• ืขื•ืžื“ื™ื ืœืจืื•ืช ืงืฉื‘ ื•ื™ื–ื•ืืœื™ ืฉืœ ื”ืจื•ื‘ื•ื˜
09:54
which has -- it looks for skin tone in HSV space,
165
594330
4000
ืฉื”ื•ื -- ื”ื•ื ืžื—ืคืฉ ืฆื‘ืขื™ื ืฉืœ ืขื•ืจ ื‘ื˜ื•ื•ื— ื”HSV (ืชื—ื•ื ื‘ื—ื™ืจื” ืฉืœ ืฆื‘ืขื™ื)
09:58
so it works across all human colorings.
166
598330
4000
ื›ืš ืฉื”ื•ื ืขื•ื‘ืจ ืขืœ ื›ืœ ื’ื•ื•ื ื™ ื”ืขื•ืจ ื”ืื ื•ืฉื™ื™ื.
10:02
It looks for highly saturated colors, from toys.
167
602330
2000
ื”ื•ื ืžื—ืคืฉ ืฆื‘ืขื™ื ืžืื•ื“ ืจื•ื•ื™ื™ื, ืžืฆืขืฆื•ืขื™ื.
10:04
And it looks for things that move around.
168
604330
2000
ื•ื”ื•ื ืžื—ืคืฉ ื“ื‘ืจื™ื ืฉื–ื–ื™ื.
10:06
And it weights those together into an attention window,
169
606330
3000
ื•ืžืฉืงืœืœ ืืช ื›ืœ ืืœื” ื™ื—ื“ ืœื—ืœื•ืŸ ืงืฉื‘,
10:09
and it looks for the highest-scoring place --
170
609330
2000
ื•ื”ื•ื ืžื—ืคืฉ ืืช ื”ืื–ื•ืจ ืขื ื”ื ื™ืงื•ื“ ื”ื’ื‘ื•ื” ื‘ื™ื•ืชืจ --
10:11
the stuff where the most interesting stuff is happening --
171
611330
2000
ื”ืžืงื•ื ื‘ื• ื”ื“ื‘ืจื™ื ื”ืžืขื ื™ื™ื ื™ื ื‘ื™ื•ืชืจ ืงื•ืจื™ื --
10:13
and that is what its eyes then segue to.
172
613330
4000
ื•ื–ื” ื”ืžืงื•ื ืืœื™ื• ืขื™ื ื™ื• ื–ื–ื•ืช ืœื‘ืกื•ืฃ.
10:17
And it looks right at that.
173
617330
2000
ื•ื”ื•ื ืžื‘ื™ื˜ ืžืžืฉ ื‘ื–ื”.
10:19
At the same time, some top-down sort of stuff:
174
619330
3000
ื‘ืื•ืชื• ื”ื–ืžืŸ, ื’ื ืชื”ืœื™ื›ื™ื ืฉืœ ืžืœืžืขืœื” ืœืžื˜ื”:
10:22
might decide that it's lonely and look for skin tone,
175
622330
3000
ื”ื•ื ื™ื›ื•ืœ ืœื”ื—ืœื™ื˜ ืฉื”ื•ื ื‘ื•ื“ื“ ื•ืœื—ืคืฉ ื’ื•ื•ื ื™ ืขื•ืจ,
10:25
or might decide that it's bored and look for a toy to play with.
176
625330
3000
ืื• ืœื”ื—ืœื™ื˜ ืฉื”ื•ื ืžืฉื•ืขืžื ื•ืœื—ืคืฉ ืฆืขืฆื•ืข ืœืฉื—ืง ืื™ืชื•.
10:28
And so these weights change.
177
628330
2000
ืื– ื”ืžืฉืงืœื™ื ื”ืืœื” ืžืฉืชื ื™ื.
10:30
And over here on the right,
178
630330
2000
ื•ื›ืืŸ ืžื™ืžื™ืŸ,
10:32
this is what we call the Steven Spielberg memorial module.
179
632330
3000
ื–ื” ืžื” ืฉืื ื—ื ื• ืงื•ืจืื™ื ืœื• ื”ืจื›ื™ื‘ ืœื–ื›ืจ ืกื˜ื™ื‘ืŸ ืกืคื™ืœื‘ืจื’,
10:35
Did people see the movie "AI"? (Audience: Yes.)
180
635330
2000
ื”ืื ืื ืฉื™ื ืจืื• ืืช ื”ืกืจื˜ "AI"?(ืงื”ืœ: "ื›ืŸ")
10:37
RB: Yeah, it was really bad, but --
181
637330
2000
ืจ.ื‘: ื›ืŸ, ื”ื•ื ื”ื™ื” ืžืžืฉ ื’ืจื•ืข, ืื‘ืœ --
10:39
remember, especially when Haley Joel Osment, the little robot,
182
639330
4000
ืืชื ื–ื•ื›ืจื™ื ื›ืฉื”ื™ื™ืœื™ ื’'ื• ืื•ืกืžื˜, ื”ืจื•ื‘ื•ื˜ ื”ืงื˜ืŸ,
10:43
looked at the blue fairy for 2,000 years without taking his eyes off it?
183
643330
4000
ื”ื‘ื™ื˜ ื‘ืคื™ื” ื”ื›ื—ื•ืœื” ื‘ืžืฉืš 2,000 ืฉื ื” ื‘ืœื™ ืœื”ืกื™ืจ ืžืžื ื” ืืช ืขื™ื ื™ื•?
10:47
Well, this gets rid of that,
184
647330
2000
ื•ื‘ื›ืŸ, ื–ื” ื ืคืชืจ ืžื–ื”,
10:49
because this is a habituation Gaussian that gets negative,
185
649330
4000
ืžืคื ื™ ืฉื–ื• ืขืงื•ืžืช ื”ื‘ื™ื˜ื•ืืฆื™ื” ื’ืื•ืกื™ื™ืื ื™ืช ืฉื ืขืฉื™ืช ืฉืœื™ืœื™ืช,
10:53
and more and more intense as it looks at one thing.
186
653330
3000
ื•ื™ื•ืชืจ ื•ื™ื•ืชืจ ื—ื–ืงื” ื›ืฉื”ื•ื ืžื‘ื™ื˜ ื‘ืžืฉื”ื• ืื—ื“.
10:56
And it gets bored, so it will then look away at something else.
187
656330
3000
ื•ื”ื•ื ืžืฉืชืขืžื, ืื– ื”ื•ื ื™ืกืชื›ืœ ืขืœ ืžืฉื”ื• ืื—ืจ.
10:59
So, once you've got that -- and here's a robot, here's Kismet,
188
659330
4000
ืื–, ื‘ืจื’ืข ืฉื”ื‘ื ืชื ืืช ื–ื” -- ื•ื”ื ื” ืจื•ื‘ื•ื˜, ื”ื ื” ืงื™ื–ืžื˜,
11:03
looking around for a toy. You can tell what it's looking at.
189
663330
4000
ืžื—ืคืฉ ื‘ืกื‘ื™ื‘ื” ืฆืขืฆื•ืข.
11:07
You can estimate its gaze direction from those eyeballs covering its camera,
190
667330
5000
ืืชื ื™ื›ื•ืœื™ื ืœื”ืขืจื™ืš ืืช ื›ื™ื•ื•ืŸ ื”ืžื‘ื˜ ืฉืœื• ืžื’ืœื’ืœื™ ื”ืขื™ืŸ ื”ืืœื” ืฉืžื›ืกื™ื ืืช ื”ืžืฆืœืžื”,
11:12
and you can tell when it's actually seeing the toy.
191
672330
3000
ื•ืืชื ื™ื›ื•ืœื™ื ืœืจืื•ืช ืžืชื™ ื”ื•ื ื‘ืขืฆื ืจื•ืื” ืืช ื”ืฆืขืฆื•ืข.
11:15
And it's got a little bit of an emotional response here.
192
675330
2000
ื•ื™ืฉ ืœื• ืชื’ื•ื‘ื” ืงืฆืช ืจื’ืฉื™ืช ืคื”.
11:17
(Laughter)
193
677330
1000
(ืฆื—ื•ืง)
11:18
But it's still going to pay attention
194
678330
2000
ืื‘ืœ ื”ื•ื ืขื“ื™ื™ืŸ ืขื•ืžื“ ืœืฉื™ื ืœื‘
11:20
if something more significant comes into its field of view --
195
680330
4000
ืื ืžืฉื”ื• ื™ื•ืชืจ ื—ืฉื•ื‘ ื™ื›ื ืก ืœืฉื“ื” ื”ืจืื™ื” ืฉืœื• --
11:24
such as Cynthia Breazeal, the builder of this robot, from the right.
196
684330
4000
ื›ืžื• ืกื™ื ืชื™ื” ื‘ืจื–ื™ืœ, ื”ื‘ื•ื ื” ืฉืœ ื”ืจื•ื‘ื•ื˜, ืžื™ืžื™ืŸ.
11:28
It sees her, pays attention to her.
197
688330
5000
ื”ื•ื ืจื•ืื” ืื•ืชื”, ื•ืชืฉื•ืžืช ื”ืœื‘ ืฉืœื• ืคื•ื ื” ืืœื™ื”.
11:33
Kismet has an underlying, three-dimensional emotional space,
198
693330
4000
ืœืงื™ื–ืžื˜ ื™ืฉ ืžืจื—ื‘ ืจื’ืฉื™ ืชืœืช ืžื™ืžื“ื™ ื‘ื‘ืกื™ืกื•,
11:37
a vector space, of where it is emotionally.
199
697330
3000
ืžืจื—ื‘ ื•ืงื˜ื•ืจื™, ืฉืœ ื”ื™ื›ืŸ ื”ื•ื ื ืžืฆื ืจื’ืฉื™ืช.
11:40
And at different places in that space, it expresses --
200
700330
5000
ื•ื‘ืžืงื•ืžื•ืช ืฉื•ื ื™ื ื‘ืžืจื—ื‘ ื”ื–ื”, ื”ื•ื ืžื‘ื™ืข --
11:46
can we have the volume on here?
201
706330
2000
ืืคืฉืจ ืœื”ื’ื‘ื™ืจ ืืช ื”ื•ื•ืœื™ื•ื ืคื”?
11:48
Can you hear that now, out there? (Audience: Yeah.)
202
708330
2000
ืืชื ื™ื›ื•ืœื™ื ืœืฉืžื•ืข ืืช ื–ื” ืขื›ืฉื™ื•, ืฉื? (ืงื”ืœ: ื›ืŸ)
11:50
Kismet: Do you really think so? Do you really think so?
203
710330
5000
ืงื™ื–ืžื˜: ืืชื ื‘ืืžืช ื—ื•ืฉื‘ื™ื ื›ื›ื”? ืืชื ื‘ืืžืช ื—ื•ืฉื‘ื™ื ื›ื›ื”?
11:57
Do you really think so?
204
717330
2000
ืืชื ื‘ืืžืช ื—ื•ืฉื‘ื™ื ื›ื›ื”?
12:00
RB: So it's expressing its emotion through its face
205
720330
3000
ืจ.ื‘: ืื– ื”ื•ื ืžื‘ื˜ื ืืช ืจื’ืฉื•ืชื™ื• ื“ืจืš ืคื ื™ื•
12:03
and the prosody in its voice.
206
723330
2000
ื•ื‘ืžืฉืงืœ ืงื•ืœื•.
12:05
And when I was dealing with my robot over here,
207
725330
4000
ื•ื›ืฉื˜ื™ืคืœืชื™ ื‘ืจื•ื‘ื•ื˜ ืฉืœื™ ื›ืืŸ,
12:09
Chris, the robot, was measuring the prosody in my voice,
208
729330
3000
ื›ืจื™ืก, ื”ืจื•ื‘ื•ื˜, ืžื“ื“ ืืช ื”ื˜ื•ืŸ ื‘ืงื•ืœื™,
12:12
and so we have the robot measure prosody for four basic messages
209
732330
5000
ืื– ื™ืฉ ืœื ื• ืืช ื”ืจื•ื‘ื•ื˜ ืžื•ื“ื“ ื˜ื•ื ื™ื ืœืืจื‘ืขื” ืžืกืจื™ื ื‘ืกื™ืกื™ื™ื
12:17
that mothers give their children pre-linguistically.
210
737330
4000
ืฉืื ื ื•ืชื ืช ืœื™ืœื“ื™ื” ืœืคื ื™ ืฉื”ื ืžื“ื‘ืจื™ื.
12:21
Here we've got naive subjects praising the robot:
211
741330
3000
ื›ืืŸ ื™ืฉ ืœื ื• ื ืกื™ื™ื ื™ื ื ืื™ื‘ื™ื ืฉืžืฉื‘ื—ื™ื ืืช ื”ืจื•ื‘ื•ื˜:
12:26
Voice: Nice robot.
212
746330
2000
ืงื•ืœ: ืจื•ื‘ื•ื˜ ื ื—ืžื“.
12:29
You're such a cute little robot.
213
749330
2000
ืืชื” ื›ื–ื” ืจื•ื‘ื•ื˜ ืงื˜ืŸ ื•ื—ืžื•ื“.
12:31
(Laughter)
214
751330
2000
(ืฆื—ื•ืง)
12:33
RB: And the robot's reacting appropriately.
215
753330
2000
ืจ.ื‘: ื•ื”ืจื•ื‘ื•ื˜ ืžื’ื™ื‘ ื‘ื”ืชืื.
12:35
Voice: ...very good, Kismet.
216
755330
4000
ืงื•ืœ: ... ื˜ื•ื‘ ืžืื•ื“, ืงื™ื–ืžื˜.
12:40
(Laughter)
217
760330
2000
(ืฆื—ื•ืง)
12:42
Voice: Look at my smile.
218
762330
2000
ืงื•ืœ: ืชื‘ื™ื˜ ื‘ื—ื™ื•ืš ืฉืœื™.
12:46
RB: It smiles. She imitates the smile. This happens a lot.
219
766330
3000
ืจ.ื‘: ื”ื•ื ืžื—ื™ื™ืš. ื”ื™ื ืžื—ืงื” ืืช ื”ื—ื™ื•ืš. ื–ื” ืงื•ืจื” ื”ืจื‘ื”.
12:49
These are naive subjects.
220
769330
2000
ืืœื” ื ืกื™ื™ื ื™ื ื ืื™ื‘ื™ื™ื.
12:51
Here we asked them to get the robot's attention
221
771330
3000
ืคื” ื‘ื™ืงืฉื ื• ืžื”ื ืœืชืคื•ืฉ ืืช ืชืฉื•ืžืช ื”ืœื‘ ืฉืœ ื”ืจื•ื‘ื•ื˜
12:54
and indicate when they have the robot's attention.
222
774330
3000
ื•ืœืกืžืŸ ืžืชื™ ื”ื ืชืคืฉื• ืืช ืชืฉื•ืžืช ืœื™ื‘ื•.
12:57
Voice: Hey, Kismet, ah, there it is.
223
777330
4000
ืงื•ืœ: ื”ื™, ืงื™ื–ืžื˜, ืื”, ื”ื ื” ื–ื”.
13:01
RB: So she realizes she has the robot's attention.
224
781330
4000
ืจ.ื‘: ืื– ื”ื™ื ืจื•ืื” ืฉื”ื™ื ืงื™ื‘ืœื” ืืช ืชืฉื•ืžืช ื”ืœื‘ ืฉืœ ื”ืจื•ื‘ื•ื˜.
13:08
Voice: Kismet, do you like the toy? Oh.
225
788330
4000
ืงื•ืœ: ืงื™ืกืžื˜, ืืชื” ืื•ื”ื‘ ืืช ื”ืฆืขืฆื•ืข? ืื•ื”.
13:13
RB: Now, here they're asked to prohibit the robot,
226
793330
2000
ืจ.ื‘: ืขื›ืฉื™ื•, ื”ื ืžืชื‘ืงืฉื™ื ืœืืกื•ืจ ืขืœ ื”ืจื•ื‘ื•ื˜,
13:15
and this first woman really pushes the robot into an emotional corner.
227
795330
4000
ื•ื”ืืฉื” ื”ืจืืฉื•ื ื” ืžืžืฉ ื“ื•ื—ืงืช ืืช ื”ืจื•ื‘ื•ื˜ ืœืคื™ื ื” ื”ืจื’ืฉื™ืช.
13:19
Voice: No. No. You're not to do that. No.
228
799330
5000
ืงื•ืœ: ืœื. ืœื. ืืกื•ืจ ืœืš ืœืขืฉื•ืช ืืช ื–ื”. ืœื.
13:24
(Laughter)
229
804330
3000
(ืฆื—ื•ืง)
13:27
Not appropriate. No. No.
230
807330
6000
ื–ื” ืœื ื‘ืกื“ืจ. ืœื. ืœื.
13:33
(Laughter)
231
813330
3000
(ืฆื—ื•ืง)
13:36
RB: I'm going to leave it at that.
232
816330
2000
ืจ.ื‘: ืื ื™ ืืฉืื™ืจ ืืช ื–ื” ื›ื›ื”.
13:38
We put that together. Then we put in turn taking.
233
818330
2000
ืื ื—ื ื• ืžื—ื‘ืจื™ื ืืช ื–ื”. ืื– ืื ื—ื ื• ืžื•ืกื™ืคื™ื ืชืงืฉื•ืจืช ืœืคื™ ื”ืชื•ืจ.
13:40
When we talk to someone, we talk.
234
820330
3000
ื›ืฉืื ื—ื ื• ืžื“ื‘ืจื™ื ืขื ืžื™ืฉื”ื•, ืื ื—ื ื• ืžื“ื‘ืจื™ื,
13:43
Then we sort of raise our eyebrows, move our eyes,
235
823330
4000
ื•ืื– ืื ื—ื ื• ืžืจื™ืžื™ื ืืช ื’ื‘ื•ืชื™ื ื•, ืžื–ื™ื–ื™ื ืืช ืขื™ื ื™ื ื•,
13:47
give the other person the idea it's their turn to talk.
236
827330
3000
ืจื•ืžื–ื™ื ืœืื“ื ืฉืชื•ืจื• ืœื“ื‘ืจ.
13:50
And then they talk, and then we pass the baton back and forth between each other.
237
830330
4000
ื•ืื– ื”ื ืžื“ื‘ืจื™ื, ื•ืื– ืื ื—ื ื• ืžืขื‘ื™ืจื™ื ืืช ื”ืžืงืœ ืžื”ืื—ื“ ืœืฉื ื™
13:54
So we put this in the robot.
238
834330
2000
ืื– ืื ื—ื ื• ืฉืžื™ื ืืช ื–ื” ื‘ืจื•ื‘ื•ื˜.
13:56
We got a bunch of naive subjects in,
239
836330
2000
ื”ื›ื ืกื ื• ื—ื‘ื•ืจืช ื ืกื™ื™ื ื™ื ื ืื™ื‘ื™ื™ื ืคื ื™ืžื”,
13:58
we didn't tell them anything about the robot,
240
838330
2000
ืœื ืืžืจื ื• ืœื”ื ืฉื•ื ื“ื‘ืจ ืขืœ ื”ืจื•ื‘ื•ื˜,
14:00
sat them down in front of the robot and said, talk to the robot.
241
840330
2000
ื”ื•ืฉื‘ื ื• ืื•ืชื ืžื•ืœ ื”ืจื•ื‘ื•ื˜ ื•ืืžืจื ื•, ื“ื‘ืจื• ืœืจื•ื‘ื•ื˜.
14:02
Now what they didn't know was,
242
842330
2000
ืขื›ืฉื™ื• ืžื” ืฉื”ื ืœื ื™ื“ืขื• ื”ื™ื”,
14:04
the robot wasn't understanding a word they said,
243
844330
2000
ื”ืจื•ื‘ื•ื˜ ืœื ื”ื‘ื™ืŸ ืžื™ืœื” ืžืžื” ืฉื”ื ืืžืจื•,
14:06
and that the robot wasn't speaking English.
244
846330
3000
ื•ื–ื” ืฉื”ืจื•ื‘ื•ื˜ ืœื ื“ื™ื‘ืจ ืื ื’ืœื™ืช.
14:09
It was just saying random English phonemes.
245
849330
2000
ื”ื•ื ืจืง ืืžืจ ืคื•ื ืžื•ืช ื‘ืื ื’ืœื™ืช.
14:11
And I want you to watch carefully, at the beginning of this,
246
851330
2000
ื•ืื ื™ ืจื•ืฆื” ืฉืชื‘ื™ื˜ื• ื”ื™ื˜ื‘, ื‘ื”ืชื—ืœื” ืฉืœ ื–ื”,
14:13
where this person, Ritchie, who happened to talk to the robot for 25 minutes --
247
853330
4000
ืคื” ื”ืื“ื ื”ื–ื”, ืจื™ืฆ'ื™, ืฉื‘ืžืงืจื” ื“ื™ื‘ืจ ืขื ื”ืจื•ื‘ื•ื˜ ื‘ืžืฉืš 25 ื“ืงื•ืช --
14:17
(Laughter)
248
857330
2000
(ืฆื—ื•ืง)
14:19
-- says, "I want to show you something.
249
859330
2000
-- ืืžืจ, "ืื ื™ ืจื•ืฆื” ืœื”ืจืื•ืช ืœืš ืžืฉื”ื•.
14:21
I want to show you my watch."
250
861330
2000
ืื ื™ ืจื•ืฆื” ืœื”ืจืื•ืช ืœืš ืืช ื”ืฉืขื•ืŸ ืฉืœื™."
14:23
And he brings the watch center, into the robot's field of vision,
251
863330
5000
ื•ื”ื•ื ืžื‘ื™ื ืืช ื”ืฉืขื•ืŸ ืœืžืจื›ื–, ืœื˜ื•ื•ื— ื”ืจืื™ื” ืฉืœ ื”ืจื•ื‘ื•ื˜,
14:28
points to it, gives it a motion cue,
252
868330
2000
ืžืฆื‘ื™ืข ืขืœื™ื•, ื ื•ืชืŸ ืœื• ืจืžื– ืชื ื•ืขื”,
14:30
and the robot looks at the watch quite successfully.
253
870330
2000
ื•ื”ืจื•ื‘ื•ื˜ ืžื‘ื™ื˜ ื‘ืฉืขื•ืŸ ื“ื™ ื‘ื”ืฆืœื—ื”.
14:32
We don't know whether he understood or not that the robot --
254
872330
3000
ืื ื—ื ื• ืœื ื™ื•ื“ืขื™ื ื‘ืืžืช ืื ื”ื•ื ืžื‘ื™ืŸ ืื• ืœื ืฉื”ืจื•ื‘ื•ื˜ --
14:36
Notice the turn-taking.
255
876330
2000
ืžื‘ื—ื™ืŸ ื‘ืชื•ืจื•ืช.
14:38
Ritchie: OK, I want to show you something. OK, this is a watch
256
878330
3000
ืจื™ืฆ'ื™: ืื•ืงื™ื™, ืื ื™ ืจื•ืฆื” ืœื”ืจืื•ืช ืœืš ืžืฉื”ื•. ืื•ืงื™ื™, ื–ื” ืฉืขื•ืŸ
14:41
that my girlfriend gave me.
257
881330
3000
ืฉื—ื‘ืจื” ืฉืœื™ ื ืชื ื” ืœื™.
14:44
Robot: Oh, cool.
258
884330
2000
ืจื•ื‘ื•ื˜: ืื•ื”, ืงื•ืœ.
14:46
Ritchie: Yeah, look, it's got a little blue light in it too. I almost lost it this week.
259
886330
4000
ืจื™ืฆ'ื™" ื›ืŸ, ืชืจืื”, ื™ืฉ ืœื• ื’ื ืื•ืจ ื›ื—ื•ืœ. ื›ืžืขื˜ ืื™ื‘ื“ืชื™ ืื•ืชื• ื”ืฉื‘ื•ืข.
14:51
(Laughter)
260
891330
4000
(ืฆื—ื•ืง)
14:55
RB: So it's making eye contact with him, following his eyes.
261
895330
3000
ืจ.ื‘: ืื– ื”ื•ื ื™ื•ืฆืจ ืงืฉืจ ืขื™ืŸ ืื™ืชื•, ืขื•ืงื‘ ืื—ืจื™ ืขื™ื ื™ื•.
14:58
Ritchie: Can you do the same thing? Robot: Yeah, sure.
262
898330
2000
ืจื™ืฆ'ื™: ืืชื” ื™ื›ื•ืœ ืœืขืฉื•ืช ืื•ืชื• ื”ื“ื‘ืจ? ืจื•ื‘ื•ื˜: ื›ืŸ, ื‘ื•ื•ื“ืื™.
15:00
RB: And they successfully have that sort of communication.
263
900330
2000
ืจ.ื‘: ื•ื”ื ืžืชืงืฉืจื™ื ื›ื›ื” ื‘ื”ืฆืœื—ื”.
15:02
And here's another aspect of the sorts of things that Chris and I were doing.
264
902330
4000
ื•ื™ืฉ ืขื•ื“ ืคืŸ ืœืกื•ื’ ื”ื“ื‘ืจื™ื ื”ืืœื” ืฉื›ืจื™ืก ื•ืื ื™ ืขื•ืฉื™ื.
15:06
This is another robot, Cog.
265
906330
2000
ื–ื” ืขื•ื“ ืจื•ื‘ื•ื˜, ืงื•ื’.
15:08
They first make eye contact, and then, when Christie looks over at this toy,
266
908330
6000
ื”ื ืงื•ื“ื ื™ืฆืจื• ืงืฉืจ ืขื™ืŸ, ื•ืื–, ื›ืฉื›ืจื™ืกื˜ื™ ืžื‘ื™ื˜ื” ืœืขื‘ืจ ื”ืฆืขืฆื•ืข ื”ื–ื”,
15:14
the robot estimates her gaze direction
267
914330
2000
ื”ืจื•ื‘ื•ื˜ ืžืขืจื™ืš ืืช ื›ื™ื•ื•ืŸ ื”ืžื‘ื˜ ืฉืœื”
15:16
and looks at the same thing that she's looking at.
268
916330
2000
ื•ืžื‘ื™ื˜ ื‘ืื•ืชื• ื”ื“ื‘ืจ ืฉื”ื™ื ืžื‘ื™ื˜ื” ื‘ื•.
15:18
(Laughter)
269
918330
1000
(ืฆื—ื•ืง)
15:19
So we're going to see more and more of this sort of robot
270
919330
3000
ืื– ืื ื—ื ื• ื ืจืื” ืขื•ื“ ื•ืขื•ื“ ืžื”ืกื•ื’ ื”ื–ื” ืฉืœ ืจื•ื‘ื•ื˜ื™ื
15:22
over the next few years in labs.
271
922330
2000
ื‘ืžืฉืš ื”ืฉื ื™ื ื”ืงืจื•ื‘ื•ืช ื‘ืžืขื‘ื“ื•ืช.
15:24
But then the big questions, two big questions that people ask me are:
272
924330
5000
ืื‘ืœ ืื– ื”ืฉืืœื” ื”ื’ื“ื•ืœื”, ืฉืชื™ ืฉืืœื•ืช ื’ื“ื•ืœื•ืช ืฉืื ืฉื™ื ืฉื•ืืœื™ื ืื•ืชื™:
15:29
if we make these robots more and more human-like,
273
929330
2000
ืื ื ืขืฉื” ืืช ื”ืจื•ื‘ื•ื˜ื™ื ื”ืืœื” ื™ื•ืชืจ ื•ื™ื•ืชืจ ื“ื•ืžื™ื ืœืื ืฉื™ื,
15:31
will we accept them, will we -- will they need rights eventually?
274
931330
5000
ื”ืื ื ืงื‘ืœ ืื•ืชื, ื”ืื -- ื”ืื ื”ื ื™ืฆื˜ืจื›ื• ื–ื›ื•ื™ื•ืช ื‘ืกื•ืคื• ืฉืœ ื“ื‘ืจ?
15:36
And the other question people ask me is, will they want to take over?
275
936330
3000
ื•ื”ืฉืืœื” ื”ืฉื ื™ื” ืฉืื ืฉื™ื ืฉื•ืืœื™ื ืื•ืชื™ ื”ื™ื, ื”ืื ื”ื ื™ืจืฆื• ืœื”ืฉืชืœื˜?
15:39
(Laughter)
276
939330
1000
(ืฆื—ื•ืง)
15:40
And on the first -- you know, this has been a very Hollywood theme
277
940330
3000
ื•ื‘ื ื•ื’ืข ืœืจืืฉื•ื ื” -- ืืชื ื™ื•ื“ืขื™ื, ื–ื” ื”ื™ื” ื ื•ืฉื ืžืื•ื“ ื”ื•ืœื™ื•ื•ื“ื™
15:43
with lots of movies. You probably recognize these characters here --
278
943330
3000
ืขื ื”ืจื‘ื” ืกืจื˜ื™ื. ืืชื ื‘ื˜ื— ืžื–ื”ื™ื ื›ืžื” ืžื”ื“ืžื•ื™ื•ืช ื”ืืœื” --
15:46
where in each of these cases, the robots want more respect.
279
946330
4000
ืฉื‘ื›ืœ ืื—ื“ ืžื”ืžืงืจื™ื ื”ืืœื”, ื”ืจื•ื‘ื•ื˜ื™ื ืจืฆื• ื™ื•ืชืจ ื›ื‘ื•ื“.
15:50
Well, do you ever need to give robots respect?
280
950330
3000
ื•ื‘ื›ืŸ, ื”ืื ืื™ ืคืขื ื ืฆื˜ืจืš ืœืชืช ืœืจื•ื‘ื•ื˜ื™ื ื›ื‘ื•ื“?
15:54
They're just machines, after all.
281
954330
2000
ื”ื ืจืง ืžื›ื•ื ื•ืช, ืื—ืจื™ ื”ื›ืœ.
15:56
But I think, you know, we have to accept that we are just machines.
282
956330
4000
ืื‘ืœ ืื ื™ ื—ื•ืฉื‘, ืืชื ื™ื•ื“ืขื™ื, ืื ื—ื ื• ื—ื™ื™ื‘ื™ื ืœืงื‘ืœ ืฉืื ื—ื ื• ืจืง ืžื›ื•ื ื•ืช.
16:00
After all, that's certainly what modern molecular biology says about us.
283
960330
5000
ืื—ืจื™ ื”ื›ืœ, ื–ื” ื‘ื”ื—ืœื˜ ืžื” ืฉื‘ื™ื•ืœื•ื’ื™ื” ืžื•ืœืงื•ืœืจื™ืช ืžื•ื“ืจื ื™ืช ืื•ืžืจืช ืขืœื™ื ื•.
16:05
You don't see a description of how, you know,
284
965330
3000
ืืชื ืœื ืจื•ืื™ื ืชืื•ืจ ืฉืœ ืื™ืš, ืืชื ื™ื•ื“ืขื™ื,
16:08
Molecule A, you know, comes up and docks with this other molecule.
285
968330
4000
ืžื•ืœืงื•ืœื” ื, ืืชื ื™ื•ื“ืขื™ื, ื‘ืื” ื•ืžืชื—ื‘ืจืช ืขื ืžื•ืœืงื•ืœื” ืื—ืจืช.
16:12
And it's moving forward, you know, propelled by various charges,
286
972330
3000
ื•ื”ื™ื ื–ื–ื” ืงื“ื™ืžื”, ืืชื ื™ื•ื“ืขื™ื, ืžื•ื ืขืช ืขืœ ื™ื“ื™ ืžืชื—ื™ื ืฉื•ื ื™ื,
16:15
and then the soul steps in and tweaks those molecules so that they connect.
287
975330
4000
ื•ืื– ื”ื ืฉืžื” ื‘ืื” ื•ืžืกื“ืจืช ืืช ื”ืžื•ืœืงื•ืœื•ืช ื›ืš ืฉื”ืŸ ื™ืชื—ื‘ืจื•.
16:19
It's all mechanistic. We are mechanism.
288
979330
3000
ื–ื” ื”ื›ืœ ืžื›ืื ื™ืกื˜ื™. ืื ื—ื ื• ืžื›ื•ื ื•ืช.
16:22
If we are machines, then in principle at least,
289
982330
3000
ืื ืื ื—ื ื• ืžื›ื•ื ื•ืช, ืื– ื‘ืขื™ืงืจื•ืŸ ืœืคื—ื•ืช,
16:25
we should be able to build machines out of other stuff,
290
985330
4000
ืื ื—ื ื• ืฆืจื™ื›ื™ื ืœื”ื™ื•ืช ืžืกื•ื’ืœื™ื ืœื‘ื ื•ืช ืžื›ื•ื ื•ืช ืžื“ื‘ืจื™ื ืื—ืจื™ื,
16:29
which are just as alive as we are.
291
989330
4000
ืฉื”ืŸ ืœืคื—ื•ืช ื—ื™ื•ืช ื›ืžื•ื ื•.
16:33
But I think for us to admit that,
292
993330
2000
ืื‘ืœ ืื ื™ ื—ื•ืฉื‘ ืฉื‘ืฉื‘ื™ืœื ื• ืœื”ื•ื“ื•ืช ื‘ื–ื”,
16:35
we have to give up on our special-ness, in a certain way.
293
995330
3000
ืื ื—ื ื• ืฆืจื™ื›ื™ื ืœื•ื•ืชืจ ืขืœ ื”ืžื™ื•ื—ื“ื•ืช ืฉืœื ื•, ื‘ืฆื•ืจื” ืžืกื•ื™ื™ืžืช.
16:38
And we've had the retreat from special-ness
294
998330
2000
ื•ื”ื™ื™ืชื” ืœื ื• ืืช ื”ื ืกื™ื’ื” ืžืžื™ื•ื—ื“ื•ืช
16:40
under the barrage of science and technology many times
295
1000330
3000
ืชื—ืช ืžืชืงืคืช ื”ืžื“ืข ื•ื”ื˜ื›ื ื•ืœื•ื’ื™ื” ื”ืจื‘ื” ืคืขืžื™ื
16:43
over the last few hundred years, at least.
296
1003330
2000
ื‘ืžื”ืœืš ื›ืžื” ืžืื•ืช ื”ืฉื ื™ื ื”ืื—ืจื•ื ื•ืช, ืœืคื—ื•ืช.
16:45
500 years ago we had to give up the idea
297
1005330
2000
ืœืคื ื™ 500 ืฉื ื” ื”ื™ื™ื ื• ืฆืจื™ื›ื™ื ืœื•ื•ืชืจ ืขืœ ื”ืจืขื™ื•ืŸ
16:47
that we are the center of the universe
298
1007330
3000
ืฉืื ื—ื ื• ืžืจื›ื– ื”ื™ืงื•ื
16:50
when the earth started to go around the sun;
299
1010330
2000
ื›ืฉื›ื“ื•ืจ ื”ืืจืฅ ื”ืชื—ื™ืœ ืœื”ืกืชื•ื‘ื‘ ืกื‘ื™ื‘ ื”ืฉืžืฉ;
16:52
150 years ago, with Darwin, we had to give up the idea we were different from animals.
300
1012330
5000
ืœืคื ื™ 150 ืฉื ื”, ืขื ื“ืืจื•ื•ื™ืŸ, ื”ื™ื™ื ื• ืฆืจื™ื›ื™ื ืœื•ื•ืชืจ ืขืœ ื”ืจืขื™ื•ืŸ ืฉืื ื—ื ื• ืฉื•ื ื™ื ืžื”ื—ื™ื•ืช.
16:57
And to imagine -- you know, it's always hard for us.
301
1017330
3000
ื•ืœื“ืžื™ื™ืŸ -- ืืชื ื™ื•ื“ืขื™ื, ื–ื” ืชืžื™ื“ ืงืฉื” ื‘ืฉื‘ื™ืœื ื•,
17:00
Recently we've been battered with the idea that maybe
302
1020330
3000
ืœืื—ืจื•ื ื” ื”ื•ืขืžื“ื ื• ื‘ืคื ื™ ื”ืจืขื™ื•ืŸ ืฉืื•ืœื™
17:03
we didn't even have our own creation event, here on earth,
303
1023330
2000
ืืคื™ืœื• ืœื ื”ื™ื” ืœื ื• ืืช ืืจื•ืข ื”ื‘ืจื™ืื” ืฉืœื ื•, ื›ืืŸ ืขืœ ื›ื“ื•ืจ ื”ืืจืฅ,
17:05
which people didn't like much. And then the human genome said,
304
1025330
3000
ืฉืื ืฉื™ื ืœื ืžืžืฉ ืื”ื‘ื•. ื•ืื– ื”ื’ื ื•ื ื”ืื ื•ืฉื™ ืืžืจ,
17:08
maybe we only have 35,000 genes. And that was really --
305
1028330
3000
ืื•ืœื™ ื™ืฉ ืœื ื• ืจืง 35,000 ื’ื ื™ื. ื•ื–ื” ื”ื™ื” ืžืžืฉ --
17:11
people didn't like that, we've got more genes than that.
306
1031330
3000
ืื ืฉื™ื ืœื ืื”ื‘ื• ืืช ื–ื”, ื™ืฉ ืœื ื• ื™ื•ืชืจ ื’ื ื™ื ืžื–ื”.
17:14
We don't like to give up our special-ness, so, you know,
307
1034330
3000
ืื ื—ื ื• ืœื ืื•ื”ื‘ื™ื ืœื•ื•ืชืจ ืขืœ ื”ืžื™ื•ื—ื“ื•ืช ืฉืœื ื•, ืื–, ืืชื ื™ื•ื“ืขื™ื,
17:17
having the idea that robots could really have emotions,
308
1037330
2000
ื”ืจืขื™ื•ืŸ ืฉืœืจื•ื‘ื•ื˜ื™ื ื‘ืืžืช ื™ื›ื•ืœื™ื ืœื”ื™ื•ืช ืจื’ืฉื•ืช,
17:19
or that robots could be living creatures --
309
1039330
2000
ืื• ืฉืจื•ื‘ื•ื˜ื™ื ื™ื›ื•ืœื™ื ืœื”ื™ื•ืช ื™ืฆื•ืจื™ื ื—ื™ื™ื --
17:21
I think is going to be hard for us to accept.
310
1041330
2000
ืื ื™ ื—ื•ืฉื‘ ืฉื–ื” ื™ื”ื™ื” ืงืฉื” ืœื ื• ืœืงื‘ืœ.
17:23
But we're going to come to accept it over the next 50 years or so.
311
1043330
4000
ืื‘ืœ ืื ื—ื ื• ืขื•ืžื“ื™ื ืœืงื‘ืœ ืืช ื–ื” ื‘ืžืฉืš ื” 50 ืฉื ื™ื ื”ื‘ืื•ืช.
17:27
And the second question is, will the machines want to take over?
312
1047330
3000
ื•ื”ืฉืืœื” ื”ืฉื ื™ื” ื”ื™ื, ื”ืื ื”ืžื›ื•ื ื•ืช ื™ืจืฆื• ืœื”ืฉืชืœื˜?
17:30
And here the standard scenario is that we create these things,
313
1050330
5000
ื•ื›ืืŸ ื”ืชืกืจื™ื˜ ื”ืกื˜ื ื“ืจื˜ื™ ื”ื•ื ืฉืื ื—ื ื• ื™ื•ืฆืจื™ื ืืช ื”ื“ื‘ืจื™ื ื”ืืœื”,
17:35
they grow, we nurture them, they learn a lot from us,
314
1055330
3000
ื”ื ื’ื“ืœื™ื, ืื ื—ื ื• ืžื˜ืคื—ื™ื ืื•ืชื, ื”ื ืœื•ืžื“ื™ื ื”ืจื‘ื” ืžืื™ืชื ื•,
17:38
and then they start to decide that we're pretty boring, slow.
315
1058330
4000
ื•ืื– ื”ื ืžืชื—ื™ืœื™ื ืœื”ื—ืœื™ื˜ ืฉืื ื—ื ื• ื“ื™ ืžืฉืขืžืžื™ื, ืื™ื˜ื™ื™ื.
17:42
They want to take over from us.
316
1062330
2000
ื”ื ืจื•ืฆื™ื ืœื”ืฉืชืœื˜.
17:44
And for those of you that have teenagers, you know what that's like.
317
1064330
3000
ื•ืœืืœื” ืžื›ื ืฉื™ืฉ ื‘ื ื™ ืขืฉืจื”, ืืชื ื™ื•ื“ืขื™ื ืื™ืš ื–ื”.
17:47
(Laughter)
318
1067330
1000
(ืฆื—ื•ืง)
17:48
But Hollywood extends it to the robots.
319
1068330
3000
ืื‘ืœ ื”ื•ืœื™ื•ื•ื“ ืžืจื—ื™ื‘ื” ืืช ื–ื” ืœืจื•ื‘ื•ื˜ื™ื.
17:51
And the question is, you know,
320
1071330
3000
ื•ื”ืฉืืœื” ื”ื™ื, ืืชื ื™ื•ื“ืขื™ื,
17:54
will someone accidentally build a robot that takes over from us?
321
1074330
4000
ื”ืื ืžื™ืฉื”ื• ื‘ื˜ืขื•ืช ื™ื‘ื ื” ืจื•ื‘ื•ื˜ ืฉืžืฉืชืœื˜ ืขืœื™ื ื•?
17:58
And that's sort of like this lone guy in the backyard,
322
1078330
3000
ื•ื–ื” ืกื•ื’ ื”ืื“ื ื”ืžืชื‘ื•ื“ื“ ื‘ื—ืฆืจ ื”ืื—ื•ืจื™ืช,
18:01
you know -- "I accidentally built a 747."
323
1081330
3000
ืืชื ื™ื•ื“ืขื™ื -- "ื‘ื˜ืขื•ืช ื‘ื ื™ืชื™ 747."
18:04
I don't think that's going to happen.
324
1084330
2000
ืื ื™ ืœื ื—ื•ืฉื‘ ืฉื–ื” ื™ืงืจื”.
18:06
And I don't think --
325
1086330
2000
ื•ืื ื™ ืœื ื—ื•ืฉื‘ --
18:08
(Laughter)
326
1088330
1000
(ืฆื—ื•ืง)
18:09
-- I don't think we're going to deliberately build robots
327
1089330
3000
-- ืื ื™ ืœื ื—ื•ืฉื‘ ืฉืื ื—ื ื• ื‘ื›ื•ื•ื ื” ื”ื•ืœื›ื™ื ืœื‘ื ื•ืช ืจื•ื‘ื•ื˜ื™ื
18:12
that we're uncomfortable with.
328
1092330
2000
ืฉืœื ื ื•ื— ืœื ื• ืื™ืชื.
18:14
We'll -- you know, they're not going to have a super bad robot.
329
1094330
2000
ื•ื‘ื›ืŸ -- ืืชื ื™ื•ื“ืขื™ื, ืœื ืขื•ืžื“ ืœื”ื™ื•ืช ืœื”ื ืจื•ื‘ื•ื˜ ืกื•ืคืจ ืจืข.
18:16
Before that has to come to be a mildly bad robot,
330
1096330
3000
ืœืคื ื™ ื–ื” ืฆืจื™ืš ืœื”ื’ื™ืข ื”ืจื•ื‘ื•ื˜ ื”ืงืฆืช ืจืข,
18:19
and before that a not so bad robot.
331
1099330
2000
ื•ืœืคื ื™ ื–ื” ืจื•ื‘ื•ื˜ ืœื ื›ืœ ื›ืš ื ื•ืจื.
18:21
(Laughter)
332
1101330
1000
(ืฆื—ื•ืง)
18:22
And we're just not going to let it go that way.
333
1102330
2000
ื•ืื ื—ื ื• ืคืฉื•ื˜ ืœื ื ื™ืชืŸ ืœื–ื” ืœื”ืชืคืชื— ืœื›ื™ื•ื•ืŸ ื”ื–ื”.
18:24
(Laughter)
334
1104330
1000
(ืฆื—ื•ืง)
18:25
So, I think I'm going to leave it at that: the robots are coming,
335
1105330
6000
ืื–, ืื ื™ ื—ื•ืฉื‘ ืฉืื ื™ ืืกื™ื™ื ื‘ื–ื”: ื”ืจื•ื‘ื•ื˜ื™ื ื‘ืื™ื,
18:31
we don't have too much to worry about, it's going to be a lot of fun,
336
1111330
3000
ืื™ืŸ ืœื ื• ืžืžืฉ ืกื™ื‘ื” ืœื“ืื•ื’, ื–ื” ื™ื”ื™ื” ื›ื™ืฃ ื’ื“ื•ืœ,
18:34
and I hope you all enjoy the journey over the next 50 years.
337
1114330
4000
ื•ืื ื™ ืžืงื•ื•ื” ืฉื›ื•ืœื›ื ืชื”ื ื• ืžื”ืžืกืข ื‘50 ื”ืฉื ื™ื ื”ื‘ืื•ืช.
18:38
(Applause)
338
1118330
2000
(ืžื—ื™ืื•ืช ื›ืคื™ื™ื)
ืขืœ ืืชืจ ื–ื”

ืืชืจ ื–ื” ื™ืฆื™ื’ ื‘ืคื ื™ื›ื ืกืจื˜ื•ื ื™ YouTube ื”ืžื•ืขื™ืœื™ื ืœืœื™ืžื•ื“ ืื ื’ืœื™ืช. ืชื•ื›ืœื• ืœืจืื•ืช ืฉื™ืขื•ืจื™ ืื ื’ืœื™ืช ื”ืžื•ืขื‘ืจื™ื ืขืœ ื™ื“ื™ ืžื•ืจื™ื ืžื”ืฉื•ืจื” ื”ืจืืฉื•ื ื” ืžืจื—ื‘ื™ ื”ืขื•ืœื. ืœื—ืฅ ืคืขืžื™ื™ื ืขืœ ื”ื›ืชื•ื‘ื™ื•ืช ื‘ืื ื’ืœื™ืช ื”ืžื•ืฆื’ื•ืช ื‘ื›ืœ ื“ืฃ ื•ื™ื“ืื• ื›ื“ื™ ืœื”ืคืขื™ืœ ืืช ื”ืกืจื˜ื•ืŸ ืžืฉื. ื”ื›ืชื•ื‘ื™ื•ืช ื’ื•ืœืœื•ืช ื‘ืกื ื›ืจื•ืŸ ืขื ื”ืคืขืœืช ื”ื•ื•ื™ื“ืื•. ืื ื™ืฉ ืœืš ื”ืขืจื•ืช ืื• ื‘ืงืฉื•ืช, ืื ื ืฆื•ืจ ืื™ืชื ื• ืงืฉืจ ื‘ืืžืฆืขื•ืช ื˜ื•ืคืก ื™ืฆื™ืจืช ืงืฉืจ ื–ื”.

https://forms.gle/WvT1wiN1qDtmnspy7