What's it like to be a robot? | Leila Takayama

92,734 views ・ 2018-02-16

TED


請雙擊下方英文字幕播放視頻。

譯者: Lilian Chiu 審譯者: Song Yu
00:12
You only get one chance to make a first impression,
0
12760
2656
你只有一次機會建立第一印象,
00:15
and that's true if you're a robot as well as if you're a person.
1
15440
3176
不論你是機器人或人類,都是如此。
00:18
The first time that I met one of these robots
2
18640
3016
當我初次見到一隻這種機器人,
00:21
was at a place called Willow Garage in 2008.
3
21680
2896
是 2008 年, 在一個叫柳樹車庫公司的地方。
00:24
When I went to visit there, my host walked me into the building
4
24600
3016
當我去那裡造訪時, 東道主陪我走進大樓,
00:27
and we met this little guy.
5
27640
1576
我們見到了這個小傢伙。
00:29
He was rolling into the hallway,
6
29240
1656
它在門廳跑來跑去,
00:30
came up to me, sat there,
7
30920
1816
跑到我面前,坐在那裡,
00:32
stared blankly past me,
8
32760
2256
面無表情地望穿我,
00:35
did nothing for a while,
9
35040
1656
好一陣子什麼都沒做,
00:36
rapidly spun his head around 180 degrees
10
36720
1936
快速把它的頭扭轉了180 度,
00:38
and then ran away.
11
38680
1536
然後跑走了。
00:40
And that was not a great first impression.
12
40240
2176
那不是很好的第一印象。
00:42
The thing that I learned about robots that day
13
42440
2176
我那天學到一件關於機器人的事,
00:44
is that they kind of do their own thing,
14
44640
2176
就是它們會做自己的事,
00:46
and they're not totally aware of us.
15
46840
2136
並且完全沒有意識到我們的存在。
00:49
And I think as we're experimenting with these possible robot futures,
16
49000
3239
我想,當我們在實驗這些 可能的未來機器人時,
00:52
we actually end up learning a lot more about ourselves
17
52263
2673
我們其實最後學到很多 關於我們自己的事,
00:54
as opposed to just these machines.
18
54960
1656
不只學到關於這些機器的事。
00:56
And what I learned that day
19
56640
1336
我那天學到的是,
00:58
was that I had pretty high expectations for this little dude.
20
58000
3416
我對這個小傢伙有很高的期望。
01:01
He was not only supposed to be able to navigate the physical world,
21
61440
3176
它不僅應該要能夠引導實體世界,
01:04
but also be able to navigate my social world --
22
64640
2656
也應該要能引導我的社交世界。
01:07
he's in my space; it's a personal robot.
23
67320
2176
它在我的空間中,它是個人機器人,
01:09
wWhy didn't it understand me?
24
69520
2016
為什麼它不了解我?
01:11
My host explained to me,
25
71560
1256
東道主向我解釋:
01:12
"Well, the robot is trying to get from point A to point B,
26
72840
3176
「嗯,機器人試著從A點到達B點,
01:16
and you were an obstacle in his way,
27
76040
1776
而你是擋在它路上的障礙物,
01:17
so he had to replan his path,
28
77840
2016
所以它得重新規劃它的路徑,
01:19
figure out where to go,
29
79880
1256
搞清楚要走去哪裡,
01:21
and then get there some other way,"
30
81160
1696
然後走其他的路到達目的地。」
01:22
which was actually not a very efficient thing to do.
31
82880
2456
這其實並不是很有效益的做法。
01:25
If that robot had figured out that I was a person, not a chair,
32
85360
3256
如果那隻機器人能理解 我是一個人,不是一張椅子,
01:28
and that I was willing to get out of its way
33
88640
2096
那我其實願意讓開,
01:30
if it was trying to get somewhere,
34
90760
1656
讓它能到它的目的地,
01:32
then it actually would have been more efficient
35
92440
2216
這樣就會用更有效益的方式
01:34
at getting its job done
36
94680
1256
完成它的工作,
01:35
if it had bothered to notice that I was a human
37
95960
2216
如果它花點功夫注意到我是一個人,
01:38
and that I have different affordances than things like chairs and walls do.
38
98200
3576
且我有著和椅子及牆壁 不同的能力,就可以做到這點。
01:41
You know, we tend to think of these robots as being from outer space
39
101800
3216
我們傾向於認為 這些機器人是來自外太空,
01:45
and from the future and from science fiction,
40
105040
2136
來自未來,來自科幻小說,
01:47
and while that could be true,
41
107200
1416
雖然那是有可能的,
01:48
I'd actually like to argue that robots are here today,
42
108640
2656
但其實我認為, 機器人現今就在這裡了,
01:51
and they live and work amongst us right now.
43
111320
2776
它們現在就在人類當中生活、工作。
01:54
These are two robots that live in my home.
44
114120
2856
這兩隻機器人住在我家。
01:57
They vacuum the floors and they cut the grass
45
117000
2496
它們會吸地板和除草,
01:59
every single day,
46
119520
1216
每天都做,
02:00
which is more than I would do if I actually had time to do these tasks,
47
120760
3376
就算我有時間做這些工作, 我也無法做到這麼多,
02:04
and they probably do it better than I would, too.
48
124160
2336
且它們可能也做得比我好。
02:06
This one actually takes care of my kitty.
49
126520
2496
這隻機器人會照顧我的貓。
02:09
Every single time he uses the box, it cleans it,
50
129040
2576
每當牠使用貓砂時,它就會清理,
02:11
which is not something I'm willing to do,
51
131640
1976
這就不是我願意做的事了,
02:13
and it actually makes his life better as well as mine.
52
133640
2576
它讓牠和我的生活都變得更好。
02:16
And while we call these robot products --
53
136240
2416
雖然我們稱這些機器人為產品,
02:18
it's a "robot vacuum cleaner, it's a robot lawnmower,
54
138680
2696
它是打掃機器人、 它是除草機器人、
02:21
it's a robot littler box,"
55
141400
1495
它是貓砂機器人,
02:22
I think there's actually a bunch of other robots hiding in plain sight
56
142919
4137
我認為,在我們視線能及之處 還隱藏了一堆其他的機器人,
02:27
that have just become so darn useful
57
147080
1856
只是它們變得太有用、太平凡,
02:28
and so darn mundane
58
148960
1456
以致於我們為它們另外命名, 比如「洗碗機」,對吧?
02:30
that we call them things like, "dishwasher," right?
59
150440
2496
02:32
They get new names.
60
152960
1216
它們得到新的名字。
02:34
They don't get called robot anymore
61
154200
1696
它們不再被稱為機器人,
02:35
because they actually serve a purpose in our lives.
62
155920
2416
因為,在我們的人生中, 它們有目的、用途。
02:38
Similarly, a thermostat, right?
63
158360
1496
同樣的,自動調溫器,對吧?
02:39
I know my robotics friends out there
64
159880
1776
我知道外頭那些機器人朋友們
02:41
are probably cringing at me calling this a robot,
65
161680
2336
大概對我稱呼它們為機器人 感到難為情,
02:44
but it has a goal.
66
164040
1256
但它有個目標,
02:45
Its goal is to make my house 66 degrees Fahrenheit,
67
165320
2896
它的目標就是讓我的房子 維持在華氏 66 度。
02:48
and it senses the world.
68
168240
1256
它會感知這個世界,
02:49
It knows it's a little bit cold,
69
169520
1576
知道有一點冷,
02:51
it makes a plan and then it acts on the physical world.
70
171120
2616
它會做計畫,然後 在實體世界採取行動。
02:53
It's robotics.
71
173760
1256
這就是機器人。
02:55
Even if it might not look like Rosie the Robot,
72
175040
2576
即使它看起來不像 《傑森一家》裡的機器人,
02:57
it's doing something that's really useful in my life
73
177640
2936
但在我的人生中, 它所做的事非常有用,
03:00
so I don't have to take care
74
180600
1376
它讓我不用去費心
03:02
of turning the temperature up and down myself.
75
182000
2576
自己把溫度調高調低。
03:04
And I think these systems live and work amongst us now,
76
184600
3816
我認為,這些系統在 我們人類當中生活和工作,
03:08
and not only are these systems living amongst us
77
188440
2336
且,不僅這些系統生活在我們當中,
03:10
but you are probably a robot operator, too.
78
190800
2656
你可能也是個機器人操作員。
03:13
When you drive your car,
79
193480
1256
當你駕駛你的車時,
03:14
it feels like you are operating machinery.
80
194760
2216
感覺就像你在操作機械。
03:17
You are also going from point A to point B,
81
197000
2816
你也是在從A點前往B點,
03:19
but your car probably has power steering,
82
199840
2216
但你的車可能有動力方向盤,
03:22
it probably has automatic braking systems,
83
202080
2696
它可能有自動煞車系統,
03:24
it might have an automatic transmission and maybe even adaptive cruise control.
84
204800
3736
它可能有自排變速箱, 甚至有主動車距控制巡航系統。
03:28
And while it might not be a fully autonomous car,
85
208560
2936
雖然它可能不是完全自主的車,
03:31
it has bits of autonomy,
86
211520
1296
它還是有一些自主性,
03:32
and they're so useful
87
212840
1336
且這些自主性很有用,
03:34
and they make us drive safer,
88
214200
1816
讓我們能更安全地開車,
03:36
and we just sort of feel like they're invisible-in-use, right?
89
216040
3656
我們只會感覺到,在使用它們時, 它們好像隱形了,對嗎?
03:39
So when you're driving your car,
90
219720
1576
當你在駕駛你的車時,
03:41
you should just feel like you're going from one place to another.
91
221320
3096
你應該只會感覺到你是 從一個地方到另一個地方。
03:44
It doesn't feel like it's this big thing that you have to deal with and operate
92
224440
3736
感覺並不像是件需要你 去處理、操作、使用
這些控制功能的大事,
03:48
and use these controls
93
228200
1256
03:49
because we spent so long learning how to drive
94
229480
2176
因為我們花了太多時間學習駕駛,
03:51
that they've become extensions of ourselves.
95
231680
2696
以致於它們已經成了 我們自己的延伸體。
03:54
When you park that car in that tight little garage space,
96
234400
2696
當你把車停在那狹小的車庫空間中,
你知道車子的邊角在什麼位置。
03:57
you know where your corners are.
97
237120
1576
03:58
And when you drive a rental car that maybe you haven't driven before,
98
238720
3256
當你開一輛你可能以前 沒開過的出租車時,
04:02
it takes some time to get used to your new robot body.
99
242000
3056
要花一點時間去習慣 你的機器人身體。
04:05
And this is also true for people who operate other types of robots,
100
245080
3976
操作其他種類的機器人也是這樣的,
04:09
so I'd like to share with you a few stories about that.
101
249080
2600
所以我想要和各位 分享幾個相關的故事。
04:12
Dealing with the problem of remote collaboration.
102
252240
2336
處理遠端協同作業的問題。
04:14
So, at Willow Garage I had a coworker named Dallas,
103
254600
2576
在柳樹車庫,我有個同事叫達拉斯,
04:17
and Dallas looked like this.
104
257200
1576
達拉斯看起來是這樣子。
04:18
He worked from his home in Indiana in our company in California.
105
258800
4056
在我們加州的公司中, 他在印第安納州的家中工作。
04:22
He was a voice in a box on the table in most of our meetings,
106
262880
2936
在我們大部分的會議中, 他是桌上盒子裡傳出來的聲音,
04:25
which was kind of OK except that, you know,
107
265840
2215
這樣是還好,不過,你們知道的,
04:28
if we had a really heated debate and we didn't like what he was saying,
108
268079
3377
如果我們發生火熱的辯論, 且我們不喜歡他的說法時,
04:31
we might just hang up on him.
109
271480
1416
我們可能就會把他掛斷。
04:32
(Laughter)
110
272920
1015
(笑聲)
04:33
Then we might have a meeting after that meeting
111
273959
2217
在那場會議後, 我們會再開一場會議,
04:36
and actually make the decisions in the hallway afterwards
112
276200
2696
且之後就在走廊做決策,
04:38
when he wasn't there anymore.
113
278920
1416
當他不在那裡的時候。
04:40
So that wasn't so great for him.
114
280360
1576
那對他而言並非好事。
04:41
And as a robotics company at Willow,
115
281960
1736
柳樹車庫是間機器人公司,
04:43
we had some extra robot body parts laying around,
116
283720
2336
在公司裡滿地都是 機器人的身體部件,
所以,達拉斯和他的伙伴柯特 組裝了這個東西,
04:46
so Dallas and his buddy Curt put together this thing,
117
286080
2496
04:48
which looks kind of like Skype on a stick on wheels,
118
288600
2936
看起來像是 Skype 接著一根棍子且下面有輪子,
04:51
which seems like a techy, silly toy,
119
291560
1736
它看來是個很蠢的科技玩具,
04:53
but really it's probably one of the most powerful tools
120
293320
2776
但其實,它可能是我所見過 所有為了遠端協同作業所製出
04:56
that I've seen ever made for remote collaboration.
121
296120
2480
最強大工具中的其中之一。
04:59
So now, if I didn't answer Dallas' email question,
122
299160
3496
所以,現在,如果我沒回覆 達拉斯在電子郵件中問的問題,
05:02
he could literally roll into my office,
123
302680
2216
他可以直接進入我的辦公室,
05:04
block my doorway and ask me the question again --
124
304920
2576
擋住我的門,再問我一次那個問題
05:07
(Laughter)
125
307520
1016
(笑聲)
05:08
until I answered it.
126
308560
1216
直到我回答為止。
05:09
And I'm not going to turn him off, right? That's kind of rude.
127
309800
2976
我不會把他關機, 對吧?那樣有點失禮。
05:12
Not only was it good for these one-on-one communications,
128
312800
2696
它不只很適合用在一對一的溝通上,
05:15
but also for just showing up at the company all-hands meeting.
129
315520
2936
也很適合出席公司的全員大會。
05:18
Getting your butt in that chair
130
318480
1696
把你的屁股坐到椅子上,
05:20
and showing people that you're present and committed to your project
131
320200
3216
讓人們看到你有出席, 有投入你的專案計畫,
05:23
is a big deal
132
323440
1256
這是很重要的,
05:24
and can help remote collaboration a ton.
133
324720
2176
且能大大協助遠端協同作業。
05:26
We saw this over the period of months and then years,
134
326920
2856
在數個月、接著數年的期間, 我們都看到這個狀況,
05:29
not only at our company but at others, too.
135
329800
2160
不只在我們的公司裡, 也在其他公司裡。
05:32
The best thing that can happen with these systems
136
332720
2336
對於這些系統,最棒的情況就是
05:35
is that it starts to feel like you're just there.
137
335080
2336
你開始覺得你就是在那裡。
05:37
It's just you, it's just your body,
138
337440
1696
那就是你,那就是你的身體,
05:39
and so people actually start to give these things personal space.
139
339160
3096
所以人們開始留 個人空間給這些東西。
05:42
So when you're having a stand-up meeting,
140
342280
1976
當你參加一場站立會議時,
05:44
people will stand around the space
141
344280
1656
人們在那空間中站立的位置,
05:45
just as they would if you were there in person.
142
345960
2216
就跟真人開會時的情況一樣。
一切都很美好, 直到當機,就不好了。
05:48
That's great until there's breakdowns and it's not.
143
348200
2576
05:50
People, when they first see these robots,
144
350800
1976
當人們初次看到這些機器人時,
05:52
are like, "Wow, where's the components? There must be a camera over there,"
145
352800
3576
通常會:「哇,零件在哪裡? 這裡一定有個攝影機。」
05:56
and they start poking your face.
146
356400
1576
他們開始戳你的臉。
05:58
"You're talking too softly, I'm going to turn up your volume,"
147
358000
2936
「你說話太輕柔了, 我要把你的音量調高。」
06:00
which is like having a coworker walk up to you and say,
148
360960
2616
這感覺就像有個同事 走到你面前,說:
「你說話太輕柔了, 我來把你的臉調一下。」
06:03
"You're speaking too softly, I'm going to turn up your face."
149
363600
2896
那挺尷尬的,且不太好,
06:06
That's awkward and not OK,
150
366520
1256
06:07
and so we end up having to build these new social norms
151
367800
2616
所以我們最後就建立了使用 這些系統時的新社交規範。
06:10
around using these systems.
152
370440
2216
06:12
Similarly, as you start feeling like it's your body,
153
372680
3416
同樣的,當你開始覺得 它是你身體的一部分時,
06:16
you start noticing things like, "Oh, my robot is kind of short."
154
376120
3696
你就會開始注意到一些事: 「喔,我的機器人有點矮。」
06:19
Dallas would say things to me -- he was six-foot tall --
155
379840
2656
達拉斯會對我說些事。他有六呎高。
06:22
and we would take him via robot to cocktail parties and things like that,
156
382520
3456
而我們會透過機器人帶他去 雞尾酒派對之類的活動,
06:26
as you do,
157
386000
1216
跟你會做的一樣。
06:27
and the robot was about five-foot-tall, which is close to my height.
158
387240
3336
而機器人大約五呎高,和我差不多。
06:30
And he would tell me,
159
390600
1216
他會告訴我:
06:31
"You know, people are not really looking at me.
160
391840
2536
「你知道嗎,人們沒有真正看著我。
06:34
I feel like I'm just looking at this sea of shoulders,
161
394400
2696
我覺得我好像是看著一片肩膀海,
06:37
and it's just -- we need a taller robot."
162
397120
1976
那就是──我們需要更高的機器人。」
06:39
And I told him,
163
399120
1256
而我告訴他:
06:40
"Um, no.
164
400400
1296
「呃,不。
06:41
You get to walk in my shoes for today.
165
401720
1936
今天,你要嚐嚐當我的滋味。
06:43
You get to see what it's like to be on the shorter end of the spectrum."
166
403680
3536
你可以體會一下身為 比較矮的族群是什麼感覺。」
06:47
And he actually ended up building a lot of empathy for that experience,
167
407240
3376
結果他從那次經驗 建立起了相當的同理心,
06:50
which was kind of great.
168
410640
1256
這樣挺好的。
06:51
So when he'd come visit in person,
169
411920
1656
所以當他親自來造訪時,
06:53
he no longer stood over me as he was talking to me,
170
413600
2416
他若要跟我說話, 也不會站得比我高,
06:56
he would sit down and talk to me eye to eye,
171
416040
2096
他會坐下來, 和我能真正面對面談話,
06:58
which was kind of a beautiful thing.
172
418160
1736
這是件很美好的事。
06:59
So we actually decided to look at this in the laboratory
173
419920
2656
所以,我們決定在 實驗室中研究這一點,
07:02
and see what others kinds of differences things like robot height would make.
174
422600
3656
看看像類似機器人身高 這種特性會造什麼其他的差異。
07:06
And so half of the people in our study used a shorter robot,
175
426280
2856
在我們的研究中,半數 受試者使用較矮的機器人,
07:09
half of the people in our study used a taller robot
176
429160
2416
另外半數則用較高的機器人,
07:11
and we actually found that the exact same person
177
431600
2256
結果發現,當同一個人
07:13
who has the exact same body and says the exact same things as someone,
178
433880
3336
用同樣的身體,說同樣的話,
07:17
is more persuasive and perceived as being more credible
179
437240
2616
如果用比較高的機器人, 就會比較有說服力,
07:19
if they're in a taller robot form.
180
439880
1656
且被認為比較可靠。
07:21
It makes no rational sense,
181
441560
1816
這沒有理性的解釋,
但那就是為什麼我們要研究心理學。
07:23
but that's why we study psychology.
182
443400
1696
07:25
And really, you know, the way that Cliff Nass would put this
183
445120
2856
其實,用克利福德那斯的說法,
07:28
is that we're having to deal with these new technologies
184
448000
3016
我們得要處理這些嶄新的技術,
07:31
despite the fact that we have very old brains.
185
451040
2736
儘管我們的大腦是老舊的。
07:33
Human psychology is not changing at the same speed that tech is
186
453800
2976
人類心理學的改變速度沒有科技快,
07:36
and so we're always playing catch-up,
187
456800
1816
所以我們一直在努力追趕,
07:38
trying to make sense of this world
188
458640
1656
試圖把這個有很多自動的東西 跑來跑去的世界給合理化。
07:40
where these autonomous things are running around.
189
460320
2336
07:42
Usually, things that talk are people, not machines, right?
190
462680
2736
通常,會說話的就是人, 不是機器,對嗎?
07:45
And so we breathe a lot of meaning into things like just height of a machine,
191
465440
4576
所以我們把很多的意義帶到 像機器高度這樣的事物中,
07:50
not a person,
192
470040
1256
而非人,
07:51
and attribute that to the person using the system.
193
471320
2360
然後將之歸因給使用系統的人。
07:55
You know, this, I think, is really important
194
475120
2216
當談到機器人學的時候,
07:57
when you're thinking about robotics.
195
477360
1736
我認為這點十分重要。
07:59
It's not so much about reinventing humans,
196
479120
2096
它的重點並不是在重新發明人類,
08:01
it's more about figuring out how we extend ourselves, right?
197
481240
3136
比較是在於我們要如何 延伸我們自己,對嗎?
08:04
And we end up using things in ways that are sort of surprising.
198
484400
2976
結果是,我們會用 蠻讓人訝異的方式來使用東西。
08:07
So these guys can't play pool because the robots don't have arms,
199
487400
4256
這些人無法玩撞球, 因為機器人沒有手臂,
08:11
but they can heckle the guys who are playing pool
200
491680
2336
但他們能和玩撞球的那些人起鬨,
08:14
and that can be an important thing for team bonding,
201
494040
3176
那對於團隊連結來說 可能是件重要的事,
08:17
which is kind of neat.
202
497240
1296
這樣挺好的。
08:18
People who get really good at operating these systems
203
498560
2496
非常擅長操作這些系統的人
08:21
will even do things like make up new games,
204
501080
2056
能做到像是創造新遊戲這類的事,
比如半夜玩機器人足球,
08:23
like robot soccer in the middle of the night,
205
503160
2136
把垃圾筒推來推去。
08:25
pushing the trash cans around.
206
505320
1456
08:26
But not everyone's good.
207
506800
1576
但並非每個人都很擅長。
08:28
A lot of people have trouble operating these systems.
208
508400
2496
有些人在操作這些 系統時會遇到問題。
08:30
This is actually a guy who logged into the robot
209
510920
2256
這個人登入了機器人,
而他的眼球向左轉了九十度。
08:33
and his eyeball was turned 90 degrees to the left.
210
513200
2376
08:35
He didn't know that,
211
515600
1256
他自己並沒察覺,
08:36
so he ended up just bashing around the office,
212
516880
2176
結果他在辦公室裡頭亂撞,
撞到別人的桌子,弄得非常尷尬,
08:39
running into people's desks, getting super embarrassed,
213
519080
2616
因此而大笑。他的音量太高了。
08:41
laughing about it -- his volume was way too high.
214
521720
2336
而照片裡的這個人告訴我:
08:44
And this guy here in the image is telling me,
215
524080
2136
08:46
"We need a robot mute button."
216
526240
2096
「我們需要機器人靜音按鈕。」
08:48
And by that what he really meant was we don't want it to be so disruptive.
217
528360
3496
他那麼說的意思是,我們不希望 機器人會這樣引起混亂。
08:51
So as a robotics company,
218
531880
1616
所以,既然我們是機器人公司,
08:53
we added some obstacle avoidance to the system.
219
533520
2456
我們就在系統上加裝了 障礙閃避功能。
08:56
It got a little laser range finder that could see the obstacles,
220
536000
3056
機器人有了一個小雷射測距儀, 能夠看見障礙物,
08:59
and if I as a robot operator try to say, run into a chair,
221
539080
3136
如果我在操作機器人時, 試圖比如撞向一張椅子,
09:02
it wouldn't let me, it would just plan a path around,
222
542240
2496
它不會讓我這麼做,它會規劃繞路,
09:04
which seems like a good idea.
223
544760
1856
這看似是個好主意。
09:06
People did hit fewer obstacles using that system, obviously,
224
546640
3176
很顯然,用了那個系統之後, 人們比較少撞到障礙物,
09:09
but actually, for some of the people,
225
549840
2096
但,對一些人而言,
09:11
it took them a lot longer to get through our obstacle course,
226
551960
2856
他們要花比較長的時間 通過障礙物課程,
09:14
and we wanted to know why.
227
554840
1560
我們想要知道為什麼。
09:17
It turns out that there's this important human dimension --
228
557080
3056
結果發現,有個很重要的人類維度,
09:20
a personality dimension called locus of control,
229
560160
2296
一個人格維度,叫做「控制點」,
09:22
and people who have a strong internal locus of control,
230
562480
3136
如果一個人有很強的內在控制點,
09:25
they need to be the masters of their own destiny --
231
565640
3056
他會想要主宰自己的命運,
09:28
really don't like giving up control to an autonomous system --
232
568720
3096
他們不喜歡把控制權交給自動系統
09:31
so much so that they will fight the autonomy;
233
571840
2136
以致會去對抗自動化:
09:34
"If I want to hit that chair, I'm going to hit that chair."
234
574000
3096
「如果我想要撞到椅子, 我就要撞到椅子。」
09:37
And so they would actually suffer from having that autonomous assistance,
235
577120
3616
自動化的協助反而會 讓這類人很辛苦,
09:40
which is an important thing for us to know
236
580760
2576
我們能知道這點是很重要的,
09:43
as we're building increasingly autonomous, say, cars, right?
237
583360
3296
畢竟我們在建立越來越 自動化的…比如汽車,對吧?
09:46
How are different people going to grapple with that loss of control?
238
586680
3360
要如何讓不同型的人使用而不失控?
09:50
It's going to be different depending on human dimensions.
239
590880
2696
不同的人類維度會有不同的方式。
09:53
We can't treat humans as if we're just one monolithic thing.
240
593600
3496
我們不能把人看成是單一的東西。
09:57
We vary by personality, by culture,
241
597120
2416
我們有不同的人格、文化,
09:59
we even vary by emotional state moment to moment,
242
599560
2456
我們甚至在不同時刻 會有不同的情緒狀態,
10:02
and being able to design these systems,
243
602040
1976
如果要能夠設計這些系統,
10:04
these human-robot interaction systems,
244
604040
2296
這些人類與機器人的互動系統,
10:06
we need to take into account the human dimensions,
245
606360
2736
我們就得要考量人類維度,
10:09
not just the technological ones.
246
609120
1720
不只是技術維度。
10:11
Along with a sense of control also comes a sense of responsibility.
247
611640
4296
隨著控制感而來的,就是責任感。
10:15
And if you were a robot operator using one of these systems,
248
615960
2856
如果你是個在使用 這些系統的機器人操作員,
10:18
this is what the interface would look like.
249
618840
2056
介面看起來就是這樣子的。
10:20
It looks a little bit like a video game,
250
620920
1936
它看起來有一點像電玩遊戲,
10:22
which can be good because that's very familiar to people,
251
622880
2976
這是好事,因為人們熟悉它;
10:25
but it can also be bad
252
625880
1216
但也可能是壞事,
10:27
because it makes people feel like it's a video game.
253
627120
2456
因為它會讓人們覺得 它是個電玩遊戲。
10:29
We had a bunch of kids over at Stanford play with the system
254
629600
2856
我們找了一群小朋友 到史丹佛來玩這個系統,
10:32
and drive the robot around our office in Menlo Park,
255
632480
2456
操作我們在門洛公園 辦公室裡的機器人,
10:34
and the kids started saying things like,
256
634960
1936
小朋友們開始說像這樣的話:
「如果你撞到那邊的那個人, 得十分,那個人則是二十分。」
10:36
"10 points if you hit that guy over there. 20 points for that one."
257
636920
3176
他們會在走廊上追著目標跑。
10:40
And they would chase them down the hallway.
258
640120
2016
(笑聲)
10:42
(Laughter)
259
642160
1016
10:43
I told them, "Um, those are real people.
260
643200
1936
我告訴他們:「呃,那些是真人。
10:45
They're actually going to bleed and feel pain if you hit them."
261
645160
3296
如果你們撞到他們, 他們真的會流血且會痛。」
10:48
And they'd be like, "OK, got it."
262
648480
1616
他們才說:「好,了解。」
10:50
But five minutes later, they would be like,
263
650120
2056
但五分鐘之後,他們又會說:
10:52
"20 points for that guy over there, he just looks like he needs to get hit."
264
652200
3616
「撞到那邊那個人有二十分, 他看起來需要被撞一下。」
10:55
It's a little bit like "Ender's Game," right?
265
655840
2136
這有一點像《戰爭遊戲》,對嗎?
在另一端有一個真實的世界,
10:58
There is a real world on that other side
266
658000
1936
10:59
and I think it's our responsibility as people designing these interfaces
267
659960
3416
我認為我們身為人的責任, 是要設計這些介面
11:03
to help people remember
268
663400
1256
來協助人們記得
11:04
that there's real consequences to their actions
269
664680
2256
他們的行為會造成真實的後果,
11:06
and to feel a sense of responsibility
270
666960
2296
讓他們在操作這些 越來越自動的東西時,
11:09
when they're operating these increasingly autonomous things.
271
669280
3280
能夠有責任感。
11:13
These are kind of a great example
272
673840
2296
這些是很棒的例子,
11:16
of experimenting with one possible robotic future,
273
676160
3256
說明對一種可能的 機器人未來所做的實驗,
11:19
and I think it's pretty cool that we can extend ourselves
274
679440
3856
我覺得我們能將自己 延伸出去,是挺酷的事,
11:23
and learn about the ways that we extend ourselves
275
683320
2336
同時也能學習我們 將自己延伸出去的方式,
11:25
into these machines
276
685680
1216
延伸至機器中,
11:26
while at the same time being able to express our humanity
277
686920
2696
同時還能表現出我們的人性
11:29
and our personality.
278
689640
1216
以及我們的個性。
11:30
We also build empathy for others
279
690880
1576
我們也會建立對他人的同理心,
11:32
in terms of being shorter, taller, faster, slower,
280
692480
3216
理解那些比較矮、比較高、 比較快、比較慢,
11:35
and maybe even armless,
281
695720
1416
甚至沒有手臂的人,
11:37
which is kind of neat.
282
697160
1336
這樣挺好的。
11:38
We also build empathy for the robots themselves.
283
698520
2536
我們也會建立 對機器人本身的同理心。
11:41
This is one of my favorite robots.
284
701080
1656
這是我最喜歡的機器人之一。
11:42
It's called the Tweenbot.
285
702760
1456
它叫 Tweenbot。
11:44
And this guy has a little flag that says,
286
704240
1976
這傢伙有一面小旗子,上面寫著:
11:46
"I'm trying to get to this intersection in Manhattan,"
287
706240
2576
「我想前往曼哈頓的 這個十字路口。」
11:48
and it's cute and rolls forward, that's it.
288
708840
2776
它很可愛,會向前跑,就這樣。
11:51
It doesn't know how to build a map, it doesn't know how to see the world,
289
711640
3456
它不知道如何建立地圖, 它不知道如何看世界,
它只會尋求協助。
11:55
it just asks for help.
290
715120
1256
11:56
The nice thing about people
291
716400
1336
而人們很棒的一點,
11:57
is that it can actually depend upon the kindness of strangers.
292
717760
3096
就是它真的可以仰賴陌生人的善心。
12:00
It did make it across the park to the other side of Manhattan --
293
720880
3896
它真的穿過公園到了 曼哈頓的另一端。
12:04
which is pretty great --
294
724800
1256
這很棒,
因為人們會把它拿起來, 轉向正確的方向。
12:06
just because people would pick it up and point it in the right direction.
295
726080
3456
(笑聲)
12:09
(Laughter)
296
729560
936
那樣很棒,對嗎?
12:10
And that's great, right?
297
730520
1256
12:11
We're trying to build this human-robot world
298
731800
2696
我們在試著建立 人類──機器人的世界,
12:14
in which we can coexist and collaborate with one another,
299
734520
3416
在這個世界中,我們彼此能 共同存在且和諧地共事,
12:17
and we don't need to be fully autonomous and just do things on our own.
300
737960
3376
我們不用完全自動化, 也不用只靠自己做事。
12:21
We actually do things together.
301
741360
1496
而是真正能夠一起做事。
12:22
And to make that happen,
302
742880
1256
要實現這個理想,
我們其實需要其他人的協助, 像是藝術家、設計師、
12:24
we actually need help from people like the artists and the designers,
303
744160
3256
政策制訂者、法律學者、
12:27
the policy makers, the legal scholars,
304
747440
1856
心理學家、社會學家、人類學家 ──
12:29
psychologists, sociologists, anthropologists --
305
749320
2216
12:31
we need more perspectives in the room
306
751560
1816
我們需要更多的觀點參與討論,
12:33
if we're going to do the thing that Stu Card says we should do,
307
753400
2976
如果我們要做史都華卡德 說我們應該做的事情,
12:36
which is invent the future that we actually want to live in.
308
756400
3936
也就是:發明一個會讓我們 想要住在其中的未來。
12:40
And I think we can continue to experiment
309
760360
2656
我認為我們能持續做實驗,
12:43
with these different robotic futures together,
310
763040
2176
將不同的機器人未來一起拿來實驗,
12:45
and in doing so, we will end up learning a lot more about ourselves.
311
765240
4680
這麼做,我們最後就能學到 很多關於我們自己的事。
12:50
Thank you.
312
770720
1216
謝謝。
12:51
(Applause)
313
771960
2440
(掌聲)
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隱私政策

eng.lish.video

Developer's Blog