请双击下面的英文字幕来播放视频。
翻译人员: Jennifer Huang
校对人员: Lexi Ding
00:13
Hi, I'm Jeff.
0
13329
1583
大家好,我是杰夫。
00:15
I lead AI Research and Health at Google.
1
15746
2833
我领导着谷歌的
人工智能研究与健康部门。
00:18
I joined Google more than 20 years ago,
2
18621
1875
我在二十多年前加入谷歌,
00:20
when we were all wedged
into a tiny office space,
3
20538
3166
那时候我们还挤在
一间狭小的办公室里,
00:23
above what's now a T-Mobile store
in downtown Palo Alto.
4
23704
2750
就位于现在帕罗阿托市中心的
一家德国电信商店上面。
00:27
I've seen a lot of computing
transformations in that time,
5
27163
2875
在那段时间,我见证了
许多计算机技术的发展,
00:30
and in the last decade, we've seen AI
be able to do tremendous things.
6
30079
3459
在过去十年,我们见证了
人工智能完成许多了不起的事情。
00:34
But we're still doing it
all wrong in many ways.
7
34746
2542
但我们仍在许多方面都没有做对。
00:37
That's what I want
to talk to you about today.
8
37329
2209
这是我今天想讨论的主题。
00:39
But first, let's talk
about what AI can do.
9
39579
2209
不过首先,我们先来聊聊
人工智能可以做什么。
00:41
So in the last decade,
we've seen tremendous progress
10
41829
3292
在过去的十年间,
我们看到了人工智能
00:45
in how AI can help computers see,
understand language,
11
45121
4333
在帮助计算机识别物体、
理解语言和谈话方面
00:49
understand speech better than ever before.
12
49496
2833
取得的巨大进步。
00:52
Things that we couldn't do
before, now we can do.
13
52329
2583
以往的天方夜谭
现在一一成为现实。
00:54
If you think about computer vision alone,
14
54954
2333
就单单拿计算机视觉来说,
00:57
just in the last 10 years,
15
57329
1583
在过去的十年中,
00:58
computers have effectively
developed the ability to see;
16
58912
2792
电脑快速地发展出了‘看’的能力;
01:01
10 years ago, they couldn't see,
now they can see.
17
61746
2500
十年前,它们看不到,
但现在可以了。
01:04
You can imagine this has had
a transformative effect
18
64287
2500
这在计算机运用上
01:06
on what we can do with computers.
19
66787
2209
具有变革性的影响。
让我们来看几个由这些功能促成的
01:08
So let's look at a couple
of the great applications
20
68996
2500
了不起的实际应用。
01:11
enabled by these capabilities.
21
71496
2000
我们可以通过机器学习
更准确地预测洪水,
01:13
We can better predict flooding,
keep everyone safe,
22
73537
2417
保证所有人的安全。
01:15
using machine learning.
23
75996
1500
我们可以翻译一百多种语言
以便更顺畅地交流,
01:17
We can translate over 100 languages
so we all can communicate better,
24
77538
3250
还可以更精准地预测和诊断疾病,
01:20
and better predict and diagnose disease,
25
80788
2000
01:22
where everyone gets
the treatment that they need.
26
82788
2333
让所有人得到所需的治疗。
01:25
So let's look at two key components
27
85163
1916
让我们来看看
01:27
that underlie the progress
in AI systems today.
28
87121
2333
构成当代人工智能系统
基础的两个关键元素。
01:30
The first is neural networks,
29
90121
1542
首先是神经网络,
01:31
a breakthrough approach to solving
some of these difficult problems
30
91704
3417
它是解决这些难题的一项重大突破,
01:35
that has really shone
in the last 15 years.
31
95121
2833
它在过去的十五年已经大放异彩。
01:37
But they're not a new idea.
32
97954
1625
但它并不是一个新的点子。
01:39
And the second is computational power.
33
99621
1833
第二个是运算能力。
01:41
It actually takes a lot
of computational power
34
101454
2167
驱动神经网络运作
01:43
to make neural networks
able to really sing,
35
103663
2083
实际需要大量的运算能力,
01:45
and in the last 15 years,
we’ve been able to halve that,
36
105746
3375
在过去的十五年,
我们做到了使其减半,
01:49
and that's partly what's enabled
all this progress.
37
109121
2458
那也是整个人工智能
得以发展至此的原因之一。
01:51
But at the same time,
I think we're doing several things wrong,
38
111954
3459
但同时,我认为我们做错了几件事,
01:55
and that's what I want
to talk to you about
39
115413
2000
那也是我想在这次演讲的最后
01:57
at the end of the talk.
40
117454
1167
和大家谈论的。
01:58
First, a bit of a history lesson.
41
118663
1958
首先,给大家上点历史课。
02:00
So for decades,
42
120663
1208
数十年间,
02:01
almost since the very
beginning of computing,
43
121913
2125
几乎从计算机科学最早出现,
02:04
people have wanted
to be able to build computers
44
124079
2250
人们就想建造
02:06
that could see, understand language,
understand speech.
45
126329
3917
可以识别语言及理解谈话的电脑。
02:10
The earliest approaches
to this, generally,
46
130579
2000
最初的方法一般是
02:12
people were trying to hand-code
all the algorithms
47
132621
2333
人们手动写下
02:14
that you need to accomplish
those difficult tasks,
48
134996
2458
完成难题所需的算法。
02:17
and it just turned out
to not work very well.
49
137454
2375
但成效一般。
02:19
But in the last 15 years,
a single approach
50
139829
3250
但在过去的十五年间,
一个方法出其不意地
一次性解决了所有难题:
02:23
unexpectedly advanced all these different
problem spaces all at once:
51
143079
4167
02:27
neural networks.
52
147954
1542
神经网络。
02:29
So neural networks are not a new idea.
53
149538
1833
神经网络并非一个新想法。
02:31
They're kind of loosely based
54
151371
1417
它们大致上基于
02:32
on some of the properties
that are in real neural systems.
55
152788
3041
一些真实神经系统的特性。
02:35
And many of the ideas
behind neural networks
56
155829
2084
大部分神经网络背后的理念
02:37
have been around since the 1960s and 70s.
57
157954
2250
出现于1960和70年代。
02:40
A neural network is what it sounds like,
58
160204
2167
神经网络如同其字面意思一样,
02:42
a series of interconnected
artificial neurons
59
162413
2708
是一连串互相连接的神经元。
02:45
that loosely emulate the properties
of your real neurons.
60
165121
3000
它们大致上效仿了
人体真正神经元的特性。
02:48
An individual neuron
in one of these systems
61
168163
2166
这种系统中的一个独立神经元,
02:50
has a set of inputs,
62
170371
1417
拥有一组输入信息,
02:51
each with an associated weight,
63
171788
2041
每组输入信息有对应的比重,
02:53
and the output of a neuron
64
173871
1833
神经元的信息输出
02:55
is a function of those inputs
multiplied by those weights.
65
175704
3209
就等于那些输入信息
乘以它们对应的比重。
02:59
So pretty simple,
66
179288
1208
所以其实挺简单的,
03:00
and lots and lots of these work together
to learn complicated things.
67
180538
3666
无数神经元协同运作
就可以学习复杂的东西。
03:04
So how do we actually learn
in a neural network?
68
184496
2833
所以我们到底是
如何在神经网络中学习的?
03:07
It turns out the learning process
69
187371
1708
其实,在学习过程中,
03:09
consists of repeatedly making
tiny little adjustments
70
189079
2792
比重在不断被微调,
03:11
to the weight values,
71
191913
1208
03:13
strengthening the influence
of some things,
72
193121
2042
增强一些东西的影响,
03:15
weakening the influence of others.
73
195204
1959
削弱其他的影响。
03:17
By driving the overall system
towards desired behaviors,
74
197204
3917
在驱使整个系统去往
理想运作模式的过程中,
03:21
these systems can be trained
to do really complicated things,
75
201163
2916
可以训练其完成
一些非常复杂的事情,
03:24
like translate
from one language to another,
76
204121
2875
例如翻译语言,
03:27
detect what kind
of objects are in a photo,
77
207038
3083
识别照片中的物体,
03:30
all kinds of complicated things.
78
210121
1875
任何类型的复杂工作。
03:32
I first got interested in neural networks
79
212038
2000
我对神经网络的兴趣,
03:34
when I took a class on them
as an undergraduate in 1990.
80
214079
3042
始于1990年本科阶段时
学到的一门相关课程。
03:37
At that time,
81
217163
1125
那时,
03:38
neural networks showed
impressive results on tiny problems,
82
218329
3792
神经网络在精细问题的解决上
取得了惊人的成果,
03:42
but they really couldn't scale to do
real-world important tasks.
83
222121
4375
但还达不到完成真实世界中
重要工作的程度。
03:46
But I was super excited.
84
226538
1500
但我非常兴奋。
03:48
(Laughter)
85
228079
2500
(笑声)
03:50
I felt maybe we just needed
more compute power.
86
230579
2417
我觉得我们可能只是
需要更强的运算能力。
03:52
And the University of Minnesota
had a 32-processor machine.
87
232996
3625
明尼苏达大学当时
有一个32位处理器。
03:56
I thought, "With more compute power,
88
236621
1792
我想:“如果有更强的运算能力,
我们真能用神经网络干点大事。”
03:58
boy, we could really make
neural networks really sing."
89
238413
3000
04:01
So I decided to do a senior thesis
on parallel training of neural networks,
90
241454
3584
所以我决定以神经网络的并行训练
作为我毕业论文的课题,
04:05
the idea of using processors in a computer
or in a computer system
91
245079
4000
理念是将电脑或电脑系统中
所有的处理器
04:09
to all work toward the same task,
92
249079
2125
运用到同一件任务上,
04:11
that of training neural networks.
93
251204
1584
用来训练神经网络。
04:12
32 processors, wow,
94
252829
1292
32位处理器,哇,
04:14
we’ve got to be able
to do great things with this.
95
254163
2833
我们肯定能用它做点大事。
04:17
But I was wrong.
96
257496
1167
但我错了。
04:20
Turns out we needed about a million times
as much computational power
97
260038
3333
我发现如果想用神经网络
做些引人注目的事情,
04:23
as we had in 1990
98
263371
1375
04:24
before we could actually get
neural networks to do impressive things.
99
264788
3333
我们所需的算力大概是
90年代算力的一百万倍。
04:28
But starting around 2005,
100
268121
2417
但从大概2005年开始,
04:30
thanks to the computing progress
of Moore's law,
101
270579
2500
多亏了摩尔定律,
04:33
we actually started to have
that much computing power,
102
273121
2625
我们真的开始拥有这么多运算能力了,
04:35
and researchers in a few universities
around the world started to see success
103
275746
4250
世界上一些大学里的研究员们
开始成功用神经网络完成各种任务。
04:40
in using neural networks for a wide
variety of different kinds of tasks.
104
280038
4083
04:44
I and a few others at Google
heard about some of these successes,
105
284121
3583
我和其他几个在谷歌的同事
听闻了这些成功事例,
04:47
and we decided to start a project
to train very large neural networks.
106
287746
3333
于是决定启动一个项目,
训练大型神经网络。
04:51
One system that we trained,
107
291079
1459
其中一个系统,
04:52
we trained with 10 million
randomly selected frames
108
292538
3541
我们用油管视频里随机截取的
一千万帧照片对其进行训练。
04:56
from YouTube videos.
109
296079
1292
04:57
The system developed the capability
110
297371
1750
这个系统发展出了能够识别
所有不同种类物体的能力。
04:59
to recognize all kinds
of different objects.
111
299121
2583
05:01
And it being YouTube, of course,
112
301746
1542
然后因为是油管的关系,
05:03
it developed the ability
to recognize cats.
113
303329
2500
所以它发展出了识别猫的能力。
05:05
YouTube is full of cats.
114
305829
1292
油管上全是猫。
05:07
(Laughter)
115
307163
1416
(笑声)
05:08
But what made that so remarkable
116
308621
2208
但让它如此引人注目的是
05:10
is that the system was never told
what a cat was.
117
310871
2417
从未有人告诉过这个系统
猫到底是什么。
05:13
So using just patterns in data,
118
313704
2500
仅仅依靠数据的形态规律,
05:16
the system honed in on the concept
of a cat all on its own.
119
316204
3625
它就能自己琢磨出来猫究竟是什么。
05:20
All of this occurred at the beginning
of a decade-long string of successes,
120
320371
3750
所有的这一切标志着
一个长达十年的,
05:24
of using neural networks
for a huge variety of tasks,
121
324121
2500
在谷歌及其它地方成功运用神经网络
完成繁杂任务的开始。
05:26
at Google and elsewhere.
122
326663
1166
05:27
Many of the things you use every day,
123
327871
2167
这些你日常使用的东西,
05:30
things like better speech
recognition for your phone,
124
330079
2500
例如手机更准确的语音识别系统,
05:32
improved understanding
of queries and documents
125
332579
2209
提升对问题和文本的理解力,
05:34
for better search quality,
126
334829
1459
优化搜索结果,
05:36
better understanding of geographic
information to improve maps,
127
336329
3042
对地理信息更完善的理解等等
提升地图及其他方面。
05:39
and so on.
128
339413
1166
05:40
Around that time,
129
340621
1167
在那个时候,
05:41
we also got excited about how we could
build hardware that was better tailored
130
341788
3750
我们还对如何打造一个更适合
神经网络运算所需的计算机硬件感兴趣。
05:45
to the kinds of computations
neural networks wanted to do.
131
345579
2792
神经网络运算有两个特性。
05:48
Neural network computations
have two special properties.
132
348371
2667
05:51
The first is they're very tolerant
of reduced precision.
133
351079
2625
第一个是它们对精准度要求很低。
05:53
Couple of significant digits,
you don't need six or seven.
134
353746
2750
几个有效位就够了,
不需要六七个那么多。
05:56
And the second is that all the
algorithms are generally composed
135
356496
3458
第二个是所有算法都普遍由多个
不同的矩阵和向量的运算组成。
05:59
of different sequences
of matrix and vector operations.
136
359996
3458
06:03
So if you can build a computer
137
363496
1750
所以如果你可以造一台
擅长低精准度矩阵及向量运算
06:05
that is really good at low-precision
matrix and vector operations
138
365246
3792
06:09
but can't do much else,
139
369079
1709
但除此之外没啥用的电脑,
06:10
that's going to be great
for neural-network computation,
140
370829
2625
它会非常适用于神经网络运算,
虽然你无法用它做太多别的事。
06:13
even though you can't use it
for a lot of other things.
141
373496
2667
如果你造出类似的东西,
人们会发现它的妙用。
06:16
And if you build such things,
people will find amazing uses for them.
142
376163
3291
这是我们制作的
第一个成品,TPU v1。
06:19
This is the first one we built, TPU v1.
143
379496
2083
06:21
"TPU" stands for Tensor Processing Unit.
144
381621
2542
“TPU”是张量处理器的意思。
06:24
These have been used for many years
behind every Google search,
145
384204
3042
多年来,这一技术运用于谷歌搜索,
06:27
for translation,
146
387246
1167
翻译,
06:28
in the DeepMind AlphaGo matches,
147
388454
1917
以及AlphaGo围棋比赛,
06:30
so Lee Sedol and Ke Jie
maybe didn't realize,
148
390413
2583
所以李世石和柯洁可能没意识到,
06:33
but they were competing
against racks of TPU cards.
149
393038
2708
他们其实是在和TPU架构比赛。
06:35
And we've built a bunch
of subsequent versions of TPUs
150
395788
2583
后来我们制作了很多个
TPU的后续版本,
06:38
that are even better and more exciting.
151
398371
1875
他们更强大也更振奋人心。
06:40
But despite all these successes,
152
400288
2083
但即使我们取得了这些成功,
06:42
I think we're still doing
many things wrong,
153
402371
2208
我觉得我们仍然做错了很多事,
接下来我将讲三件我们做错的事情,
06:44
and I'll tell you about three
key things we're doing wrong,
154
404621
2792
以及如何修正他们。
06:47
and how we'll fix them.
155
407454
1167
第一个是,现如今的大部分神经网络
06:48
The first is that most
neural networks today
156
408663
2083
06:50
are trained to do one thing,
and one thing only.
157
410746
2250
只被训练进行单一种类的任务。
06:53
You train it for a particular task
that you might care deeply about,
158
413038
3208
你训练它去做一件你很关心的事情,
06:56
but it's a pretty heavyweight activity.
159
416288
1916
但这是一项非常繁重的工作。
06:58
You need to curate a data set,
160
418246
1667
你需要搜索数据组,
06:59
you need to decide
what network architecture you'll use
161
419913
3000
选择这个问题所需的网络架构,
07:02
for this problem,
162
422954
1167
07:04
you need to initialize the weights
with random values,
163
424121
2708
接着随机分配起始比重,
07:06
apply lots of computation
to make adjustments to the weights.
164
426871
2875
然后为调整比重进行大量运算。
07:09
And at the end, if you’re lucky,
you end up with a model
165
429746
2750
到最后,如果你幸运的话,
07:12
that is really good
at that task you care about.
166
432538
2291
可以得到一个非常适用于
你关心的问题的模型。
07:14
But if you do this over and over,
167
434871
1583
但如果你一直这样做,
07:16
you end up with thousands
of separate models,
168
436496
2667
到最后会得到几千个独立的模型,
07:19
each perhaps very capable,
169
439204
1792
每个可能都很有用,
07:21
but separate for all the different
tasks you care about.
170
441038
2791
但都只针对某个单一类型的问题。
07:23
But think about how people learn.
171
443829
1667
想一想人类是怎样学习的。
07:25
In the last year, many of us
have picked up a bunch of new skills.
172
445538
3166
去年,我们中的很多人
都掌握了一些新的技能。
07:28
I've been honing my gardening skills,
173
448746
1792
我沉浸于钻研园艺,
07:30
experimenting with vertical
hydroponic gardening.
174
450579
2375
尝试垂直水培园艺。
07:32
To do that, I didn't need to relearn
everything I already knew about plants.
175
452954
3792
我无需为此重新学习一遍
我已经掌握的有关植物的知识。
07:36
I was able to know
how to put a plant in a hole,
176
456746
3375
我知道怎么把植物放进洞里,
07:40
how to pour water, that plants need sun,
177
460163
2291
怎么浇水,以及植物需要光照,
07:42
and leverage that
in learning this new skill.
178
462496
3583
我只需要整合这些知识
用以学习新的技术。
07:46
Computers can work
the same way, but they don’t today.
179
466079
3042
电脑也可以这样运作,
但现如今它们并没有。
07:49
If you train a neural
network from scratch,
180
469163
2583
如果你从头开始训练一个神经网络,
07:51
it's effectively like forgetting
your entire education
181
471746
3542
每次学习新的东西
都像是让你忘掉
之前学习的所有知识。
07:55
every time you try to do something new.
182
475288
1875
07:57
That’s crazy, right?
183
477163
1000
这太疯狂了对吧!
07:58
So instead, I think we can
and should be training
184
478788
3708
所以比起这样,我认为我们可以,
08:02
multitask models that can do
thousands or millions of different tasks.
185
482538
3791
也应该去训练能完成成千上万种
不同任务的多任务处理模型。
08:06
Each part of that model would specialize
in different kinds of things.
186
486329
3375
此模型的每个部分都有自己的专长。
08:09
And then, if we have a model
that can do a thousand things,
187
489704
2792
然后,假设我们有一个
可以完成一千种任务的模型,
08:12
and the thousand and first
thing comes along,
188
492538
2166
这时候第一千零一种任务来了,
08:14
we can leverage
the expertise we already have
189
494746
2125
我们就可以整合系统中已有的
和这个新任务相关的知识,
08:16
in the related kinds of things
190
496913
1541
08:18
so that we can more quickly be able
to do this new task,
191
498496
2792
从而更快速的完成这项新任务,
08:21
just like you, if you're confronted
with some new problem,
192
501288
2791
就像你本人一样,
如果你面临了一些新问题,
08:24
you quickly identify
the 17 things you already know
193
504121
2625
你会快速识别17件自己已知
并能够帮助解决这些新问题的知识。
08:26
that are going to be helpful
in solving that problem.
194
506746
2542
08:29
Second problem is that most
of our models today
195
509329
2709
第二个问题是,
大部分现今的模型
只能应对一种形态的数据,
08:32
deal with only a single
modality of data --
196
512079
2125
08:34
with images, or text or speech,
197
514204
3084
图片,文字或语音,
08:37
but not all of these all at once.
198
517329
1709
但无法做到一网打尽。
08:39
But think about how you
go about the world.
199
519079
2042
但想一想你如何在这世上生活。
08:41
You're continuously using all your senses
200
521121
2333
你不断地动用你所有的感官
08:43
to learn from, react to,
201
523454
3083
去学习,去做出反应,
08:46
figure out what actions
you want to take in the world.
202
526579
2667
去搞清楚现在应该做什么。
08:49
Makes a lot more sense to do that,
203
529287
1667
这样显然更加合理,
08:50
and we can build models in the same way.
204
530954
2000
我们也可以用同样的方式建造模型。
08:52
We can build models that take in
these different modalities of input data,
205
532954
4042
我们可以建造一个可以接收
所有不同种类数据的模型,
文字,图像,语音,
08:57
text, images, speech,
206
537037
1750
08:58
but then fuse them together,
207
538829
1417
然后把它们融合在一起,
09:00
so that regardless of whether the model
sees the word "leopard,"
208
540329
3833
这样无论这个模型看到文字“豹子”,
09:04
sees a video of a leopard
or hears someone say the word "leopard,"
209
544204
4083
看到豹子的视频,还是
听到有人说出“豹子”这个词
09:08
the same response
is triggered inside the model:
210
548329
2250
它都会触发同样的反应:
09:10
the concept of a leopard
211
550621
1750
一个豹子的概念
09:12
can deal with different
kinds of input data,
212
552412
2250
可以应对很多种不同的数据输入项,
09:14
even nonhuman inputs,
like genetic sequences,
213
554662
3000
甚至是非人工的输入项,
例如基因序列,
09:17
3D clouds of points,
as well as images, text and video.
214
557662
3209
3D点云数据,当然也包括
图片,文字和影像。
09:20
The third problem
is that today's models are dense.
215
560912
3625
第三个问题是现今的模型是稠密的。
09:24
There's a single model,
216
564579
1417
拿一个模型来举例,
09:25
the model is fully activated
for every task,
217
565996
2375
进行任何一项任务时
这个模型都需要被完全激活,
09:28
for every example
that we want to accomplish,
218
568412
2125
任何一个我们想要完成的事情,
09:30
whether that's a really simple
or a really complicated thing.
219
570537
2917
无论困难与否。
09:33
This, too, is unlike
how our own brains work.
220
573496
2666
这再一次有悖人脑运作的习惯。
09:36
Different parts of our brains
are good at different things,
221
576204
3000
大脑不同的区块擅长不同的工作,
09:39
and we're continuously calling
upon the pieces of them
222
579246
3291
我们会不断的运用
与需要完成的任务相关的区块。
09:42
that are relevant for the task at hand.
223
582579
2167
09:44
For example, nervously watching
a garbage truck
224
584787
2334
比如,紧张地看着一辆垃圾车
朝自家车的方向倒车,
09:47
back up towards your car,
225
587162
1875
09:49
the part of your brain that thinks
about Shakespearean sonnets
226
589037
2917
你大脑中负责思考莎士比亚的
十四行诗的区块应该不会被激活。
09:51
is probably inactive.
227
591996
1250
09:53
(Laughter)
228
593246
1625
(笑声)
09:54
AI models can work the same way.
229
594912
1750
人工智能模型也可以这样运作。
09:56
Instead of a dense model,
230
596662
1292
比起稠密模型,
09:57
we can have one
that is sparsely activated.
231
597954
2250
我们可以制造一个
激活反应较稀松的模型。
10:00
So for particular different tasks,
we call upon different parts of the model.
232
600204
4250
在做不同任务的时候,
我们运用模型中不同的区块。
10:04
During training, the model can also learn
which parts are good at which things,
233
604496
4375
在训练时,模型也可以学习
哪个区块擅长哪个领域,
10:08
to continuously identify what parts
it wants to call upon
234
608871
3791
从而不断地识别需要运用
哪些区块来完成新的任务。
10:12
in order to accomplish a new task.
235
612662
1834
10:14
The advantage of this is we can have
a very high-capacity model,
236
614496
3541
此类模型的优势在于
虽然容量大但十分高效,
10:18
but it's very efficient,
237
618037
1250
10:19
because we're only calling
upon the parts that we need
238
619329
2583
因为我们只会用到
完成某项任务所需的区块。
10:21
for any given task.
239
621954
1208
10:23
So fixing these three things, I think,
240
623162
2000
因此我认为,解决了这三个问题,
10:25
will lead to a more powerful AI system:
241
625162
2209
人工智能的能力将更上一层楼:
10:27
instead of thousands of separate models,
242
627412
2000
比起几千个独立的模型,
[独立模型/通用模型]
10:29
train a handful of general-purpose models
243
629412
2000
训练几个能够应对
成千上万件事情的通用模型。
10:31
that can do thousands
or millions of things.
244
631454
2083
10:33
Instead of dealing with single modalities,
245
633579
2042
比起接收单一的数据形态,
10:35
deal with all modalities,
246
635662
1334
让它处理所有的数据形态
并把它们整合到一起。
10:36
and be able to fuse them together.
247
636996
1708
10:38
And instead of dense models,
use sparse, high-capacity models,
248
638746
3458
比起稠密模型,
运用稀松且高容量的模型,
10:42
where we call upon the relevant
bits as we need them.
249
642246
2958
让我们能按需激活指定区块。
10:45
We've been building a system
that enables these kinds of approaches,
250
645246
3416
我们已经在制造
符合以上条件的模型了,
10:48
and we’ve been calling
the system “Pathways.”
251
648704
2542
我们给它取了个名字
叫“Pathways”。
10:51
So the idea is this model
will be able to do
252
651287
3084
我们的理念是
10:54
thousands or millions of different tasks,
253
654412
2084
这个模型可以完成
成千上万种不同类型的任务,
10:56
and then, we can incrementally
add new tasks,
254
656537
2250
然后我们可以逐步增加新的任务,
10:58
and it can deal
with all modalities at once,
255
658787
2125
它也可以同时处理各种形态的数据,
11:00
and then incrementally learn
new tasks as needed
256
660954
2958
然后逐步学习新技能,
11:03
and call upon the relevant
bits of the model
257
663954
2083
并按需为不同任务启动不同区块。
11:06
for different examples or tasks.
258
666037
1709
11:07
And we're pretty excited about this,
259
667787
1750
我们对此感到非常兴奋,
11:09
we think this is going
to be a step forward
260
669537
2042
我们认为这将是人工智能
系统建造迈出的重要一步。
11:11
in how we build AI systems.
261
671621
1333
11:12
But I also wanted
to touch on responsible AI.
262
672954
3708
不过我也想浅谈一下
什么是可靠的人工智能。
11:16
We clearly need to make sure
that this vision of powerful AI systems
263
676662
4875
我们一定要确保这样强大的
人工智能系统造福所有人。
11:21
benefits everyone.
264
681579
1167
11:23
These kinds of models raise
important new questions
265
683496
2458
此类模型将一些重要的
新问题带进大众的视野,
11:25
about how do we build them with fairness,
266
685954
2458
我们应该如何确保系统于所有人而言
11:28
interpretability, privacy and security,
267
688454
3208
都具有公平性、可解释性、
私密性以及安全性。
11:31
for all users in mind.
268
691662
1459
11:33
For example, if we're going
to train these models
269
693621
2291
例如,如果我们要训练
这些模型完成成千上万种任务,
11:35
on thousands or millions of tasks,
270
695954
2125
11:38
we'll need to be able to train them
on large amounts of data.
271
698079
2875
我们需要极大量的数据。
11:40
And we need to make sure that data
is thoughtfully collected
272
700996
3250
我们一定要确保
这些数据的采集经过深思熟虑,
11:44
and is representative of different
communities and situations
273
704287
3667
它们需要代表世界上
不同的社群和情况。
11:47
all around the world.
274
707954
1167
11:49
And data concerns are only
one aspect of responsible AI.
275
709579
4000
数据上的担忧仅仅是
可靠的人工智能这个议题的一部分。
11:53
We have a lot of work to do here.
276
713621
1583
我们要做的还有很多。
11:55
So in 2018, Google published
this set of AI principles
277
715246
2666
2018年,谷歌发表了在开发
此类科技时应注意的人工智能守则。
11:57
by which we think about developing
these kinds of technology.
278
717912
3500
12:01
And these have helped guide us
in how we do research in this space,
279
721454
3625
这帮助指导了我们
如何在这个领域从事研究,
12:05
how we use AI in our products.
280
725121
1833
以及如何在产品中使用人工智能。
12:06
And I think it's a really helpful
and important framing
281
726996
2750
我认为这对思考复杂且有深度的问题,
12:09
for how to think about these deep
and complex questions
282
729746
2875
及在社会中运用人工智能
都有帮助且很重要。
12:12
about how we should
be using AI in society.
283
732621
2833
12:15
We continue to update these
as we learn more.
284
735454
3625
我们在学习过程中不断更新该守则,
12:19
Many of these kinds of principles
are active areas of research --
285
739079
3458
很多此类准则都是
现在的热点研究领域,
12:22
super important area.
286
742537
1542
它们非常重要。
12:24
Moving from single-purpose systems
that kind of recognize patterns in data
287
744121
4125
由只能识别数据中的
图案规律的单一用途系统,
12:28
to these kinds of general-purpose
intelligent systems
288
748246
2750
到对世界有更深理解的
通用智能系统的转变,
12:31
that have a deeper
understanding of the world
289
751037
2292
12:33
will really enable us to tackle
290
753329
1583
真的可以赋予我们能力
去解决人类面临的重大问题。
12:34
some of the greatest problems
humanity faces.
291
754954
2500
12:37
For example,
292
757454
1167
比如,
12:38
we’ll be able to diagnose more disease;
293
758662
2375
我们可以诊断更多疾病,
12:41
we'll be able to engineer better medicines
294
761079
2000
我们可以通过给这些模型灌输化学
和物理知识从而设计出更好的药品,
12:43
by infusing these models
with knowledge of chemistry and physics;
295
763079
3083
12:46
we'll be able to advance
educational systems
296
766204
2667
我们可以通过个性化定制的补习服务
来优化现有的教育系统,
12:48
by providing more individualized tutoring
297
768871
2041
12:50
to help people learn
in new and better ways;
298
770912
2375
让人们用更先进的方法来学习,
12:53
we’ll be able to tackle
really complicated issues,
299
773287
2375
我们可以解决十分复杂的难题,
12:55
like climate change,
300
775704
1208
例如全球变暖,
12:56
and perhaps engineering
of clean energy solutions.
301
776954
2792
或是帮助清洁能源的发展。
12:59
So really, all of these kinds of systems
302
779787
2667
所以其实,所有此类系统
13:02
are going to be requiring
the multidisciplinary expertise
303
782496
2875
都需要世界各地的
多学科专家共同协作。
13:05
of people all over the world.
304
785371
1875
13:07
So connecting AI
with whatever field you are in,
305
787287
3542
将人工智能和你所在的领域相结合,
13:10
in order to make progress.
306
790829
1750
从而推动产业的进程。
13:13
So I've seen a lot
of advances in computing,
307
793579
2292
我看到了许多计算机科学的优势,
13:15
and how computing, over the past decades,
308
795912
2292
以及在过去的几十年中计算机科学
如何帮助几百万人更好地理解世界。
13:18
has really helped millions of people
better understand the world around them.
309
798204
4167
13:22
And AI today has the potential
to help billions of people.
310
802412
3000
今天的人工智能拥有
帮助数十亿人的潜力。
13:26
We truly live in exciting times.
311
806204
2125
我们真的生活在
一个振奋人心的时代。
13:28
Thank you.
312
808746
1166
谢谢。
13:29
(Applause)
313
809912
7000
(掌声)
13:39
Chris Anderson: Thank you so much.
314
819829
1667
克里斯·安德森:非常感谢。
13:41
I want to follow up on a couple things.
315
821537
2375
我还有几个问题。
13:44
This is what I heard.
316
824454
2792
据我所知,
13:47
Most people's traditional picture of AI
317
827287
4125
大部分人对于人工智能的印象
13:51
is that computers recognize
a pattern of information,
318
831412
3125
是电脑识别出信息的模式,
13:54
and with a bit of machine learning,
319
834579
2125
运用一些机器学习的技术,
13:56
they can get really good at that,
better than humans.
320
836704
2625
就可以比人类更擅长某项工作。
13:59
What you're saying is those patterns
321
839329
1792
你所说的模式已经不再是
人工智能所处理的原子层面了,
14:01
are no longer the atoms
that AI is working with,
322
841121
2875
14:04
that it's much richer-layered concepts
323
844037
2542
而是一种层次更加丰富的概念
比如组成一只美洲豹的,
14:06
that can include all manners
of types of things
324
846621
3458
14:10
that go to make up a leopard, for example.
325
850121
3000
各种形式的各种信息。
14:13
So what could that lead to?
326
853121
2916
那这会产生什么样的结果呢?
14:16
Give me an example
of when that AI is working,
327
856079
2750
请举个例子,当人们使用人工智能
14:18
what do you picture happening in the world
328
858829
2042
你预想接下来五到十年中
世界上会发生什么有趣的事情?
14:20
in the next five or 10 years
that excites you?
329
860912
2209
14:23
Jeff Dean: I think
the grand challenge in AI
330
863537
2500
杰夫·迪恩:我认为人工智能
面临的最大的挑战是
14:26
is how do you generalize
from a set of tasks
331
866079
2375
如何尽可能简单且高效地
14:28
you already know how to do
332
868496
1416
14:29
to new tasks,
333
869954
1208
运用已知的知识来完成新任务。
14:31
as easily and effortlessly as possible.
334
871204
2542
14:33
And the current approach of training
separate models for everything
335
873746
3250
现在我们训练一个通用模型的方法
14:36
means you need lots of data
about that particular problem,
336
876996
3166
需要大量的与该问题
相关的数据支持,
14:40
because you're effectively trying
to learn everything
337
880162
2542
因为我们在试图从零开始高效地学习
世界上包括那个问题在内的所有知识。
14:42
about the world
and that problem, from nothing.
338
882746
2541
14:45
But if you can build these systems
339
885287
1667
不过如果你可以建造出一个
14:46
that already are infused with how to do
thousands and millions of tasks,
340
886996
4458
能完成成千上万种任务的系统,
14:51
then you can effectively
teach them to do a new thing
341
891496
3791
你就只需要用较少的数据
去教会它一个新的东西。
14:55
with relatively few examples.
342
895329
1500
14:56
So I think that's the real hope,
343
896871
2083
我认为这是一个希望,
14:58
that you could then have a system
where you just give it five examples
344
898996
4083
只需给这个系统五个
和你想完成的任务相关的事例,
15:03
of something you care about,
345
903121
1541
15:04
and it learns to do that new task.
346
904704
1917
它就能学会这个新技能。
15:06
CA: You can do a form
of self-supervised learning
347
906662
2375
克里斯:你只需要一点点‘肥料’
就能打造一个自我监督学习系统。
15:09
that is based on remarkably
little seeding.
348
909079
2542
15:11
JD: Yeah, as opposed to needing
10,000 or 100,000 examples
349
911662
2875
杰夫:是的,比起用一万或者十万个
事例来搞明白世界上所有的事情。
15:14
to figure everything in the world out.
350
914579
1833
15:16
CA: Aren't there kind of terrifying
unintended consequences
351
916454
2792
克里斯:它可能带来什么
非常恐怖的后果吗?
15:19
possible, from that?
352
919287
1334
15:20
JD: I think it depends
on how you apply these systems.
353
920621
2625
杰夫:我认为这取决于
你如何使用此类系统。
15:23
It's very clear that AI
can be a powerful system for good,
354
923287
4167
显然人工智能可以是
普惠大众的一大利器,
15:27
or if you apply it in ways
that are not so great,
355
927496
2333
也可能因为不当的使用
而产生负面影响。
15:29
it can be a negative consequence.
356
929871
3375
15:33
So I think that's why it's important
to have a set of principles
357
933287
3042
所以我认为这体现出了
行为准则的重要性,
15:36
by which you look at potential uses of AI
358
936371
2250
在你放眼人工智能的无限可能时,
15:38
and really are careful and thoughtful
about how you consider applications.
359
938662
5125
谨慎且缜密地思考它的用途。
15:43
CA: One of the things
people worry most about
360
943787
2209
克里斯:其中一个人们最担心的点是
15:45
is that, if AI is so good at learning
from the world as it is,
361
945996
3583
如果人工智能如此擅长学习,
15:49
it's going to carry forward
into the future
362
949621
2875
它可能对未来世界的格局产生影响,
15:52
aspects of the world as it is
that actually aren't right, right now.
363
952537
4375
但这在现在看来还不太能被接受。
15:56
And there's obviously been
a huge controversy about that
364
956954
2708
显然最近针对谷歌也有
一个非常大的争议。
15:59
recently at Google.
365
959662
2417
16:02
Some of those principles
of AI development,
366
962121
3083
一些之前提到的人工智能开发准则
16:05
you've been challenged that you're not
actually holding to them.
367
965246
4708
被质疑并未被遵守。
16:10
Not really interested to hear
about comments on a specific case,
368
970329
3250
倒不用说一些具体事例,
16:13
but ... are you really committed?
369
973621
2708
不过……你们真的有遵守吗?
16:16
How do we know that you are
committed to these principles?
370
976329
2750
我们如何知道谷歌
是否遵守这些条规?
16:19
Is that just PR, or is that real,
at the heart of your day-to-day?
371
979079
4417
它只是一个公关行为,
还是它真的时刻存于你们心中。
16:23
JD: No, that is absolutely real.
372
983496
1541
杰夫:毋庸置疑是真的。
16:25
Like, we have literally hundreds of people
373
985079
2125
我们差不多有几百人
16:27
working on many of these
related research issues,
374
987204
2333
在研究与此相关的议题,
16:29
because many of those
things are research topics
375
989537
2417
因为很多这种准则
本身就是研究议题。
16:31
in their own right.
376
991996
1166
16:33
How do you take data from the real world,
377
993162
2250
你如何在现实世界中提取数据,
16:35
that is the world as it is,
not as we would like it to be,
378
995412
5000
我们需要还原世界本来的样子,
而非我们想让它变成的样子,
16:40
and how do you then use that
to train a machine-learning model
379
1000412
3209
接着是如何用这些数据
训练一个机器学习模型,
16:43
and adapt the data bit of the scene
380
1003621
2500
调整或在原有基础上增加一些数据,
16:46
or augment the data with additional data
381
1006162
2500
16:48
so that it can better reflect
the values we want the system to have,
382
1008704
3292
从而使系统反映出我们想要的价值,
16:51
not the values that it sees in the world?
383
1011996
2208
而非那些它在世界中观察到的。
16:54
CA: But you work for Google,
384
1014204
2417
克里斯:但你在谷歌工作,
16:56
Google is funding the research.
385
1016662
2084
谷歌在为这些研究提供资金。
16:59
How do we know that the main values
that this AI will build
386
1019371
4458
我们怎么知道开发这些人工智能
最大的价值在于全人类的福祉,
17:03
are for the world,
387
1023871
1208
17:05
and not, for example, to maximize
the profitability of an ad model?
388
1025121
4916
而不是,比如说使一个
广告模型利益最大化?
17:10
When you know everything
there is to know about human attention,
389
1030037
3084
当你知道了有关
人类关注点的所有信息,
17:13
you're going to know so much
390
1033121
1500
你对于我们内心的扭曲
和阴暗就了解的太多了。
17:14
about the little wriggly,
weird, dark parts of us.
391
1034621
2458
17:17
In your group, are there rules
about how you hold off,
392
1037121
6167
在你的团队中,有没有阻止
不当行为的规则,
在商业压力前,有没有
竖起一面政教之墙,
17:23
church-state wall
between a sort of commercial push,
393
1043329
3750
17:27
"You must do it for this purpose,"
394
1047079
2292
要求“你必须这样做,”
17:29
so that you can inspire
your engineers and so forth,
395
1049413
2458
从而激励你的工程师们之类的,
17:31
to do this for the world, for all of us.
396
1051913
1916
为世界,为我们考虑。
17:33
JD: Yeah, our research group
does collaborate
397
1053871
2125
杰夫:是的,我们的研究团队
会和谷歌的其它一些团队合作,
17:36
with a number of groups across Google,
398
1056038
1833
包括广告部门,
搜索部门,地图部门,
17:37
including the Ads group,
the Search group, the Maps group,
399
1057913
2750
我们有这样的合作,
但也有许多纯粹的公开发表的文献。
17:40
so we do have some collaboration,
but also a lot of basic research
400
1060704
3209
17:43
that we publish openly.
401
1063954
1542
我们去年发表了一千多份论文,
17:45
We've published more
than 1,000 papers last year
402
1065538
3333
涉及各种领域,也包含了
许多你之前提到的议题,
17:48
in different topics,
including the ones you discussed,
403
1068871
2583
例如公平性,
机器学习模型的可解释性,
17:51
about fairness, interpretability
of the machine-learning models,
404
1071454
3042
那些很重要的东西,
17:54
things that are super important,
405
1074538
1791
17:56
and we need to advance
the state of the art in this
406
1076371
2417
且我们需要优化它的最高标准,
17:58
in order to continue to make progress
407
1078829
2209
从而继续确保这些模型
在安全且可靠的情况下被开发。
18:01
to make sure these models
are developed safely and responsibly.
408
1081079
3292
18:04
CA: It feels like we're at a time
when people are concerned
409
1084788
3041
克里斯:我感觉我们处在一个人们
十分忌惮大型科技公司能力的时代,
18:07
about the power of the big tech companies,
410
1087829
2042
18:09
and it's almost, if there was ever
a moment to really show the world
411
1089871
3750
如果能有机会真正向世界展示
18:13
that this is being done
to make a better future,
412
1093663
3333
这是为了建造更好的未来,
18:17
that is actually key to Google's future,
413
1097038
2791
这是谷歌未来的关键,
18:19
as well as all of ours.
414
1099871
1750
也是所有人的。
18:21
JD: Indeed.
415
1101663
1166
杰夫:完全正确。
18:22
CA: It's very good to hear you
come and say that, Jeff.
416
1102871
2583
克里斯:非常高兴
听到你这么说,杰夫。
非常感谢你来到TED。
18:25
Thank you so much for coming here to TED.
417
1105454
2042
杰夫:谢谢
18:27
JD: Thank you.
418
1107538
1166
(掌声)
18:28
(Applause)
419
1108704
1167
New videos
Original video on YouTube.com
关于本网站
这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。