What Makes Us Human in the Age of AI? A Psychologist and a Technologist Answer | TED Intersections
48,718 views ・ 2024-09-10
请双击下面的英文字幕来播放视频。
00:00
Brian S. Lowery: If you could produce
a more immersive social experience,
0
79
4046
翻译人员: Eric Ma
校对人员: Manlin Fang
Brian:如果你能创造出
更身临其境的社交验,
00:04
now everybody's having
their individual social experiences.
1
4167
3670
那么,现在每个人都有
自己的社交体验。
00:07
Like now what I worry
about with AI, with VR,
2
7837
4087
就像我现在
担心的人工智能、虚拟现实、
00:11
with all these kind of technologies
that are expanding,
3
11966
4171
所有这些正在扩展的技术一样,
00:16
we all inhabit our own singular world.
4
16179
2919
我们都生活在我们自己的单一世界中。
00:19
That is more frightening to me
than like, you know,
5
19140
3503
对我来说,这比我们都经历了同样
的事情更可怕,你懂的。
00:22
that we all converged
in the same experience.
6
22685
2336
00:25
[Intersections]
7
25063
3086
[十字路口]
(由TED提供)
00:34
[Brian S. Lowery: Social psychologist]
8
34363
3003
[Brian S. Lowery:社会心理学家]
00:37
[Kylan Gibbs: Technologist]
9
37950
3504
[Kylan Gibbs:技术专家]
00:41
BSL: So what makes a human a human?
10
41454
2502
BSL:那么是什么让人类成为人呢?
00:45
(Laughter)
11
45625
2335
(笑声)
00:48
Kylan Gibbs: It’s one of
those questions, isn’t it?
12
48002
2419
这是其中一个问题,
不是吗?
00:50
I mean, there's like,
two ways I would look at it.
13
50588
2377
我的意思是,我大概有
两种看法。
00:52
One is from my personal life
and one is from my work life.
14
52965
3045
一种来自我的个人生活,
另一种来自我的工作。
00:56
One thing that's interesting is like,
15
56010
1794
一件有趣的事是,
00:57
there's been points
when I've been spending
16
57845
2044
有时候我每天花
00:59
four to five hours a day
interacting with AI.
17
59889
2211
四到五个小时与人工智能互动。
01:02
And the interesting thing that happens
in that is the things that you notice,
18
62100
3628
其中,有趣的事情是当你第一次
开始与它互动时,
01:05
when you first start interacting with it,
oh, this is really realistic.
19
65728
3337
你注意到,
哦,这真的很现实。
01:09
Similarly when people first had
black and white TV and they're like,
20
69107
3211
同样,当人们第一次拥有
黑白电视时,他们会想,
01:12
wow, this is like real life.
21
72360
1376
哇,这就像现实生活一样。
01:13
But then as you get used to it,
you start to kind of realize,
22
73736
2878
但是当你习惯了之后,
你就会开始意识到
01:16
the things that make it less authentic.
23
76614
1877
那些让它变得不那么真实的东西。
01:18
And I think something
that I realized with AI is
24
78491
2294
而且我认为,通过人工智能,
01:20
there's certain ways that we interact
that are just more spontaneous.
25
80785
3253
我意识到我们的某些互动
方式更加自发。
01:24
There's something about the predictability
of AI that teaches you
26
84080
3086
人工智能的可预测性可以教会你
01:27
about the spontaneity of being human.
27
87166
1794
人类的自发性。
01:28
The ways they communicate,
the naturalness, the contextual awareness.
28
88960
3253
他们的沟通方式、
自然性、情境意识。
01:32
These little things that all add up.
29
92255
1835
这些小东西加起来了。
01:34
That's on the technical side.
30
94132
1418
那是在技术方面。
01:35
On the other side,
31
95550
1167
另一方面,
01:36
there's something of just the shared
experience of being human
32
96759
2920
我认为人类的共同
经历实际上
01:39
that actually I think differentiates it
from other animals’ experience.
33
99679
3336
使它与其他动物
的经历有所不同。
01:43
You know, you have
a traumatic moment in your life,
34
103057
2628
你的生活中
有一个痛苦的时刻,
01:45
and then you start to resonate
with other people's.
35
105685
2919
然后你开始引起他人的共鸣。
01:48
I feel like every time I've had
something nearly catastrophic,
36
108604
2962
我觉得每当我遇到
近乎灾难性事情时,
01:51
it opened up a new door of empathy,
and then you start to be like,
37
111607
3212
它都会打开一扇新的同理心之门,
然后你会开始觉得,
01:54
oh man, that really hurt, you know?
38
114819
1960
天啊,真的很痛苦,你知道吗?
01:56
Or like when you cry
about something, you're like, wow.
39
116821
2586
或者就像当你为某件
事哭的时候,你会想,哇。
01:59
And then you start to remember,
like this is what usually happens to me.
40
119407
3462
然后你开始记住,
就像我经常发生的事情一样。
02:02
I start crying about something,
41
122869
1543
我开始为某件事哭泣,
02:04
and then I think about all the things
that I did for my mom or my grandma
42
124412
3462
然后我想起我为妈妈
或奶奶所做的一切
02:07
and the things that they felt.
43
127874
1459
以及他们的感受。
02:09
And I feel like there's something
in that kind of like shared experience
44
129333
3420
而且我觉得,
在那种共同的体验中
02:12
where we have these things
that differentiate us,
45
132753
2294
有一些东西
可以使我们与众不同,
02:15
we’re all different people.
46
135047
1335
我们都是不同的人。
02:16
But there’s something
about those common feelings
47
136382
2294
但是,这些共同的感受
02:18
that it all kind of arises from that.
48
138718
1793
在某种程度上是由此产生的。
02:20
Anyway, that's one thought.
49
140511
1293
无论如何,这是一个想法。
02:21
BSL: I love that answer, and I want to say
that you're not normal in that way.
50
141846
3670
BSL:我喜欢这个答案,我想
说你这样不正常。
02:25
Here's why I don't think you're normal.
51
145558
1877
这就是为什么我认为你不正常:
02:27
People anthropomorphize anything.
52
147435
1585
人们将任何事物拟人化。
02:29
It doesn't have to even
be that good, right?
53
149061
2086
甚至不必那么好,
对吧?
02:31
It doesn't have to be
anywhere near as good AI
54
151147
2169
人们不必像对待人
一样对待它好,
02:33
for people to treat it
like it has some human character,
55
153316
2669
不一定要把它当成
一个真人一样对待
02:35
people treat their cars
like they're people.
56
155985
2085
人们想对待一个人
一样对待他们的车
02:38
So I'm surprised that when
you interact with it a lot,
57
158070
3504
因此,令我惊讶的是
当你经常与它互动时,
02:41
it feels less real to you.
58
161616
2919
对你来说感觉不那么真实。
02:44
KG: There's something about resolution.
59
164535
3587
KG:分辨率有些问题。
02:48
It's like the way of seeing the world
and you kind of increase this.
60
168164
3211
这就像看待世界的方式,
你在某种程度上增加了这个水平。
02:51
It's like the same reason
you can't look at TV that's not 4K now.
61
171375
3587
这就像 你现在不能看不是
4K 的电视的原因一样。
02:54
And it's someone I think who worked
on early VR was saying, you know,
62
174962
4171
我想是开发早期虚拟
现实的人在说,对吧,
02:59
the interesting thing about it was
when you stepped out of it,
63
179175
2919
它的有趣之处在于,
当你走出虚拟现实时,
03:02
you're like, oh, the real world
is actually amazing.
64
182136
2461
你会想,哦,现实世界
真的很神奇。
03:04
And it's actually really hard
to recreate that in technology.
65
184597
2878
实际上,要在技术中
重现这一点真的很困难。
03:07
And I think the same is true for AI,
it's like maybe for some people,
66
187475
3295
而且我认为人工智能也是如此,
就像某些人一样,
03:10
when they interact with it,
67
190770
1293
当他们与人工智能互动时,
03:12
the thing that they see
is some commonality.
68
192104
2086
他们看到
的是某种共同点。
03:14
But the thing that I always notice is,
69
194190
1835
但我经常注意到的是,
03:16
this is very different from
the conversations I have with my parents.
70
196025
3253
这与我和父母
的对话有很大的不同。
03:19
Even when it says something similar,
there’s something off.
71
199320
2961
即使它说了类似的话,
也有一些不对劲。
03:22
It's those little things,
72
202323
1209
正是这些小东西,
03:23
that's like I think what,
over time, will add up
73
203574
2252
就像我认为,
随着时间的推移,
03:25
as people use AI more,
is they’ll start to recognize,
74
205826
2503
随着人们越来越多地使用人工智能,
他们会开始识别,
03:28
and I can't even point at them like,
what are those nuances, though,
75
208329
3253
我甚至无法指着它们,但是,
03:31
that make us human?
76
211582
1168
那些使我们成为人类的细微差别
是什么?
03:32
BSL: You just know it
when you see it and you're like,
77
212792
2544
BSL:
当你看到它时你就会知道,
03:35
and it's missing an AI.
78
215378
1167
而且它缺少人工智能。
03:36
I mean that's also interesting
79
216545
1460
我的意思是这也很有趣,
03:38
because what you just suggested
is that the more people use AI,
80
218005
3170
因为你刚才想说的是,
使用人工智能的人越多,
03:41
the less real it's going
to feel to people.
81
221175
2586
人们的感觉
就越不真实。
03:43
Do you think that's
what's going to happen?
82
223761
2044
你认为会发生这样
的事情吗?
03:46
KG: I mean, there's probably
another case, you know,
83
226389
2460
KG:我的意思是,可能还有
另一种情况,你知道,
03:48
it's the same way as, you know,
84
228849
1502
就像你的Instagram
03:50
your Instagram and Facebook feed
isn't a real conversation.
85
230351
2794
和Facebook推送一样
不是真正的对话。
03:53
There are certainly, kids especially,
who would look at those kinds of feeds
86
233145
3587
当然,尤其是孩子们,
他们会看着这样的社交媒体
03:56
and feel like, oh, that's a real
representation of my friends
87
236774
2878
然后觉得,哦,那是我朋友们
的真实写照
03:59
or my favorite celebrities
or whatever I actually think,
88
239652
2628
或者是我最喜欢的名人
或者我真正想到的任何东西,
04:02
when it's like completely --
I shouldn't say completely --
89
242280
2752
而这完全是错误的-
我不应该完全说
04:05
largely false.
90
245032
1168
在很大程度上是错误的。
04:06
And I do think something similar
will happen with AI,
91
246242
2502
而且我确实认为人工智能
也会发生类似
04:08
where some people for sure
will almost be encaptured.
92
248744
3295
的情况,有些人肯定
会几乎被抓住。
04:12
And they will believe
93
252039
1210
他们会相信
04:13
that that's the most
realistic thing that exists
94
253249
2252
这是现存的最
现实的东西,
04:15
and then start to compare people to that.
95
255543
1960
然后开始将人与之进行比较。
04:17
But I think that, you know,
if you have that degree of empathy,
96
257545
2961
但我认为,你知道,
如果你有这种程度的同理心,
04:20
you'll be like, oh,
there's something off here.
97
260548
2210
你会想,哦,这里
有一些东西。
04:22
It's the same way even if you use
a Zoom call, there's something off.
98
262758
3295
即使你使用 Zoom通话,
也是如此,也会有一些问题。
04:26
It's hard to pick it up.
99
266053
1168
很难把它捡起来。
04:27
But like, I’m not picking up
all the signals,
100
267221
2127
但是,比如,我不会接收
所有的信号
04:29
and it's the very little nuances
101
269348
1543
,你可能也会
巧妙地听到
04:30
that you probably
just subtly pick up as well.
102
270891
2169
这些细微的细微差别。
04:33
BSL: So you don't think
that the technology
103
273060
2002
BSL:那么你
认为这项技术的发展速度还不
04:35
is going to advance quickly enough,
104
275104
1710
够快,它能以足够快的速度
04:36
where it’ll overcome those little things
fast enough to capture all of us?
105
276814
3587
克服这些足以俘获我们
所有人的小问题吗?
04:40
You're not worried about that?
106
280443
1459
你不担心吗?
04:41
KG: I am definitely worried about that.
107
281902
1877
KG:我对此肯定很担心。
04:43
Mainly because because I think
for most people it's easy, right?
108
283779
4171
主要是因为我认为
对大多数人来说这很容易,对吧?
04:47
So the thing about AI is it's so beholden,
109
287992
3045
因此,人工智能的问题在于
它非常受其束缚,
04:51
at least if you think
about like, the chatbot styles,
110
291078
2503
至少如果你
想象聊天机器人的风格,
04:53
it's so beholden to what we want.
111
293581
1585
它是如此受我们想要的东西的束缚。
04:55
And that's kind of
like what people, I think,
112
295166
2169
我认为,
04:57
a lot of people want
in their life or they need,
113
297335
2252
这有点 像人们
在生活中想要或需要
04:59
is the sense of control.
114
299587
1168
的是控制感。
05:00
And the AI gives you the sense that, like,
115
300796
2002
人工智能给你一种感觉,比如,
05:02
I can control this anthropomorphic thing.
116
302840
2377
我可以控制这个拟人化的东西。
05:05
And honestly, one of my fears
is that people get used to that.
117
305217
3337
老实说,我的恐惧之一
是人们会习惯这种情况。
05:08
And what does it mean when
I get used to interacting with something
118
308554
3170
当 我习惯于与只受
我的观点和兴趣约束
05:11
that is beholden to only
my views and interests,
119
311766
2252
的事物互动,然后去
05:14
and then I go and interact with a human
who has their own interests?
120
314060
3211
和一个有自己兴趣
的人互动时,这意味着什么?
05:17
BSL: Do you think people want control?
121
317313
1835
BSL:你认为人们想要控制权吗?
05:19
I think people want to be controlled.
122
319148
1793
我认为人们想要被控制。
05:23
KG: Maybe it's a form of control, though.
123
323277
1960
KG:不过,也许这是一种控制形式。
05:25
To be controlled
is a predictability, I guess.
124
325279
2169
我猜被控制
是一种可预测性。
05:27
BSL: Yeah, people want the world
to make sense, don't you think?
125
327448
3128
BSL:是的,人们
希望世界变得有意义,你不觉得吗?
05:30
KG: Yes, yes, I think they also
want the world to be ...
126
330576
3712
KG:是的,是的,我想他们也
希望世界成为...
05:34
There's something about, like,
preferring predictability over optimality.
127
334288
4379
比如说,
更倾向于可预测性而不是最优性。
05:38
So, like, I've even felt it
when you have, you know,
128
338709
2836
所以,比如,
当你有心理健康的时刻时,我甚至
05:41
a mental health moment,
129
341587
1627
有过这样的感觉,
05:43
you have friends who have
mental health moments.
130
343255
2253
你的朋友有
心理健康的时刻。
05:45
The things that I've
always seen as interesting
131
345549
2211
我一直认为有趣的事情
05:47
is your brain and your mind
prefer to stay in a state that's familiar,
132
347760
3420
是你的大脑和大脑
宁愿保持一种熟悉的状态,
05:51
even if it's worse.
133
351222
1168
即使情况更糟。
05:52
So if you're in like a depressed state,
134
352390
1876
因此,如果你处于抑郁状态,
05:54
you almost would rather like stick in that
than break outside of it, right?
135
354308
3545
你几乎宁愿坚持下去,
也不愿逃脱困境,对吧?
05:57
So there's something
about things that are familiar
136
357853
2628
因此,有些
东西是熟悉的,
06:00
rather than actually better.
137
360523
1501
而不是更好的。
06:02
And I don't know, there's a bias
towards, you know,
138
362066
2586
而且我不知道,你知道,
那时人们会有偏见
06:04
you almost identifying then
with those kinds of states.
139
364693
2712
你几乎认同这几种状态。
06:07
BSL: Yeah, there's research on this.
140
367405
1751
BSL:是的,有关于这个的研究。
06:09
One, it's called the status quo bias.
141
369198
1793
第一,它被称为现状偏差。
06:10
People like things that are already there.
142
370991
2002
人们喜欢已经存在的东西。
06:13
And two, people like to have what they
believe about themselves affirmed
143
373035
3420
第二,如果人们真的相信他们,
06:16
if they really believe them,
even if they're not positive.
144
376455
3045
即使他们不是积极的,他们也喜欢
让他们对自己的看法得到肯定。
06:19
So that is true.
145
379500
2169
所以这是真的。
06:21
So, like, what does that look like in AI?
146
381669
2461
那么,比如,在人工智能
中会是什么样子?
06:24
(Laughter)
147
384130
1710
(笑声)
06:28
KG: I mean, it's definitely interesting
to me that people seem to love like,
148
388384
4129
KG:我的意思是,对我来说,
很有意思的是,人们看起来很喜欢,
06:32
you talk to a lot of these things
and they sound like computers
149
392555
3003
你说很多这样的东西,
它们听起来像计算机,
06:35
and they sound like AI,
but people love it
150
395558
1960
听起来像人工智能,
但人们喜欢它,
06:37
because it's kind of
familiar, it's controllable.
151
397560
2335
因为它有点
熟悉,是可以控制的。
06:39
If you start to add lots of personalities
and these kinds of things,
152
399937
3212
如果你开始添加很多个性
之类的东西,
06:43
it makes sense in context,
but I found it interesting
153
403190
2503
这在情境中是有道理的,
但我发现有趣的是,
06:45
that as we started developing
these AI systems
154
405693
2169
当我们开始开发这些
可供人们互动的人工智能系统时,
06:47
that people interact with,
155
407903
1252
06:49
they all have this kind of similar voice.
156
409196
1961
它们都有类似的声音。
06:51
And it's a very "AI voice."
157
411157
1293
而且这是一种非常
“人工智能的声音”。
06:52
You can kind of tell
that you're talking to an AI.
158
412491
2336
你可以分辨出你
在和人工智能说话。
06:54
Maybe that’s intentional.
159
414869
1209
也许这是故意的。
06:56
But there is something there, where like,
I think people have a preference
160
416078
4463
但也有一些东西,比如,
我认为人们更
07:00
to getting what they want
from humans, from humans,
161
420541
2419
愿意从人类到人类、
07:02
and from AI, from AI.
162
422960
2002
从人工智能到人工智能
那里得到他们想要的东西。
07:04
But that could blend,
163
424962
1502
但这可能会混为一谈,
07:06
there's already lots of, you know,
164
426505
1877
你知道,在某些
07:08
people in certain demographics
who spend a lot of time on the internet
165
428424
3336
人群中,已经有很多
人花了很多时间在互联网上,
07:11
and they start to identify,
166
431760
1377
他们开始意识到,
07:13
that's their favorite form
of interacting with people.
167
433137
2544
这是他们最喜欢的与人互动
的方式。
07:15
And so I do think
that there's a reality where,
168
435723
2335
因此,我确实
认为,
07:18
as we move into the future,
169
438100
1293
在我们迈向未来的过程中,
07:19
there will be people who bias
towards that for whatever reasons.
170
439435
3003
出于各种原因,
会有一些人倾向于那样做。
07:22
Whether it's the comfort of knowing
that someone's not judging them,
171
442480
3211
不管是
知道别人没有评判他们的安慰,
07:25
whether it's like the format
that it speaks to you with,
172
445733
2669
还是像它对你说话的格式,
07:28
that will kind of bias towards
preferring those types of interactions.
173
448444
3628
这都会偏向于偏向
于偏爱这些类型的互动。
07:32
But on the other hand, I always think
174
452448
1793
但另一方面,我一直认为会
07:34
there’ll be a distribution of people,
175
454241
2795
有人分布,
07:37
and you'll have some people
who really don't like it.
176
457036
3044
而且会有一些
人真的不喜欢。
07:40
And, you know, like I was saying,
177
460080
2128
而且,你知道,就像我说的那样,
07:42
the more that I interact with it now,
I find it almost painful
178
462249
2920
我现在与它的互动越多,
我就会觉得几乎很痛苦,
07:45
because I just pick up
on so many of these issues
179
465169
2294
因为我刚刚发现了很多这样的问题
07:47
that you're like, I can't even use it
at a certain point.
180
467505
2919
你就是这样,我甚至无法
在某个时候使用它。
07:50
And, you know, you'd think
that, like, you know,
181
470424
2461
而且,你会
这么想,比如,你知道,
07:52
I’m in the AI space,
and I write 20-page docs.
182
472927
3128
我在人工智能领域,
我写的是20页的文档。
07:56
I don't need AI for a single bit of it
183
476096
1919
我一点也不需要人工智能,
07:58
because it does remove that voice.
184
478057
2878
因为它确实会移除那个声音。
08:00
And I do also wonder, though,
185
480976
2044
我也想知道,
08:03
as people interact with it more,
186
483062
1877
随着人们与人工智能互动的增多,他们
08:04
will they either identify the differences
187
484980
2253
会识别出差异
08:07
or start to conform to the things
that they're trained with AI.
188
487274
3003
还是开始遵循他们
用人工智能训练的内容。
08:10
It's the same as if you interact
with your partner for example, right?
189
490319
3295
例如,就像你
与伴侣互动一样,对吧?
08:13
You start to be biased
by the communication
190
493656
2294
你开始对沟通产生偏见
,
08:15
because you're talking so much.
191
495991
1502
因为你说话太多了。
08:17
BSL: You mean they're training you?
192
497493
1710
BSL:你的意思是他们在训练你?
08:19
KG: They're training you.
193
499203
1293
KG:他们在训练你。
08:20
Your partner is probably like, you know,
194
500496
2127
你的伴侣可能,对吧,
08:22
they have a preferred way
of communicating.
195
502665
2043
有一种首选
的沟通方式。
08:24
You get used to it, these kinds of things.
196
504708
2044
你已经习惯了,诸如此类的事情。
08:26
So I do wonder if, as people
interact with AI more,
197
506752
2419
因此,我确实想知道,随着人们
越来越多地与人工智能互动
08:29
that they'll kind of all converge.
198
509171
1627
他们是否会融为一体。
08:30
That's probably one
of my biggest fears actually of AI.
199
510839
2586
实际上,这可能是我
对人工智能最大的恐惧之一。
08:33
BSL: I'm concerned
about the exact opposite.
200
513425
2086
BSL:我担心的情况恰恰相反。
08:35
I'm going to shift a little bit.
201
515511
1543
我要稍微换一下话题
08:37
So when we talk about AI,
what you're describing,
202
517054
2503
因此,当我们谈论人工智能时,
你所描述的,通常
08:39
it's usually like dyadic interactions.
203
519598
1835
就像二元交互。
08:41
Like, I'm interacting
with one AI, one agent.
204
521433
3671
比如,我正在与一个人工智能、
一个代理进行互动。
08:45
But really what people do is interact
with multiple people, right?
205
525145
3587
但实际上,人们要做的是
与多人互动,对吧?
08:48
You interact in some community
or some small group setting.
206
528774
3462
你在某个社区
或某个小组环境中进行互动。
08:52
And I'm surprised that there's
not more of that in AI.
207
532236
2586
令我惊讶的是,人工智能中
没有更多这样的东西。
08:54
So you're also in gaming.
208
534863
1210
所以你也在玩游戏。
08:56
I don't really game,
209
536073
1668
我真的不玩游戏,
08:57
but my understanding
is that a lot of the gaming
210
537741
2253
但我的理解
是,很多游戏
09:00
is about connecting with the people,
and it's a community kind of experience.
211
540035
3671
都是关于与人建立联系的,
这是一种社区式的体验。
09:03
So there's two things.
212
543706
1418
所以有两件事。
09:05
One, I'm really surprised
that AI seems so focused on these,
213
545124
3253
第一,我真的很惊讶
人工智能似乎如此专注于这些,
09:08
like, one-on-one interactions
214
548419
1668
一对一的互动
09:10
as opposed to like, multiple AI agents
215
550129
2794
而不是像多个 AI 代理
09:12
creating a more immersive
social experience.
216
552923
2836
创造更身临其境的
社交体验。
09:15
KG: I love you brought it up
because that's really what we do.
217
555759
2920
KG:我喜欢你提出这个问题,
因为我们确实是这么做的。
09:18
BSL: Good, so that's one.
218
558721
1710
BSL:很好,原来如此。
09:20
Other thing, like, the reason I worry
less about convergence
219
560472
2920
另一件事,比如,我之所以
不那么担心融合,
09:23
and more about divergence
220
563434
1251
而更多地担心分歧,
09:24
is if you could produce
a more immersive social experience,
221
564727
4129
是因为如果你能创造出
更身临其境的社交体验,那么
09:28
now everybody’s having their
individual social experiences.
222
568856
3670
现在每个人都有自己的
社交体验。
09:32
Like now, what I worry
about with AI, with VR,
223
572568
4087
就像现在一样,我
担心的人工智能、
09:36
with all these kind of technologies
that are expanding,
224
576697
5589
虚拟现
实、所有这些都在扩展技术,
09:42
what we can control
about our social environment,
225
582328
2377
我们可以控制
我们的社交环境,
09:44
about our physical perceptions
in the environment,
226
584747
2544
对环境
中的身体感知,
09:47
is that we all inhabit
our own singular world.
227
587333
3712
是我们都生活
在自己的单一世界中。
09:51
That is more frightening
to me than like, you know,
228
591045
3545
对我来说,这比我们都聚集在同样
09:54
that we all converged
in the same experience.
229
594632
2335
的经历 中更可怕。
09:57
KG: Well, my mom’s a grade-seven teacher,
230
597009
2085
KG:嗯,我妈妈是一名
七年级的老师
09:59
and the one thing that she said
is really interesting
231
599094
2503
她说的一件事非常有趣,
10:01
is if you went back like 20 years,
232
601639
1626
就是如果你回到 20 年前,
10:03
everybody was watching the same TV shows,
233
603265
1960
每个人都在看同样的电视节目,
10:05
and they come to class
and they'd all be talking about it.
234
605267
2753
他们来上课
然后都在谈论这件事。
10:08
And now everybody watches
their own favorite YouTube channel.
235
608062
2877
现在,每个人都在观看
自己喜欢的YouTube频道。
10:10
And it's the siloing of reality.
236
610939
1627
这是现实的孤立。
10:12
Like, what we do is when we work
with games, for example,
237
612608
2878
比如,我们做的是当我们
处理游戏时,例如,
10:15
one of the interesting things is like,
as people play through games,
238
615486
3211
有趣的事情之一就是,
当人们玩游戏时,
10:18
it's basically the same thing.
239
618739
1460
它基本上是同一回事。
10:20
You could have a million people
go through a game,
240
620199
2377
你可以让一百万人
玩一场游戏,
10:22
and it’s some differences
241
622576
1335
这有一些区别,
10:23
but you're largely going
to hit the same points.
242
623911
2252
但你基本上会
遇到同样的分数。
10:26
And so one of the things
that we think about is,
243
626205
2252
因此,我们考虑的一件事是
10:28
what does that mean for agency?
244
628457
1501
这对代理机构意味着什么?
10:29
The way we interact with media
245
629958
1460
我们与媒体互动的方式
10:31
changes the way that we feel
agency in the world.
246
631460
2294
改变了我们对世界的
代理感的方式。
10:33
So if we see inert media
that we can't change,
247
633754
2169
因此,如果我们看到我们
无法改变的惰性媒体,
10:35
it also gives you this sense
that you can't change the world.
248
635964
2878
它也会给你
一种你无法改变世界的感觉。
10:38
And so to your point,
249
638842
1168
因此,就你而言,
我们想在游戏中做
10:40
one of the things that we want to do
with games is, how do you make it
250
640010
3295
的一件事是,如何
10:43
so that each person can actually
influence that outcome?
251
643305
2628
让每个人都能真正
影响游戏结果?
10:45
And as you add more agents into that,
252
645974
1794
当你在其中添加更多代理时,
10:47
that you see, OK, I interact with this one
and it has a cascade effect.
253
647768
3337
你会看到,好吧,我和这个
代理互动,它会产生级联效应。
10:51
I love it.
254
651146
1210
我喜欢它。
10:52
I mean, even in some
of the stuff we've done here,
255
652356
2377
我的意思是,即使在我们所做
的一些事情中,
10:54
the magic actually happens
when you do have those agents interacting,
256
654775
3253
魔法实际上也发生在你与
这些代理进行互动的时候,
10:58
because then you’re also not just seeing
like that one-to-one interaction
257
658028
3504
因为那样你看到的
不仅仅是那种一对一的互动,
11:01
but the emergent effect
of basically that interaction.
258
661573
2545
而是基本上这种互动
的突发效果。
11:04
And another thing is, if your main
controls that you have in the computer
259
664118
3461
还有一件事是,如果你在
电脑中的主要控制
11:07
is like point-and-click
or, in games, jump and shoot,
260
667579
2962
是点击或者在
游戏中是跳跃和射击,
11:10
we're trying to see like,
what does it mean if social skills
261
670582
2837
我们正在尝试看看,
如果像这样的社交技巧
11:13
like interaction like this,
262
673419
1918
这样的互动的方式
11:15
are the ways that you actually interact
with the games, the technology
263
675379
3337
与游戏、技术
11:18
and the agents.
264
678757
1168
和代理进行实际互动意味着
什么。
11:19
That’s a very different way of conversing
or of dialogue than button presses.
265
679967
3879
这是一种与按下按钮截然不同
的对话或对话方式。
11:23
And I think that changes the way
that you sense agents in the world.
266
683846
3336
而且我认为这改变了
你感知世界代理人的方式。
11:27
Because I think the way that most people
change the world is by speaking
267
687224
3420
因为我认为大多数人改变世界
的方式
11:30
and interacting and interacting
with other humans,
268
690644
2419
是与其他人说话和互动,
11:33
not by pressing buttons.
269
693063
1335
而不是按下按钮。
11:34
I mean, arguably it's the case in some.
270
694398
2711
我的意思是,可以说,
在某些情况下是这样。
11:37
BSL: You know, the other thing
that's interesting to me is
271
697151
2752
BSL:你知道,我感兴趣
的另一件事是
11:39
I don't think people have an understanding
272
699945
2127
我认为人们不了解他们所
11:42
of the systems they exist in, right?
273
702072
2169
处的系统,对吧?
11:44
People think about themselves as existing
in like individual relationships,
274
704283
3545
人们认为自己存在 于个人关系中,
11:47
and they have a harder time
understanding system affects
275
707870
2878
他们更难
理解系统的影响,
11:50
like I affect you,
which affects your partner,
276
710748
2335
就像我影响你一样,
它会影响你的伴侣
11:53
which affects your partner's
parents, right?
277
713125
2169
影响你的伴侣的
父母,对吧?
11:55
That is a harder thing to grasp.
278
715335
2169
这是一件很难理解的事情。
11:57
But I think there's something
that's fundamentally human about that.
279
717546
3212
但我认为,这从根本上
讲是人性化的。
12:00
Like you are also impacted by all
these different things going on,
280
720758
3503
就像你也受到
所有这些不同的事情的影响
12:04
like, we had the person come
and put on our makeup,
281
724303
2419
一样,比如,我们让那个人来
化妆,
12:06
and now I'm looking beautiful
282
726722
1418
现在我看起来很漂亮
12:08
and it's affecting
everybody else around me.
283
728140
2085
这也影响着我周围的其他人。
12:10
(Laughter)
284
730225
1085
(笑声)
12:11
KG: It's glowing.
285
731351
1168
KG:它在发光。
12:12
BSL: Exactly.
286
732519
1168
BSL:没错。
12:13
How does that fit in?
287
733729
1835
这怎么适合?
12:15
I just haven't heard people
talk about it in that way,
288
735606
2544
我只是没有听过人们以这种方式
谈论它,
12:18
which is surprising to me,
because that, I think,
289
738150
2294
这让我感到惊讶,
因为我认为
12:20
is what fundamentally makes humans human.
290
740486
3295
这是使人类成为人类的根本原因。
12:23
It's interaction
291
743781
2085
这是互动
12:25
and complex social situations.
292
745908
1668
和复杂的社交场合。
12:27
KG: And these like, nested systems.
293
747618
1710
KG:比如嵌套系统。
12:29
And like, they all affect
each other, right?
294
749328
2085
比如,它们都
相互影响,对吧?
12:31
You think that your small activity
295
751455
1918
你认为你的小活动
12:33
doesn't affect whatever
higher-level political stuff,
296
753415
2503
不会影响任何
更高层次的政治问题,
12:35
but it's all aggregate.
297
755959
1210
但都是综合性的。
12:37
And it's all interlinking as well.
298
757169
1627
而且这一切也是相互关联的。
12:38
I mean, it's like the AI thing
is interesting too,
299
758837
2503
我的意思是,人工智能的东西
也很有趣,
12:41
because I often hear people talk
about it as like this evolution.
300
761340
3170
因为我经常听到人们把它当作
这样的进化来谈论。
12:44
You have like, you know, singular cells
to monkeys to humans to AI.
301
764510
4045
你知道,从单一细胞到猴子
然后到人类最后到人工智能
12:48
Whereas like, you could flip it,
where it's like more like, you know,
302
768597
3295
但是,你可以把它翻过来
,你知道,它更像是
12:51
cells to organs to human to AI.
303
771892
1543
细胞到器官,再到人再到人工智能。
12:53
It's a system overarching
that just because it's trained on us
304
773477
3545
这是一个总体系统 ,
只要它对我们进行了训练,
12:57
and we do these things, that we actually
influence that system,
305
777022
2961
我们做了这些事情,我们就会真正
影响这个系统,
13:00
now that people are interacting with it,
it has this interplay.
306
780025
3086
既然人们正在与之互动,
它就会产生这种相互作用。
13:03
And that's interesting too,
when it becomes like, you know,
307
783946
2836
这也很有趣,
当它变成,
13:06
AI isn't this singular entity.
308
786824
1459
人工智能不是这个单一的实体时。
13:08
It is more of an institution
or a system almost
309
788325
2711
它更像是一个几乎可以
13:11
that is kind of overarching
everything else.
310
791078
2127
支配其他 一切的机构
或系统。
13:13
BSL: And it's also weird because it's
like our vision of who we are.
311
793247
3295
BSL:这也很奇怪,因为这
就像我们对自己的看法。
13:16
So when we talk about AGI,
312
796542
1376
因此,当我们谈论 AGI 时,
13:17
it's like we don't even know
what intelligence is
313
797960
2460
就像我们甚至不知道
什么是智能
13:20
and we think we're going to produce
something that has it.
314
800420
3129
我们认为我们会生产出具有智能
的东西。
13:23
It's just an interesting situation
where we talk about it, as you said,
315
803590
3837
这只是一个有趣的情况,
我们谈论它,正如你所说,
13:27
it's natural evolution,
316
807469
1668
它是自然进化,
13:29
but in fact we’re creating it,
317
809137
1627
但实际上我们正在创造它
13:30
and it's not clear that we know exactly
what it is we're creating.
318
810764
4171
尚不清楚我们是否确切地知道
我们在创造什么。
13:34
KG: I actually think that one
of the most interesting things
319
814977
2836
KG:实际上,我认为
最有趣的事情之一
13:37
is that we're starting to work
on AI at a point where,
320
817813
2544
是我们开始
研究人工智能的时刻,
13:40
like, I still think we're figuring out,
you know, ourselves.
321
820399
2836
比如,我仍然认为我们
自己正在弄清楚。
13:43
Neuroscience is very early on in its days,
322
823235
3754
神经科学还处于起步阶段
13:47
and yet we're creating things
323
827030
1418
但是我们正在
创造类似的东西
13:48
that are like, based
on our own intelligence,
324
828490
2127
基于自己的智力
13:50
and we don't really understand
even what's going on inside.
325
830617
2795
甚至我们根本不了解
内部发生的事情。
13:53
And so to your point on,
what are the effects?
326
833412
2169
那么就你而言,效果
是什么?
13:55
We don't really know yet.
327
835581
1876
我们还不知道。
13:57
Every year a new paper comes out
328
837499
2044
每年都会发表一篇新论文
13:59
and changes how people think
about child rearing.
329
839585
2627
改变人们对抚养孩子的看法。
14:02
Like how to bring up a child well,
like all those kinds of things.
330
842212
3295
比如何好好抚养孩子,
就像所有这些事情。
14:05
And now we're creating systems
that will, you know,
331
845549
2419
你知道,现在我们正在创建的系统
将在
14:07
kind of be overarching other humans.
332
847968
2127
某种程度上支配其他人类。
14:10
What does that mean, I don't know.
333
850137
1626
那是什么意思,我不知道。
14:11
I do actually think,
334
851805
1210
实际上,我确实在想,
14:13
I happen to be in AI,
we happen to be at this point in time.
335
853015
3044
我碰巧在人工智能领域,
我们恰好处于这个时刻。
14:16
But if we could pause for a second,
336
856101
1710
但我认为,如果我们能停顿一秒钟
14:17
I think it would be good, another
few centuries of figuring out what we are
337
857811
3587
那就太好了,在我们创造出
符合我们想象的东西
14:21
and understanding that a little bit better
338
861440
2002
之前,再
弄清楚自己是什么样子
14:23
before we created something
that was in our image.
339
863442
2335
并更好地理解
这一点。
14:25
Because, we’re kind of just, you know,
340
865777
1836
因为,我们有点,你知道吧,
14:27
it's kind of like taking a photograph
and like painting it, right?
341
867654
3128
有点像拍照
然后画画,对吧?
14:30
You're not actually getting
the person and painting it.
342
870782
2586
实际上,你没有找到
人物然后画它。
14:33
There's something about the life
that's missing there.
343
873368
2545
那里的生命
中有些缺失。
14:35
So I do agree.
344
875954
1293
所以我确实同意。
14:37
I think that we're honestly
kind of premature,
345
877289
2294
说实话,我认为
我们现在还为时过早,
14:39
but I think it's just how, I guess,
346
879583
2210
但我认为,我猜,
14:41
you know, life goes that things
come out when they naturally should.
347
881835
3962
你知道,生活就是这样的,
事物自然应该出现。
14:45
BSL: So, I mean, you work in AI,
348
885839
1752
BSL:所以,我的意思是,
你在人工智能领域工作,
14:47
so what's the most
exciting thing for you in AI?
349
887633
3128
那么在人工智能领域对你来说最
令人兴奋的事情是什么?
14:50
What's your hope for it?
350
890802
2169
你对此有什么希望?
14:54
KG: I think it's kind of back
to that agency question.
351
894598
2586
KG:我认为这又回
到了那个代理的问题。
14:57
So I mean, you know,
you read a news story,
352
897184
4004
所以我的意思是,
你知道,
15:01
you read a book, you watch a movie,
you watch a TV show,
353
901188
2669
你读新闻报道,读书,看电影,
看电视节目,
15:03
this is specific to, like, my domain,
354
903899
2502
这是我的领域所特有的,
15:06
like there's something
about the communication
355
906443
2294
比如
15:08
that we're having right now where, like,
356
908779
1918
说,我们现在的沟通,比如,
15:10
I'm adapting to the things that you say,
to your body language,
357
910697
2962
我正在适应你所说的话,
你的肢体语言,
15:13
all those kinds of things, right?
358
913700
1585
诸如此类,对吧?
15:15
To, like, the people in the room
we have here, all these things.
359
915285
3045
对我们房间里的人来说,
所有这些东西。
15:18
And so when you have ...
360
918372
1167
所以当你有...
15:20
AI able to, sort of, help that adaptation
so that you have that agency
361
920624
3837
人工智能能够在某种程度
上帮助你进行适应,这样你就能
15:24
in the things that you interact with.
362
924461
1793
在与之互动的事物中
拥有这样的代理权。
15:26
I don't necessarily believe
in fully personalized media
363
926254
2586
我不一定相信
完全个性化的媒体,
15:28
because I think we need like
a shared social context.
364
928882
2503
因为我认为我们需要
一个共享的社交环境。
15:31
The reason we watch a movie or a TV show
365
931385
2335
我们看电影或电视节目的原因
15:33
is because then we can all
talk about it, right?
366
933720
2253
是因为那样我们就可以
谈论了,对吧?
15:35
But there is something about the fact
367
935973
1793
但是有一个事实是
15:37
that we're all interacting
with these internet objects.
368
937808
2586
我们都在
与这些互联网对象进行交互。
15:40
And so the way that technology feels,
you're on a screen, it doesn't change.
369
940435
3629
因此,科技的感觉,
你在屏幕上,它不会改变。
15:44
You're in a movie, it doesn't change.
370
944064
1793
你在看电影,情况不会改变。
15:45
You're watching Netflix, it doesn't change
depending on what you do.
371
945857
3212
你在看 Netflix,它不会
因你所做的事情而改变。
15:49
And I think that changes the way
we see our own agency in the world.
372
949069
3211
而且我认为这改变了我们在世界上
看待自己的代理机构的方式。
15:52
And so I hope with AI that one
of the things that it does
373
952322
3170
因此,我希望通过人工智能
它能做的一件事
15:55
is kind of opens this door to agency
374
955492
3003
就是为代理机构打开这扇大门
15:58
in the way that we interact with media
and technology in general,
375
958537
3253
以我们与媒体
和技术互动的总体方式,
16:01
such that we do notice that effect
that you have on systems.
376
961790
2920
这样我们就能注意到
你对系统的影响。
16:04
Because even if it's small, right,
377
964710
1793
因为即使它很小,对吧
16:06
where I take a certain action
and it completely changes an app
378
966545
2919
我采取某种行动,
它会彻底改变应用程序
16:09
or it changes an experience,
379
969464
1377
或改变体验,
16:10
maybe that helps us learn that we have
an effect in the social systems
380
970841
4462
也许这可以帮助我们了解我们
对社交系统
16:15
as well that we're affecting.
381
975303
1419
也有影响,我们正在影响什么。
16:16
So something to that effect.
382
976722
1376
所以大意如此。
16:18
BSL: So you want to make
our agency more transparent.
383
978098
2920
BSL:所以你想让
我们的机构更加透明。
16:21
And do you think it does that,
384
981018
1501
你认为它能做到吗,
16:22
because right now I'm not sure
it doesn't obfuscate our agency.
385
982561
3003
因为我现在不确定
它不会混淆我们的机构。
16:25
KG: No I don't necessarily know.
386
985564
2085
KG:不,我不一定知道。
16:27
I agree, I mean this is why I think also
387
987649
2961
我同意,我的意思是
这就是为什么我认为
16:30
media and games is, you know,
the domain I mainly focus on.
388
990610
3045
媒体和游戏也是我主要关注
的领域。
16:33
And I think it's interesting, especially
because young people use it a lot.
389
993655
3545
而且我认为这很有趣,尤其是
因为年轻人经常使用它。
16:37
And so I've heard very veteran
game developers say
390
997242
3086
因此,我听过非常资深的
游戏开发者说人们
16:40
how people interact with games
391
1000370
1460
如何与游戏互动
16:41
kind of trains kids how they
should interact with the world.
392
1001830
2836
就像训练孩子
如何与世界互动。
16:44
So even people who tend to be
professional players in different games
393
1004708
3253
因此,即使是那些在不同游戏中
倾向于成为职业玩家的人
16:47
have different psychological profiles
394
1007961
2252
也有不同的心理特征,
16:50
because they bias towards certain ways
of interacting and seeing the world.
395
1010213
3587
因为他们偏向于某些
互动和看待世界的方式。
16:53
The same way, I guess, if you trained
in something academic, right,
396
1013800
3170
我猜也一样,如果你接受过学术
方面的培训,对吧,
16:57
you have a different way of viewing it.
397
1017012
1877
你有不同的看法。
16:58
And so if we make games and media in a way
398
1018889
3295
因此,如果我们以一种
让你也能感受到
17:02
that you feel that sort
of social impact as well,
399
1022184
3128
这种社会影响的方式
制作游戏和媒体,
17:05
maybe, maybe it opens the door to like,
another realm of understanding.
400
1025353
3629
也许,它会打开通往
另一个理解领域的大门。
17:09
But, yeah, I agree that like a lot
of the systems that we have today
401
1029024
5213
但是,是的,我同意,就像我们
今天的许多系统一样,可能还会
17:14
give you maybe a false sense
also of agency
402
1034279
2044
给你一种虚假的代理
17:16
where like we were talking
about the AI systems,
403
1036364
2253
意识,就像我们在 谈论人工智能
系统一样,你实际上感觉自己
17:18
where you actually feel
like you're controlling this thing,
404
1038658
2795
在控制这个东西,
17:21
whereas maybe it's also biasing,
you know, and "controlling,"
405
1041453
3420
而 你知道的,它可能还有偏见,和
“控制”,也会对你
17:24
having some influence over you as well.
406
1044873
1877
产生一定的影响。
17:26
BSL: So where do you think
things are going?
407
1046792
2085
BSL:那么你认为
事情会走向何方?
17:28
So there's obviously a huge
race among some very,
408
1048877
2336
因此,一些资源
17:31
very well-resourced
organizations over AI, right?
409
1051254
4588
非常充足 的组织之间显然存在着
围绕人工智能的激烈竞争,对吧?
17:35
You know, Microsoft, Google,
I mean, are the biggest maybe.
410
1055884
3086
例如,微软、
谷歌可能是最大的。
17:40
And they are very quickly
going to need to monetize it
411
1060931
3253
而且他们
很快就需要通过它获利,
17:44
because this is what those companies
are designed to do.
412
1064226
3253
因为这是
这些公司的设计目标。
17:47
Like what do you foresee?
413
1067521
2502
比如你能预见什么?
17:50
Because I just look
at social media as an example.
414
1070065
2961
因为我只是
以社交媒体为例。
17:53
I think, at the time
when it first came out,
415
1073068
4129
我认为,
当它刚问世时,
17:57
people were really excited,
as a new way to connect with people
416
1077197
3003
人们真的很兴奋,作为一种与人
建立联系的新方式,也是一种
18:00
and a way to stay connected to people
417
1080200
1960
与你认识的人保持联系的方式,
18:02
you know, you couldn't otherwise;
418
1082160
2461
否则你无法做到的;
18:04
catch up with people
you lost contact with,
419
1084663
2210
赶上失去联系的人 ,
18:06
that sort of thing.
420
1086873
1627
诸如此类的事情。
18:08
And it changed into something else.
421
1088542
1960
然后它变成了另外一回事。
18:12
In large part because
of the way it was monetized,
422
1092295
2378
这在很大程度上是由于
其获利方式,
18:14
like going to ads, focus on attention.
423
1094673
2002
例如投放广告,因此
将注意力集中在注意力上。
18:18
What's the trajectory of AI?
424
1098844
2585
人工智能的轨迹是什么?
18:22
KG: You know, I'm taking guesses.
425
1102055
2544
KG:你知道,我在猜一猜。
18:24
BSL: Yeah, of course,
we're all taking guesses,
426
1104641
2252
BSL:是的,当然,
我们都在猜测,
18:26
I won’t hold you to it, don’t worry.
427
1106935
1752
我不会强迫你去猜的,别担心。
18:28
KG: I think that the reality is,
428
1108728
2962
KG:我认为现实是,
18:31
we were kind of mentioning before
about the challenges of scale.
429
1111731
3045
我们之前曾提到
过规模的挑战。
18:34
And when you invest tens
of billions of dollars in something,
430
1114818
3587
而且,当你在某件
事上投资数百亿美元时,
18:38
you need scale.
431
1118405
1501
你需要规模。
18:39
And I think that's one of --
the way that AI is developed
432
1119906
3087
我认为这是其中之一 ——
人工智能 的开发方式
18:43
and specifically even the types
of models we're using,
433
1123034
2545
特别是我们正在
使用的模型类型,
18:45
the economic model of it,
434
1125620
1210
它的经济模型,
18:46
which is effectively
the more compute you have,
435
1126872
2210
它实际上是你的计算越多,
18:49
the better models you can create.
436
1129082
1585
你可以创造的模型越好。
18:50
The better models you can create,
the more usage you get.
437
1130667
2711
您可以创建的模型越好,获得
的使用量就越多。
18:53
The more usage you get, the better.
438
1133378
1710
使用量越多越好。
18:55
So it has somewhat of a, honestly,
like monopolistic tendency, I think,
439
1135088
3670
因此,说实话,
我认为它
18:58
in the way that actually even
like the architectures
440
1138800
2503
在某种程度上有一种垄断倾向,
其运作方式甚至就像它的架构
19:01
and the economy of it works.
441
1141303
1710
和经济一样。
19:03
And so
442
1143013
1626
因此,
19:04
I think it's almost inevitable
that whatever AI systems are produced
443
1144639
3254
我认为,
无论这些大型组织生产
19:07
by these large organizations
444
1147893
1376
什么人工智能系统,都
19:09
will be pushed to scale
as quickly as possible.
445
1149311
2752
将尽快扩大规模,
这是几乎不可避免的。
19:12
And there's some pluses in that
where like, you know, sure,
446
1152063
3754
这其中有一些优点,
比如,你知道,当然,
19:15
they're building in feedback loops,
people can give their input, it biases it.
447
1155859
3712
他们正在建立反馈循环,人们
可以提出自己的意见,这会造成偏见。
19:19
But also at the same time,
448
1159613
1251
但同时,当一个模型适合
十亿人时,这意味着
19:20
what does it mean when a single model
is fit to a billion people, right?
449
1160906
3712
什么,对吧?
19:24
So that's kind of what I meant
about the converging effect
450
1164659
2753
所以这就是我所
说的聚合效应的意思,
19:27
where, what happens when we are pushed
to produce something
451
1167412
2920
当我们被迫生产
19:30
that fits to a billion people?
452
1170373
1627
适合十亿人的产品时会发生什么?
19:32
There's a lot of diversity in there.
453
1172042
1751
里面有很多多样性。
19:33
And so, you know, we create these scaled
systems that are fitting with the whole,
454
1173793
4588
因此,你知道,我们创建了这些
与整体相适应的缩放系统,
19:38
like, trying to fit the whole planet.
455
1178423
1794
比如试图适应整个地球。
19:40
Does that work?
456
1180258
1168
那行得通吗?
19:41
And so I think what will, you know,
457
1181468
2002
所以我想会怎样,你知道,
19:43
we're going to go through this
phase where like, yeah,
458
1183511
2545
我们将经历这个
阶段,比如,是的,
19:46
you're going to have a billion people
interacting the same AI.
459
1186056
2961
你将有十亿人使用相同的人工智能
进行交互。
19:49
And I don't know what
the effect of that will be.
460
1189017
2294
而且我不知道那
会产生什么影响。
19:51
Even the monetization models now
are kind of you pay
461
1191353
2460
即使是现在的
盈利模式也是你
19:53
to use these kinds of things,
which are maybe OK,
462
1193855
2628
为使用这类东西付出代价的,
这也许没问题,
19:56
but certainly ads will probably
enter the equation.
463
1196524
2420
但广告肯定会被纳入
其中。
19:58
Also, what happens when you want attention
464
1198944
2002
另外,当你需要关注
20:00
and AI is much better at that
than the algorithms
465
1200987
2378
和人工智能时发生的事情比你在
20:03
you even have on YouTube and Instagram.
466
1203365
1877
YouTube和Instagram
上使用的算法要好得多。
20:05
And you can start
to capture that attention.
467
1205283
2086
然后你就可以开始
吸引这种注意力了。
20:07
And so I certainly think it's going
to be an interesting little bit here now,
468
1207369
3628
因此,我当然认为现在这里会
是一件有趣的事情,
20:10
as we see these huge organizations
spending tens of billions of dollars
469
1210997
3921
因为我们看到这些大型组织
花费了数百亿美元
20:14
and the choices that they make
to then monetize that,
470
1214918
2961
他们做出了哪些选择,
然后通过这种方式获利,
20:17
and what that means
for how AI proliferates.
471
1217921
3420
以及这对
人工智能的扩散意味着什么。
20:21
I know a lot of the folks
in the organizations,
472
1221341
3045
我认识组织
中的很多人
20:24
and their interests have
never been in that domain.
473
1224427
2753
他们的兴趣
从未出现在这个领域。
20:27
But at the same time,
you're beholden, you know,
474
1227222
2252
但同时,你知道,
你受股市利益的束缚,
20:29
to stock market interests and whatever
it is, then what happens?
475
1229474
3295
不管
它是什么,然后会发生什么?
20:32
It shifts it, right?
476
1232811
1418
它会改变它,对吧?
20:34
We're in a capitalist world.
477
1234271
1376
我们处在一个资本主义世界中。
20:35
And that's kind of like, you know,
478
1235647
1668
这就有点像
20:37
what ultimately will change
the incentives.
479
1237315
2461
最终会改变激励措施
的因素。
20:39
So yeah it's interesting.
480
1239818
1585
所以是的,这很有趣。
20:41
I mean I am interested in,
coming from your background,
481
1241444
2920
我的意思是,从你的背景
来看,我很感兴趣,
20:44
you have a very different stance on it.
482
1244406
1918
你对此的立场截然不同。
20:46
But, you know, it's all this
AI stuff is interesting.
483
1246324
2544
但是,你知道,所有这些
人工智能的东西都很有趣。
20:48
But, you know, when you think,
almost to your first question,
484
1248868
2878
但是,你知道,当你思考
几乎是你的第一个问题时,
20:51
what makes us human and like as people,
485
1251746
2836
是什么让我们成为人类,
就像人一样,
20:54
just technology in general
and specifically with AI,
486
1254624
3379
只是一般
的技术,特别是人工智能,
20:58
like where can people find
the meaning in their life,
487
1258044
4004
比如人们在哪里可以找到
生活中的意义,
21:02
the values that they find true?
488
1262090
3086
他们认为真实的价值观?
21:05
And how will that change, do you think,
489
1265677
1877
你觉得那会有什么变化
21:07
I guess, with like the advent
of these new technologies?
490
1267554
2627
伴随这些新技术的出现
一样
21:10
Or how have you seen it change
491
1270181
2211
或者,你是否看到它伴随着
我们已经看到的技术变为现实变化
21:12
with the technologies we've already
seen come to life?
492
1272392
2669
21:16
BSL: This is going to sound
like a funny answer.
493
1276479
2253
BSL:这听起来
像是一个有趣的答案。
21:18
I think people are too worked up
about technology, personally.
494
1278732
2919
就我个人而言,我认为人们
对技术太着迷了。
21:21
I mean, you know,
we had this conversation.
495
1281693
2210
我的意思是,你知道,
我们进行了这样的交谈。
21:23
I think, you know, people have been
using technology since we've been human.
496
1283945
4171
我认为,自从我们成为人类以来,
人们就一直在使用技术。
21:28
So paper was a huge invention.
497
1288158
2460
所以纸是一项伟大的发明。
21:30
Talked about this.
498
1290618
1210
谈过这个。
21:31
The printing press, huge invention.
499
1291870
1710
印刷机,伟大的发明。
21:33
Computer, huge invention.
500
1293580
1460
计算机,伟大的发明。
21:35
Internet, huge invention.
501
1295081
1252
互联网,伟大的发明。
21:36
AI, great, another huge invention.
502
1296333
1835
AI,太棒了,又一个伟大的发明。
21:38
And through all of that, I think
503
1298209
2795
综上所述,我认为
21:41
what you see in a lot
of the biggest technologies
504
1301046
2294
你在许多
最大的技术中看到
21:43
is the desire to connect
with other human beings.
505
1303381
2419
的是
与其他人建立联系的愿望。
21:45
I think what fundamentally makes us human
is our connection to other human beings,
506
1305800
4129
我认为使我们成为人类的根本原因
是我们与其他人的联系,
21:49
our ability to engage with other human
beings, and like consciousness
507
1309971
3295
我们与其他人
互动的能力,以及意识
21:53
and all these other things I think
are necessary preconditions.
508
1313266
3879
和所有其他我认为
是必要的先决条件的东西。
21:57
But really, what makes us human
is connections with other humans.
509
1317187
3253
但实际上,使我们成为人类
的是与其他人的联系。
22:00
And that is incredibly complex.
510
1320482
4129
这非常复杂。
22:04
And I don't think we're close in terms
of technology of replicating that.
511
1324652
3713
而且我认为我们能创造足以复制
上述的方面的科技还差很多。
22:08
I mean, even what you described
it's like you have this feeling of like,
512
1328948
3421
我的意思是,即使是你所描述的
那样,你也有这样的感觉,
22:12
this isn't right, this is off.
513
1332369
1835
这是不对的,这是不对的。
22:14
And even if you felt like it was right,
514
1334245
1877
而且,即使你觉得这是对的,
22:16
it still would be off in ways
you didn't quite get.
515
1336164
2586
它仍然会以
你不太明白的方式失效。
22:18
I don't think we're close.
516
1338792
2502
我认为我们离得很远。
22:21
Though because it's designed to pull
our attention away from other things,
517
1341336
4171
但是,由于它旨在将我们的注意力
从其他事情上移开,因此
22:25
I think it impedes our ability
518
1345507
5547
我认为它阻碍了我们
做我们所有人想做的事情,
22:31
to do what we all kind of want to do,
519
1351096
3920
即相互互动的能力。
22:35
which is interact with each other.
520
1355016
2628
22:37
And it might change the way
we interact with each other
521
1357977
4505
而且它可能会改变
我们彼此互动的方式
22:42
in a way that might feel less fulfilling.
522
1362524
3169
以使人感到不那么充实的途径。
22:46
And I think you see some of that
in social interactions now.
523
1366069
4463
我想你现在
在社交互动中看到了其中的一部分。
22:50
Some of that I mean, recently maybe,
524
1370573
3546
我的意思是,最近,
22:54
COVID was an issue.
525
1374119
2043
新冠病毒可能是一个问题。
22:56
But, you know, people feeling less
comfortable in face-to-face interactions.
526
1376204
4213
但是,你知道,人们在面对面的
互动中感觉不太舒服。
23:00
Like people dating,
527
1380458
1293
就像人们约会一样,
23:01
there's no serendipity in hanging out
and you meet who you meet.
528
1381793
3045
出去玩没有偶然的机会,
你会遇到你遇到的人。
23:04
It's like you're using an algorithm
to try to present to you options.
529
1384879
4255
就像你在使用算法来尝试
向你呈现选项一样。
23:09
That's a very different world.
530
1389175
1794
那是一个截然不同的世界。
23:11
So even that's prior to AI.
531
1391010
2294
因此,即便如此,
也是在人工智能出现之前。
23:14
And I don't know how AI
is going to further influence that.
532
1394013
3796
而且我不知道人工智能
将如何进一步影响这一点。
23:17
KG: And I guess just
even like the general point,
533
1397809
2669
KG:我想
即使和一般观点一样,你认为
23:20
how core do you think
534
1400478
2378
人际关系的核心需求
23:22
the need for connection is
535
1402856
1501
有多重要
23:24
in the sense that, you know,
536
1404399
1376
因为你知道,
23:25
I've heard some parents say
that, through COVID,
537
1405775
2252
听过一些家长
说,经历过新冠病毒,
23:28
their kids went through
a major change, you know,
538
1408069
2544
他们的孩子经历了重大变化,
23:30
these regressions,
539
1410655
1543
退化
23:32
their different habits
and these kinds of things
540
1412198
2920
他们不同的习惯
和诸如此类的事情,
23:35
because they weren't
connecting with people.
541
1415160
2085
他们没有与人
建立联系。
23:37
And then it's taken years
to overcome that.
542
1417287
2169
然后花了好几年时间才克
服这个问题。
23:39
So I do also wonder, like, you know,
543
1419456
1918
所以我也想知道,
23:41
whether it's through technology
or things like COVID
544
1421416
2461
无论是通过技术
还是新冠病毒之类的东西,
23:43
or just like circumstances,
545
1423877
1418
还是像环境一样,我们
23:45
could we lose that need for connection?
546
1425295
2002
能否失去对连接的需求?
23:47
Or even if we need it,
547
1427297
1960
或者,即使我们需要它,
23:49
you know, we might lose the desire for it
548
1429299
1960
我们可能会失去对它的渴望,
23:51
and feel emotional trauma as a result,
but still not go for it.
549
1431301
3920
并因此感到情感创伤,
但仍然无法去做。
23:55
Like, how core do you think it is?
550
1435263
1627
比如,你认为它有多核心?
23:56
And do you think we're safe
in that kind of need?
551
1436931
3629
而且你认为我们在这种需求
中安全吗?
24:01
BSL: So I'm going to give you
the most extreme answer,
552
1441769
2545
BSL:所以我要给你一个最极端
的答案,我认为
24:04
which is I think the true one.
553
1444355
1460
这是真实的答案。
24:05
That you will cease to be human
554
1445857
1793
你将不再是人类
24:07
if you don't have a need
for human connection.
555
1447650
2169
如果你不需要人际关系。
24:09
Like I think you will be
a physical person,
556
1449861
2002
就像我认为你将成为
一个自然人一样,
24:11
but you will literally break down
as a human being.
557
1451905
3712
但你实际上会像人
一样崩溃。
24:15
And this is why in part --
558
1455658
1961
这就是为什么
24:17
Social isolation
559
1457660
1752
社交隔离
24:19
or solitary confinement
is considered inhumane.
560
1459454
3211
或单独监禁在某种程度上被认为
是不人道的。
24:22
Because people literally break down,
561
1462707
2711
因为人们真的会崩溃,所以
24:25
you will start to have hallucinations.
562
1465460
2044
你会开始产生幻觉。
24:27
You will break down mentally
and physically absent human connection.
563
1467545
5631
如果没有人际关系,你将在精神
和身体上崩溃。
24:33
So I don’t think there’s any possibility,
in my mind, of losing the need.
564
1473635
4421
因此,
在我看来,不可能失去需求。
24:38
Like, you may get less than you need,
565
1478056
2085
比如,你得到的可能会比你需要的少
24:40
and that will have negative
consequences for you.
566
1480183
3128
,这将对你
产生负面影响。
24:44
But I'm not worried about people
not wanting to be around people.
567
1484562
3420
但我并不担心人们
不想和别人在一起。
24:48
KG: Are you worried that, like,
568
1488024
1710
KG:你是否担心
24:49
things like social media or AI
or any of these things
569
1489734
3086
诸如社交媒体
或人工智能之类的东西
24:52
could give you that sense
that you're fulfilling that need,
570
1492862
2795
会让你感觉
到你在满足需求,
24:55
but not actually fulfilling it?
571
1495657
1543
但实际上并没有满足需求?
24:57
It's totally true, right?
572
1497242
1251
这是完全正确的,对吧?
24:58
Solitary confinement is a great example
because we need it.
573
1498535
4129
单独监禁就是一个很好的例子,
因为我们需要它。
25:02
We absolutely lose our sanity
as well as, you know, our well-being.
574
1502664
4421
你知道,我们绝对会失去理智,
也失去了身心健康。
25:07
But maybe we can, you know,
575
1507085
2377
但是,也许我们可以,
25:09
technology can manufacture
the sense that we're fulfilling it.
576
1509504
3253
技术可以制造
出我们正在实现它的感觉。
25:12
And then over time,
577
1512757
1210
然后随着时间的推移,
25:13
we see these mental health crises
evolve as a result?
578
1513967
2502
我们会看到这些心理健康危机因此
而演变吗?
25:16
BSL: Yeah, that's a good question.
579
1516511
1626
BSL:是的,这是个好问题。
25:18
I think it's unlikely, but I don't know.
580
1518179
2127
我认为这不太可能,但我不知道。
25:20
Honestly I don't know.
581
1520306
1669
老实说我不知道。
25:23
I'll talk about meaning for a second.
582
1523726
1836
我一会儿再谈意义。
25:25
And I think of that as fundamentally tied
583
1525603
1961
我认为这与我们与他人建立联系
的需求息息相关
25:27
to our need for connection
to other people.
584
1527564
2085
25:29
I think sometimes we confuse,
for example, our need for meaning,
585
1529649
4004
例如,我认为有时候我们会将
对意义的需求
25:33
with a desire for personal achievement.
586
1533695
2252
与对个人成就的渴望混淆。
25:35
That we chase personal achievement
587
1535989
2210
我们追求个人成就
25:38
and what we're trying
to do is generate meaning.
588
1538241
2919
我们
想要做的是创造意义。
25:41
So I think we can be confused
589
1541828
3044
因此,我认为我们可能会感到困惑
25:44
and we can have those needs displaced
into less productive routes.
590
1544914
5506
我们可能会将这些需求转移
到效率较低的道路上。
25:51
But I don't think it's going away.
591
1551504
1752
但我认为它不会消失。
25:53
But, you know, I don't know
that it's the healthiest.
592
1553256
3587
但是,你知道,我不
知道它是最健康的。
25:57
KG: No, I'm totally aligned.
593
1557760
1669
KG:不,我完全一致。
25:59
Thank you, Brian,
that was an awesome conversation.
594
1559721
2460
谢谢你,布莱恩,
这是一次很棒的谈话。
26:02
BSL: It was great to talk to you.
595
1562181
1585
BSL:很高兴和你交谈。
26:03
Really fantastic and
super informative. Thank you.
596
1563766
2545
真的很棒而且内容
非常丰富。谢谢。
New videos
Original video on YouTube.com
关于本网站
这个网站将向你介绍对学习英语有用的YouTube视频。你将看到来自世界各地的一流教师教授的英语课程。双击每个视频页面上显示的英文字幕,即可从那里播放视频。字幕会随着视频的播放而同步滚动。如果你有任何意见或要求,请使用此联系表与我们联系。