The pattern behind self-deception | Michael Shermer

740,786 views ・ 2010-06-14

TED


請雙擊下方英文字幕播放視頻。

譯者: Hsin Cheng Lin 審譯者: Lin Su-Wei(林書暐)
00:16
So since I was here last in '06,
0
16260
3000
自從上一次,06年來這裡之後,
00:19
we discovered that global climate change
1
19260
2000
我們發現全球暖化
00:21
is turning out to be a pretty serious issue,
2
21260
2000
變成一個相當嚴肅的議題。
00:23
so we covered that fairly extensively
3
23260
2000
所以我們在"懷疑論者"雜誌裡,
00:25
in Skeptic magazine.
4
25260
2000
對暖化議題有相當深入的報導。
00:27
We investigate all kinds
5
27260
2000
我們調查了各式各樣
00:29
of scientific and quasi-scientific controversies,
6
29260
3000
科學以及偽科學的爭論。
00:32
but it turns out we don't have to worry about any of this
7
32260
2000
不過看來我們不用這麼擔心,
00:34
because the world's going to end in 2012.
8
34260
2000
反正2012年就是世界末日了。
00:36
Another update:
9
36260
2000
另外一個新聞則是,
00:38
You will recall I introduced you guys
10
38260
2000
你們可能記得我曾介紹過的
00:40
to the Quadro Tracker.
11
40260
2000
Quadro探測器。
00:42
It's like a water dowsing device.
12
42260
2000
有點像探測水源的裝置,
00:44
It's just a hollow piece of plastic with an antenna that swivels around.
13
44260
3000
在中空的塑膠上面接著一根會轉動的天線。
00:47
And you walk around, and it points to things.
14
47260
2000
當你移動的時候,它會指向某些東西,
00:49
Like if you're looking for marijuana in students' lockers,
15
49260
3000
比方說當你想在學生置物櫃裡找大麻時,
00:52
it'll point right to somebody.
16
52260
2000
它就會指著某個人...
00:54
Oh, sorry. (Laughter)
17
54260
2000
喔,抱歉。 (笑聲)
00:56
This particular one that was given to me
18
56260
2000
而我拿到的這隻
00:58
finds golf balls,
19
58260
2000
則是專門用來找高爾夫球的。
01:00
especially if you're at a golf course
20
60260
2000
特別是當你在高爾夫球場上,
01:02
and you check under enough bushes.
21
62260
3000
為了找球翻遍樹叢之後。
01:05
Well, under the category of "What's the harm of silly stuff like this?"
22
65260
3000
在那些你覺得無傷大雅的玩意兒之中,
01:08
this device, the ADE 651,
23
68260
3000
這個裝置,ADE651
01:11
was sold to the Iraqi government
24
71260
3000
被伊拉克政府
01:14
for 40,000 dollars apiece.
25
74260
2000
用四萬美金一組的價錢給買下。
01:16
It's just like this one, completely worthless,
26
76260
2000
就跟我手上的一樣,完全不值一文,
01:18
in which it allegedly worked by "electrostatic
27
78260
2000
並且被說成是利用"靜電-
01:20
magnetic ion attraction,"
28
80260
3000
磁離子引力"來操作。
01:24
which translates to
29
84260
2000
或許翻成
01:26
"pseudoscientific baloney" -- would be the nice word --
30
86260
3000
"偽科學的胡扯"可能比較好,
01:29
in which you string together a bunch of words that sound good,
31
89260
2000
把一堆華麗的名詞串在一起,
01:31
but it does absolutely nothing.
32
91260
2000
但是沒有任何意義。
01:33
In this case, at trespass points,
33
93260
3000
在這個例子裡,
01:36
allowing people to go through
34
96260
2000
若是要讓人們通過地雷區,
01:38
because your little tracker device said they were okay,
35
98260
3000
只靠這種"探測器"的判斷,
01:41
actually cost lives.
36
101260
3000
是會賠上性命的。
01:44
So there is a danger to pseudoscience,
37
104260
2000
因此相信這類偽科學,
01:46
in believing in this sort of thing.
38
106260
3000
是有潛在危險的。
01:49
So what I want to talk about today is belief.
39
109260
3000
所以今天我想談談信念。
01:52
I want to believe,
40
112260
2000
我想要相信,
01:54
and you do too.
41
114260
2000
而你們也是。
01:56
And in fact, I think my thesis here is that
42
116260
2000
而事實上,我想我今天的論點是,
01:58
belief is the natural state of things.
43
118260
2000
信念是自然而然形成的,
02:00
It is the default option. We just believe.
44
120260
2000
是預設的選項。盡管信就是了。
02:02
We believe all sorts of things.
45
122260
2000
我們相信形形色色的事物。
02:04
Belief is natural;
46
124260
2000
信念是天生的,
02:06
disbelief, skepticism, science, is not natural.
47
126260
2000
疑問、懷疑論、科學則否。
02:08
It's more difficult.
48
128260
2000
它們難以接受,
02:10
It's uncomfortable to not believe things.
49
130260
2000
因為抱持疑問會讓人不快。
02:12
So like Fox Mulder on "X-Files,"
50
132260
3000
像"X檔案"裡的福克斯穆德,
02:15
who wants to believe in UFOs? Well, we all do,
51
135260
3000
誰想要相信UFO? 我們都想。
02:18
and the reason for that is because
52
138260
2000
這是因為
02:20
we have a belief engine in our brains.
53
140260
3000
我們的大腦裡有處理信念的機制
02:23
Essentially, we are pattern-seeking primates.
54
143260
3000
事實上,我們這種靈長類喜歡尋找模式(pattern-seeking)。
02:26
We connect the dots: A is connected to B; B is connected to C.
55
146260
3000
我們找尋事物的關聯:由A到B,由B到C,
02:29
And sometimes A really is connected to B,
56
149260
3000
有時候A和B的確互為因果,
02:32
and that's called association learning.
57
152260
2000
即所謂的關聯式學習。
02:34
We find patterns, we make those connections,
58
154260
3000
我們尋找模式,發現事物的脈絡。
02:37
whether it's Pavlov's dog here
59
157260
2000
不管是巴甫洛夫的狗,
02:39
associating the sound of the bell with the food,
60
159260
3000
把鈴聲與餵食兩件事連結起來,
02:42
and then he salivates to the sound of the bell,
61
162260
2000
於是一聽到鈴聲就流口水。
02:44
or whether it's a Skinnerian rat,
62
164260
2000
或是斯金納的老鼠,
02:46
in which he's having an association
63
166260
2000
把自身的行為,
02:48
between his behavior and a reward for it,
64
168260
2000
與得到的獎賞之間聯繫起來,
02:50
and therefore he repeats the behavior.
65
170260
2000
因而重複同樣的行為。
02:52
In fact, what Skinner discovered
66
172260
2000
斯金納發現,
02:54
is that, if you put a pigeon in a box like this,
67
174260
3000
如果把鴿子放進這樣的一個箱子裡,
02:57
and he has to press one of these two keys,
68
177260
2000
讓它按下兩個按鍵中的其中一個。
02:59
and he tries to figure out what the pattern is,
69
179260
2000
它會試著猜測其中的模式,
03:01
and you give him a little reward in the hopper box there --
70
181260
2000
然後透過箱子拿到一點小小的獎賞。
03:03
if you just randomly assign rewards
71
183260
3000
若是你隨機給予獎勵,
03:06
such that there is no pattern,
72
186260
2000
即使沒有任何的規則可尋,
03:08
they will figure out any kind of pattern.
73
188260
2000
它們也會想出五花八門的方式。
03:10
And whatever they were doing just before they got the reward,
74
190260
2000
它們會記得嘗到甜頭之前所做的動作,
03:12
they repeat that particular pattern.
75
192260
2000
並且一再重複同樣動作。
03:14
Sometimes it was even spinning around twice counterclockwise,
76
194260
3000
有時候甚至會是逆時針轉兩圈、
03:17
once clockwise and peck the key twice.
77
197260
3000
順時針轉一圈然後啄兩下按鍵。
03:20
And that's called superstition,
78
200260
2000
然後迷信就產生了。
03:22
and that, I'm afraid,
79
202260
2000
而不幸的是,
03:24
we will always have with us.
80
204260
2000
我們也是如此。
03:26
I call this process "patternicity" --
81
206260
2000
我將這個過程稱作"模式化"。
03:28
that is, the tendency to find meaningful patterns
82
208260
2000
意思是從毫無意義的雜訊中,
03:30
in both meaningful and meaningless noise.
83
210260
3000
尋找有意義模式的傾向。
03:33
When we do this process, we make two types of errors.
84
213260
3000
當這樣做的時候,我們可能會犯兩種類型的錯誤:
03:36
A Type I error, or false positive,
85
216260
2000
類型一,也稱作『錯誤接受』,
03:38
is believing a pattern is real
86
218260
2000
是當相信某種模式是正確的,
03:40
when it's not.
87
220260
2000
而實際上不是。(誤認不存在的模式)
03:42
Our second type of error is a false negative.
88
222260
2000
第二種類型則是『錯誤拒絕』,
03:44
A Type II error is not believing
89
224260
2000
是當某種模式實際上是正確的,
03:46
a pattern is real when it is.
90
226260
3000
卻拒絕相信。(忽略模式的存在)
03:49
So let's do a thought experiment.
91
229260
2000
讓我們來點思想實驗:
03:51
You are a hominid three million years ago
92
231260
2000
你是生活在三百萬年前的原始人,
03:53
walking on the plains of Africa.
93
233260
3000
在非洲大草原上走著,
03:56
Your name is Lucy, okay?
94
236260
2000
你的名字是...露西。
03:58
And you hear a rustle in the grass.
95
238260
2000
你聽到草叢中一陣響動,
04:00
Is it a dangerous predator,
96
240260
2000
這會是危險的掠食者,
04:02
or is it just the wind?
97
242260
2000
或者只是一陣風?
04:04
Your next decision could be the most important one of your life.
98
244260
3000
你接下來的判斷可能就是這一生中最重要的決定。
04:07
Well, if you think that the rustle in the grass is a dangerous predator
99
247260
3000
如果你誤把草叢中的聲響當做危險的掠食者,
04:10
and it turns out it's just the wind,
100
250260
2000
結果只是一陣風。
04:12
you've made an error in cognition,
101
252260
2000
你犯下錯誤認知,
04:14
made a Type I error, false positive.
102
254260
2000
類型一的誤判。
04:16
But no harm. You just move away.
103
256260
2000
但是沒差,你只是閃邊。
04:18
You're more cautious. You're more vigilant.
104
258260
2000
變得更加小心謹慎。
04:20
On the other hand, if you believe that the rustle in the grass is just the wind,
105
260260
2000
另一方面,如果你把草叢中的聲響當做風聲,
04:22
and it turns out it's a dangerous predator,
106
262260
3000
結果是危險的掠食者--
04:25
you're lunch.
107
265260
2000
你就成了午餐。
04:27
You've just won a Darwin award.
108
267260
2000
你拿到所謂"達爾文的獎賞",
04:29
You've been taken out of the gene pool.
109
269260
2000
從基因庫中被抹去。
04:31
Now the problem here is that
110
271260
2000
問題來了,
04:33
patternicities will occur whenever the cost
111
273260
2000
模式化會發生在
04:35
of making a Type I error
112
275260
2000
當類型一錯誤的代價,
04:37
is less than the cost of making a Type II error.
113
277260
2000
低於類型二錯誤時。
04:39
This is the only equation in the talk by the way.
114
279260
2000
這是這個演講裡唯一的一條公式。
04:41
We have a pattern detection problem
115
281260
2000
我們有辨識模式的困難,
04:43
that is assessing the difference between a Type I and a Type II error
116
283260
3000
在於評估類型一和類型二錯誤的時候,
04:46
is highly problematic,
117
286260
2000
沒有辦法準確的區別,
04:48
especially in split-second, life-and-death situations.
118
288260
3000
尤其是當生死關頭的那一瞬間。
04:51
So the default position
119
291260
2000
所以我們的預設反應
04:53
is just: Believe all patterns are real --
120
293260
2000
被設定成"相信所有的模式都是真的"。
04:55
All rustles in the grass are dangerous predators
121
295260
3000
"草叢中的響動都是危險的掠食者"
04:58
and not just the wind.
122
298260
2000
"不會只是風聲而已"。
05:00
And so I think that we evolved ...
123
300260
2000
我想經過演化的歷程,
05:02
there was a natural selection for the propensity for our belief engines,
124
302260
3000
信任機制的傾向經過自然的挑選。
05:05
our pattern-seeking brain processes,
125
305260
2000
我們尋求模式的大腦進化成,
05:07
to always find meaningful patterns
126
307260
2000
總是去找尋有意義的模式。
05:09
and infuse them with these sort of
127
309260
2000
並且將之連結到對掠食者的恐懼,
05:11
predatory or intentional agencies that I'll come back to.
128
311260
3000
或是等下會提到的 意圖化的形象。
05:14
So for example, what do you see here?
129
314260
2000
舉例來說,你們看到什麼?
05:16
It's a horse head, that's right.
130
316260
2000
馬的頭部,沒錯。
05:18
It looks like a horse. It must be a horse.
131
318260
2000
看起來像匹馬,這個肯定是馬。
05:20
That's a pattern.
132
320260
2000
那就是模式。
05:22
And is it really a horse?
133
322260
2000
不過真的是匹馬嗎?
05:24
Or is it more like a frog?
134
324260
3000
還是更像隻青蛙?
05:27
See, our pattern detection device,
135
327260
2000
我們偵測模式的機制,
05:29
which appears to be located in the anterior cingulate cortex --
136
329260
3000
位於大腦的前扣帶皮層,
05:32
it's our little detection device there --
137
332260
3000
我們小小的偵測裝置,
05:35
can be easily fooled, and this is the problem.
138
335260
2000
很容易就會被誤導,問題就出在這兒。
05:37
For example, what do you see here?
139
337260
2000
比方說,這是什麼?
05:39
Yes, of course, it's a cow.
140
339260
3000
當然了,是一隻牛。
05:42
Once I prime the brain -- it's called cognitive priming --
141
342260
3000
一旦我給了大腦提示 --稱作認知啟動--
05:45
once I prime the brain to see it,
142
345260
2000
當我讓大腦開始辨識它,
05:47
it pops back out again even without the pattern that I've imposed on it.
143
347260
3000
即使不用提示,大腦也會一再的嘗試辨認。
05:50
And what do you see here?
144
350260
2000
這次你看到了什麼?
05:52
Some people see a Dalmatian dog.
145
352260
2000
有些人看到一隻大麥町狗。
05:54
Yes, there it is. And there's the prime.
146
354260
2000
沒有錯,這就是提示。
05:56
So when I go back without the prime,
147
356260
2000
即使把提示拿掉,
05:58
your brain already has the model
148
358260
2000
腦中的樣板還是在,
06:00
so you can see it again.
149
360260
2000
所以你仍然看的到。
06:02
What do you see here?
150
362260
3000
這是什麼?
06:05
Planet Saturn. Yes, that's good.
151
365260
2000
土星,很好。
06:07
How about here?
152
367260
3000
這個呢?
06:10
Just shout out anything you see.
153
370260
3000
看到什麼就出個聲。
06:14
That's a good audience, Chris.
154
374260
2000
很棒的觀眾,克里斯。
06:16
Because there's nothing in this. Well, allegedly there's nothing.
155
376260
3000
因為裡面什麼都沒有,據說是沒有。
06:19
This is an experiment done by Jennifer Whitson
156
379260
3000
這是珍妮佛惠特森在德州大學奧斯汀分校,
06:22
at U.T. Austin
157
382260
2000
所做的實驗。
06:24
on corporate environments
158
384260
2000
探討在企業環境下,
06:26
and whether feelings of uncertainty and out of control
159
386260
3000
那些無法確定或是不受控制的感覺,
06:29
makes people see illusory patterns.
160
389260
2000
是否會使人看到虛幻的模式。
06:31
That is, almost everybody sees the planet Saturn.
161
391260
3000
也就是說,幾乎所有人都看的到土星的圖樣。
06:34
People that are put in a condition of feeling out of control
162
394260
3000
但一旦人們處於自覺失控的狀況下,
06:37
are more likely to see something in this,
163
397260
2000
他們就越有可能從圖中看出什麼,
06:39
which is allegedly patternless.
164
399260
3000
即使這裡沒有任何模式可尋。
06:42
In other words, the propensity to find these patterns
165
402260
3000
換句話說,當狀況失去控制的時候,
06:45
goes up when there's a lack of control.
166
405260
3000
嘗試尋找模式的傾向就越強。
06:48
For example, baseball players are notoriously superstitious
167
408260
3000
例如,棒球選手是出了名的迷信
06:51
when they're batting,
168
411260
2000
不過只在打擊的時候。
06:53
but not so much when they're fielding.
169
413260
2000
守備的時候就不是這麼一回事。
06:55
Because fielders are successful
170
415260
2000
因為守備成功率,
06:57
90 to 95 percent of the time.
171
417260
2000
通常有九成到九成五左右。
06:59
The best batters fail seven out of 10 times.
172
419260
3000
即使最好的打者,十次裡也會有七次失誤。
07:02
So their superstitions, their patternicities,
173
422260
2000
所以他們的迷信,模式化的程度,
07:04
are all associated with feelings of lack of control
174
424260
3000
跟這種無法掌控的感覺
07:07
and so forth.
175
427260
2000
有很大的關聯。
07:10
What do you see in this particular one here, in this field?
176
430260
3000
這次你們看到了什麼?在這個區域裡。
07:13
Anybody see an object there?
177
433260
2000
有人看到什麼嗎?
07:15
There actually is something here,
178
435260
2000
其實有些東西,
07:17
but it's degraded.
179
437260
2000
只是被模糊處理過了。
07:19
While you're thinking about that,
180
439260
2000
當你們在嘗試的時候,
07:21
this was an experiment done by Susan Blackmore,
181
441260
2000
這是蘇珊布萊克摩爾,一位心理學家
07:23
a psychologist in England,
182
443260
2000
在英國所做的實驗。
07:25
who showed subjects this degraded image
183
445260
2000
他們讓受試者觀看模糊處理過的圖片,
07:27
and then ran a correlation between
184
447260
2000
接著分析兩者間的關聯性:
07:29
their scores on an ESP test:
185
449260
2000
他們的ESP(第六感)測驗分數--
07:31
How much did they believe in the paranormal,
186
451260
2000
對科學無法解釋的事件、
07:33
supernatural, angels and so forth.
187
453260
3000
超自然、天使等等,相信程度的多寡--
07:36
And those who scored high on the ESP scale,
188
456260
3000
在ESP量表上得分越高的人,
07:39
tended to not only see
189
459260
2000
就越容易,
07:41
more patterns in the degraded images
190
461260
2000
不僅看到更多的樣式,
07:43
but incorrect patterns.
191
463260
2000
也辨識出不正確的樣式。
07:45
Here is what you show subjects.
192
465260
2000
這是你讓受試者看到的,
07:47
The fish is degraded 20 percent, 50 percent
193
467260
3000
魚的圖片經過百分之二十,和五十的模糊化處理。
07:50
and then the one I showed you,
194
470260
2000
以及我放的這張,
07:52
70 percent.
195
472260
2000
百分之七十。
07:54
A similar experiment was done by another [Swiss] psychologist
196
474260
2000
另一位英國(應為瑞士籍)心理學家,
07:56
named Peter Brugger,
197
476260
2000
彼得布魯格也做過類似的實驗。
07:58
who found significantly more meaningful patterns
198
478260
3000
他發現連接左邊視覺區的右腦半球,
08:01
were perceived on the right hemisphere,
199
481260
2000
比起左腦,
08:03
via the left visual field, than the left hemisphere.
200
483260
3000
更容易察覺有意義的圖樣。
08:06
So if you present subjects the images such
201
486260
2000
因此如果你給受測者看的圖片,
08:08
that it's going to end up on the right hemisphere instead of the left,
202
488260
3000
偏重於用右腦來處理,
08:11
then they're more likely to see patterns
203
491260
2000
那麼比起需要用左腦處理的圖片,
08:13
than if you put it on the left hemisphere.
204
493260
2000
他們會更容易發現其中的模式。
08:15
Our right hemisphere appears to be
205
495260
2000
看來我們的右腦
08:17
where a lot of this patternicity occurs.
206
497260
2000
負責了大多數的模式化行為。
08:19
So what we're trying to do is bore into the brain
207
499260
2000
所以我們試著觀察大腦內部,
08:21
to see where all this happens.
208
501260
2000
找出相關的區域。
08:23
Brugger and his colleague, Christine Mohr,
209
503260
3000
布魯格和他的同事,克莉絲汀莫爾
08:26
gave subjects L-DOPA.
210
506260
2000
提供受測者左旋多巴胺(L-DOPA)。
08:28
L-DOPA's a drug, as you know, given for treating Parkinson's disease,
211
508260
3000
你們或許聽過左旋多巴胺是用來治療帕金森氏症,
08:31
which is related to a decrease in dopamine.
212
511260
3000
與患者的多巴胺含量降低有關,
08:34
L-DOPA increases dopamine.
213
514260
2000
而這種藥能夠提升多巴胺的含量。
08:36
An increase of dopamine caused
214
516260
2000
多巴胺含量的提升
08:38
subjects to see more patterns
215
518260
2000
則使得受測者看到更多圖樣,
08:40
than those that did not receive the dopamine.
216
520260
2000
比未使用的受測者還多。
08:42
So dopamine appears to be the drug
217
522260
2000
也就是說多巴胺應該和模式化認知
08:44
associated with patternicity.
218
524260
2000
有相當程度的關聯。
08:46
In fact, neuroleptic drugs
219
526260
2000
事實上,一些鎮靜劑
08:48
that are used to eliminate psychotic behavior,
220
528260
2000
被用來抑制精神病相關症狀,如
08:50
things like paranoia, delusions
221
530260
2000
妄想症、錯覺,
08:52
and hallucinations,
222
532260
2000
以及幻覺,
08:54
these are patternicities.
223
534260
2000
這些都是模式化行為。
08:56
They're incorrect patterns. They're false positives. They're Type I errors.
224
536260
3000
只是那些是不正確的模式。類型一的誤判。
08:59
And if you give them drugs
225
539260
2000
當患者使用
09:01
that are dopamine antagonists,
226
541260
2000
抑制多巴胺的藥劑,
09:03
they go away.
227
543260
2000
這些症狀就會消失。
09:05
That is, you decrease the amount of dopamine,
228
545260
2000
多巴胺的含量一降低,
09:07
and their tendency to see
229
547260
2000
患者原本容易看到不存在的模式,
09:09
patterns like that decreases.
230
549260
2000
這樣的傾向也隨著降低。
09:11
On the other hand, amphetamines like cocaine
231
551260
3000
另一種情況,安非他命類藥劑如可卡因,
09:14
are dopamine agonists.
232
554260
2000
是多巴胺的促進劑。
09:16
They increase the amount of dopamine.
233
556260
2000
這類藥劑會提升體內的多巴胺分泌量,
09:18
So you're more likely to feel in a euphoric state,
234
558260
3000
讓你容易覺得飄飄欲仙,
09:21
creativity, find more patterns.
235
561260
2000
文思泉湧,看到更多東西。
09:23
In fact, I saw Robin Williams recently
236
563260
2000
我最近就和羅賓威廉斯討論到
09:25
talk about how he thought he was much funnier
237
565260
2000
他嗑藥的時候--過去他有這方面的問題--
09:27
when he was doing cocaine, when he had that issue, than now.
238
567260
3000
覺得那時可卡因讓他比現在更加風趣。
09:30
So perhaps more dopamine
239
570260
2000
或許多巴胺的量越多,
09:32
is related to more creativity.
240
572260
2000
能夠讓你更有創造力。
09:34
Dopamine, I think, changes
241
574260
2000
我認為多巴胺會改變
09:36
our signal-to-noise ratio.
242
576260
2000
我們對信號/雜訊的認知。
09:38
That is, how accurate we are
243
578260
2000
也就是讓我們找出模式
09:40
in finding patterns.
244
580260
2000
的準確程度。
09:42
If it's too low, you're more likely to make too many Type II errors.
245
582260
3000
如果準確度太低,你會傾向於犯下更多類型二的錯誤。
09:45
You miss the real patterns. You don't want to be too skeptical.
246
585260
2000
你會忽略實際存在的模式,你不想變得太疑神疑鬼,
09:47
If you're too skeptical, you'll miss the really interesting good ideas.
247
587260
3000
果真如此,你可能會錯過那些真正有趣的想法。
09:51
Just right, you're creative, and yet you don't fall for too much baloney.
248
591260
3000
恰到好處的話,你會很有創意。也不會被人耍著玩。
09:54
Too high and maybe you see patterns everywhere.
249
594260
3000
太過度的話,你會發現到處都是模式。
09:57
Every time somebody looks at you, you think people are staring at you.
250
597260
3000
別人只是看你一眼,你卻覺得對方在瞪你。
10:00
You think people are talking about you.
251
600260
2000
你覺得大家都在討論你。
10:02
And if you go too far on that, that's just simply
252
602260
2000
若是這樣的狀況太過嚴重,
10:04
labeled as madness.
253
604260
2000
你就會被當作瘋子。
10:06
It's a distinction perhaps we might make
254
606260
2000
我們或許可以從兩位諾貝爾獎得主,
10:08
between two Nobel laureates, Richard Feynman
255
608260
2000
理查費曼和約翰奈許之間
10:10
and John Nash.
256
610260
2000
看出這樣的差異。
10:12
One sees maybe just the right number
257
612260
2000
其中一人看出的模式,
10:14
of patterns to win a Nobel Prize.
258
614260
2000
剛好足以讓他得到諾貝爾獎。
10:16
The other one also, but maybe too many patterns.
259
616260
2000
另外一人看到的可能太多了一點,
10:18
And we then call that schizophrenia.
260
618260
3000
我們會說這是精神分裂症。
10:21
So the signal-to-noise ratio then presents us with a pattern-detection problem.
261
621260
3000
所以辨別信號/雜訊的能力顯示出模式偵測的問題,
10:24
And of course you all know exactly
262
624260
2000
當然你們都知道
10:26
what this is, right?
263
626260
2000
這是什麼,對吧。
10:28
And what pattern do you see here?
264
628260
2000
這裡你看到了什麼?
10:30
Again, I'm putting your anterior cingulate cortex to the test here,
265
630260
3000
我正在考驗你大腦的前扣帶皮層,
10:33
causing you conflicting pattern detections.
266
633260
3000
混淆你眼中所看到的模式。
10:36
You know, of course, this is Via Uno shoes.
267
636260
2000
當然,這是VIA UNO的鞋子,
10:38
These are sandals.
268
638260
3000
兩款涼鞋。
10:41
Pretty sexy feet, I must say.
269
641260
3000
不得不承認這腳很性感,
10:44
Maybe a little Photoshopped.
270
644260
2000
搞不好有修過圖。
10:46
And of course, the ambiguous figures
271
646260
2000
當然還有模稜兩可的圖形,
10:48
that seem to flip-flop back and forth.
272
648260
2000
看上去不停的變動著。
10:50
It turns out what you're thinking about a lot
273
650260
2000
也就是說,一個人腦袋裡裝了什麼,
10:52
influences what you
274
652260
2000
就會影響你
10:54
tend to see.
275
654260
2000
所看到的。
10:56
And you see the lamp here, I know.
276
656260
2000
我知道你們看到的是檯燈,
10:58
Because the lights on here.
277
658260
3000
因為燈是亮的。
11:01
Of course, thanks to the environmentalist movement
278
661260
2000
感謝環保人士的積極運動,
11:03
we're all sensitive to the plight of marine mammals.
279
663260
3000
我們才能對海洋生物的困境有所了解。
11:06
So what you see in this particular ambiguous figure
280
666260
3000
所以我們才能在這張曖昧的圖片裡看到...
11:09
is, of course, the dolphins, right?
281
669260
2000
海豚,當然了。
11:11
You see a dolphin here,
282
671260
2000
你看這邊有一隻,
11:13
and there's a dolphin,
283
673260
2000
那邊一隻,
11:15
and there's a dolphin.
284
675260
2000
又一隻。
11:17
That's a dolphin tail there, guys.
285
677260
3000
拜託,那是海豚尾巴好嗎。
11:20
(Laughter)
286
680260
3000
(笑聲)
11:25
If we can give you conflicting data, again,
287
685260
3000
如果你得到互相衝突的資訊,
11:28
your ACC is going to be going into hyperdrive.
288
688260
3000
你的前扣帶皮層就會像進入加速狀態一樣。
11:31
If you look down here, it's fine. If you look up here, then you get conflicting data.
289
691260
3000
底下這裡看起來很正常,但是往上一看,便會察覺矛盾。
11:34
And then we have to flip the image
290
694260
2000
直到我們將圖片翻轉,
11:36
for you to see that it's a set up.
291
696260
2000
你才看得出這是刻意設計的。
11:40
The impossible crate illusion.
292
700260
2000
"不存在的箱子"的幻覺。
11:42
It's easy to fool the brain in 2D.
293
702260
2000
平面影像很容易騙過大腦,
11:44
So you say, "Aw, come on Shermer, anybody can do that
294
704260
2000
你說"得了吧,大家都做得到,
11:46
in a Psych 101 text with an illusion like that."
295
706260
2000
每本心理學入門書都有一張這種圖。"
11:48
Well here's the late, great Jerry Andrus'
296
708260
2000
這張是最近由傑里安德勒斯所做的,
11:50
"impossible crate" illusion in 3D,
297
710260
3000
立體版本的"不存在的箱子"。
11:53
in which Jerry is standing inside
298
713260
2000
而且傑里看起來像是
11:55
the impossible crate.
299
715260
2000
站在箱子的裡面。
11:57
And he was kind enough to post this
300
717260
2000
他非常大方的上傳了解答,
11:59
and give us the reveal.
301
719260
2000
讓我們能看出箇中巧妙。
12:01
Of course, camera angle is everything. The photographer is over there,
302
721260
3000
當然,重點在於相機的角度,攝影師在那裡,
12:04
and this board appears to overlap with this one, and this one with that one, and so on.
303
724260
3000
使得這塊板子看來像是疊在另一塊上方,以此類推。
12:07
But even when I take it away,
304
727260
2000
即使我已經讓你看過解答,
12:09
the illusion is so powerful because of how are brains are wired
305
729260
2000
大腦尋找特定模式的運作方式,
12:11
to find those certain kinds of patterns.
306
731260
3000
還是讓幻覺的印象非常強烈。
12:14
This is a fairly new one
307
734260
2000
這張比較新一點。
12:16
that throws us off because of the conflicting patterns
308
736260
2000
比較兩張照片中不同角度的矛盾,
12:18
of comparing this angle with that angle.
309
738260
3000
會讓我們搞不清楚。
12:21
In fact, it's the exact same picture side by side.
310
741260
3000
其實這是兩張完全相同的照片。
12:24
So what you're doing is comparing that angle
311
744260
2000
問題在於沒注意到比對的對象,
12:26
instead of with this one, but with that one.
312
746260
2000
兩者間的角度有所不同,
12:28
And so your brain is fooled.
313
748260
2000
你的大腦就被騙了。
12:30
Yet again, your pattern detection devices are fooled.
314
750260
2000
你的模式辨識機制又再一次被愚弄了。
12:32
Faces are easy to see
315
752260
2000
臉孔很容易辨識,
12:34
because we have an additional evolved
316
754260
2000
因為在大腦的顳葉裡,
12:36
facial recognition software
317
756260
2000
我們擁有獨立進化過的,
12:38
in our temporal lobes.
318
758260
3000
臉部辨識軟體。
12:41
Here's some faces on the side of a rock.
319
761260
3000
這些是岩石上出現的臉孔,
12:44
I'm actually not even sure if this is -- this might be Photoshopped.
320
764260
3000
我甚至不確定這是不是修過圖,這個可能是。
12:47
But anyway, the point is still made.
321
767260
2000
無論如何,重點還是一樣。
12:49
Now which one of these looks odd to you?
322
769260
2000
現在哪一張看起來怪怪的?
12:51
In a quick reaction, which one looks odd?
323
771260
2000
用直覺來作答。
12:53
The one on the left. Okay. So I'll rotate it
324
773260
2000
左邊這張,好,我把它轉過來,
12:55
so it'll be the one on the right.
325
775260
2000
也就是變成右手邊的這張,
12:57
And you are correct.
326
777260
2000
你們是對的。
12:59
A fairly famous illusion -- it was first done with Margaret Thatcher.
327
779260
3000
相當有名的幻象,柴契爾夫人是第一個被用上的,
13:02
Now, they trade up the politicians every time.
328
782260
2000
每隔一陣子就換一位政治人物。
13:04
Well, why is this happening?
329
784260
2000
為什麼會這樣?
13:06
Well, we know exactly where it happens,
330
786260
2000
恩,我們確切知道這在哪裡發生,
13:08
in the temporal lobe, right across, sort of above your ear there,
331
788260
3000
在顳葉裡,偏右,大概在耳朵上方。
13:11
in a little structure called the fusiform gyrus.
332
791260
3000
有一個組織叫做梭狀回,
13:14
And there's two types of cells that do this,
333
794260
2000
裡頭有兩種細胞,
13:16
that record facial features either globally,
334
796260
3000
不管是動用整群細胞來紀錄臉部的特徵,
13:19
or specifically these large, rapid-firing cells,
335
799260
2000
或是這一類大型,反應快速的細胞。
13:21
first look at the general face.
336
801260
2000
先看出臉部的大概,
13:23
So you recognize Obama immediately.
337
803260
2000
讓你能立刻認出歐巴馬。
13:25
And then you notice something quite
338
805260
2000
接著你會注意到有點不對勁,
13:27
a little bit odd about the eyes and the mouth.
339
807260
2000
眼睛和嘴巴的部分有些奇怪,
13:29
Especially when they're upside down,
340
809260
2000
尤其是當它們上下顛倒的時候。
13:31
you're engaging that general facial recognition software there.
341
811260
3000
那就是你正在使用臉部辨識的軟體。
13:34
Now I said back in our little thought experiment,
342
814260
3000
現在我們回到之前的思想實驗。
13:37
you're a hominid walking on the plains of Africa.
343
817260
2000
你是正在非洲草原上行走的原始人,
13:39
Is it just the wind or a dangerous predator?
344
819260
3000
心裡想著是風聲,還是危險的掠食者?
13:42
What's the difference between those?
345
822260
2000
兩者間的差異在哪?
13:44
Well, the wind is inanimate;
346
824260
2000
風聲是沒有生命的,
13:46
the dangerous predator is an intentional agent.
347
826260
2000
而掠食者則代表了一個擁有意圖的形象。
13:48
And I call this process agenticity.
348
828260
2000
我把這個過程稱作形象化,
13:50
That is the tendency to infuse patterns
349
830260
2000
也就是傾向於將觀察到的模式賦予意義、
13:52
with meaning, intention and agency,
350
832260
2000
目的以及形象--
13:54
often invisible beings from the top down.
351
834260
3000
常被理解為從上方俯瞰,不可見的存在--
13:57
This is an idea that we got
352
837260
2000
這個想法是從另一位TED演講者,
13:59
from a fellow TEDster here, Dan Dennett,
353
839260
2000
丹尼特而來的。
14:01
who talked about taking the intentional stance.
354
841260
2000
他提到抱持"有目的的立場"。
14:03
So it's a type of that expanded to explain, I think, a lot of different things:
355
843260
3000
所以我想,這個說法的衍伸,可以解釋很多事物,
14:06
souls, spirits, ghosts, gods, demons, angels,
356
846260
3000
靈魂、精靈、鬼魂、神祇、魔鬼、天使、
14:09
aliens, intelligent designers,
357
849260
2000
外星人、智慧設計者、
14:11
government conspiracists
358
851260
2000
政府陰謀論者,
14:13
and all manner of invisible agents
359
853260
2000
以及各種不可見的形象,
14:15
with power and intention, are believed
360
855260
2000
擁有力量和目的,人們相信
14:17
to haunt our world and control our lives.
361
857260
2000
它們會在人間作祟或控制我們的生活。
14:19
I think it's the basis of animism
362
859260
2000
我想這就是泛靈論,
14:21
and polytheism and monotheism.
363
861260
3000
以及一神論和多神論的源頭。
14:24
It's the belief that aliens are somehow
364
864260
2000
人們相信外星人
14:26
more advanced than us, more moral than us,
365
866260
2000
比我們更進步,更具道德感,
14:28
and the narratives always are
366
868260
2000
而故事旁白總是會說,
14:30
that they're coming here to save us and rescue us from on high.
367
870260
3000
它們是從天而降來拯救我們的。
14:33
The intelligent designer's always portrayed
368
873260
2000
"智慧設計者"則被描繪成
14:35
as this super intelligent, moral being
369
875260
3000
無比聰明,道德高尚的存在,
14:38
that comes down to design life.
370
878260
2000
降臨塵世以創造萬物。
14:40
Even the idea that government can rescue us --
371
880260
2000
即使是政府可以救助我們,
14:42
that's no longer the wave of the future,
372
882260
2000
這種已經退流行的想法,
14:44
but that is, I think, a type of agenticity:
373
884260
2000
我仍然認為是某種形象化的行為。
14:46
projecting somebody up there,
374
886260
2000
幻想某個高高在上,
14:48
big and powerful, will come rescue us.
375
888260
2000
全能而偉大的傢伙會來拯救我們。
14:50
And this is also, I think, the basis of conspiracy theories.
376
890260
2000
換個角度說,我認為這也是陰謀論的源頭:
14:52
There's somebody hiding behind there pulling the strings,
377
892260
3000
某個藏鏡人在背後操控一切,
14:55
whether it's the Illuminati
378
895260
2000
例如光明會,
14:57
or the Bilderbergers.
379
897260
2000
或是畢德堡集團。
14:59
But this is a pattern detection problem, isn't it?
380
899260
2000
但是我們面對的是模式辨認的問題,
15:01
Some patterns are real and some are not.
381
901260
2000
有些模式是真的,有些則否。
15:03
Was JFK assassinated by a conspiracy or by a lone assassin?
382
903260
3000
約翰甘迺迪是被暗中殺害,或者只是一位刺客?
15:06
Well, if you go there -- there's people there on any given day --
383
906260
3000
如果你到這個地方,那裡一年到頭都有人,
15:09
like when I went there, here -- showing me where the different shooters were.
384
909260
3000
像我去的那次,有人指給我看不同槍手的位置。
15:12
My favorite one was he was in the manhole.
385
912260
3000
我最愛的是躲在人孔蓋下面的那個,
15:15
And he popped out at the last second, took that shot.
386
915260
3000
殺手在關鍵時刻從底下蹦出來,開槍殺了甘迺迪。
15:18
But of course, Lincoln was assassinated by a conspiracy.
387
918260
2000
當然我們知道林肯是被陰謀殺害的,
15:20
So we can't just uniformly dismiss
388
920260
2000
所以我們也不能單純只是忽略,
15:22
all patterns like that.
389
922260
2000
所有可能的說法。
15:24
Because, let's face it, some patterns are real.
390
924260
2000
因為事實上,有些模式是真的,
15:26
Some conspiracies really are true.
391
926260
2000
有些陰謀是千真萬確的。
15:30
Explains a lot, maybe.
392
930260
2000
原來如此,是吧。
15:32
And 9/11 has a conspiracy theory. It is a conspiracy.
393
932260
3000
911事件也有陰謀論說法。
15:35
We did a whole issue on it.
394
935260
2000
我們做了一整期的專題來討論。
15:37
Nineteen members of Al Queda plotting to fly planes into buildings
395
937260
2000
十九名基地組織的成員計畫用飛機衝撞大樓,
15:39
constitutes a conspiracy.
396
939260
2000
算是某種陰謀。
15:41
But that's not what the "9/11 truthers" think.
397
941260
2000
但是"911真相調查會"可不這麼想,
15:43
They think it was an inside job by the Bush administration.
398
943260
3000
他們認為是小布希政府在幕後操作...
15:46
Well, that's a whole other lecture.
399
946260
2000
光這個主題就可以另外講一整場。
15:48
You know how we know that 9/11
400
948260
2000
不過你會說,我們怎麼知道911事件
15:50
was not orchestrated by the Bush administration?
401
950260
2000
不是布希政府自導自演的?
15:52
Because it worked.
402
952260
2000
因為它成功了!
15:54
(Laughter)
403
954260
3000
(笑聲)
15:57
(Applause)
404
957260
3000
(掌聲)
16:00
So we are natural-born dualists.
405
960260
2000
我們是天生的二元論者。
16:02
Our agenticity process comes from
406
962260
2000
我們創造形象的能力,來自於
16:04
the fact that we can enjoy movies like these.
407
964260
2000
一種讓我們能夠享受這種電影的天賦。
16:06
Because we can imagine, in essence,
408
966260
2000
因為我們能夠想像事物的本質,
16:08
continuing on.
409
968260
2000
並且舉一反三。
16:10
We know that if you stimulate the temporal lobe,
410
970260
2000
我們知道如果刺激顳葉,
16:12
you can produce a feeling of out-of-body experiences,
411
972260
2000
就可以模擬靈魂出竅的經驗、
16:14
near-death experiences,
412
974260
2000
瀕死體驗等,
16:16
which you can do by just touching an electrode to the temporal lobe there.
413
976260
3000
你只需要用電極刺激顳葉裡的特定位置。
16:19
Or you can do it through loss of consciousness,
414
979260
2000
或者是利用離心機加速,
16:21
by accelerating in a centrifuge.
415
981260
2000
讓人失去知覺。
16:23
You get a hypoxia, or a lower oxygen.
416
983260
3000
當你缺氧,血液含氧量降低,
16:26
And the brain then senses
417
986260
2000
大腦就會接著
16:28
that there's an out-of-body experience.
418
988260
2000
產生脫離身體的感覺。
16:30
You can use -- which I did, went out and did --
419
990260
2000
你可以利用 --我曾經嘗試過--
16:32
Michael Persinger's God Helmet,
420
992260
2000
麥可佩辛格的"上帝的頭盔",
16:34
that bombards your temporal lobes with electromagnetic waves.
421
994260
2000
它利用電磁波轟炸你的顳葉,
16:36
And you get a sense of out-of-body experience.
422
996260
3000
讓你能夠體驗靈魂出竅的感受。
16:39
So I'm going to end here with a short video clip
423
999260
2000
最後我想利用一段影片
16:41
that sort of brings all this together.
424
1001260
2000
來作個總結。
16:43
It's just a minute and a half.
425
1003260
2000
只有一分半鐘的短片,
16:45
It ties together all this into the power of expectation and the power of belief.
426
1005260
3000
展現出「期望」和「信念」的力量有多麼強大。
16:48
Go ahead and roll it.
427
1008260
2000
請撥放影片。
16:50
Narrator: This is the venue they chose for their fake auditions
428
1010260
3000
旁白:這裡是他們用來假裝面試的地方,
16:53
for an advert for lip balm.
429
1013260
2000
面試內容是一個護唇膏的廣告。
16:55
Woman: We're hoping we can use part of this
430
1015260
2000
我們希望可以使用一部分內容,
16:57
in a national commercial, right?
431
1017260
2000
放在全國性的廣告上。
16:59
And this is test on some lip balms
432
1019260
2000
測試內容是這裡所放的
17:01
that we have over here.
433
1021260
2000
幾種護唇膏。
17:03
And these are our models who are going to help us,
434
1023260
2000
我們請了幾位模特兒來幫忙,
17:05
Roger and Matt.
435
1025260
2000
羅傑和麥特。
17:07
And we have our own lip balm,
436
1027260
2000
這是我們的護唇膏,
17:09
and we have a leading brand.
437
1029260
2000
以及幾款市面上的領導品牌。
17:11
Would you have any problem
438
1031260
2000
如果需要你和我們的模特兒接吻,
17:13
kissing our models to test it?
439
1033260
2000
你會不會有困難?
17:15
Girl: No.
440
1035260
2000
女孩:不會。
17:17
Woman: You wouldn't? (Girl: No.) Woman: You'd think that was fine.
441
1037260
2000
沒有問題吧?(女孩:沒有。)你可以接受?
17:19
Girl: That would be fine. (Woman: Okay.)
442
1039260
2000
女孩:沒問題。(好的。)
17:21
So this is a blind test.
443
1041260
3000
這是一個蒙眼測試。
17:24
I'm going to ask you to go ahead
444
1044260
2000
我會請你上前,
17:26
and put a blindfold on.
445
1046260
2000
用眼罩遮住。
17:29
Kay, now can you see anything? (Girl: No.)
446
1049260
3000
好,你現在還看的到嗎?(女孩:看不到。)
17:32
Pull it so you can't even see down. (Girl: Okay.)
447
1052260
2000
往下拉一點,這樣才不會看到下方。(女孩:好。)
17:34
Woman: It's completely blind now, right?
448
1054260
2000
現在你完全看不見了。
17:36
Girl: Yes. (Woman: Okay.)
449
1056260
2000
女孩:是的。(很好。)
17:38
Now, what I'm going to be looking for in this test
450
1058260
3000
這個測試的目的是要觀察護唇膏
17:41
is how it protects your lips,
451
1061260
3000
對嘴唇的保護作用,
17:44
the texture, right,
452
1064260
2000
觸感,
17:46
and maybe if you can discern any flavor or not.
453
1066260
3000
你也許能夠辨別的出某種味道。
17:49
Girl: Okay. (Woman: Have you ever done a kissing test before?)
454
1069260
3000
女孩:了解。(你有過接吻測試的經驗嗎?)
17:52
Girl: No.
455
1072260
2000
女孩:沒有。
17:54
Woman: Take a step here.
456
1074260
2000
往前站一點。
17:56
Okay, now I'm going to ask you to pucker up.
457
1076260
2000
好,現在請你把嘴唇噘起來,
17:58
Pucker up big and lean in just a little bit, okay?
458
1078260
3000
盡量噘起來,然後稍微往前傾,很好。
18:06
(Music)
459
1086260
4000
(音樂)
18:10
(Laughter)
460
1090260
5000
(笑聲)
18:19
(Laughter)
461
1099260
3000
(笑聲)
18:30
Woman: Okay.
462
1110260
2000
好。
18:32
And, Jennifer, how did that feel?
463
1112260
2000
珍妮,感覺如何?
18:34
Jennifer: Good.
464
1114260
2000
珍妮:不錯。
18:36
(Laughter)
465
1116260
7000
(笑聲)
18:43
Girl: Oh my God!
466
1123260
2000
女孩:喔,天哪。
18:45
(Laughter)
467
1125260
4000
(笑聲)
18:50
Michael Shermer: Thank you very much. Thank you. Thanks.
468
1130260
3000
非常感謝各位,謝謝
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隱私政策

eng.lish.video

Developer's Blog