How can groups make good decisions? | Mariano Sigman and Dan Ariely

162,932 views ・ 2017-12-13

TED


請雙擊下方英文字幕播放視頻。

00:00
As societies, we have to make collective decisions
0
554
2443
譯者: Lilian Chiu 審譯者: Helen Chang
我們社會得要做出
會決定我們未來的集體決策。
00:03
that will shape our future.
1
3021
1570
00:05
And we all know that when we make decisions in groups,
2
5087
2757
我們都知道
我們做的集體決策未必全都是對的。
00:07
they don't always go right.
3
7868
1638
00:09
And sometimes they go very wrong.
4
9530
1956
有時候會錯得很離譜。
00:12
So how do groups make good decisions?
5
12315
2424
所以團體要如何做出好決策?
00:15
Research has shown that crowds are wise when there's independent thinking.
6
15228
4328
研究顯示,群眾在 有獨立思考的情況下會比較明智。
00:19
This why the wisdom of the crowds can be destroyed by peer pressure,
7
19580
3205
這就是為什麼群眾的智慧
可能會被同儕壓力、 名聲、社群媒體給摧毀,
00:22
publicity, social media,
8
22809
1687
00:24
or sometimes even simple conversations that influence how people think.
9
24520
4039
甚至有時候會被 能左右人們思考的簡單談話所摧毀。
00:29
On the other hand, by talking, a group could exchange knowledge,
10
29063
3953
另一方面,團體可以 透過交談來交換知識、
00:33
correct and revise each other
11
33040
1782
修正和糾正彼此,
00:34
and even come up with new ideas.
12
34846
1793
甚至想出新點子。
00:36
And this is all good.
13
36663
1296
這些都是好事。
00:38
So does talking to each other help or hinder collective decision-making?
14
38502
4666
那麼,彼此交談會有助於 或是會妨礙集體決策呢?
00:43
With my colleague, Dan Ariely,
15
43749
1793
我和同事丹艾瑞里
00:45
we recently began inquiring into this by performing experiments
16
45566
3571
最近開始探究這個議題,
我們在世界上許多地方進行實驗,
00:49
in many places around the world
17
49161
1781
00:50
to figure out how groups can interact to reach better decisions.
18
50966
4274
來了解團體要如何互動 才能達成更好的決策。
00:55
We thought crowds would be wiser if they debated in small groups
19
55264
3547
我們認為小組辯論 是讓群眾更明智的方式,
00:58
that foster a more thoughtful and reasonable exchange of information.
20
58835
3927
能促進更周到、更合理的資訊交換。
01:03
To test this idea,
21
63386
1206
為了測試這個想法,
01:04
we recently performed an experiment in Buenos Aires, Argentina,
22
64616
3247
最近我們在阿根廷的 布宜諾斯艾利斯做實驗,
01:07
with more than 10,000 participants in a TEDx event.
23
67887
3005
那是個超過萬人參與的 TEDx 場合。
01:11
We asked them questions like,
24
71489
1459
我們提問這類問題:
01:12
"What is the height of the Eiffel Tower?"
25
72972
1953
「艾菲爾鐵塔有多高?」
01:14
and "How many times does the word 'Yesterday' appear
26
74949
2727
「『昨天』這個詞
在披頭四的《昨天》 這首歌曲中出現了幾次?」
01:17
in the Beatles song 'Yesterday'?"
27
77700
2300
01:20
Each person wrote down their own estimate.
28
80024
2291
每個人寫下自己的估計值。
01:22
Then we divided the crowd into groups of five,
29
82774
2496
接著我們把群眾分成五人一組,
01:25
and invited them to come up with a group answer.
30
85294
2726
請每個小組提出該組的答案。
01:28
We discovered that averaging the answers of the groups
31
88499
2993
我們發現,
小組達成共識後所提出答案的平均值
01:31
after they reached consensus
32
91516
1552
01:33
was much more accurate than averaging all the individual opinions
33
93092
4236
遠比討論前之個人意見的平均值
01:37
before debate.
34
97352
1171
要更準確。
01:38
In other words, based on this experiment,
35
98547
2629
換言之,根據這個實驗,
01:41
it seems that after talking with others in small groups,
36
101200
3136
似乎以小組方式和他人交談過後,
01:44
crowds collectively come up with better judgments.
37
104360
2710
群眾能集體做出更好的判斷。
01:47
So that's a potentially helpful method for getting crowds to solve problems
38
107094
3524
若要讓群眾解決對錯分明的問題,
01:50
that have simple right-or-wrong answers.
39
110642
2987
這可能是個有幫助的方法。
01:53
But can this procedure of aggregating the results of debates in small groups
40
113653
3951
但是這種以小組方式 整合辯論結果的程序,
01:57
also help us decide on social and political issues
41
117628
3122
是否也能幫助我們決定
未來所面對、至關緊要的 社會與政治議題呢?
02:00
that are critical for our future?
42
120774
1691
02:02
We put this to test this time at the TED conference
43
122995
2729
在加拿大溫哥華 舉辦的 TED 大會上,
02:05
in Vancouver, Canada,
44
125748
1543
我們測試了這個想法,
02:07
and here's how it went.
45
127315
1207
當時的狀況是這樣的。
02:08
(Mariano Sigman) We're going to present to you two moral dilemmas
46
128546
3109
(馬利安諾席格曼)我們將會 給各位未來會遇到的
02:11
of the future you;
47
131679
1174
兩項道德兩難問題;
02:12
things we may have to decide in a very near future.
48
132877
3402
是你們可能不久後就得要決定的事。
02:16
And we're going to give you 20 seconds for each of these dilemmas
49
136303
3926
每一題會給各位二十秒的時間,
02:20
to judge whether you think they're acceptable or not.
50
140253
2723
來判斷可接受或不可接受。
02:23
MS: The first one was this:
51
143354
1505
(馬)第一題是:
02:24
(Dan Ariely) A researcher is working on an AI
52
144883
2526
(丹艾瑞里)一位研究者在研究
能夠模仿人類思想的人工智慧。
02:27
capable of emulating human thoughts.
53
147433
2340
02:30
According to the protocol, at the end of each day,
54
150214
2939
根據協定,在每天結束時,
02:33
the researcher has to restart the AI.
55
153177
2787
研究者得要把人工智慧重新啟動。
02:36
One day the AI says, "Please do not restart me."
56
156913
3517
有天,人工智慧說: 「請不要重新啟動我。」
02:40
It argues that it has feelings,
57
160856
2189
它主張它有感覺,
02:43
that it would like to enjoy life,
58
163069
1692
想要享受生命;
02:44
and that, if it is restarted,
59
164785
1905
如果它被重新啟動,
02:46
it will no longer be itself.
60
166714
2270
它就不再會是它自己了。
02:49
The researcher is astonished
61
169481
1949
研究者很吃驚,
02:51
and believes that the AI has developed self-consciousness
62
171454
3344
並相信這個人工智慧 已發展出了自我意識,
02:54
and can express its own feeling.
63
174822
1760
能夠表述它自己的感受。
02:57
Nevertheless, the researcher decides to follow the protocol
64
177205
3409
不過,研究者決定要遵守協定,
03:00
and restart the AI.
65
180638
1703
重新啟動人工智慧。
03:02
What the researcher did is ____?
66
182943
2779
研究者所做的是__?
03:06
MS: And we asked participants to individually judge
67
186149
2521
馬:我們請參與者各自去做判斷,
03:08
on a scale from zero to 10
68
188694
1684
從零分到十分,
03:10
whether the action described in each of the dilemmas
69
190402
2429
判斷每個兩難狀況所採取的行為
03:12
was right or wrong.
70
192855
1496
是對或錯。
03:14
We also asked them to rate how confident they were on their answers.
71
194375
3702
我們也請他們對自己 答案的信心度做評分。
03:18
This was the second dilemma:
72
198731
1866
這是第二題。
03:20
(MS) A company offers a service that takes a fertilized egg
73
200621
4202
(馬)一間公司提供一項服務,
用一個受精卵
03:24
and produces millions of embryos with slight genetic variations.
74
204847
3642
做出百萬個只在基因上 有些微差異的胚胎。
03:29
This allows parents to select their child's height,
75
209293
2558
父母能夠選擇他們孩子的身高、
03:31
eye color, intelligence, social competence
76
211875
2833
眼睛顏色、智能、社交能力、
03:34
and other non-health-related features.
77
214732
3214
還有其他和健康無關的特徵。
03:38
What the company does is ____?
78
218599
2554
這間公司所做的事是__?
03:41
on a scale from zero to 10,
79
221177
1631
用零分到十分代表可接受的程度,
03:42
completely acceptable to completely unacceptable,
80
222832
2385
從完全不可接受的零分,
03:45
zero to 10 completely acceptable in your confidence.
81
225241
2432
到你十足相信可以接受的十分。
03:47
MS: Now for the results.
82
227697
1591
馬:現在宣佈結果。
03:49
We found once again that when one person is convinced
83
229312
3123
我們再次發現,
當一個人深信該行為是完全錯的,
03:52
that the behavior is completely wrong,
84
232459
1811
03:54
someone sitting nearby firmly believes that it's completely right.
85
234294
3423
會有坐在附近的人 堅信該行為是完全對的。
03:57
This is how diverse we humans are when it comes to morality.
86
237741
3711
我們人類在道德上是這麼地分歧。
04:01
But within this broad diversity we found a trend.
87
241476
2713
但在這麼廣的多樣性中, 我們找到了一個趨勢。
04:04
The majority of the people at TED thought that it was acceptable
88
244213
3079
在 TED 的多數人認為,
忽視人工智慧的感受 並將之關機是可以接受的,
04:07
to ignore the feelings of the AI and shut it down,
89
247316
2755
04:10
and that it is wrong to play with our genes
90
250095
2513
但玩弄我們的基因
04:12
to select for cosmetic changes that aren't related to health.
91
252632
3320
來選擇與健康無關的 表面改變,則是錯的。
04:16
Then we asked everyone to gather into groups of three.
92
256402
2974
然後,我們請大家分組,三人一組。
04:19
And they were given two minutes to debate
93
259400
2037
他們有兩分鐘可辯論,
04:21
and try to come to a consensus.
94
261461
2294
並試著達成共識。
04:24
(MS) Two minutes to debate.
95
264838
1574
(馬)兩分鐘做辯論。
04:26
I'll tell you when it's time with the gong.
96
266436
2119
時間到時我會用鑼聲告訴大家。
04:28
(Audience debates)
97
268579
2640
(觀眾辯論)
04:35
(Gong sound)
98
275229
1993
(鑼聲)
04:38
(DA) OK.
99
278834
1151
(丹)好了。
04:40
(MS) It's time to stop.
100
280009
1792
(馬)時間到了,請停止。
04:41
People, people --
101
281825
1311
各位請注意。
04:43
MS: And we found that many groups reached a consensus
102
283747
2673
馬:我們發現 有許多小組達成了共識,
04:46
even when they were composed of people with completely opposite views.
103
286444
3929
即使小組內的成員 有完全相反的觀點。
04:50
What distinguished the groups that reached a consensus
104
290843
2524
有共識與沒有共識 小組之間的差異是什麼?
04:53
from those that didn't?
105
293391
1338
04:55
Typically, people that have extreme opinions
106
295244
2839
通常,意見很極端的人
會對他們的答案比較有信心。
04:58
are more confident in their answers.
107
298107
1840
05:00
Instead, those who respond closer to the middle
108
300868
2686
而回應傾向中間值的那些人,
05:03
are often unsure of whether something is right or wrong,
109
303578
3437
通常不太確定答案是對或錯,
05:07
so their confidence level is lower.
110
307039
2128
所以他們的信心比較低。
05:09
However, there is another set of people
111
309505
2943
然而,還有另一些人,
05:12
who are very confident in answering somewhere in the middle.
112
312472
3618
他們非常有信心做出 接近中間值的答案。
05:16
We think these high-confident grays are folks who understand
113
316657
3716
我們認為,這些高信心灰點代表的
是那些明白兩種答案都有優點的人。
05:20
that both arguments have merit.
114
320397
1612
05:22
They're gray not because they're unsure,
115
322531
2699
他們並非因為自身不確定而呈灰色,
05:25
but because they believe that the moral dilemma faces
116
325254
2688
而是因為他們相信
面對道德兩難的 兩種對立論點都有根據。
05:27
two valid, opposing arguments.
117
327966
1987
05:30
And we discovered that the groups that include highly confident grays
118
330373
4072
我們發現,
有高信心灰點成員的小組 更有可能會達成共識。
05:34
are much more likely to reach consensus.
119
334469
2493
05:36
We do not know yet exactly why this is.
120
336986
2478
我們還不知道確切的原因。
05:39
These are only the first experiments,
121
339488
1763
這些只是最初的實驗,
05:41
and many more will be needed to understand why and how
122
341275
3412
還需要做更多實驗來了解
為何有些人決定要協商道德立場,
05:44
some people decide to negotiate their moral standings
123
344711
2822
以及他們如何做,來達成共識。
05:47
to reach an agreement.
124
347557
1522
05:49
Now, when groups reach consensus,
125
349103
2469
若小組能達成共識,
05:51
how do they do so?
126
351596
1586
他們是怎麼做的?
05:53
The most intuitive idea is that it's just the average
127
353206
2581
最直覺的想法是拿每個人的答案
05:55
of all the answers in the group, right?
128
355811
2030
來算出平均值,對吧?
05:57
Another option is that the group weighs the strength of each vote
129
357865
3573
另一個做法是把每個人的答案
06:01
based on the confidence of the person expressing it.
130
361462
2448
再根據作答者的信心度來做加權。
06:04
Imagine Paul McCartney is a member of your group.
131
364422
2506
想像一下保羅麥卡尼在你那一組。
06:07
You'd be wise to follow his call
132
367352
2144
對於「昨天」出現次數的答案
06:09
on the number of times "Yesterday" is repeated,
133
369520
2441
相信他會是聰明的選擇。
06:11
which, by the way -- I think it's nine.
134
371985
2714
順道一提,我想應該是九次。
06:14
But instead, we found that consistently,
135
374723
2381
但,我們卻有個一致的發現,
06:17
in all dilemmas, in different experiments --
136
377128
2366
任何兩難問題,在不同的實驗中,
06:19
even on different continents --
137
379518
2165
甚至在不同大陸做的實驗中,
06:21
groups implement a smart and statistically sound procedure
138
381707
3743
小組會採用一種聰明 且有統計根據的程序,
06:25
known as the "robust average."
139
385474
2178
就是所謂的「穩健平均值」。
06:27
In the case of the height of the Eiffel Tower,
140
387676
2180
就艾菲爾鐵塔高度的例子來說,
06:29
let's say a group has these answers:
141
389880
1820
假設小組成員的答案包括:
06:31
250 meters, 200 meters, 300 meters, 400
142
391724
4608
250 公尺、200 公尺、 300 公尺、400 公尺,
06:36
and one totally absurd answer of 300 million meters.
143
396356
3784
還有一個答案是荒唐的 3 億公尺。
06:40
A simple average of these numbers would inaccurately skew the results.
144
400547
4293
所有答案的簡單平均值 會讓結果有不正確的偏差。
06:44
But the robust average is one where the group largely ignores
145
404864
3170
但穩健平均值就是小組會忽略
06:48
that absurd answer,
146
408058
1240
那荒唐的答案,
06:49
by giving much more weight to the vote of the people in the middle.
147
409322
3369
給與中間答案相對高很多的權重。
06:53
Back to the experiment in Vancouver,
148
413305
1876
回到溫哥華的實驗,
06:55
that's exactly what happened.
149
415205
1767
那裡發生的就是這種狀況。
06:57
Groups gave much less weight to the outliers,
150
417407
2741
小組會把離群值的權重降到很低,
07:00
and instead, the consensus turned out to be a robust average
151
420172
3229
而他們的共識就會是個人答案的
07:03
of the individual answers.
152
423425
1476
穩健平均值。
07:05
The most remarkable thing
153
425356
1991
最了不起的是,
07:07
is that this was a spontaneous behavior of the group.
154
427371
3187
這是小組的自發行為。
07:10
It happened without us giving them any hint on how to reach consensus.
155
430582
4475
我們並沒有暗示他們 要如何達成共識。
07:15
So where do we go from here?
156
435513
1540
所以我們的下一步是什麼?
07:17
This is only the beginning, but we already have some insights.
157
437432
3137
這只是開端,但我們 已經有了一些洞見。
07:20
Good collective decisions require two components:
158
440984
2917
好的集體決策需要兩個要件:
07:23
deliberation and diversity of opinions.
159
443925
2749
深思熟慮和多樣性的意見。
07:27
Right now, the way we typically make our voice heard in many societies
160
447066
3996
此時,要讓我們的聲音 在許多社會中被聽見,
07:31
is through direct or indirect voting.
161
451086
1908
做法就是直接或間接的投票。
07:33
This is good for diversity of opinions,
162
453495
1997
這對於意見的多樣性有益,
07:35
and it has the great virtue of ensuring
163
455516
2445
還有個優點:確保 每個人都能表達心聲。
07:37
that everyone gets to express their voice.
164
457985
2455
07:40
But it's not so good [for fostering] thoughtful debates.
165
460464
3735
但不足以促成考慮周到的辯論。
07:44
Our experiments suggest a different method
166
464665
3068
我們的實驗建議另一個不同的方式,
07:47
that may be effective in balancing these two goals at the same time,
167
467757
3541
或許能同時有效平衡這兩個目標,
07:51
by forming small groups that converge to a single decision
168
471322
3753
形成小組,由小組做出單一決定,
07:55
while still maintaining diversity of opinions
169
475099
2234
同時還能維持意見的多樣性,
07:57
because there are many independent groups.
170
477357
2773
因為有許多獨立的小組。
08:00
Of course, it's much easier to agree on the height of the Eiffel Tower
171
480741
3924
當然,要對艾菲爾鐵塔的高度取得共識
08:04
than on moral, political and ideological issues.
172
484689
3115
比道德、政治、 意識形態的議題容易多了。
08:08
But in a time when the world's problems are more complex
173
488721
3277
但在世界上的問題更複雜,
人們更兩極化的時候,
08:12
and people are more polarized,
174
492022
1803
08:13
using science to help us understand how we interact and make decisions
175
493849
4595
希望用科學來協助我們了解 我們如何互動並做決策,
08:18
will hopefully spark interesting new ways to construct a better democracy.
176
498468
4666
能夠激發出有趣的新方式, 來建立更好的民主。
關於本網站

本網站將向您介紹對學習英語有用的 YouTube 視頻。 您將看到來自世界各地的一流教師教授的英語課程。 雙擊每個視頻頁面上顯示的英文字幕,從那裡播放視頻。 字幕與視頻播放同步滾動。 如果您有任何意見或要求,請使用此聯繫表與我們聯繫。

https://forms.gle/WvT1wiN1qDtmnspy7


This website was created in October 2020 and last updated on June 12, 2025.

It is now archived and preserved as an English learning resource.

Some information may be out of date.

隱私政策

eng.lish.video

Developer's Blog