The pattern behind self-deception | Michael Shermer

729,912 views ・ 2010-06-14

TED


Please double-click on the English subtitles below to play the video.

Prevodilac: Tamara Jovanović Lektor: Jelena Nedjic
00:16
So since I was here last in '06,
0
16260
3000
Od kada sam poslednji put bio ovde 2006.,
00:19
we discovered that global climate change
1
19260
2000
otkrili smo da je promena globalne klime
00:21
is turning out to be a pretty serious issue,
2
21260
2000
prerasla u prilično ozbiljan problem,
00:23
so we covered that fairly extensively
3
23260
2000
tako da smo pisali opsežno o tome
00:25
in Skeptic magazine.
4
25260
2000
u časopisu "Skeptic".
00:27
We investigate all kinds
5
27260
2000
Istražujemo sve tipove
00:29
of scientific and quasi-scientific controversies,
6
29260
3000
naučnih i kvazi-naučnih kontroverzi,
00:32
but it turns out we don't have to worry about any of this
7
32260
2000
ali ispostavilo se da ne moramo da brinemo ni oko čega,
00:34
because the world's going to end in 2012.
8
34260
2000
jer će 2012. godine biti smak sveta.
00:36
Another update:
9
36260
2000
Još jedno obaveštenje:
00:38
You will recall I introduced you guys
10
38260
2000
sećate se da sam vam predstavio
00:40
to the Quadro Tracker.
11
40260
2000
"Quadro Tracker".
00:42
It's like a water dowsing device.
12
42260
2000
Podseća na rašlje za traženje vode.
00:44
It's just a hollow piece of plastic with an antenna that swivels around.
13
44260
3000
To je samo prazno parče plastike sa antenom koja se okreće.
00:47
And you walk around, and it points to things.
14
47260
2000
Hodate naokolo i usmeravate je ka stvarima.
00:49
Like if you're looking for marijuana in students' lockers,
15
49260
3000
Kao da tražite marihuanu u ormarićima učenika,
00:52
it'll point right to somebody.
16
52260
2000
pokazaće pravo na nekog.
00:54
Oh, sorry. (Laughter)
17
54260
2000
O, izvinite. (Smeh)
00:56
This particular one that was given to me
18
56260
2000
Konkretno ovaj koji mi je poklonjen,
00:58
finds golf balls,
19
58260
2000
pronalazi loptice za golf,
01:00
especially if you're at a golf course
20
60260
2000
naročito ako ste na terenu za golf
01:02
and you check under enough bushes.
21
62260
3000
i proverite ispod dovoljno žbunova.
01:05
Well, under the category of "What's the harm of silly stuff like this?"
22
65260
3000
Pa, u kategoriju "Kakvu štetu mogu naneti ove blesave stvarčice?"
01:08
this device, the ADE 651,
23
68260
3000
spada ovaj uređaj, ADE 651,
01:11
was sold to the Iraqi government
24
71260
3000
koji je prodat iračkoj vladi
01:14
for 40,000 dollars apiece.
25
74260
2000
za 40 000 dolara po uređaju.
01:16
It's just like this one, completely worthless,
26
76260
2000
Poput ovog ovde je, potpuno beskoristan,
01:18
in which it allegedly worked by "electrostatic
27
78260
2000
navodno ga pokreće "elektrostatična
01:20
magnetic ion attraction,"
28
80260
3000
magnetna jonska privlačnost",
01:24
which translates to
29
84260
2000
što znači
01:26
"pseudoscientific baloney" -- would be the nice word --
30
86260
3000
"pseudo-naučna budalaština" -- bila bi to odgovarajuća reč --
01:29
in which you string together a bunch of words that sound good,
31
89260
2000
natrpate u nazivu gomilu zvučnih reči,
01:31
but it does absolutely nothing.
32
91260
2000
ali sam uređaj nema nikakvu funkciju.
01:33
In this case, at trespass points,
33
93260
3000
U ovom slučaju, na tačkama prelaska,
01:36
allowing people to go through
34
96260
2000
omogućava ljudima da prođu,
01:38
because your little tracker device said they were okay,
35
98260
3000
jer je vaš mali uređaj za praćenje rekao da su okej,
01:41
actually cost lives.
36
101260
3000
što je zapravo koštalo života.
01:44
So there is a danger to pseudoscience,
37
104260
2000
Tako da postoji opasnost u pseudo-nauci,
01:46
in believing in this sort of thing.
38
106260
3000
u verovanju u ovakve stvari.
01:49
So what I want to talk about today is belief.
39
109260
3000
Ono o čemo želim danas da govorim je verovanje.
01:52
I want to believe,
40
112260
2000
Želim da verujem,
01:54
and you do too.
41
114260
2000
kao i vi.
01:56
And in fact, I think my thesis here is that
42
116260
2000
Zapravo, ja stojim iza stava
01:58
belief is the natural state of things.
43
118260
2000
da je verovanje prirodno stanje stvari.
02:00
It is the default option. We just believe.
44
120260
2000
To se podrazumeva. Jednostavno verujemo.
02:02
We believe all sorts of things.
45
122260
2000
Verujemo u svakakve stvari.
02:04
Belief is natural;
46
124260
2000
Verovanje je prirodno;
02:06
disbelief, skepticism, science, is not natural.
47
126260
2000
neverovanje, skepticizam, nauka, nisu prirodni.
02:08
It's more difficult.
48
128260
2000
Teži su.
02:10
It's uncomfortable to not believe things.
49
130260
2000
Neugodnije je ne verovati u stvari.
02:12
So like Fox Mulder on "X-Files,"
50
132260
3000
Recimo, u stilu Foksa Moldera (Fox Mulder) u "Dosijeu Iks",
02:15
who wants to believe in UFOs? Well, we all do,
51
135260
3000
ko želi da veruje u NLO-e? Pa, svi mi to želimo,
02:18
and the reason for that is because
52
138260
2000
a razlog je to što
02:20
we have a belief engine in our brains.
53
140260
3000
imamo mehanizam za verovanje u mozgu.
02:23
Essentially, we are pattern-seeking primates.
54
143260
3000
U suštini, mi smo primati koji traže šeme.
02:26
We connect the dots: A is connected to B; B is connected to C.
55
146260
3000
Povezujemo tačke: A je povezano sa B; B je povezano sa C.
02:29
And sometimes A really is connected to B,
56
149260
3000
Ponekad je A zaista povezano sa B,
02:32
and that's called association learning.
57
152260
2000
i to se zove asocijativno učenje.
02:34
We find patterns, we make those connections,
58
154260
3000
Nalazimo šeme, uspostavljamo veze,
02:37
whether it's Pavlov's dog here
59
157260
2000
bez obzira da li je to Pavlovljev pas
02:39
associating the sound of the bell with the food,
60
159260
3000
koji povezuje zvuk zvona sa hranom
02:42
and then he salivates to the sound of the bell,
61
162260
2000
i onda luči pljuvačku na zvuk zvona,
02:44
or whether it's a Skinnerian rat,
62
164260
2000
ili Skinerov pacov,
02:46
in which he's having an association
63
166260
2000
koji je utvrdio vezu
02:48
between his behavior and a reward for it,
64
168260
2000
između svog ponašanja i nagrade,
02:50
and therefore he repeats the behavior.
65
170260
2000
pa zato ponavlja ponašanje.
02:52
In fact, what Skinner discovered
66
172260
2000
Zapravo, ono što je Skiner otkrio
02:54
is that, if you put a pigeon in a box like this,
67
174260
3000
je da, ako stavite goluba u ovakvu kutiju
02:57
and he has to press one of these two keys,
68
177260
2000
i on mora da pritisne jedan od ova dva tastera,
02:59
and he tries to figure out what the pattern is,
69
179260
2000
pokušava da otkrije šemu
03:01
and you give him a little reward in the hopper box there --
70
181260
2000
za dobijanje male nagrade u toj kutiji --
03:03
if you just randomly assign rewards
71
183260
3000
a vi samo nasumično dodeljujete nagrade,
03:06
such that there is no pattern,
72
186260
2000
tako da ne postoji šema,
03:08
they will figure out any kind of pattern.
73
188260
2000
otkriće bilo kakvu šemu.
03:10
And whatever they were doing just before they got the reward,
74
190260
2000
Šta god da su radili pre dobijanja nagrade,
03:12
they repeat that particular pattern.
75
192260
2000
ponavljali bi tu određenu šemu.
03:14
Sometimes it was even spinning around twice counterclockwise,
76
194260
3000
Ponekad se golub okretao dvaput u smeru suprotnom od kazaljke na satu,
03:17
once clockwise and peck the key twice.
77
197260
3000
jednom u smeru kazaljke i kljucnuo taster dvaput.
03:20
And that's called superstition,
78
200260
2000
To se zove sujeverje,
03:22
and that, I'm afraid,
79
202260
2000
to je, bojim se,
03:24
we will always have with us.
80
204260
2000
nešto što ćemo uvek imati u sebi.
03:26
I call this process "patternicity" --
81
206260
2000
Zovem ovaj proces "stvaranje šema" --
03:28
that is, the tendency to find meaningful patterns
82
208260
2000
to jest, skolonost ka pronalaženju smislenih obrazaca
03:30
in both meaningful and meaningless noise.
83
210260
3000
kako u logičnom, tako i u nelogičnom skupu.
03:33
When we do this process, we make two types of errors.
84
213260
3000
U toku ovog procesa, pravimo dve vrste grešaka.
03:36
A Type I error, or false positive,
85
216260
2000
Greška prvog tipa ili lažni pozitiv
03:38
is believing a pattern is real
86
218260
2000
je verovanje da je šema stvarna,
03:40
when it's not.
87
220260
2000
kada zapravo nije.
03:42
Our second type of error is a false negative.
88
222260
2000
Drugi tip greške je lažni negativ.
03:44
A Type II error is not believing
89
224260
2000
Greška drugog tipa je verovanje
03:46
a pattern is real when it is.
90
226260
3000
da obrazac nije stvaran, kada to jeste.
03:49
So let's do a thought experiment.
91
229260
2000
Hajde da napravimo misaoni eksperiment.
03:51
You are a hominid three million years ago
92
231260
2000
Vi ste hominid od pre tri miliona godina,
03:53
walking on the plains of Africa.
93
233260
3000
koji šeta afričkom ravnicom.
03:56
Your name is Lucy, okay?
94
236260
2000
Zovete se Lusi, okej?
03:58
And you hear a rustle in the grass.
95
238260
2000
Čujete šuškanje u travi.
04:00
Is it a dangerous predator,
96
240260
2000
Da li je to opasan grabljivac
04:02
or is it just the wind?
97
242260
2000
ili samo vetar?
04:04
Your next decision could be the most important one of your life.
98
244260
3000
Sledeća odluka koju napravite može biti najznačajnija u životu.
04:07
Well, if you think that the rustle in the grass is a dangerous predator
99
247260
3000
Ako mislite da je šuškanje poteklo od grabljivca,
04:10
and it turns out it's just the wind,
100
250260
2000
a ispostavi se da je samo vetar,
04:12
you've made an error in cognition,
101
252260
2000
napravili ste kognitivnu grešku,
04:14
made a Type I error, false positive.
102
254260
2000
grešku prvog tipa, lažni pozitiv.
04:16
But no harm. You just move away.
103
256260
2000
Ali sve je u redu. Nastavljate dalje.
04:18
You're more cautious. You're more vigilant.
104
258260
2000
Obazriviji ste. Budniji.
04:20
On the other hand, if you believe that the rustle in the grass is just the wind,
105
260260
2000
Sa druge strane, ako verujete da je to šuškanje samo vetar,
04:22
and it turns out it's a dangerous predator,
106
262260
3000
a ispostavi se da je grabljivac,
04:25
you're lunch.
107
265260
2000
postaćete ručak.
04:27
You've just won a Darwin award.
108
267260
2000
Upravo ste osvojili Darvinovu nagradu.
04:29
You've been taken out of the gene pool.
109
269260
2000
Uklonjeni ste iz genetskog fonda.
04:31
Now the problem here is that
110
271260
2000
Problem je u tome što
04:33
patternicities will occur whenever the cost
111
273260
2000
se do stvaranja šema dolazi svaki put
04:35
of making a Type I error
112
275260
2000
kada je šteta od greške prvog tipa
04:37
is less than the cost of making a Type II error.
113
277260
2000
manja od pravljenja greške drugog tipa.
04:39
This is the only equation in the talk by the way.
114
279260
2000
Uzged, ovo je jedina jednačina u ovom govoru.
04:41
We have a pattern detection problem
115
281260
2000
Imamo problem u opažanju šema,
04:43
that is assessing the difference between a Type I and a Type II error
116
283260
3000
to jest, procenjivanje razlike između greške prvog i drugog tipa
04:46
is highly problematic,
117
286260
2000
je veoma problematično,
04:48
especially in split-second, life-and-death situations.
118
288260
3000
naročito u situacijama kada je život ugrožen, koje traju delić sekunde.
04:51
So the default position
119
291260
2000
Zato, podrazumevani stav
04:53
is just: Believe all patterns are real --
120
293260
2000
je jednostavno: Verujte da su svi obrasci stvarni --
04:55
All rustles in the grass are dangerous predators
121
295260
3000
Svako šuškanje u travi je opasni grabljivac,
04:58
and not just the wind.
122
298260
2000
a ne samo vetar.
05:00
And so I think that we evolved ...
123
300260
2000
Zato mislim da smo evoluirali...
05:02
there was a natural selection for the propensity for our belief engines,
124
302260
3000
postojala je prirodna selekcija za razvijanje mehanizama verovanja,
05:05
our pattern-seeking brain processes,
125
305260
2000
za moždanim procesima koji omogućavaju traženje šema,
05:07
to always find meaningful patterns
126
307260
2000
da uvek pronađemo smislene šeme
05:09
and infuse them with these sort of
127
309260
2000
i pripišemo im ova
05:11
predatory or intentional agencies that I'll come back to.
128
311260
3000
predatorska ili namerna svojstva o kojima ću pričati kasnije.
05:14
So for example, what do you see here?
129
314260
2000
Pa, na primer, šta vidite ovde?
05:16
It's a horse head, that's right.
130
316260
2000
To je glava konja, tako je.
05:18
It looks like a horse. It must be a horse.
131
318260
2000
Izgleda kao konj. Mora da je konj.
05:20
That's a pattern.
132
320260
2000
To je šema.
05:22
And is it really a horse?
133
322260
2000
A da li je to stvarno konj?
05:24
Or is it more like a frog?
134
324260
3000
Ili je više nalik žabi?
05:27
See, our pattern detection device,
135
327260
2000
Vidite, naš uređaj za traženje šema,
05:29
which appears to be located in the anterior cingulate cortex --
136
329260
3000
koji je izgleda smešten u prednjem limbusu --
05:32
it's our little detection device there --
137
332260
3000
to je naš mali uređaj za opažanje --
05:35
can be easily fooled, and this is the problem.
138
335260
2000
može se lako zavarati i to je problem.
05:37
For example, what do you see here?
139
337260
2000
Na primer, šta vidite ovde?
05:39
Yes, of course, it's a cow.
140
339260
3000
Da, naravno, to je krava.
05:42
Once I prime the brain -- it's called cognitive priming --
141
342260
3000
Kada pripremim mozak -- to se zove kognitivna priprema --
05:45
once I prime the brain to see it,
142
345260
2000
kada pripremim mozak da to vidi,
05:47
it pops back out again even without the pattern that I've imposed on it.
143
347260
3000
slika se opet vraća, čak i bez šeme koju sam joj namenio.
05:50
And what do you see here?
144
350260
2000
A šta vidite ovde?
05:52
Some people see a Dalmatian dog.
145
352260
2000
Neki ljudi vide dalmatinca.
05:54
Yes, there it is. And there's the prime.
146
354260
2000
Da, evo ga. Evo i pripreme za sliku.
05:56
So when I go back without the prime,
147
356260
2000
Kada se vratim bez pripreme,
05:58
your brain already has the model
148
358260
2000
vaš mozak već ima model,
06:00
so you can see it again.
149
360260
2000
pa ga možete videti opet.
06:02
What do you see here?
150
362260
3000
Šta vidite ovde?
06:05
Planet Saturn. Yes, that's good.
151
365260
2000
Planetu Saturn. Da, to je dobro.
06:07
How about here?
152
367260
3000
A ovde?
06:10
Just shout out anything you see.
153
370260
3000
Samo uzviknite šta god da vidite.
06:14
That's a good audience, Chris.
154
374260
2000
To je sjajna publika, Kris.
06:16
Because there's nothing in this. Well, allegedly there's nothing.
155
376260
3000
Jer ovde nema ničeg. Pa, navodno, nema ničeg.
06:19
This is an experiment done by Jennifer Whitson
156
379260
3000
Ovo je eksperiment Dženifer Vitson (Jennifer Whitson)
06:22
at U.T. Austin
157
382260
2000
sa Teksaškog Univerziteta u Ostinu,
06:24
on corporate environments
158
384260
2000
o korporativnom okruženju
06:26
and whether feelings of uncertainty and out of control
159
386260
3000
i povezanosti osećanja nesigurnosti i manjka kontrole
06:29
makes people see illusory patterns.
160
389260
2000
sa stvaranjem imaginarnih šema.
06:31
That is, almost everybody sees the planet Saturn.
161
391260
3000
Zapravo, skoro svako vidi Saturn.
06:34
People that are put in a condition of feeling out of control
162
394260
3000
Ljudi koji su u uslovima kada osećaju manjak kontrole
06:37
are more likely to see something in this,
163
397260
2000
imaju veće šanse da vide nešto u ovome,
06:39
which is allegedly patternless.
164
399260
3000
gde navodno nema šema.
06:42
In other words, the propensity to find these patterns
165
402260
3000
Drugim rečima, sklonost ka nalaženju ovih šema
06:45
goes up when there's a lack of control.
166
405260
3000
povećava se kada se gubi kontrola.
06:48
For example, baseball players are notoriously superstitious
167
408260
3000
Na primer, igrači bejzbola su čuveni po svom sujeverju
06:51
when they're batting,
168
411260
2000
kada udaraju lopticu,
06:53
but not so much when they're fielding.
169
413260
2000
ali ne toliko kada igraju odbranu.
06:55
Because fielders are successful
170
415260
2000
Zato što su igrači odbrane uspešni
06:57
90 to 95 percent of the time.
171
417260
2000
90 do 95 odsto vremena.
06:59
The best batters fail seven out of 10 times.
172
419260
3000
Najbolji udarači uspevaju 3 od 10 puta.
07:02
So their superstitions, their patternicities,
173
422260
2000
Tako da je njihovo sujeverje, njihovo stvaranje šema,
07:04
are all associated with feelings of lack of control
174
424260
3000
povezano sa osećanjem nedostatka kontrole
07:07
and so forth.
175
427260
2000
i tako dalje.
07:10
What do you see in this particular one here, in this field?
176
430260
3000
Šta vidite baš ovde, u ovom polju?
07:13
Anybody see an object there?
177
433260
2000
Da li neko ovde vidi predmet?
07:15
There actually is something here,
178
435260
2000
Zapravo, ima nečeg ovde,
07:17
but it's degraded.
179
437260
2000
ali je slika uništena.
07:19
While you're thinking about that,
180
439260
2000
Dok razmišljate o tome,
07:21
this was an experiment done by Susan Blackmore,
181
441260
2000
ovo je eksperiment Suzan Blekmur (Susan Blackmore),
07:23
a psychologist in England,
182
443260
2000
psihologa iz Engleske,
07:25
who showed subjects this degraded image
183
445260
2000
koja je učesnicima eksperimenta pokazivala ovu uništenu sliku
07:27
and then ran a correlation between
184
447260
2000
i potom određivala korelaciju između
07:29
their scores on an ESP test:
185
449260
2000
njihovih rezultata na ESP testu:
07:31
How much did they believe in the paranormal,
186
451260
2000
Koliko veruju u paranormalno,
07:33
supernatural, angels and so forth.
187
453260
3000
natprirodno, anđele i tako dalje.
07:36
And those who scored high on the ESP scale,
188
456260
3000
Oni koji su imali najbolje rezultate na ESP testu
07:39
tended to not only see
189
459260
2000
ne samo da su uočavali
07:41
more patterns in the degraded images
190
461260
2000
više šema u uništenim slikama,
07:43
but incorrect patterns.
191
463260
2000
već i netačne šeme.
07:45
Here is what you show subjects.
192
465260
2000
Ovo se pokazuje učesnicima eksperimenta.
07:47
The fish is degraded 20 percent, 50 percent
193
467260
3000
Ova riba je uništena 20%, 50%,
07:50
and then the one I showed you,
194
470260
2000
a ova koju sam vam pokazao,
07:52
70 percent.
195
472260
2000
70 odsto.
07:54
A similar experiment was done by another [Swiss] psychologist
196
474260
2000
Sličan ogled je sproveo još jedan britanski psiholog,
07:56
named Peter Brugger,
197
476260
2000
Piter Bruger (Peter Brugger),
07:58
who found significantly more meaningful patterns
198
478260
3000
koji je pronašao da je znatno više smislenih šema
08:01
were perceived on the right hemisphere,
199
481260
2000
opaženo u desnoj hemisferi,
08:03
via the left visual field, than the left hemisphere.
200
483260
3000
putem levog vidnog polja, nego u levoj hemisferi.
08:06
So if you present subjects the images such
201
486260
2000
Tako da ako učesnicima pokažete takve slike
08:08
that it's going to end up on the right hemisphere instead of the left,
202
488260
3000
koje će završiti u desnoj hemisferi, umesto u levoj,
08:11
then they're more likely to see patterns
203
491260
2000
veća je verovatnoća da će se opaziti šeme,
08:13
than if you put it on the left hemisphere.
204
493260
2000
nego kada se obrada vrši u levoj hemisferi.
08:15
Our right hemisphere appears to be
205
495260
2000
Čini se da se u našoj desnoj hemisferi
08:17
where a lot of this patternicity occurs.
206
497260
2000
stvara najviše šema.
08:19
So what we're trying to do is bore into the brain
207
499260
2000
Ono što pokušavamo da uradimo je da uđemo duboko u mozak,
08:21
to see where all this happens.
208
501260
2000
da bismo videli gde se sve to dešava.
08:23
Brugger and his colleague, Christine Mohr,
209
503260
3000
Bruger i njegova koleginica, Kristine Mor (Christine Mohr),
08:26
gave subjects L-DOPA.
210
506260
2000
su davali učesnicima L-DOPA.
08:28
L-DOPA's a drug, as you know, given for treating Parkinson's disease,
211
508260
3000
L-DOPA se koristi, kao što znate, za lečenje Parkinsonove bolesti,
08:31
which is related to a decrease in dopamine.
212
511260
3000
koja je povezana sa padom u nivou dopamina.
08:34
L-DOPA increases dopamine.
213
514260
2000
L-DOPA povećava nivo dopamina.
08:36
An increase of dopamine caused
214
516260
2000
Rast nivoa dopamina izazvao je
08:38
subjects to see more patterns
215
518260
2000
uočavanje većeg broja šema,
08:40
than those that did not receive the dopamine.
216
520260
2000
u odnosu na tretman bez dopamina.
08:42
So dopamine appears to be the drug
217
522260
2000
Čini se da je dopamin jedinjenje
08:44
associated with patternicity.
218
524260
2000
koje je povezano sa stvaranjem šema.
08:46
In fact, neuroleptic drugs
219
526260
2000
Zapravo, antipsihotici,
08:48
that are used to eliminate psychotic behavior,
220
528260
2000
koji se koriste za lečenje psihoza,
08:50
things like paranoia, delusions
221
530260
2000
poput paranoje, diluzija
08:52
and hallucinations,
222
532260
2000
i halucinacija,
08:54
these are patternicities.
223
534260
2000
ovo su poremećaji čestog stvaranja šema.
08:56
They're incorrect patterns. They're false positives. They're Type I errors.
224
536260
3000
To su netačne šeme. To su lažni pozitivi. To su greške prvog tipa.
08:59
And if you give them drugs
225
539260
2000
Ako im date lakove,
09:01
that are dopamine antagonists,
226
541260
2000
koji su dopaminski blokatori,
09:03
they go away.
227
543260
2000
one nestaju.
09:05
That is, you decrease the amount of dopamine,
228
545260
2000
Zapravo, smanjujete nivo dopamina
09:07
and their tendency to see
229
547260
2000
i njihova sklonost da zapaze
09:09
patterns like that decreases.
230
549260
2000
takve šeme se smanjuje.
09:11
On the other hand, amphetamines like cocaine
231
551260
3000
Sa druge strane, amfetamini poput kokaina
09:14
are dopamine agonists.
232
554260
2000
su stimulansi dopamina.
09:16
They increase the amount of dopamine.
233
556260
2000
Povećavaju nivo dopamina.
09:18
So you're more likely to feel in a euphoric state,
234
558260
3000
Samim tim je veća verovatnoća da ćete osećati euforiju,
09:21
creativity, find more patterns.
235
561260
2000
biti kreativniji, naći više šema.
09:23
In fact, I saw Robin Williams recently
236
563260
2000
Ustvari, nedavno sam slušao Robina Vilijamsa (Robin Williams)
09:25
talk about how he thought he was much funnier
237
565260
2000
kako govori o tome da misli da je bio mnogo duhovitiji
09:27
when he was doing cocaine, when he had that issue, than now.
238
567260
3000
kada je konzumirao kokain, kada je imao taj problem, nego sada.
09:30
So perhaps more dopamine
239
570260
2000
Verovatno dopamin
09:32
is related to more creativity.
240
572260
2000
utiče na kreativnost.
09:34
Dopamine, I think, changes
241
574260
2000
Mislim da dopamin menja
09:36
our signal-to-noise ratio.
242
576260
2000
našu percepciju odnosa signal/šum.
09:38
That is, how accurate we are
243
578260
2000
To jest, našu preciznost
09:40
in finding patterns.
244
580260
2000
u nalaženju šema.
09:42
If it's too low, you're more likely to make too many Type II errors.
245
582260
3000
Ako ga ima malo, verovatno ćete načiniti više grešaka drugog tipa.
09:45
You miss the real patterns. You don't want to be too skeptical.
246
585260
2000
Ne vidite prave šeme. Ne želite da budete isuviše skeptični.
09:47
If you're too skeptical, you'll miss the really interesting good ideas.
247
587260
3000
Ako ste isuviše skeptični, propustićete baš interesantne,dobre ideje.
09:51
Just right, you're creative, and yet you don't fall for too much baloney.
248
591260
3000
Normalan nivo dopamina, kreativni ste, a opet, nećete poverovati u svašta.
09:54
Too high and maybe you see patterns everywhere.
249
594260
3000
Previsok nivo i možda ćete videti šeme svuda.
09:57
Every time somebody looks at you, you think people are staring at you.
250
597260
3000
Svaki put kada vas neko pogleda, mislite da ljudi zure u vas.
10:00
You think people are talking about you.
251
600260
2000
Mislite da ljudi govore o vama.
10:02
And if you go too far on that, that's just simply
252
602260
2000
Ako odete predaleko, to se jednostavno
10:04
labeled as madness.
253
604260
2000
naziva ludilom.
10:06
It's a distinction perhaps we might make
254
606260
2000
To je razlika koju možda možemo uočiti
10:08
between two Nobel laureates, Richard Feynman
255
608260
2000
između dvojice Nobelovaca, Ričarda Fejnmana (Richard Feynman)
10:10
and John Nash.
256
610260
2000
i Džona Neša (John Nash).
10:12
One sees maybe just the right number
257
612260
2000
Jedan možda zapaža baš pravi broj
10:14
of patterns to win a Nobel Prize.
258
614260
2000
šema potrebnih za osvajanje Nobelove nagrade.
10:16
The other one also, but maybe too many patterns.
259
616260
2000
Drugi isto tako, ali možda vidi isuviše šema.
10:18
And we then call that schizophrenia.
260
618260
3000
Mi to nazivamo šizofrenijom.
10:21
So the signal-to-noise ratio then presents us with a pattern-detection problem.
261
621260
3000
Zato nam odnos signal/šum skreće pažnju na problem opažanja šema.
10:24
And of course you all know exactly
262
624260
2000
Naravno, svi znate tačno
10:26
what this is, right?
263
626260
2000
šta je ovo, je l' tako?
10:28
And what pattern do you see here?
264
628260
2000
A koju šemu vidite ovde?
10:30
Again, I'm putting your anterior cingulate cortex to the test here,
265
630260
3000
Opet, testiram vaš prednji limbus,
10:33
causing you conflicting pattern detections.
266
633260
3000
izazivam konfliktna opažanja šema.
10:36
You know, of course, this is Via Uno shoes.
267
636260
2000
Znate, naravno, ovo su "Via Uno" cipele.
10:38
These are sandals.
268
638260
3000
Ovo su sandale.
10:41
Pretty sexy feet, I must say.
269
641260
3000
Prilično seksi stopala, moram priznati.
10:44
Maybe a little Photoshopped.
270
644260
2000
Možda malo fotošopirana.
10:46
And of course, the ambiguous figures
271
646260
2000
I naravno, nejasne figure
10:48
that seem to flip-flop back and forth.
272
648260
2000
koje se ređaju jedna za drugom.
10:50
It turns out what you're thinking about a lot
273
650260
2000
Ispostavlja se da ono o čemu mnogo razmišljate
10:52
influences what you
274
652260
2000
utiče na to
10:54
tend to see.
275
654260
2000
što vidite.
10:56
And you see the lamp here, I know.
276
656260
2000
Vidite lampu ovde, znam.
10:58
Because the lights on here.
277
658260
3000
Zbog svetla ovde.
11:01
Of course, thanks to the environmentalist movement
278
661260
2000
Naravno, zahvaljujući pokretu za zaštitu životne sredine,
11:03
we're all sensitive to the plight of marine mammals.
279
663260
3000
svi smo pogođeni patnjama morskih sisara.
11:06
So what you see in this particular ambiguous figure
280
666260
3000
Tako da je ono što vidite u ovom nejasnom obliku
11:09
is, of course, the dolphins, right?
281
669260
2000
naravno, delfin, zar ne?
11:11
You see a dolphin here,
282
671260
2000
Vidite delfina ovde,
11:13
and there's a dolphin,
283
673260
2000
i još jednog
11:15
and there's a dolphin.
284
675260
2000
i još jednog.
11:17
That's a dolphin tail there, guys.
285
677260
3000
Ljudi, ovo ovde je delfinov rep.
11:20
(Laughter)
286
680260
3000
(Smeh)
11:25
If we can give you conflicting data, again,
287
685260
3000
Ako vam damo protivrečne podatke, opet,
11:28
your ACC is going to be going into hyperdrive.
288
688260
3000
vaš limbus će raditi punom parom.
11:31
If you look down here, it's fine. If you look up here, then you get conflicting data.
289
691260
3000
Ako pogledate ovde dole, dobro je. Ako pogledate gore, dobijate protivrečne podatke.
11:34
And then we have to flip the image
290
694260
2000
Onda moramo da okrenemo sliku
11:36
for you to see that it's a set up.
291
696260
2000
da biste videli da je ovo nameštaljka.
11:40
The impossible crate illusion.
292
700260
2000
Iluzija nemogućeg sanduka.
11:42
It's easy to fool the brain in 2D.
293
702260
2000
Lako je zavarati mozak u dve dimenzije.
11:44
So you say, "Aw, come on Shermer, anybody can do that
294
704260
2000
Kažete: "Hej, Šermeru, svako može da stvori takvu iluziju
11:46
in a Psych 101 text with an illusion like that."
295
706260
2000
uz pomoć knjige o osnovama psihologije."
11:48
Well here's the late, great Jerry Andrus'
296
708260
2000
Evo nove, sjajne iluzije "nemogućeg sanduka" u 3D-u
11:50
"impossible crate" illusion in 3D,
297
710260
3000
Džerija Endrusa (Jerry Andrus),
11:53
in which Jerry is standing inside
298
713260
2000
u kojoj Džeri stoji
11:55
the impossible crate.
299
715260
2000
u nemogućem sanduku.
11:57
And he was kind enough to post this
300
717260
2000
Bio je dovoljno ljubazan da nam pošalje ovo
11:59
and give us the reveal.
301
719260
2000
i otkrije nam tajnu.
12:01
Of course, camera angle is everything. The photographer is over there,
302
721260
3000
Naravno, sve je u uglu kamere. Fotograf je ovamo
12:04
and this board appears to overlap with this one, and this one with that one, and so on.
303
724260
3000
i ova daska se naizgled preklapa sa onom, i ova sa onom, i tako dalje.
12:07
But even when I take it away,
304
727260
2000
Ali čak i kada to sklonim,
12:09
the illusion is so powerful because of how are brains are wired
305
729260
2000
iluzija je tako moćna zbog načina na koji je naš mozak programiran
12:11
to find those certain kinds of patterns.
306
731260
3000
da nalazi različite šeme.
12:14
This is a fairly new one
307
734260
2000
Ova je relativno nova,
12:16
that throws us off because of the conflicting patterns
308
736260
2000
koja nas potpuno iznenađuje zbog protivrečnih šemi
12:18
of comparing this angle with that angle.
309
738260
3000
poređenja jednog ugla sa drugim.
12:21
In fact, it's the exact same picture side by side.
310
741260
3000
Zapravo, to su potpuno iste slike jedna do druge.
12:24
So what you're doing is comparing that angle
311
744260
2000
Ono što radite je upoređivanje tog ugla,
12:26
instead of with this one, but with that one.
312
746260
2000
sa onim tamo, umesto sa ovim.
12:28
And so your brain is fooled.
313
748260
2000
Tako je vaš mozak zavaran.
12:30
Yet again, your pattern detection devices are fooled.
314
750260
2000
Opet, vaši uređaji za opažanje šema su zavarani.
12:32
Faces are easy to see
315
752260
2000
Lako je videti lica,
12:34
because we have an additional evolved
316
754260
2000
jer nam je evoluirao dodatni
12:36
facial recognition software
317
756260
2000
softver za prepoznavanje lica
12:38
in our temporal lobes.
318
758260
3000
u slepoočnim režnjevima.
12:41
Here's some faces on the side of a rock.
319
761260
3000
Evo nekih lica na padini stene.
12:44
I'm actually not even sure if this is -- this might be Photoshopped.
320
764260
3000
Nisam čak ni siguran oko ovog -- možda je fotošopirano.
12:47
But anyway, the point is still made.
321
767260
2000
Kako god, shvatate moju poentu.
12:49
Now which one of these looks odd to you?
322
769260
2000
Sad, koja od ovih slika vam izgleda čudno?
12:51
In a quick reaction, which one looks odd?
323
771260
2000
Brzo, koja izgleda čudno?
12:53
The one on the left. Okay. So I'll rotate it
324
773260
2000
Ona levo. Okej. Okrenuću je,
12:55
so it'll be the one on the right.
325
775260
2000
pa će biti sa desne strane.
12:57
And you are correct.
326
777260
2000
U pravu ste.
12:59
A fairly famous illusion -- it was first done with Margaret Thatcher.
327
779260
3000
Relativno poznata iluzija -- prvo je to urađeno sa Margaret Tačer (Margaret Thatcher).
13:02
Now, they trade up the politicians every time.
328
782260
2000
Sada, svaki put menjaju političare.
13:04
Well, why is this happening?
329
784260
2000
Pa, zašto se ovo dešava?
13:06
Well, we know exactly where it happens,
330
786260
2000
Znamo tačno gde se to dešava,
13:08
in the temporal lobe, right across, sort of above your ear there,
331
788260
3000
u slepoočnom režnju, odmah ovde, negde iznad uha,
13:11
in a little structure called the fusiform gyrus.
332
791260
3000
u maloj strukturi koja se zove "fusiform gyrus".
13:14
And there's two types of cells that do this,
333
794260
2000
Postoje dva tipa ćelija koje su zadužene za ovo,
13:16
that record facial features either globally,
334
796260
3000
koje pamte crte lica bilo u globalu
13:19
or specifically these large, rapid-firing cells,
335
799260
2000
ili posebno, ove velike ćelije koje brzo reaguju,
13:21
first look at the general face.
336
801260
2000
prvo gledaju lice u celosti.
13:23
So you recognize Obama immediately.
337
803260
2000
Tako odmah prepoznajete Obamu.
13:25
And then you notice something quite
338
805260
2000
Onda primećujete nešto prilično
13:27
a little bit odd about the eyes and the mouth.
339
807260
2000
čudno u vezi sa očima i usnama.
13:29
Especially when they're upside down,
340
809260
2000
Naročito kada su naopačke,
13:31
you're engaging that general facial recognition software there.
341
811260
3000
aktivirate taj softver za opšte prepoznavanje lica.
13:34
Now I said back in our little thought experiment,
342
814260
3000
Spomenuo sam ranije u našem malom misaonom ogledu,
13:37
you're a hominid walking on the plains of Africa.
343
817260
2000
vi ste hominid koji šeta afričkim ravnicama.
13:39
Is it just the wind or a dangerous predator?
344
819260
3000
Da li je to samo vetar ili opasni grabljivac?
13:42
What's the difference between those?
345
822260
2000
U čemu je razlika između njih?
13:44
Well, the wind is inanimate;
346
824260
2000
Pa, vetar je neživ;
13:46
the dangerous predator is an intentional agent.
347
826260
2000
opasni grabljivac ima nameru u svojim postupcima.
13:48
And I call this process agenticity.
348
828260
2000
Ovaj proces nazivam dodeljivanje namere.
13:50
That is the tendency to infuse patterns
349
830260
2000
To je tendencija da se dodeljuju šeme
13:52
with meaning, intention and agency,
350
832260
2000
sa smislom, namerom i delotvornošću,
13:54
often invisible beings from the top down.
351
834260
3000
često nevidljivim bićima najvišeg nivoa.
13:57
This is an idea that we got
352
837260
2000
Ovu ideju sam dobio
13:59
from a fellow TEDster here, Dan Dennett,
353
839260
2000
od Dena Deneta (Dan Dennett), takođe govornika TED-a,
14:01
who talked about taking the intentional stance.
354
841260
2000
koji je govorio o percipiranju namere.
14:03
So it's a type of that expanded to explain, I think, a lot of different things:
355
843260
3000
Mislim da je jedan tip toga proširen da bi objasnio dosta različitih stvari:
14:06
souls, spirits, ghosts, gods, demons, angels,
356
846260
3000
duše, duhove, aveti, bogove, demone, anđele,
14:09
aliens, intelligent designers,
357
849260
2000
vanzemaljce, inteligentne tvorce,
14:11
government conspiracists
358
851260
2000
vladine zaverenike
14:13
and all manner of invisible agents
359
853260
2000
i svakakve nevidljive entitete,
14:15
with power and intention, are believed
360
855260
2000
koji poseduju moć i nameru; veruje se
14:17
to haunt our world and control our lives.
361
857260
2000
da oni opsedaju naš svet i upravljaju našim životima.
14:19
I think it's the basis of animism
362
859260
2000
Mislim da je to osnova animizma,
14:21
and polytheism and monotheism.
363
861260
3000
politeizma i monoteizma.
14:24
It's the belief that aliens are somehow
364
864260
2000
To je verovanje da su vanzemaljci nekako
14:26
more advanced than us, more moral than us,
365
866260
2000
napredniji od nas, moralniji
14:28
and the narratives always are
366
868260
2000
i u pričama se uvek tvrdi
14:30
that they're coming here to save us and rescue us from on high.
367
870260
3000
da će doći sa neba ovde da bi nas sačuvali i spasili.
14:33
The intelligent designer's always portrayed
368
873260
2000
Inteligentni tvorac je uvek prikazan kao
14:35
as this super intelligent, moral being
369
875260
3000
natprosečno inteligentno, moralno biće,
14:38
that comes down to design life.
370
878260
2000
koje silazi dole da bi stvorilo život.
14:40
Even the idea that government can rescue us --
371
880260
2000
Čak je i ideja da nas vlade mogu spasiti --
14:42
that's no longer the wave of the future,
372
882260
2000
to više nije dominantna ideja budućnosti,
14:44
but that is, I think, a type of agenticity:
373
884260
2000
ali to je, mislim, tip dodeljivanja namere:
14:46
projecting somebody up there,
374
886260
2000
zamišljamo nekog tamo gore,
14:48
big and powerful, will come rescue us.
375
888260
2000
velikog i moćnog, ko će doći da nas spase.
14:50
And this is also, I think, the basis of conspiracy theories.
376
890260
2000
Smatram da je to, takođe, u osnovi teorija zavere.
14:52
There's somebody hiding behind there pulling the strings,
377
892260
3000
Neko se krije u pozadini stvari i drži sve konce u rukama,
14:55
whether it's the Illuminati
378
895260
2000
bez obzira da li su to Iluminati
14:57
or the Bilderbergers.
379
897260
2000
ili članovi bilderberške grupe.
14:59
But this is a pattern detection problem, isn't it?
380
899260
2000
Ali, to je problem u opažanju šema, zar ne?
15:01
Some patterns are real and some are not.
381
901260
2000
Neke šeme su stvarne, a druge nisu.
15:03
Was JFK assassinated by a conspiracy or by a lone assassin?
382
903260
3000
Da li su Džona Kenedija (JFK) ubili zaverenici ili jedan atentator?
15:06
Well, if you go there -- there's people there on any given day --
383
906260
3000
Pa, ako ćete se baviti time -- ima ljudi koji o tome razmišljaju svakog dana --
15:09
like when I went there, here -- showing me where the different shooters were.
384
909260
3000
kao kada sam ja bio u toj priči, ovde -- pokazuju mi gde su bili različiti atentatori.
15:12
My favorite one was he was in the manhole.
385
912260
3000
Moja omiljena priča je da je jedan od njih bio u kanalizacionom otvoru.
15:15
And he popped out at the last second, took that shot.
386
915260
3000
U poslednjem trenutku je iskočio i pucao.
15:18
But of course, Lincoln was assassinated by a conspiracy.
387
918260
2000
Ali naravno, Linkolna su ubili zaverenici.
15:20
So we can't just uniformly dismiss
388
920260
2000
Samim tim ne možemo olako da odbacimo
15:22
all patterns like that.
389
922260
2000
šeme te vrste.
15:24
Because, let's face it, some patterns are real.
390
924260
2000
Jer, kao što znate, neke šeme su stvarne.
15:26
Some conspiracies really are true.
391
926260
2000
Neke zavere zaista jesu stvarne.
15:30
Explains a lot, maybe.
392
930260
2000
To možda dosta toga objašnjava.
15:32
And 9/11 has a conspiracy theory. It is a conspiracy.
393
932260
3000
I za 11. septembar je vezana teorija zavere. To je zavera.
15:35
We did a whole issue on it.
394
935260
2000
Čitavo jedno izdanje časopisa smo posvetili tome.
15:37
Nineteen members of Al Queda plotting to fly planes into buildings
395
937260
2000
Kada devetnaest članova Al Kaide kuje plan
15:39
constitutes a conspiracy.
396
939260
2000
o udaranju u zgrade avionima, to je zavera.
15:41
But that's not what the "9/11 truthers" think.
397
941260
2000
Ali to nije ono što misle ljudi koji žele da saznaju istinu o 11. septembru.
15:43
They think it was an inside job by the Bush administration.
398
943260
3000
Misle da je to Bušova vlada uradila iza kulisa.
15:46
Well, that's a whole other lecture.
399
946260
2000
Ali, to je tema za drugo predavanje.
15:48
You know how we know that 9/11
400
948260
2000
Znate li kako znamo da 11. septembar
15:50
was not orchestrated by the Bush administration?
401
950260
2000
nije delo Bušove vlade?
15:52
Because it worked.
402
952260
2000
Zato što je uspešno izvedeno.
15:54
(Laughter)
403
954260
3000
(Smeh)
15:57
(Applause)
404
957260
3000
(Aplauz)
16:00
So we are natural-born dualists.
405
960260
2000
Mi smo od rođenja dualisti.
16:02
Our agenticity process comes from
406
962260
2000
Naša sklonost dodeljivanju namera jasna je zbog
16:04
the fact that we can enjoy movies like these.
407
964260
2000
činjenice da možemo uživati u gledanju ovakvih filmova.
16:06
Because we can imagine, in essence,
408
966260
2000
Zato što možemo zamisliti, u suštini,
16:08
continuing on.
409
968260
2000
kako će se odvijati dalje.
16:10
We know that if you stimulate the temporal lobe,
410
970260
2000
Znamo da ako stimulišete slepoočni režanj,
16:12
you can produce a feeling of out-of-body experiences,
411
972260
2000
možete izazvati osećaj vantelesnog iskustva,
16:14
near-death experiences,
412
974260
2000
iskustva bliske smrti,
16:16
which you can do by just touching an electrode to the temporal lobe there.
413
976260
3000
samo tako što ćete elektrodu postaviti na slepoočni režanj.
16:19
Or you can do it through loss of consciousness,
414
979260
2000
Ili možete to izazvati i gubitkom svesti,
16:21
by accelerating in a centrifuge.
415
981260
2000
ubrzavanjem u centrifugi.
16:23
You get a hypoxia, or a lower oxygen.
416
983260
3000
Upadate u hipoksiju, smanjen vam je nivo kiseonika u telu.
16:26
And the brain then senses
417
986260
2000
Mozak onda oseća
16:28
that there's an out-of-body experience.
418
988260
2000
vantelesno iskustvo.
16:30
You can use -- which I did, went out and did --
419
990260
2000
Možete upotrebiti -- što sam ja uradio, odvažio se i uradio --
16:32
Michael Persinger's God Helmet,
420
992260
2000
Božanski šlem Majkla Persindžera (Michael Persinger),
16:34
that bombards your temporal lobes with electromagnetic waves.
421
994260
2000
koji bombarduje vaše slepoočne režnjeve elektromagnetnim talasima.
16:36
And you get a sense of out-of-body experience.
422
996260
3000
Dobijete osećaj vantelesnog iskustva.
16:39
So I'm going to end here with a short video clip
423
999260
2000
Završiću ovim kratkim video snimkom,
16:41
that sort of brings all this together.
424
1001260
2000
koji nekako sve ovo povezuje.
16:43
It's just a minute and a half.
425
1003260
2000
Traje samo minut i po.
16:45
It ties together all this into the power of expectation and the power of belief.
426
1005260
3000
Povezuje sve ovo u moć iščekivanja i moć verovanja.
16:48
Go ahead and roll it.
427
1008260
2000
Hajde pustite ga.
16:50
Narrator: This is the venue they chose for their fake auditions
428
1010260
3000
Narator: Ovo je mesto koje su odabrali za održavanje njihovih lažnih audicija
16:53
for an advert for lip balm.
429
1013260
2000
za reklamu za preparat za usta.
16:55
Woman: We're hoping we can use part of this
430
1015260
2000
Žena: Nadamo se da ćemo moći da iskoristimo deo ovog
16:57
in a national commercial, right?
431
1017260
2000
u nacionalnoj reklami, zar ne?
16:59
And this is test on some lip balms
432
1019260
2000
Ovo je test nekih preparata za usne,
17:01
that we have over here.
433
1021260
2000
koja imamo ovde.
17:03
And these are our models who are going to help us,
434
1023260
2000
Ovo su naši modeli koji će nam pomoći,
17:05
Roger and Matt.
435
1025260
2000
Rodžer i Met.
17:07
And we have our own lip balm,
436
1027260
2000
Imamo ovde naš preparat za usne,
17:09
and we have a leading brand.
437
1029260
2000
a i najprodavaniju marku.
17:11
Would you have any problem
438
1031260
2000
Da li ti je problem
17:13
kissing our models to test it?
439
1033260
2000
da poljubiš naše modele radi testa?
17:15
Girl: No.
440
1035260
2000
Devojka: Nije.
17:17
Woman: You wouldn't? (Girl: No.) Woman: You'd think that was fine.
441
1037260
2000
Žena: Sigurno nije? (Devojka: Nije.) Žena: Misliš da je to u redu.
17:19
Girl: That would be fine. (Woman: Okay.)
442
1039260
2000
Devojka: To bi bilo u redu. (Žena: Okej.)
17:21
So this is a blind test.
443
1041260
3000
Ovo je test naslepo.
17:24
I'm going to ask you to go ahead
444
1044260
2000
Zamoliću te
17:26
and put a blindfold on.
445
1046260
2000
da staviš povez preko očiju.
17:29
Kay, now can you see anything? (Girl: No.)
446
1049260
3000
Ok, da li vidiš išta? (Devojka: Ne.)
17:32
Pull it so you can't even see down. (Girl: Okay.)
447
1052260
2000
Navuci ga da ne možeš da vidiš ni dole. (Devojka: Okej.)
17:34
Woman: It's completely blind now, right?
448
1054260
2000
Žena: Ništa ne vidiš sada, je l' da?
17:36
Girl: Yes. (Woman: Okay.)
449
1056260
2000
Devojka: Da. (Žena: Okej.)
17:38
Now, what I'm going to be looking for in this test
450
1058260
3000
E sad, ono što ću ispitati u ovom testu
17:41
is how it protects your lips,
451
1061260
3000
je kako preparat štiti usne,
17:44
the texture, right,
452
1064260
2000
teksturu, okej,
17:46
and maybe if you can discern any flavor or not.
453
1066260
3000
i da li možeš da prepoznaš neki ukus ili ne.
17:49
Girl: Okay. (Woman: Have you ever done a kissing test before?)
454
1069260
3000
Devojka: Okej. (Žena: Da li si ikada imala test ljubljenja?)
17:52
Girl: No.
455
1072260
2000
Devojka: Ne.
17:54
Woman: Take a step here.
456
1074260
2000
Žena: Stani ovde.
17:56
Okay, now I'm going to ask you to pucker up.
457
1076260
2000
Ok, sada ću te zamoliti da se napućiš.
17:58
Pucker up big and lean in just a little bit, okay?
458
1078260
3000
Lepo se napući i malo se nagni, okej?
18:06
(Music)
459
1086260
4000
(Muzika)
18:10
(Laughter)
460
1090260
5000
(Smeh)
18:19
(Laughter)
461
1099260
3000
(Smeh)
18:30
Woman: Okay.
462
1110260
2000
Žena: Okej.
18:32
And, Jennifer, how did that feel?
463
1112260
2000
Dženifer, kako je bilo?
18:34
Jennifer: Good.
464
1114260
2000
Dženifer: Dobro.
18:36
(Laughter)
465
1116260
7000
(Smeh)
18:43
Girl: Oh my God!
466
1123260
2000
Devojka: O, Bože!
18:45
(Laughter)
467
1125260
4000
(Smeh)
18:50
Michael Shermer: Thank you very much. Thank you. Thanks.
468
1130260
3000
Majkl Šermer (Michael Shermer): Hvala vam mnogo. Hvala vam. Hvala.
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7