Gero Miesenboeck reengineers a brain

52,036 views ・ 2010-11-05

TED


Dvaput kliknite na engleske titlove ispod za reprodukciju videozapisa.

Prevoditelj: Senzos Osijek Recezent: Tilen Pigac - EFZG
00:15
I have a doppelganger.
0
15260
3000
Imam dvojnika.
00:18
(Laughter)
1
18260
3000
(Smijeh)
00:21
Dr. Gero is a brilliant
2
21260
3000
Dr. Gero je briljantan,
00:24
but slightly mad scientist
3
24260
2000
ali pomalo lud znanstvenik
00:26
in the "Dragonball Z: Android Saga."
4
26260
3000
iz "Dragonball Z: Android Saga."
00:29
If you look very carefully,
5
29260
2000
Ako pozorno promotrite,
00:31
you see that his skull has been replaced
6
31260
3000
vidjet ćete da je njegova lubanja zamijenjena
00:34
with a transparent Plexiglas dome
7
34260
2000
kupolom od prozirnog pleksiglasa
00:36
so that the workings of his brain can be observed
8
36260
3000
kako bi se rad njegovog mozga mogao promatrati
00:39
and also controlled with light.
9
39260
3000
i kontrolirati pomoću svjetlosti.
00:42
That's exactly what I do --
10
42260
2000
To je upravo ono što i ja radim --
00:44
optical mind control.
11
44260
2000
vizualna kontrola uma.
00:46
(Laughter)
12
46260
2000
(Smijeh)
00:48
But in contrast to my evil twin
13
48260
2000
Ali, za razliku od mog zlog blizanca
00:50
who lusts after world domination,
14
50260
3000
koji žudi za dominacijom nad svijetom,
00:53
my motives are not sinister.
15
53260
3000
ja nemam tako zlokobne motive.
00:56
I control the brain
16
56260
2000
Kontroliram mozak
00:58
in order to understand how it works.
17
58260
2000
da bih shvatio kako radi.
01:00
Now wait a minute, you may say,
18
60260
2000
"Čekaj malo", mogli biste reći,
01:02
how can you go straight to controlling the brain
19
62260
3000
"kako odmah možeš prijeći na kontrolu mozga
01:05
without understanding it first?
20
65260
2000
prije nego što si ga razumio u potpunosti?
01:07
Isn't that putting the cart before the horse?
21
67260
3000
Nije li to trčanje pred rudo?"
01:11
Many neuroscientists agree with this view
22
71260
3000
Mnogi neuroznanstvenici se slažu s ovim
01:14
and think that understanding will come
23
74260
3000
i smatraju da će razumijevanje proizaći
01:17
from more detailed observation and analysis.
24
77260
3000
iz detaljnog promatranja i analiza.
01:20
They say, "If we could record the activity of our neurons,
25
80260
4000
Kažu: " Kada bismo mogli snimiti aktivnost neurona,
01:24
we would understand the brain."
26
84260
3000
razumjeli bismo mozak."
01:27
But think for a moment what that means.
27
87260
3000
Ali, razmislite na trenutak što to znači.
01:30
Even if we could measure
28
90260
2000
Čak kada bismo mogli mjeriti
01:32
what every cell is doing at all times,
29
92260
2000
što svaka stanica radi
01:34
we would still have to make sense
30
94260
2000
i dalje bismo morali dokučiti koji je smisao
01:36
of the recorded activity patterns,
31
96260
2000
snimljenih uzoraka aktivnosti,
01:38
and that's so difficult,
32
98260
2000
a to je veoma teško
01:40
chances are we'll understand these patterns
33
100260
2000
i vjerojatnost da bismo to razumjeli
01:42
just as little as the brains that produce them.
34
102260
3000
je jednako malena kao i mozak koji proizvodi tu aktivnost.
01:45
Take a look at what brain activity might look like.
35
105260
3000
Pogledajte kako izgleda aktivnost mozga.
01:48
In this simulation, each black dot
36
108260
2000
U ovoj simulaciji, svaka crna točkica
01:50
is one nerve cell.
37
110260
2000
je jedna živčana stanica.
01:52
The dot is visible
38
112260
2000
Točkica je vidljiva
01:54
whenever a cell fires an electrical impulse.
39
114260
2000
u trenutku kada stanica proizvodi električni impuls.
01:56
There's 10,000 neurons here.
40
116260
2000
Ovdje imamo 10.000 neurona.
01:58
So you're looking at roughly one percent
41
118260
2000
Pred vama je, ugrubo, jedan posto
02:00
of the brain of a cockroach.
42
120260
3000
mozga žohara.
02:04
Your brains are about 100 million times
43
124260
3000
Vaši mozgovi su otprilike 100 milijuna puta
02:07
more complicated.
44
127260
2000
kompleksniji.
02:09
Somewhere, in a pattern like this,
45
129260
2000
Negdje, u uzorcima poput ovih,
02:11
is you,
46
131260
2000
ste vi,
02:13
your perceptions,
47
133260
2000
vaše percepcije,
02:15
your emotions, your memories,
48
135260
3000
vaše emocije, vaša sjećanja,
02:18
your plans for the future.
49
138260
2000
vaši planovi za budućnost.
02:20
But we don't know where,
50
140260
2000
Ali, ne znamo gdje točno
02:22
since we don't know how to read the pattern.
51
142260
3000
jer još uvijek nismo dešifrirali uzorke.
02:25
We don't understand the code used by the brain.
52
145260
3000
Ne znamo koji kod mozak koristi.
02:28
To make progress,
53
148260
2000
Da bismo napredovali,
02:30
we need to break the code.
54
150260
2000
moramo probiti kod.
02:32
But how?
55
152260
2000
Ali kako?
02:35
An experienced code-breaker will tell you
56
155260
2000
Svaki iskusan haker će vam reći
02:37
that in order to figure out what the symbols in a code mean,
57
157260
3000
da mora biti u mogućnosti igrati se sa simbolima koda,
02:40
it's essential to be able to play with them,
58
160260
3000
ispremještati ih po svojoj volji kako bi
02:43
to rearrange them at will.
59
163260
2000
mogao shvatiti što koji simbol znači.
02:45
So in this situation too,
60
165260
2000
U situaciji u kojoj se mi nalazimo,
02:47
to decode the information
61
167260
2000
ako želimo dekodirati informacije
02:49
contained in patterns like this,
62
169260
2000
sadržane u uzorcima poput ovih,
02:51
watching alone won't do.
63
171260
2000
pukim promatranjem nećemo puno postići.
02:53
We need to rearrange the pattern.
64
173260
2000
Moramo ispremještati uzorke.
02:55
In other words,
65
175260
2000
Drugim riječima,
02:57
instead of recording the activity of neurons,
66
177260
2000
umjesto snimanja aktivnosti neurona,
02:59
we need to control it.
67
179260
2000
tu aktivnost moramo kontrolirati.
03:01
It's not essential that we can control
68
181260
2000
Nije nužno kontrolirati
03:03
the activity of all neurons in the brain, just some.
69
183260
3000
aktivnost svih neurona u mozgu, nekoliko će biti dovoljno.
03:06
The more targeted our interventions, the better.
70
186260
2000
Što su točnije usmjerene naše intervencije, to bolje.
03:08
And I'll show you in a moment
71
188260
2000
Za trenutak ću vam pokazati
03:10
how we can achieve the necessary precision.
72
190260
3000
kako možemo postići tu potrebnu preciznost.
03:13
And since I'm realistic, rather than grandiose,
73
193260
3000
Budući da sam više realističan, nego veličanstven,
03:16
I don't claim that the ability to control the function of the nervous system
74
196260
3000
ne tvrdim da će nam mogućnost kontrole funkcije živčanog sustava
03:19
will at once unravel all its mysteries.
75
199260
3000
razmrsiti sve misterije,
03:22
But we'll certainly learn a lot.
76
202260
3000
ali sigurno ćemo puno naučiti.
03:27
Now, I'm by no means
77
207260
2000
Dakle, ja osobno nikako nisam
03:29
the first person to realize
78
209260
2000
prva osoba koja je shvatila
03:31
how powerful a tool intervention is.
79
211260
3000
kako je intervencija na mozgu moćno oružje.
03:34
The history of attempts
80
214260
2000
Povijest pokušaja
03:36
to tinker with the function of the nervous system
81
216260
2000
kopkanja po funkcijama živčanog sustava
03:38
is long and illustrious.
82
218260
2000
je veoma duga i slikovita.
03:40
It dates back at least 200 years,
83
220260
3000
Stvar je stara najmanje 200 godina,
03:43
to Galvani's famous experiments
84
223260
2000
Galvani je radio dobro poznate eksperimente
03:45
in the late 18th century and beyond.
85
225260
3000
krajem 18. stoljeća i poslije.
03:49
Galvani showed that a frog's legs twitched
86
229260
3000
Pokazao je kako noga žabe trza
03:52
when he connected the lumbar nerve
87
232260
2000
kada se slabinski živac spoji
03:54
to a source of electrical current.
88
234260
3000
na izvor električnog napona.
03:57
This experiment revealed the first, and perhaps most fundamental,
89
237260
3000
Ovaj eksperiment nam je otkrio prvu i, možda, najosnovniju
04:00
nugget of the neural code:
90
240260
2000
tajnu koda živčanog sustava:
04:02
that information is written in the form
91
242260
2000
da je informacija zapisana u obliku
04:04
of electrical impulses.
92
244260
3000
električnog impulsa.
04:08
Galvani's approach
93
248260
2000
Galvanijev pristup ispitivanja
04:10
of probing the nervous system with electrodes
94
250260
2000
živčanog sustava pomoću elektroda
04:12
has remained state-of-the-art until today,
95
252260
3000
ostao je na snazi do danas,
04:15
despite a number of drawbacks.
96
255260
3000
unatoč nekim nedostacima.
04:18
Sticking wires into the brain is obviously rather crude.
97
258260
3000
Prikopčavanje žica na mozak je očigledno pomalo primitivno.
04:21
It's hard to do in animals that run around,
98
261260
2000
Tako nešto je teško primijeniti na životinjama koje trče uokolo,
04:23
and there is a physical limit
99
263260
2000
a postoji i fizička granica
04:25
to the number of wires
100
265260
2000
broja žica
04:27
that can be inserted simultaneously.
101
267260
3000
koje mogu biti prikopčane istovremeno.
04:30
So around the turn of the last century,
102
270260
2000
Na prijelazu u novo stoljeće,
04:32
I started to think,
103
272260
2000
počeo sam razmišljati:
04:34
"Wouldn't it be wonderful if one could take this logic
104
274260
3000
"Ne bi li bilo prekrasno kada bi netko naglavačke
04:37
and turn it upside down?"
105
277260
2000
preokrenuo vladajuću logiku?"
04:39
So instead of inserting a wire
106
279260
2000
Tako da umjesto umetanja žice
04:41
into one spot of the brain,
107
281260
3000
u neku točku na mozgu,
04:44
re-engineer the brain itself
108
284260
2000
preuredimo sam mozak
04:46
so that some of its neural elements
109
286260
3000
pa da neki od njegovih živčanih elemenata
04:49
become responsive to diffusely broadcast signals
110
289260
3000
postanu osjetljivi na difuzno odaslane signale
04:52
such as a flash of light.
111
292260
3000
poput bljeska ili svjetla.
04:55
Such an approach would literally, in a flash of light,
112
295260
3000
Takav pristup bi, doslovno brzinom munje,
04:58
overcome many of the obstacles to discovery.
113
298260
3000
savladao mnoge prepreke koje su se prepriječile na putu ka otkriću.
05:01
First, it's clearly a non-invasive,
114
301260
3000
Prvo, to je očigledno neinvazivan
05:04
wireless form of communication.
115
304260
3000
i bežičan način komunikacije.
05:07
And second, just as in a radio broadcast,
116
307260
2000
I drugo, baš kao kod radio emitiranja,
05:09
you can communicate with many receivers at once.
117
309260
3000
možeš komunicirati s mnogo slušatelja u isto vrijeme --
05:12
You don't need to know where these receivers are,
118
312260
3000
ne moraš znati gdje se oni nalaze
05:15
and it doesn't matter if these receivers move --
119
315260
2000
i nije bitno kreću li se --
05:17
just think of the stereo in your car.
120
317260
3000
samo pomislite na radio u svojem automobilu.
05:20
It gets even better,
121
320260
3000
Ali ni to nije sve,
05:23
for it turns out that we can fabricate the receivers
122
323260
3000
mogli bismo proizvoditi receptore
05:26
out of materials that are encoded in DNA.
123
326260
3000
iz materijala koji su kodirani u DNK.
05:29
So each nerve cell
124
329260
2000
Svaka živčana stanica
05:31
with the right genetic makeup
125
331260
2000
s pravim genetičkim make up-om
05:33
will spontaneously produce a receiver
126
333260
3000
bi spontano proizvela receptor
05:36
that allows us to control its function.
127
336260
3000
koji bi nam dozvoljavao da kontroliramo njenu funkciju.
05:39
I hope you'll appreciate
128
339260
2000
Nadam se da cijenite
05:41
the beautiful simplicity
129
341260
2000
prekrasnu jednostavnost
05:43
of this concept.
130
343260
2000
ovog koncepta.
05:45
There's no high-tech gizmos here,
131
345260
2000
Tu nema high-tech uređaja,
05:47
just biology revealed through biology.
132
347260
3000
samo čista biologija.
05:51
Now let's take a look at these miraculous receivers up close.
133
351260
3000
Bacimo sada pogled na te čudesne receptore izbliza.
05:54
As we zoom in on one of these purple neurons,
134
354260
3000
Ako zumiramo jedan od ovih ljubičastih neurona,
05:57
we see that its outer membrane
135
357260
2000
možemo vidjeti da je njihova vanjska membrana
05:59
is studded with microscopic pores.
136
359260
2000
okružena mikroskopskim porama.
06:01
Pores like these conduct electrical current
137
361260
2000
Te pore provode električni napon
06:03
and are responsible
138
363260
2000
i odgovorne su za
06:05
for all the communication in the nervous system.
139
365260
2000
komunikaciju u živčanom sustavu.
06:07
But these pores here are special.
140
367260
2000
No, te pore su posebne
06:09
They are coupled to light receptors
141
369260
2000
jer su povezane sa svjetlosnim receptorima
06:11
similar to the ones in your eyes.
142
371260
3000
koji sliče onima u našim očima.
06:14
Whenever a flash of light hits the receptor,
143
374260
2000
Kada bljesak ili svjetlost pogodi receptor,
06:16
the pore opens, an electrical current is switched on,
144
376260
3000
pore se otvaraju, električni napon se uključi
06:19
and the neuron fires electrical impulses.
145
379260
3000
i neuron šalje električni signal.
06:22
Because the light-activated pore is encoded in DNA,
146
382260
3000
Činjenica da su svjetlošću aktivirane pore kodirane u DNK,
06:25
we can achieve incredible precision.
147
385260
3000
omogućuje nam veliku preciznost.
06:28
This is because,
148
388260
2000
Iako svaka
06:30
although each cell in our bodies
149
390260
2000
stanica našeg tijela
06:32
contains the same set of genes,
150
392260
2000
sadrži isti set gena,
06:34
different mixes of genes get turned on and off
151
394260
2000
drukčije mješavine gena se aktiviraju i deaktiviraju
06:36
in different cells.
152
396260
2000
u različitim stanicama.
06:38
You can exploit this to make sure
153
398260
2000
Ovo možemo koristiti kako bismo bili sigurni
06:40
that only some neurons
154
400260
2000
da samo neki od neurona
06:42
contain our light-activated pore and others don't.
155
402260
3000
sadrže naše svjetlošću aktivirane pore.
06:45
So in this cartoon, the bluish white cell
156
405260
2000
U ovoj animaciji, modro bijela stanica
06:47
in the upper-left corner
157
407260
2000
u gornjem lijevom kutu
06:49
does not respond to light
158
409260
2000
ne odgovara na svjetlost
06:51
because it lacks the light-activated pore.
159
411260
3000
jer joj nedostaje svjetlošću aktivirana pora.
06:54
The approach works so well
160
414260
2000
Ovaj koncept djeluje tako dobro
06:56
that we can write purely artificial messages
161
416260
2000
da čak možemo mozgu direktno
06:58
directly to the brain.
162
418260
2000
slati umjetne naredbe.
07:00
In this example, each electrical impulse,
163
420260
2000
U ovom primjeru, svaki električni impuls,
07:02
each deflection on the trace,
164
422260
3000
svaki otklon od putanje,
07:05
is caused by a brief pulse of light.
165
425260
3000
uzrokovan je kratkotrajnim pulsirajućim svjetlom,
07:08
And the approach, of course, also works
166
428260
2000
a pristup djeluje
07:10
in moving, behaving animals.
167
430260
3000
i kod životinja koje se kreću.
07:13
This is the first ever such experiment,
168
433260
2000
Ovo je prvi eksperiment te vrste
07:15
sort of the optical equivalent of Galvani's.
169
435260
3000
i na neki je način optički ekvivalent Galvanijevog.
07:18
It was done six or seven years ago
170
438260
2000
Izvela ga je moja nekadašnja studentica
07:20
by my then graduate student, Susana Lima.
171
440260
3000
Susana Lima, prije šest ili sedam godina.
07:23
Susana had engineered the fruit fly on the left
172
443260
3000
Susana je projektirala voćnu mušicu slijeva
07:26
so that just two out of the 200,000 cells in its brain
173
446260
4000
tako da samo dvije od 200.000 stanica u njenom mozgu
07:30
expressed the light-activated pore.
174
450260
3000
imaju ekspresiju pora aktiviranih svjetlošću.
07:33
You're familiar with these cells
175
453260
2000
Upoznati ste s tim stanicama
07:35
because they are the ones that frustrate you
176
455260
2000
jer su to one koje vas frustriraju
07:37
when you try to swat the fly.
177
457260
2000
kad pokušavate zgnječiti mušicu.
07:39
They trained the escape reflex that makes the fly jump into the air
178
459260
3000
One su istrenirale refleks bijega koji djeluje tako da mušica skoči u zrak
07:42
and fly away whenever you move your hand in position.
179
462260
3000
i odleti kad god vi pomičete ruku.
07:46
And you can see here that the flash of light has exactly the same effect.
180
466260
3000
Ovdje možete vidjeti da bljesak ima identičan učinak.
07:49
The animal jumps, it spreads its wings, it vibrates them,
181
469260
3000
Životinja poskoči, raširi krila, vibrira krilima,
07:52
but it can't actually take off
182
472260
2000
ali ne može poletjeti
07:54
because the fly is sandwiched between two glass plates.
183
474260
3000
jer je zaglavljena između dvije staklene ploče.
07:58
Now to make sure that this was no reaction of the fly
184
478260
2000
Sada, kako bi bila sigurna da ovo nije bila reakcija
08:00
to a flash it could see,
185
480260
3000
na svjetlost koju mušica vidi,
08:03
Susana did a simple
186
483260
2000
Susana je izvela jednostavan,
08:05
but brutally effective experiment.
187
485260
2000
ali pomalo brutalan pokus.
08:07
She cut the heads off of her flies.
188
487260
3000
Odrubila je glave svojim mušicama.
08:11
These headless bodies can live for about a day,
189
491260
3000
Dekapitirane mušice mogu živjeti otprilike jedan dan,
08:14
but they don't do much.
190
494260
2000
ali ne mogu raditi puno toga,
08:16
They just stand around
191
496260
3000
samo stoje
08:19
and groom excessively.
192
499260
3000
i pretjerano se dotjeruju.
08:22
So it seems that the only trait that survives decapitation is vanity.
193
502260
3000
Dakle, čini se da je taština jedina osobina koja uspijeva preživjeti odrubljivanje glave.
08:25
(Laughter)
194
505260
3000
(Smijeh)
08:30
Anyway, as you'll see in a moment,
195
510260
2000
Kako god, za trenutak ćete vidjeti
08:32
Susana was able to turn on the flight motor
196
512260
3000
da je Susana uspjela uključiti motor za letenje
08:35
of what's the equivalent of the spinal cord of these flies
197
515260
3000
u tzv. ekvivalentu kralješnične moždine ovih mušica
08:38
and get some of the headless bodies
198
518260
2000
i izazvati da neka od ovih bezglavih tijela
08:40
to actually take off and fly away.
199
520260
3000
mušica polete.
08:47
They didn't get very far, obviously.
200
527260
2000
Očito, nisu daleko stigle.
08:50
Since we took these first steps,
201
530260
2000
Od kada smo mi poduzeli prve korake,
08:52
the field of optogenetics has exploded.
202
532260
3000
optogenetika se strahovito razvila
08:55
And there are now hundreds of labs
203
535260
2000
i sada postoje na stotine laboratorija
08:57
using these approaches.
204
537260
2000
koji primjenjuju ovaj pristup.
08:59
And we've come a long way
205
539260
2000
Mnogo smo napredovali
09:01
since Galvani's and Susana's first successes
206
541260
3000
od Galvanija i Susane koji su uspjeli postići
09:04
in making animals twitch or jump.
207
544260
2000
da životinje trzaju ili polete.
09:06
We can now actually interfere with their psychology
208
546260
3000
Sada možemo inteferirati sa njihovom psihologijom
09:09
in rather profound ways,
209
549260
2000
na mnogo korisnih načina,
09:11
as I'll show you in my last example,
210
551260
2000
što ću vam i pokazati u svom posljednjem primjeru,
09:13
which is directed at a familiar question.
211
553260
3000
koji je povezan sa poznatim pitanjima.
09:16
Life is a string of choices
212
556260
3000
Život je niz izbora
09:19
creating a constant pressure to decide what to do next.
213
559260
3000
koji stvaraju stalan pritisak na sljedeću odluku koju ćemo donijeti.
09:23
We cope with this pressure by having brains,
214
563260
3000
Nosimo se s tim pritiskom tako što imamo mozak
09:26
and within our brains, decision-making centers
215
566260
3000
koji je centar našeg odlučivanja
09:29
that I've called here the "Actor."
216
569260
3000
i koji ja zovem "Glumac".
09:33
The Actor implements a policy that takes into account
217
573260
3000
Glumac provodi politiku koja uzima u obzir
09:36
the state of the environment
218
576260
2000
stanje okoliša
09:38
and the context in which we operate.
219
578260
3000
i kontekst u kojem djelujemo.
09:41
Our actions change the environment, or context,
220
581260
3000
Naše djelovanje radi promjene u okolišu i kontekstu,
09:44
and these changes are then fed back into the decision loop.
221
584260
3000
a ove promjene su onda vraćaju u petlju odluke.
09:48
Now to put some neurobiological meat
222
588260
3000
Sada moramo dodati malo neurobiološkog mesa
09:51
on this abstract model,
223
591260
2000
na ovaj kostur apstraktnog modela.
09:53
we constructed a simple one-dimensional world
224
593260
2000
Konstruirali smo jednostavni jednodimenzionalni svijet
09:55
for our favorite subject, fruit flies.
225
595260
3000
za naš omiljeni subjekt, voćne mušice.
09:58
Each chamber in these two vertical stacks
226
598260
2000
Svaka komora između dvije vertikalne pregrade
10:00
contains one fly.
227
600260
2000
sadrži jednu mušicu.
10:02
The left and the right halves of the chamber
228
602260
3000
Lijeva i desna polovica komore
10:05
are filled with two different odors,
229
605260
2000
su ispunjene s dva različita mirisa.
10:07
and a security camera watches
230
607260
2000
Sigurnosna kamera nam omogućuje da promatramo
10:09
as the flies pace up and down between them.
231
609260
3000
kako mušice hodaju amo-tamo po komori.
10:12
Here's some such CCTV footage.
232
612260
2000
Ovdje imamo jednu takvu CCTV snimku.
10:14
Whenever a fly reaches the midpoint of the chamber
233
614260
3000
Kada mušica dođe do sredine komore
10:17
where the two odor streams meet,
234
617260
2000
gdje se miješaju dva mirisa,
10:19
it has to make a decision.
235
619260
2000
mora odlučiti.
10:21
It has to decide whether to turn around
236
621260
2000
Mora odlučiti hoće li se okrenuti
10:23
and stay in the same odor,
237
623260
2000
i ostati u istoj polovici
10:25
or whether to cross the midline
238
625260
2000
ili prijeći sredinu
10:27
and try something new.
239
627260
2000
i probati nešto novo.
10:29
These decisions are clearly a reflection
240
629260
3000
Ove odluke su refleksija
10:32
of the Actor's policy.
241
632260
3000
Glumčeve politike.
10:36
Now for an intelligent being like our fly,
242
636260
3000
Za pametno biće poput naše mušice,
10:39
this policy is not written in stone
243
639260
3000
ova politika nije trajno zapisana u kamenu,
10:42
but rather changes as the animal learns from experience.
244
642260
3000
već se mijenja kako životinja uči iz novih iskustava.
10:45
We can incorporate such an element
245
645260
2000
Možemo ujediniti taj element
10:47
of adaptive intelligence into our model
246
647260
3000
prilagodljive inteligencije u naš model
10:50
by assuming that the fly's brain
247
650260
2000
pretpostavljajući da mozak mušice
10:52
contains not only an Actor,
248
652260
2000
ne sadrži samo Glumca,
10:54
but a different group of cells,
249
654260
2000
već i drukčiju skupinu stanica --
10:56
a "Critic," that provides a running commentary
250
656260
3000
Kritičara koji osigurava komentar
10:59
on the Actor's choices.
251
659260
2000
na Glumčeve odluke.
11:01
You can think of this nagging inner voice
252
661260
3000
Ovaj gunđajući unutarnji glas možete zamisliti
11:04
as sort of the brain's equivalent
253
664260
2000
kao neku vrstu ekvivalenta
11:06
of the Catholic Church,
254
666260
2000
Katoličke Crkve,
11:08
if you're an Austrian like me,
255
668260
3000
ako ste Austrijanci poput mene,
11:11
or the super-ego, if you're Freudian,
256
671260
3000
super-ega ako ste frojdovac
11:14
or your mother, if you're Jewish.
257
674260
2000
ili kao vašu majku ako ste Židov.
11:16
(Laughter)
258
676260
4000
(Smijeh)
11:20
Now obviously,
259
680260
2000
Dakle očito da je
11:22
the Critic is a key ingredient
260
682260
3000
Kritičar glavni sastojak
11:25
in what makes us intelligent.
261
685260
2000
koji nas čini inteligentnima.
11:27
So we set out to identify
262
687260
2000
Odlučili smo identificirati
11:29
the cells in the fly's brain
263
689260
2000
stanice koje igraju ulogu
11:31
that played the role of the Critic.
264
691260
2000
Kritičara u mozgu mušice.
11:33
And the logic of our experiment was simple.
265
693260
3000
Logika našeg eksperimenta je bila poprilično jednostavna --
11:36
We thought if we could use our optical remote control
266
696260
3000
kada bismo rabili optički daljinski prekidač za kontrolu
11:39
to activate the cells of the Critic,
267
699260
3000
aktivnosti stanica Kritičara,
11:42
we should be able, artificially, to nag the Actor
268
702260
3000
trebali bismo moći natjerati Glumca
11:45
into changing its policy.
269
705260
2000
da promijeni svoju politiku.
11:47
In other words,
270
707260
2000
Drugim riječima,
11:49
the fly should learn from mistakes
271
709260
2000
mušica treba učiti iz grešaka
11:51
that it thought it had made
272
711260
2000
koje misli da je učinila,
11:53
but, in reality, it had not made.
273
713260
3000
ali ih zapravo nije učinila.
11:56
So we bred flies
274
716260
2000
Parili smo mušice
11:58
whose brains were more or less randomly peppered
275
718260
3000
čiji mozak je bio manje ili više nasumično obasut
12:01
with cells that were light addressable.
276
721260
2000
stanicama koje su odgovarale na svjetlost,
12:03
And then we took these flies
277
723260
2000
a onda smo tim mušicama
12:05
and allowed them to make choices.
278
725260
2000
dali da donose odluke.
12:07
And whenever they made one of the two choices,
279
727260
2000
Kad god su one izabrale jedan
12:09
chose one odor,
280
729260
2000
od dva mirisa,
12:11
in this case the blue one over the orange one,
281
731260
2000
u ovom slučaju plavi umjesto narančastog,
12:13
we switched on the lights.
282
733260
2000
mi bismo upalili svjetlo.
12:15
If the Critic was among the optically activated cells,
283
735260
3000
Ako je Kritičar bio među optički podražljivim stanicama,
12:18
the result of this intervention
284
738260
2000
rezultat ove intervencije
12:20
should be a change in policy.
285
740260
3000
bi trebalo biti mijenjanje politike --
12:23
The fly should learn to avoid
286
743260
2000
mušica bi trebala naučiti izbjegavati
12:25
the optically reinforced odor.
287
745260
2000
optički pojačan miris.
12:27
Here's what happened in two instances:
288
747260
3000
Evo što se dogodilo u dva slučaja:
12:30
We're comparing two strains of flies,
289
750260
3000
uspoređivali smo dva soja mušica,
12:33
each of them having
290
753260
2000
a svaka od njih je imala
12:35
about 100 light-addressable cells in their brains,
291
755260
2000
otprilike 100 svjetlošću podražljivih stanica u mozgu --
12:37
shown here in green on the left and on the right.
292
757260
3000
to je ovo zeleno, na lijevoj i desnoj strani.
12:40
What's common among these groups of cells
293
760260
3000
Ove dvije grupe stanica imaju zajedničko svojstvo
12:43
is that they all produce the neurotransmitter dopamine.
294
763260
3000
da proizvode neurotransmiter dopamin.
12:46
But the identities of the individual
295
766260
2000
Ali identitet zasebnih neurona
12:48
dopamine-producing neurons
296
768260
2000
koji luče dopamin
12:50
are clearly largely different on the left and on the right.
297
770260
3000
je očigledno različit na lijevoj i desnoj strani.
12:53
Optically activating
298
773260
2000
Svjetlosna aktivacija
12:55
these hundred or so cells
299
775260
2000
tih stotinu stanica
12:57
into two strains of flies
300
777260
2000
u dva soja mušice
12:59
has dramatically different consequences.
301
779260
2000
ima dramatično različite posljedice.
13:01
If you look first at the behavior
302
781260
2000
Ako prvo promotrimo ponašanje
13:03
of the fly on the right,
303
783260
2000
desne mušice,
13:05
you can see that whenever it reaches the midpoint of the chamber
304
785260
3000
možemo vidjeti da uvijek kada dođe na sredinu komore
13:08
where the two odors meet,
305
788260
2000
gdje se dva mirisa spajaju,
13:10
it marches straight through, as it did before.
306
790260
3000
maršira ravno kroz komoru, baš kao što je činila i prije.
13:13
Its behavior is completely unchanged.
307
793260
2000
Njezino ponašanje je ostalo nepromijenjeno.
13:15
But the behavior of the fly on the left is very different.
308
795260
3000
Ponašanje lijeve mušice je sasvim različito.
13:18
Whenever it comes up to the midpoint,
309
798260
3000
Kada dođe na sredinu komore,
13:21
it pauses,
310
801260
2000
ona stane
13:23
it carefully scans the odor interface
311
803260
2000
i pažljivo promotri granicu
13:25
as if it was sniffing out its environment,
312
805260
2000
kao da njuši okolinu
13:27
and then it turns around.
313
807260
2000
i onda se okrene.
13:29
This means that the policy that the Actor implements
314
809260
3000
Ovo znači da politika koju Glumac primjenjuje
13:32
now includes an instruction to avoid the odor
315
812260
2000
sada uključuje i uputu da se izbjegava miris
13:34
that's in the right half of the chamber.
316
814260
3000
na desnoj polovini komore.
13:37
This means that the Critic
317
817260
2000
To znači da je u toj životinji
13:39
must have spoken in that animal,
318
819260
2000
progovorio Kritičar
13:41
and that the Critic must be contained
319
821260
2000
koji mora biti sadržan
13:43
among the dopamine-producing neurons on the left,
320
823260
3000
među neuronima koji proizvode dopamin na lijevoj,
13:46
but not among the dopamine producing neurons on the right.
321
826260
3000
ali ne i među neuronima koji proizvode dopamin na desnoj strani.
13:49
Through many such experiments,
322
829260
3000
Nakon mnogo ovakvih eksperimenata,
13:52
we were able to narrow down
323
832260
2000
suzili smo
13:54
the identity of the Critic
324
834260
2000
identitet Kritičara
13:56
to just 12 cells.
325
836260
2000
na samo 12 stanica.
13:58
These 12 cells, as shown here in green,
326
838260
3000
Tih 12 stanica, kao što možete vidjeti ovo zeleno,
14:01
send the output to a brain structure
327
841260
2000
šalju izlazne informacije strukturi u mozgu
14:03
called the "mushroom body,"
328
843260
2000
koja se naziva "gljivasto tijelo" (Corpora pedunculata)
14:05
which is shown here in gray.
329
845260
2000
koja je prikazana sivo.
14:07
We know from our formal model
330
847260
2000
Iz prethodnog modela znamo
14:09
that the brain structure
331
849260
2000
da je Glumac izvršna struktura
14:11
at the receiving end of the Critic's commentary is the Actor.
332
851260
3000
mozga nakon primanja Kritičarovog komentara.
14:14
So this anatomy suggests
333
854260
2000
Anatomija nalaže
14:16
that the mushroom bodies have something to do
334
856260
3000
da gljivasta tijela imaju neku ulogu
14:19
with action choice.
335
859260
2000
u izboru pokreta.
14:21
Based on everything we know about the mushroom bodies,
336
861260
2000
Ako uzmemo u obzir sve što znamo o gljivastim tijelima,
14:23
this makes perfect sense.
337
863260
2000
ovo ima smisla.
14:25
In fact, it makes so much sense
338
865260
2000
Zapravo, ovo toliko ima smisla
14:27
that we can construct an electronic toy circuit
339
867260
3000
da čak možemo izraditi električnu igračku
14:30
that simulates the behavior of the fly.
340
870260
3000
koja simulira ponašanje mušice.
14:33
In this electronic toy circuit,
341
873260
3000
Neuroni gljivastog tijela
14:36
the mushroom body neurons are symbolized
342
876260
2000
su simbolizirani
14:38
by the vertical bank of blue LEDs
343
878260
3000
okomitim nizom plavih LED dioda
14:41
in the center of the board.
344
881260
3000
na sredini ploče.
14:44
These LED's are wired to sensors
345
884260
2000
Te LED diode su prikopčane na senzore
14:46
that detect the presence of odorous molecules in the air.
346
886260
3000
koji detektiraju prisutnost molekula mirisa u zraku.
14:50
Each odor activates a different combination of sensors,
347
890260
3000
Svaki miris aktivira drugu kombinaciju senzora,
14:53
which in turn activates
348
893260
2000
a koja zauzvrat aktivira
14:55
a different odor detector in the mushroom body.
349
895260
3000
drugi detektor mirisa u gljivastom tijelu.
14:58
So the pilot in the cockpit of the fly,
350
898260
2000
Tako nam pilot u kokpitu mušice,
15:00
the Actor,
351
900260
2000
Glumac,
15:02
can tell which odor is present
352
902260
2000
može reći koji je miris prisutan
15:04
simply by looking at which of the blue LEDs lights up.
353
904260
4000
jednostavno gledajući koja od plavih LED dioda je upaljena.
15:09
What the Actor does with this information
354
909260
2000
Što će Glumac učiniti s tim informacijama
15:11
depends on its policy,
355
911260
2000
ovisi o njegovoj politici
15:13
which is stored in the strengths of the connection,
356
913260
2000
koja je pohranjena u jačini veze
15:15
between the odor detectors
357
915260
2000
između detektora mirisa
15:17
and the motors
358
917260
2000
i motora
15:19
that power the fly's evasive actions.
359
919260
2000
koji mušici omogućuje kretanje.
15:22
If the connection is weak, the motors will stay off
360
922260
2000
Ako je veza slaba, motor će ostati isključen
15:24
and the fly will continue straight on its course.
361
924260
3000
i mušica će nastaviti svoju putanju ravno po komori.
15:27
If the connection is strong, the motors will turn on
362
927260
3000
Ako je veza jaka, motor će se upaliti
15:30
and the fly will initiate a turn.
363
930260
3000
i mušica će se okrenuti na suprotnu stranu.
15:33
Now consider a situation
364
933260
2000
Sada uzmimo u obzir situaciju
15:35
in which the motors stay off,
365
935260
2000
kada motor ostaje isključen,
15:37
the fly continues on its path
366
937260
3000
mušica nastavlja ići svojim putem
15:40
and it suffers some painful consequence
367
940260
2000
i doživljava neke bolne posljedice,
15:42
such as getting zapped.
368
942260
2000
na primjer smrt.
15:44
In a situation like this,
369
944260
2000
U situaciji poput ove,
15:46
we would expect the Critic to speak up
370
946260
2000
očekujemo da će Kritičar progovoriti
15:48
and to tell the Actor
371
948260
2000
i reći Glumcu
15:50
to change its policy.
372
950260
2000
da promijeni politiku.
15:52
We have created such a situation, artificially,
373
952260
3000
Mi smo umjetno kreirali ovu situaciju,
15:55
by turning on the critic with a flash of light.
374
955260
3000
uključujući Kritičara pomoću bljeska svjetla.
15:58
That caused a strengthening of the connections
375
958260
3000
To je prouzrokovalo pojačavanje veze
16:01
between the currently active odor detector
376
961260
3000
između trenutno aktivnog detektora mirisa
16:04
and the motors.
377
964260
2000
i motora.
16:06
So the next time
378
966260
2000
Sljedeći put
16:08
the fly finds itself facing the same odor again,
379
968260
3000
kada mušica bude u dodiru s istim mirisom,
16:11
the connection is strong enough to turn on the motors
380
971260
3000
veza će biti dovoljno jaka da pokrene motor
16:14
and to trigger an evasive maneuver.
381
974260
3000
pomoću kojega će mušica izvesti manevar odstupanja.
16:19
I don't know about you,
382
979260
3000
Ne znam za vas,
16:22
but I find it exhilarating to see
383
982260
3000
ali meni je prilično uzbudljivo gledati
16:25
how vague psychological notions
384
985260
3000
kako nejasni psihološki pojmovi
16:28
evaporate and give rise
385
988260
2000
isparavaju i predaju primat
16:30
to a physical, mechanistic understanding of the mind,
386
990260
3000
fizičkom, mehaničkom razumijevanju uma
16:33
even if it's the mind of the fly.
387
993260
3000
pa makar to bio i um mušice.
16:36
This is one piece of good news.
388
996260
3000
Ovo je samo prva dobra vijest.
16:39
The other piece of good news,
389
999260
2000
Druga je,
16:41
for a scientist at least,
390
1001260
2000
barem za znanstvenike,
16:43
is that much remains to be discovered.
391
1003260
3000
da ostaje još mnogo toga što treba biti otkriveno.
16:46
In the experiments I told you about,
392
1006260
2000
U eksperimentima o kojima sam vam pričao,
16:48
we have lifted the identity of the Critic,
393
1008260
3000
formirali smo identitet Kritičara,
16:51
but we still have no idea
394
1011260
2000
ali i dalje nemamo pojma
16:53
how the Critic does its job.
395
1013260
2000
kako on radi svoj posao.
16:55
Come to think of it, knowing when you're wrong
396
1015260
2000
Razmislite malo o tome, dosta je teško znati da si u krivu,
16:57
without a teacher, or your mother, telling you,
397
1017260
3000
ako ti to netko, učiteljica
17:00
is a very hard problem.
398
1020260
2000
ili majka, ne kaže.
17:02
There are some ideas in computer science
399
1022260
2000
Postoje neke ideje u informatičkoj znanosti
17:04
and in artificial intelligence
400
1024260
2000
i umjetnoj inteligenciji
17:06
as to how this might be done,
401
1026260
2000
kako je sve to moguće,
17:08
but we still haven't solved
402
1028260
2000
ali zasad nismo riješili
17:10
a single example
403
1030260
2000
niti jedan primjer
17:12
of how intelligent behavior
404
1032260
3000
kako se inteligentno ponašanje
17:15
springs from the physical interactions
405
1035260
2000
aktivira iz fizičkih interakcija
17:17
in living matter.
406
1037260
2000
u živoj materiji.
17:19
I think we'll get there in the not too distant future.
407
1039260
3000
Mislim da ćemo do toga doći u bliskoj budućnosti.
17:22
Thank you.
408
1042260
2000
Hvala.
17:24
(Applause)
409
1044260
4000
(Pljesak)
O ovoj web stranici

Ova stranica će vas upoznati s YouTube videozapisima koji su korisni za učenje engleskog jezika. Vidjet ćete lekcije engleskog koje vode vrhunski profesori iz cijelog svijeta. Dvaput kliknite na engleske titlove prikazane na svakoj video stranici da biste reproducirali video s tog mjesta. Titlovi se pomiču sinkronizirano s reprodukcijom videozapisa. Ako imate bilo kakvih komentara ili zahtjeva, obratite nam se putem ovog obrasca za kontakt.

https://forms.gle/WvT1wiN1qDtmnspy7