Sebastian Deterding: What your designs say about you

33,106 views ・ 2012-05-31

TED


Dvaput kliknite na engleske titlove ispod za reprodukciju videozapisa.

Prevoditelj: Senzos Osijek Recezent: Suzana Baric
00:15
We are today talking about moral persuasion:
0
15651
2586
Danas govorimo o moralnom uvjeravanju.
00:18
What is moral and immoral in trying to change people's behaviors
1
18261
3982
Što je moralno i nemoralno
u pokušaju da se promijeni ljudsko ponašanje
00:22
by using technology and using design?
2
22267
2453
korištenjem tehnologije i dizajna?
00:24
And I don't know what you expect,
3
24744
1832
Ne znam što očekujete,
00:26
but when I was thinking about that issue,
4
26600
1953
ali kad sam razmišljao o tom problemu,
00:28
I early on realized what I'm not able to give you are answers.
5
28577
4039
rano sam shvatio
da su ono što vam ne mogu dati odgovori.
00:33
I'm not able to tell you what is moral or immoral,
6
33203
2771
Nisam u mogućnosti reći vam što je moralno, a što nemoralno
00:35
because we're living in a pluralist society.
7
35998
2547
jer živimo u pluralističkom društvu.
00:38
My values can be radically different from your values,
8
38569
4242
Moje vrijednosti mogu biti radikalno drugačije
od vaših vrijednosti.
00:42
which means that what I consider moral or immoral based on that
9
42835
3177
Što znači da ono što ja smatram moralnim ili nemoralnim prema tome
00:46
might not necessarily be what you consider moral or immoral.
10
46036
3612
ne mora nužno biti ono što vi smatrate moralnim ili nemoralnim.
Također, shvatio sam da postoji jedna stvar koju vam mogu dati.
00:50
But I also realized there is one thing that I could give you,
11
50029
3214
A to je što je ovaj tip iza mene dao svijetu –
00:53
and that is what this guy behind me gave the world --
12
53267
2533
00:55
Socrates.
13
55824
1150
Sokrat.
00:56
It is questions.
14
56998
1395
To su pitanja.
00:58
What I can do and what I would like to do with you
15
58417
2633
Što mogu i što bih htio napraviti s vama jest dati vam,
kao to početno pitanje,
01:01
is give you, like that initial question,
16
61074
1945
skupinu pitanja
01:03
a set of questions to figure out for yourselves,
17
63043
3410
kako biste sami shvatili,
01:06
layer by layer, like peeling an onion,
18
66477
3641
sloj po sloj,
poput guljenja luka,
da dođete do srži onoga u što vjerujete
01:10
getting at the core of what you believe is moral or immoral persuasion.
19
70142
5020
da je moralno ili nemoralno uvjeravanje.
01:15
And I'd like to do that with a couple of examples of technologies
20
75559
4041
To bih volio napraviti s nekoliko primjera
tehnologije u kojoj su ljudi koristili elemente igre
01:19
where people have used game elements to get people to do things.
21
79624
4980
kako bi natjerali ljude da nešto učine.
Tako je ovo prvo, vrlo jednostavno, vrlo jasno pitanje
01:25
So it's at first a very simple, very obvious question
22
85343
3040
01:28
I would like to give you:
23
88407
1196
koje vam želim dati:
01:29
What are your intentions if you are designing something?
24
89627
2687
Koje su vaše namjere ako dizajnirate nešto?
01:32
And obviously, intentions are not the only thing,
25
92668
3373
I naravno da namjere nisu jedina stvar
pa evo još jedan primjer jedne od ovih aplikacija.
01:36
so here is another example for one of these applications.
26
96065
3178
01:39
There are a couple of these kinds of Eco dashboards right now --
27
99267
3087
Postoji par tih vrsta eko-kontrolnih ploča u ovom trenu –
kontrolne ploče ugrađene su u aute
01:42
dashboards built into cars --
28
102378
1460
01:43
which try to motivate you to drive more fuel-efficiently.
29
103862
2832
i pokušavaju vas motivirati da vozite sa što manjom potrošnjom goriva.
01:46
This here is Nissan's MyLeaf,
30
106718
1773
Ovo ovdje je Nissanov MyLeaf,
01:48
where your driving behavior is compared with the driving behavior
31
108515
3064
koji uspoređuje vaše ponašanje u vožnji
s ponašanjem drugih ljudi
01:51
of other people,
32
111603
1151
01:52
so you can compete for who drives a route the most fuel-efficiently.
33
112778
3277
kako biste se mogli natjecati tko vozi
s najmanjom potrošnjom.
Ove stvari su jako učinkovite, ispada
01:56
And these things are very effective, it turns out --
34
116079
2477
da su toliko učinkovite da motiviraju ljude
01:58
so effective that they motivate people to engage in unsafe driving behaviors,
35
118580
3939
da se upuste u nesigurno ponašanje tijekom vožnje –
02:02
like not stopping at a red light,
36
122543
1762
poput nezaustavljanja na crvenom semaforskom svjetlu.
02:04
because that way you have to stop and restart the engine,
37
124329
2715
Jer na taj način moraju stati i ponovno pokrenuti motor,
a to bi potrošilo dosta goriva, zar ne?
02:07
and that would use quite some fuel, wouldn't it?
38
127068
2723
02:10
So despite this being a very well-intended application,
39
130338
4409
Tako da, iako je ovo vrlo dobronamjerna aplikacija,
02:14
obviously there was a side effect of that.
40
134771
2375
naravno da je bilo nuspojava.
02:17
Here's another example for one of these side effects.
41
137170
2548
Evo još jednog primjera tih nuspojava.
02:19
Commendable: a site that allows parents to give their kids little badges
42
139742
4784
Pohvalno:
stranica koja dopušta roditeljima da daju svojoj djeci male značke
02:24
for doing the things that parents want their kids to do,
43
144550
2658
za stvari koje roditelji žele da njihova djeca rade –
poput vezanja cipela.
02:27
like tying their shoes.
44
147232
1316
02:28
And at first that sounds very nice,
45
148572
2244
Isprva to zvuči jako zgodno,
02:30
very benign, well-intended.
46
150840
2150
vrlo benigno, dobronamjerno.
Ali ispada da ako ispitate istraživanja ljudskog stanja uma,
02:33
But it turns out, if you look into research on people's mindset,
47
153014
3770
02:36
caring about outcomes,
48
156808
1487
brigu o ishodima,
02:38
caring about public recognition,
49
158319
1773
brigu o javnom priznanju,
02:40
caring about these kinds of public tokens of recognition
50
160116
3861
brigu o ovakvom obliku javnih znakova priznanja,
nije nužno jako korisno
02:44
is not necessarily very helpful
51
164001
1994
02:46
for your long-term psychological well-being.
52
166019
2307
za vaše dugoročno psihološko blagostanje.
02:48
It's better if you care about learning something.
53
168350
2676
Bolje je ako vam je stalo do toga da naučite nešto.
02:51
It's better when you care about yourself
54
171050
1905
Bolje je kada se brinete za sebe,
02:52
than how you appear in front of other people.
55
172979
2574
umjesto kako izgledate drugim ljudima.
Tako da ta vrsta motivacijskog oruđa koje se koristi
02:56
So that kind of motivational tool that is used actually, in and of itself,
56
176021
5041
sama po sebi
ima dugotrajne nuspojave
03:01
has a long-term side effect,
57
181086
1908
03:03
in that every time we use a technology
58
183018
1810
u tome što svaki put kad koristimo tehnologiju
03:04
that uses something like public recognition or status,
59
184852
3174
koja koristi neku vrstu javnog priznanja ili statusa,
03:08
we're actually positively endorsing this
60
188050
2376
zapravo pozitivno potičemo to
03:10
as a good and normal thing to care about --
61
190450
3410
kao dobru i normalnu stvar o kojoj treba voditi brigu –
03:13
that way, possibly having a detrimental effect
62
193884
2864
i na taj način postoji mogućnost štetnog učinka
03:16
on the long-term psychological well-being of ourselves as a culture.
63
196772
3855
na dugotrajno psihološko blagostanje nas samih kao kulture.
03:20
So that's a second, very obvious question:
64
200651
2644
To je drugo, vrlo očito pitanje:
03:23
What are the effects of what you're doing --
65
203319
2311
Koji su učinci onog što radimo?
03:25
the effects you're having with the device, like less fuel,
66
205654
4124
Učinci koje imate s uređajem,
kao što je manja potrošnja goriva
03:29
as well as the effects of the actual tools you're using
67
209802
2705
te učinci stvarnih oruđa koja koristite
03:32
to get people to do things --
68
212531
1674
kako biste natjerali ljude da učine nešto –
03:34
public recognition?
69
214229
1571
javno priznanje.
03:35
Now is that all -- intention, effect?
70
215824
2921
Je li to sve – namjera, učinak?
03:38
Well, there are some technologies which obviously combine both.
71
218769
3134
Postoje neke tehnologije
koje očito spajaju oboje.
03:41
Both good long-term and short-term effects
72
221927
2787
I dobre dugotrajne i kratkotrajne učinke
03:44
and a positive intention like Fred Stutzman's "Freedom,"
73
224738
2809
i pozitivnu namjeru poput Slobode Freda Stutzmana,
03:47
where the whole point of that application is --
74
227571
2207
gdje je cijela bit aplikacije,
pa, obično smo bombardirani
03:49
well, we're usually so bombarded with constant requests by other people,
75
229802
3725
pozivima i zahtjevima drugih ljudi,
03:53
with this device,
76
233551
1156
a s ovim uređajem možete ugasiti povezivost računala s internetom
03:54
you can shut off the Internet connectivity of your PC of choice
77
234731
3480
na predodređeno vrijeme
03:58
for a pre-set amount of time,
78
238235
1448
03:59
to actually get some work done.
79
239707
1971
kako biste stigli čak napraviti i nešto posla.
04:01
And I think most of us will agree that's something well-intended,
80
241702
3151
Mislim da će se većina nas složiti
da je to nešto dobronamjerno
04:04
and also has good consequences.
81
244877
2220
i ima pozitivne posljedice.
Riječima Michela Foucaulta,
04:07
In the words of Michel Foucault,
82
247121
1646
04:08
it is a "technology of the self."
83
248791
1940
"To je tehnologija samog sebe."
04:10
It is a technology that empowers the individual
84
250755
2837
To je tehnologija koja pojedincu daje moć
04:13
to determine its own life course,
85
253616
1815
odrediti svoj životni smjer,
04:15
to shape itself.
86
255455
1519
oblikovati sam sebe.
04:17
But the problem is, as Foucault points out,
87
257410
2984
Ali problem je,
kao što Foucault ističe,
04:20
that every technology of the self
88
260418
1785
što svaka tehnologija sebe
04:22
has a technology of domination as its flip side.
89
262227
3445
ima tehnologiju dominacije kao svoju drugu stranu.
04:25
As you see in today's modern liberal democracies,
90
265696
4603
Kao što vidite u današnjim modernim liberalnim demokracijama,
društvo, država,
04:30
the society, the state, not only allows us to determine our self,
91
270323
4683
dopušta nam ne samo da se sami odredimo,
04:35
to shape our self,
92
275030
1151
04:36
it also demands it of us.
93
276205
1991
da se oblikujemo, već to od nas i zahtijeva.
Zahtijeva da se usavršimo,
04:38
It demands that we optimize ourselves,
94
278220
1961
04:40
that we control ourselves,
95
280205
1840
da se kontroliramo,
da kontinuirano upravljamo nama samima
04:42
that we self-manage continuously,
96
282069
2711
04:44
because that's the only way in which such a liberal society works.
97
284804
3893
jer je to jedini način
na koji takvo liberalno društvo funkcionira.
04:48
These technologies want us to stay in the game
98
288721
4268
Ove tehnologije žele da ostanemo u igri
04:53
that society has devised for us.
99
293013
2773
koju je društvo smislilo za nas.
04:55
They want us to fit in even better.
100
295810
2287
Žele da se još bolje uklopimo.
04:58
They want us to optimize ourselves to fit in.
101
298121
2578
Žele da se usavršimo kako bismo se uklopili.
Ne želim reći da je to nužno loša stvar.
05:01
Now, I don't say that is necessarily a bad thing;
102
301628
3079
Samo mislim da nas ovaj primjer
05:05
I just think that this example points us to a general realization,
103
305293
4328
upućuje na opću spoznaju
05:09
and that is: no matter what technology or design you look at,
104
309645
3803
da bez obzira koju tehnologiju ili dizajn pogledate,
05:13
even something we consider as well-intended
105
313472
3021
čak i nešto što smatramo i dobronamjernim i da ima dobre učinke –
05:16
and as good in its effects as Stutzman's Freedom,
106
316517
2966
poput Stutzmanove Slobode –
05:19
comes with certain values embedded in it.
107
319507
2707
dolazi s određenim vrijednostima usađenima u sebi.
05:22
And we can question these values.
108
322238
1936
Možemo ispitati te vrijednosti.
Možemo pitati: Je li dobro
05:24
We can question: Is it a good thing
109
324198
1944
što svi mi kontinuirano usavršavamo sami sebe
05:26
that all of us continuously self-optimize ourselves
110
326166
3484
05:29
to fit better into that society?
111
329674
2011
kako bismo se bolje uklopili u to društvo?
05:31
Or to give you another example:
112
331709
1492
Ili da vam dam drugi primjer,
05:33
What about a piece of persuasive technology
113
333225
2476
što je s dijelom uvjeravajuće tehnologije
koja uvjerava muslimanke da nose svoje marame?
05:35
that convinces Muslim women to wear their headscarves?
114
335725
3191
05:38
Is that a good or a bad technology
115
338940
2052
Je li to dobra ili loša tehnologija
u svojim namjerama ili učincima?
05:41
in its intentions or in its effects?
116
341016
2563
05:43
Well, that basically depends on the kind of values you bring to bear
117
343603
4254
To ovisi
o vrsti vrijednosti koju možete primijeniti
05:47
to make these kinds of judgments.
118
347881
2237
kako biste donijeli ovakve procjene.
To je treće pitanje:
05:50
So that's a third question:
119
350142
1528
05:51
What values do you use to judge?
120
351694
1528
Koje vrijednosti koristite pri procjeni?
05:53
And speaking of values:
121
353848
1341
I govoreći o vrijednostima,
05:55
I've noticed that in the discussion about moral persuasion online
122
355213
3356
primijetio sam da u raspravi o moralnoj uvjerljivosti online
05:58
and when I'm talking with people,
123
358593
1637
i kad pričam s ljudima,
06:00
more often than not, there is a weird bias.
124
360254
2667
često postoji čudna subjektivnost.
I ta subjektivnost je da pitamo,
06:03
And that bias is that we're asking:
125
363463
2896
06:06
Is this or that "still" ethical?
126
366383
2813
je li ovo ili ono "još uvijek" etično?
06:09
Is it "still" permissible?
127
369220
2650
Je li "još uvijek" dopušteno?
06:11
We're asking things like:
128
371894
1198
Pitamo stvari poput:
Je li ovaj obrazac za Oxfam donaciju –
06:13
Is this Oxfam donation form,
129
373116
2189
06:15
where the regular monthly donation is the preset default,
130
375329
3048
gdje je redovna mjesečna donacija unaprijed zadana
06:18
and people, maybe without intending it,
131
378401
2079
i ljudi su, možda bez namjere,
06:20
are encouraged or nudged into giving a regular donation
132
380504
3812
na taj način ohrabreni ili potaknuti
dati redovnu donaciju umjesto jednokratne –
06:24
instead of a one-time donation,
133
384340
1489
06:25
is that "still' permissible?
134
385853
1343
je li to dopušteno?
Je li još uvijek etično?
06:27
Is it "still" ethical?
135
387220
1363
06:28
We're fishing at the low end.
136
388607
1479
Na samom smo dnu.
Ali, pitanje
06:30
But in fact, that question, "Is it 'still' ethical?"
137
390879
2474
"Je li to još etično?"
je samo jedan način gledanja na etiku.
06:33
is just one way of looking at ethics.
138
393377
1781
06:35
Because if you look at the beginning of ethics in Western culture,
139
395182
4892
Ako pogledate na početak etike
u zapadnoj kulturi,
vidite ideju vrlo drugačiju
06:40
you see a very different idea of what ethics also could be.
140
400098
3532
od onoga što bi etika još mogla biti.
06:43
For Aristotle, ethics was not about the question,
141
403970
3997
Za Aristotela etika nije bila o pitanju
06:47
"Is that still good, or is it bad?"
142
407991
2272
"Je li to još uvijek dobro ili je loše?"
06:50
Ethics was about the question of how to live life well.
143
410287
3428
Etika je bila o pitanju kako dobro živjeti život.
On je to stavio u riječ "arete"
06:54
And he put that in the word "arête,"
144
414216
2181
koju sa starogrčkog prevodimo kao "vrlina".
06:56
which we, from [Ancient Greek], translate as "virtue."
145
416421
2755
Ali zapravo znači izvrsnost.
06:59
But really, it means "excellence."
146
419200
1640
07:00
It means living up to your own full potential as a human being.
147
420864
5431
Znači živjeti ispunjavajući vlastiti potencijal
ljudskog bića.
07:06
And that is an idea that, I think,
148
426937
1656
To je ideja koju je,
mislim, Paul Richard Buchanan odlično oblikovao u nedavnom eseju
07:08
Paul Richard Buchanan put nicely in a recent essay,
149
428617
2697
07:11
where he said, "Products are vivid arguments
150
431338
2143
u kojem kaže: "Proizvodi su živahni argumenti
07:13
about how we should live our lives."
151
433505
2123
o tome kako bismo trebali živjeti svoje živote."
Naši dizajni nisu etički ili neetički
07:16
Our designs are not ethical or unethical
152
436086
2577
07:18
in that they're using ethical or unethical means of persuading us.
153
438687
4590
u tome što koriste etičke ili neetičke mjere kako bi nas uvjerili.
07:23
They have a moral component
154
443661
1570
Imaju moralni sastojak
u obliku vizije i težnje dobrom životu
07:25
just in the kind of vision and the aspiration of the good life
155
445255
4227
kojeg nam predstavljaju.
07:29
that they present to us.
156
449506
1348
07:31
And if you look into the designed environment around us
157
451441
3503
Ako pogledate u oblikovan okoliš oko nas
07:34
with that kind of lens,
158
454968
1172
s takvom lećom,
07:36
asking, "What is the vision of the good life
159
456164
2453
pitajući: "Što je vizija dobrog života
07:38
that our products, our design, present to us?",
160
458641
2738
koju naši proizvodi, naš dizajn, nama predstavljaju?",
07:41
then you often get the shivers,
161
461403
2276
onda vas često prođu trnci
07:43
because of how little we expect of each other,
162
463703
2328
zbog toga kako malo očekujemo jedni od drugih,
07:46
of how little we actually seem to expect of our life,
163
466055
3890
kako malo očekujemo
od našeg života i toga kako dobar život izgleda.
07:49
and what the good life looks like.
164
469969
2034
To je četvrto pitanje s kojim bih vas ostavio:
07:53
So that's a fourth question I'd like to leave you with:
165
473110
3023
07:56
What vision of the good life do your designs convey?
166
476157
4362
Koju viziju dobrog života
naši dizajni prenose?
08:01
And speaking of design,
167
481249
1372
I govoreći o dizajnu,
08:02
you'll notice that I already broadened the discussion,
168
482645
4190
možete primjetiti da sam već proširio raspravu
08:06
because it's not just persuasive technology that we're talking about here,
169
486859
4444
jer to nije samo uvjeravajuća tehnologija o kojoj pričamo ovdje,
to je svaki dio dizajna koji pustimo u svijet.
08:11
it's any piece of design that we put out here in the world.
170
491327
4077
08:15
I don't know whether you know
171
495428
1400
Ne znam znate li
08:16
the great communication researcher Paul Watzlawick who, back in the '60s,
172
496852
3555
velikog istraživača komunikacije Paula Watzlawicka
koji je u 60-ima ustvrdio
da ne možemo nekomunicirati.
08:20
made the argument that we cannot not communicate.
173
500431
2511
Čak i ako odlučimo biti tihi,
08:22
Even if we choose to be silent, we chose to be silent,
174
502966
2601
odlučili smo biti tihi. Komuniciramo nešto svojom odlukom da budemo tihi.
08:25
and we're communicating something by choosing to be silent.
175
505591
2956
08:28
And in the same way that we cannot not communicate,
176
508571
2735
Isto kao što ne možemo nekomunicirati,
ne možemo ne uvjeravati.
08:31
we cannot not persuade:
177
511330
1531
08:32
whatever we do or refrain from doing,
178
512885
2011
Što god da učinimo ili se suzdržavamo učiniti,
08:34
whatever we put out there as a piece of design, into the world,
179
514920
4365
što god donesemo kao dio dizajna
u svijet
ima uvjeravajući element.
08:39
has a persuasive component.
180
519309
2047
08:41
It tries to affect people.
181
521380
1873
Pokušava utjecati na ljude.
08:43
It puts a certain vision of the good life out there in front of us,
182
523277
3747
Stavlja određenu viziju dobrog života
ispred nas.
To je ono što Peter-Paul Verbeek,
08:47
which is what Peter-Paul Verbeek,
183
527048
1625
08:48
the Dutch philosopher of technology, says.
184
528697
2716
nizozemski filozof tehnologije, kaže.
Bez obzira na to namjeravamo li mi to ili ne,
08:51
No matter whether we as designers intend it or not,
185
531437
3948
08:55
we materialize morality.
186
535409
2142
kao dizajneri mi matezijaliziramo moral.
08:57
We make certain things harder and easier to do.
187
537575
2803
Određene stvari činimo težima ili lakšima za napraviti.
09:00
We organize the existence of people.
188
540402
2211
Organiziramo postojanje ljudi.
09:02
We put a certain vision
189
542637
1151
Stavljamo određenu viziju toga što je dobro ili loše
09:03
of what good or bad or normal or usual is
190
543812
3404
ili normalno ili uobičajeno pred ljude
09:07
in front of people,
191
547240
1151
09:08
by everything we put out there in the world.
192
548415
2400
svime što donesemo u svijet.
Čak i nešto bezopasno poput kompleta školskih stolica
09:11
Even something as innocuous as a set of school chairs
193
551247
3086
09:14
is a persuasive technology,
194
554357
2047
je uvjeravajuća tehnologija.
Zato što predstavlja i materijalizira
09:16
because it presents and materializes a certain vision of the good life --
195
556428
4690
određenu viziju dobrog života –
dobrog života u kojem poučavanje i učenje i slušanje ovisi
09:21
a good life in which teaching and learning and listening
196
561142
2857
09:24
is about one person teaching, the others listening;
197
564023
3099
o jednoj osobi koja poučava, drugi je slušaju,
u kojem se sjedi kad se uči,
09:27
in which it is about learning-is-done-while-sitting;
198
567146
4053
u kojem učite za sebe,
09:31
in which you learn for yourself;
199
571223
1595
09:32
in which you're not supposed to change these rules,
200
572842
2420
u kojem ne biste trebali mijenjati ova pravila
09:35
because the chairs are fixed to the ground.
201
575286
2447
jer su stolice prikovane za tlo.
09:38
And even something as innocuous as a single-design chair,
202
578888
2911
Čak i nešto bezopasno poput jednog dizajna stolice –
09:41
like this one by Arne Jacobsen,
203
581823
1572
poput ove Arne Jacobsen –
09:43
is a persuasive technology,
204
583419
1776
je uvjeravajuća tehnologija.
Zato što komunicira ideju dobrog života.
09:45
because, again, it communicates an idea of the good life:
205
585219
3043
09:48
a good life -- a life that you, as a designer, consent to by saying,
206
588735
4925
Dobar život –
život na koji kažete da pristajete kao dizajner
govoreći: "U dobrom životu,
09:53
"In a good life, goods are produced as sustainably or unsustainably
207
593684
3507
roba proizvodi održivo ili neodrživo kao ova stolica.
09:57
as this chair.
208
597215
1611
09:58
Workers are treated as well or as badly
209
598850
1977
Prema radnicima se postupa jednako dobro ili loše
10:00
as the workers were treated that built that chair."
210
600851
2572
kao s radnicima koji su stvorili tu stolicu.".
10:03
The good life is a life where design is important
211
603762
2301
Dobar život je život gdje je dizajn važan
jer je netko očito odvojio vrijeme i potrošio novac
10:06
because somebody obviously took the time and spent the money
212
606087
2921
na taj oblik dobrodizajnirane stolice,
10:09
for that kind of well-designed chair;
213
609032
1784
10:10
where tradition is important,
214
610840
1404
gdje je tradicija važna
jer je to tradicionalni klasik
10:12
because this is a traditional classic and someone cared about this;
215
612268
3166
i netko je vodio brigu o tome,
i gdje je nešto poput izražene potrošnje,
10:15
and where there is something as conspicuous consumption,
216
615458
2690
gdje je u redu i normalno
10:18
where it is OK and normal to spend a humongous amount of money
217
618172
2956
potrošiti ogromne količine novca na takvu stolicu
10:21
on such a chair,
218
621152
1151
da biste pokazali drugim ljudima koji je vaš socijalni status.
10:22
to signal to other people what your social status is.
219
622327
2573
10:24
So these are the kinds of layers, the kinds of questions
220
624924
3317
Ovo su vrste slojeva, vrste pitanja
10:28
I wanted to lead you through today;
221
628265
1970
koja sam htio da vas vode danas –
pitanje: Koje su namjere
10:30
the question of: What are the intentions that you bring to bear
222
630259
3034
koje primjenjujete kad dizajnirate nešto?
10:33
when you're designing something?
223
633317
1560
10:34
What are the effects, intended and unintended, that you're having?
224
634901
3252
Koji su učinci, namjerni ili nenamjerni, koje imate?
Koje su vrijednosti koje koristite
10:38
What are the values you're using to judge those?
225
638177
2801
pri procjeni?
Koje su vrline, težnje
10:41
What are the virtues, the aspirations
226
641002
1973
10:42
that you're actually expressing in that?
227
642999
2028
koje zapravo izražavate time?
10:45
And how does that apply,
228
645400
1857
I kako se to primjenjuje,
10:47
not just to persuasive technology,
229
647281
1986
ne samo kod uvjeravajuće tehnologije,
10:49
but to everything you design?
230
649291
2048
već kod svega što dizajnirate?
10:51
Do we stop there?
231
651912
1291
Stajemo li ovdje?
10:53
I don't think so.
232
653815
1167
Ne bih rekao.
10:55
I think that all of these things are eventually informed
233
655291
4438
Mislim da su sve ove stvari konačno informirane
10:59
by the core of all of this,
234
659753
1423
biti svega ovog –
11:01
and this is nothing but life itself.
235
661200
3091
i ovo nije ništa nego sam život.
11:04
Why, when the question of what the good life is
236
664594
2717
Zašto, kad pitanje što je dobar život
11:07
informs everything that we design,
237
667335
2339
informira sve što dizajniramo,
11:09
should we stop at design and not ask ourselves:
238
669698
2794
trebamo li stati na dizajnu i ne pitati se
11:12
How does it apply to our own life?
239
672516
1990
kako to primjeniti u vlastitom životu?
11:14
"Why should the lamp or the house be an art object,
240
674881
2712
"Zašto bi svjetiljka ili kuća trebale biti umjetnički predmet,
11:17
but not our life?"
241
677617
1176
ali ne i naš život?",
11:18
as Michel Foucault puts it.
242
678817
1483
kako to kaže Michel Foucault.
11:20
Just to give you a practical example of Buster Benson.
243
680696
3611
Samo da vam dam praktični primjer Bustera Bensona.
11:24
This is Buster setting up a pull-up machine
244
684331
2258
Ovdje Buster postavlja prečku za zgibove
11:26
at the office of his new start-up, Habit Labs,
245
686613
2612
u uredu svog novog Habit Labsa
gdje pokušavaju stvoriti druge aplikacije
11:29
where they're trying to build other applications like "Health Month"
246
689249
3215
poput Mjeseca zdravlja za ljude.
11:32
for people.
247
692488
1158
I zašto gradi nešto ovakvo?
11:33
And why is he building a thing like this?
248
693670
1962
11:35
Well, here is the set of axioms
249
695656
2033
Postoji sklop aksioma
11:37
that Habit Labs, Buster's start-up, put up for themselves
250
697713
3398
koje je Habit Labs, Busterova nova tvrtka, postavila za sebe
kako bi radili zajedno kao tim
11:41
on how they wanted to work together as a team
251
701135
2705
11:43
when they're building these applications --
252
703864
2029
pri smišljanju ovih aplikacija –
11:45
a set of moral principles they set themselves
253
705917
2189
sklop moralnih principa koje sami postavljaju
za zajednički rad –
11:48
for working together --
254
708130
1361
11:49
one of them being,
255
709515
1238
i jedan od njih je:
11:50
"We take care of our own health and manage our own burnout."
256
710777
3087
"Vodimo brigu o našem zdravlju i upravljamo vlastitim izgaranjem."
Jer na kraju, kako možemo pitati sebe
11:54
Because ultimately, how can you ask yourselves
257
714221
3500
11:57
and how can you find an answer on what vision of the good life
258
717745
3930
i kako možemo naći odgovor
na to što je vizija dobrog života
12:01
you want to convey and create with your designs
259
721699
3169
koju želimo prenijeti i stvoriti svojim dizajnima
12:04
without asking the question:
260
724892
1730
bez da se pitamo
12:06
What vision of the good life do you yourself want to live?
261
726646
3832
koja je vizija dobrog života
koji sami želimo živjeti?
I s tim završavam, hvala vam.
12:11
And with that,
262
731018
1174
12:12
I thank you.
263
732945
1412
12:14
(Applause)
264
734381
4156
(Pljesak)
O ovoj web stranici

Ova stranica će vas upoznati s YouTube videozapisima koji su korisni za učenje engleskog jezika. Vidjet ćete lekcije engleskog koje vode vrhunski profesori iz cijelog svijeta. Dvaput kliknite na engleske titlove prikazane na svakoj video stranici da biste reproducirali video s tog mjesta. Titlovi se pomiču sinkronizirano s reprodukcijom videozapisa. Ako imate bilo kakvih komentara ili zahtjeva, obratite nam se putem ovog obrasca za kontakt.

https://forms.gle/WvT1wiN1qDtmnspy7