Beware online "filter bubbles" | Eli Pariser

1,585,118 views ・ 2011-05-02

TED


Mesedez, egin klik bikoitza beheko ingelesezko azpitituluetan bideoa erreproduzitzeko.

Translator: Jose Luis Ulazia Reviewer: olatz mercader
00:15
Mark Zuckerberg,
0
15260
2000
Mark Zuckerberg,
00:17
a journalist was asking him a question about the news feed.
1
17260
3000
kazetari bat ari zitzaion galdezka albiste banatze zerbitzuaz.
00:20
And the journalist was asking him,
2
20260
2000
Eta kazetariak zion,
00:22
"Why is this so important?"
3
22260
2000
Zergatik da hau hain garrantzitsua?
00:24
And Zuckerberg said,
4
24260
2000
Eta Zuckerberg-ek zion,
00:26
"A squirrel dying in your front yard
5
26260
2000
"Zure baratzean katagorri bat hil zorian egotea
00:28
may be more relevant to your interests right now
6
28260
3000
garrantzitsuago izan daiteke zuretzako
00:31
than people dying in Africa."
7
31260
3000
Afrikan jendea hiltzen egotea baino."
00:34
And I want to talk about
8
34260
2000
Ba hortaz hitzegin nahi dut,
00:36
what a Web based on that idea of relevance might look like.
9
36260
3000
nolakoa izan daiteken garrantziaren ideiari lotutako Web-a.
00:40
So when I was growing up
10
40260
2000
Ni hazi nintzen lekuan
00:42
in a really rural area in Maine,
11
42260
2000
Maine-ko nekazal ingurune batean,
00:44
the Internet meant something very different to me.
12
44260
3000
Internetak oso bestelako gauza bat susposatzen zuen.
00:47
It meant a connection to the world.
13
47260
2000
Munduarekin lotura bat zen.
00:49
It meant something that would connect us all together.
14
49260
3000
Denok elkarrekin konektatzen gintuen zerbait.
00:52
And I was sure that it was going to be great for democracy
15
52260
3000
Eta ziur nengoen onuragarria izango zela demokraziarentzat
00:55
and for our society.
16
55260
3000
eta gure gizartearentzat.
00:58
But there's this shift
17
58260
2000
Baina aldaketa egon da
01:00
in how information is flowing online,
18
60260
2000
online-ko informazio fluxuaren baitan,
01:02
and it's invisible.
19
62260
3000
eta ikustezina da.
01:05
And if we don't pay attention to it,
20
65260
2000
Eta ez badiogu arretarik jartzen,
01:07
it could be a real problem.
21
67260
3000
arazo erreal bat izan daiteke.
01:10
So I first noticed this in a place I spend a lot of time --
22
70260
3000
Hortaz konturatu nintzenean ohiko lekuan nengoen --
01:13
my Facebook page.
23
73260
2000
nire Facebook-eko orrian.
01:15
I'm progressive, politically -- big surprise --
24
75260
3000
Ni progresista naiz, politikoki -- sorpresa handia --
01:18
but I've always gone out of my way to meet conservatives.
25
78260
2000
baina beti egon naiz konserbadoreen ideiak entzuteko prest.
01:20
I like hearing what they're thinking about;
26
80260
2000
Gustoko dut zer dioten entzutea;
01:22
I like seeing what they link to;
27
82260
2000
Atsegin dut ikustea zer lotura dituzten;
01:24
I like learning a thing or two.
28
84260
2000
Eta gauza bat edo beste ikastea.
01:26
And so I was surprised when I noticed one day
29
86260
3000
Beraz harrrituta gelditu nintzen egun batez ohartzean
01:29
that the conservatives had disappeared from my Facebook feed.
30
89260
3000
kontserbadoreak nire Facebook-eko albisteetatik desagertu zirela.
01:33
And what it turned out was going on
31
93260
2000
Eta gertatu zen
01:35
was that Facebook was looking at which links I clicked on,
32
95260
4000
Facebook begira neukala zein loturatan klikatzen nuen,
01:39
and it was noticing that, actually,
33
99260
2000
eta ohartu zen, gaur egun,
01:41
I was clicking more on my liberal friends' links
34
101260
2000
maizago klikatzen nituela nire lagun liberalen estekak
01:43
than on my conservative friends' links.
35
103260
3000
nire lagun kontserbadoreenak baino.
01:46
And without consulting me about it,
36
106260
2000
Eta niri kontsulta egin gabe
01:48
it had edited them out.
37
108260
2000
handik ezabatu egin zituen.
01:50
They disappeared.
38
110260
3000
Desargertu egin ziren.
01:54
So Facebook isn't the only place
39
114260
2000
Baina Facebook ez da bakarra erabiltzen
01:56
that's doing this kind of invisible, algorithmic
40
116260
2000
ikustezinak diren era honetako algoritmikoak
01:58
editing of the Web.
41
118260
3000
Weba editatzeko.
02:01
Google's doing it too.
42
121260
2000
Google ere horretan ari da.
02:03
If I search for something, and you search for something,
43
123260
3000
Nik zerbait bilatzen badut, eta zuk zerbait bilatzen baduzu,
02:06
even right now at the very same time,
44
126260
2000
oraitxe bertan, momentu berean,
02:08
we may get very different search results.
45
128260
3000
oso emaitza desberdinak lortu ditzakegu.
02:11
Even if you're logged out, one engineer told me,
46
131260
3000
Baita logeatzen ez bagara ere, ingeniari batek esan zidanez,
02:14
there are 57 signals
47
134260
2000
57 seinale ditu
02:16
that Google looks at --
48
136260
3000
Google-ek kontutan--
02:19
everything from what kind of computer you're on
49
139260
3000
hasi zer nolako ordenagailurekin ari zaren
02:22
to what kind of browser you're using
50
142260
2000
edo eta nolako nabigatzailea erabiltzen duzun
02:24
to where you're located --
51
144260
2000
edo nondik ari zaren--
02:26
that it uses to personally tailor your query results.
52
146260
3000
erabiltzen ditu zure bilaketaren emaitzak lortzeko.
02:29
Think about it for a second:
53
149260
2000
Pentsatu hortaz segundu batez:
02:31
there is no standard Google anymore.
54
151260
4000
ez dago Google standarrik inoiz gehiago.
02:35
And you know, the funny thing about this is that it's hard to see.
55
155260
3000
Eta kuriosoena da hori ikustea oso zaila dela.
02:38
You can't see how different your search results are
56
158260
2000
Ezin duzu ikusi zure bilaketaren emaitzak zenbateraino diren desberdinak
02:40
from anyone else's.
57
160260
2000
beste edozeinenarekin alderatuz.
02:42
But a couple of weeks ago,
58
162260
2000
Baina orain dela pare bat aste,
02:44
I asked a bunch of friends to Google "Egypt"
59
164260
3000
lagun talde bati eskatu nien "Egypt" Googleatzeko
02:47
and to send me screen shots of what they got.
60
167260
3000
eta bidaltzeko emaitzaren pantaila-irudi bat.
02:50
So here's my friend Scott's screen shot.
61
170260
3000
Hona nire lagun Scott-en irudia.
02:54
And here's my friend Daniel's screen shot.
62
174260
3000
Eta hemen nire lagun Dani-rena.
02:57
When you put them side-by-side,
63
177260
2000
Elkarren ondoan jartzen badira,
02:59
you don't even have to read the links
64
179260
2000
estekak irakurri ere egin beharrik ez dago
03:01
to see how different these two pages are.
65
181260
2000
konturatzeko zeinen desberdinak diren bi orriak.
03:03
But when you do read the links,
66
183260
2000
Baina estekak irakurtzen badituzu,
03:05
it's really quite remarkable.
67
185260
3000
azpimarrztekoa da,
03:09
Daniel didn't get anything about the protests in Egypt at all
68
189260
3000
Danielek ez zuela ezertxo ere lortu Egiptoko protestei buruz
03:12
in his first page of Google results.
69
192260
2000
Google-n emaitzen lehenengo orrian.
03:14
Scott's results were full of them.
70
194260
2000
Scott-en kasuan berriz hartaz beteak zeuden.
03:16
And this was the big story of the day at that time.
71
196260
2000
Eta hura zen garai hartako historia garrantzitsuena.
03:18
That's how different these results are becoming.
72
198260
3000
Halakoxe desberdinak izan daitezke emaitzak.
03:21
So it's not just Google and Facebook either.
73
201260
3000
Eta ez dira bakarrik Google eta Facebook.
03:24
This is something that's sweeping the Web.
74
204260
2000
Hau web-ean erabat gertatzen ari da.
03:26
There are a whole host of companies that are doing this kind of personalization.
75
206260
3000
Empresa asko ari dira era honetako pertsonalizazioak egiten.
03:29
Yahoo News, the biggest news site on the Internet,
76
209260
3000
Yahoo News, Interneteko berri emaile garrantzitsuena,
03:32
is now personalized -- different people get different things.
77
212260
3000
personalizatua dago - jende desberdinak emaitza desberdinak.
03:36
Huffington Post, the Washington Post, the New York Times --
78
216260
3000
Huffington Post, Washington Post, New York Times --
03:39
all flirting with personalization in various ways.
79
219260
3000
denak pertsonalizazioan ari dira neurri batean.
03:42
And this moves us very quickly
80
222260
3000
Eta honek oso azkar garamatza
03:45
toward a world in which
81
225260
2000
mundu batetara non
03:47
the Internet is showing us what it thinks we want to see,
82
227260
4000
Internetak erakusten digun, bere ustez guk ikusi nahi duguna,
03:51
but not necessarily what we need to see.
83
231260
3000
baina ez derrigorrez guk ikusi beharko genukeena.
03:54
As Eric Schmidt said,
84
234260
3000
Eric Schmidt-ek zion bezala,
03:57
"It will be very hard for people to watch or consume something
85
237260
3000
"Oso zaila izango zaio inori zeozer ikusi ala erostea
04:00
that has not in some sense
86
240260
2000
era batera ala bestera
04:02
been tailored for them."
87
242260
3000
beraientzako egokitua izan ez bada."
04:05
So I do think this is a problem.
88
245260
2000
Nire ustez hau arazo bat da.
04:07
And I think, if you take all of these filters together,
89
247260
3000
zeren filtro guziak batera hartzen badira,
04:10
you take all these algorithms,
90
250260
2000
algoritmo guzti hauek batera,
04:12
you get what I call a filter bubble.
91
252260
3000
iragazki bubuila bat lortzen dugu.
04:16
And your filter bubble is your own personal,
92
256260
3000
Eta zure iragazki burbuila, zeure
04:19
unique universe of information
93
259260
2000
informazio unibertso bakarra da
04:21
that you live in online.
94
261260
2000
online bizi duzuna.
04:23
And what's in your filter bubble
95
263260
3000
Eta burbuila hortan dagoena
04:26
depends on who you are, and it depends on what you do.
96
266260
3000
nor zaren eta zer egiten duzun arabera izango da.
04:29
But the thing is that you don't decide what gets in.
97
269260
4000
Baina ez du norberak erabakitzen zer sartzen den bertan.
04:33
And more importantly,
98
273260
2000
Eta garrantzitsuago dena,
04:35
you don't actually see what gets edited out.
99
275260
3000
ez duzu ikusten zer izan den bertatik ezabatua.
04:38
So one of the problems with the filter bubble
100
278260
2000
Burbuila honen arazoetako bat
04:40
was discovered by some researchers at Netflix.
101
280260
3000
Netflix-eko ikertzaile batzuk aurkitu zuten.
04:43
And they were looking at the Netflix queues, and they noticed something kind of funny
102
283260
3000
Netflix-eko ilarei begira hari zirela, gauza kurioso batetaz ohartu ziren
04:46
that a lot of us probably have noticed,
103
286260
2000
gutako asko ere agian ohartuko ginen,
04:48
which is there are some movies
104
288260
2000
alegia, filma batzuk badira
04:50
that just sort of zip right up and out to our houses.
105
290260
3000
gure zerrendatan bapatean azaldu eta desagertzen direnak.
04:53
They enter the queue, they just zip right out.
106
293260
3000
Ilaran sartu bezain pronto irtetzen direnak.
04:56
So "Iron Man" zips right out,
107
296260
2000
Horrela "Iron Man" bapatean kanpora doa,
04:58
and "Waiting for Superman"
108
298260
2000
eta "Waiting Superman"-ek
05:00
can wait for a really long time.
109
300260
2000
luze itxaron beharko.
05:02
What they discovered
110
302260
2000
Aurkitu zuten
05:04
was that in our Netflix queues
111
304260
2000
gure Netflix-eko zerrendatan
05:06
there's this epic struggle going on
112
306260
3000
halako borroka epikoak gerta ohi direla
05:09
between our future aspirational selves
113
309260
3000
gure etorkizuneko aspirazioak batetik
05:12
and our more impulsive present selves.
114
312260
3000
eta gure egunerokoak bestetik.
05:15
You know we all want to be someone
115
315260
2000
Jakina, denok nahiko genuke izan
05:17
who has watched "Rashomon,"
116
317260
2000
"Rashomon" ikusi duen norbait.
05:19
but right now
117
319260
2000
baina oraingoz
05:21
we want to watch "Ace Ventura" for the fourth time.
118
321260
3000
"Ace Ventura" ikusi nahi dugu laugarren aldiz.
05:24
(Laughter)
119
324260
3000
(farreak)
05:27
So the best editing gives us a bit of both.
120
327260
2000
Horrela edizio onenak bietatik zerbait ematen digu.
05:29
It gives us a little bit of Justin Bieber
121
329260
2000
Justin Bieber-etik pixka bat
05:31
and a little bit of Afghanistan.
122
331260
2000
eta Afghanistanetik beste pixka bat.
05:33
It gives us some information vegetables;
123
333260
2000
Barazki albisteak batetik,
05:35
it gives us some information dessert.
124
335260
3000
postre albisteak bestetik.
05:38
And the challenge with these kinds of algorithmic filters,
125
338260
2000
Eta halako algoritmo iragazleek duten erronka,
05:40
these personalized filters,
126
340260
2000
halako iragazki personalizatuak,
05:42
is that, because they're mainly looking
127
342260
2000
zera da, oinarrian begiratzen ari direlako
05:44
at what you click on first,
128
344260
4000
zer klikatzen duzun lehenen,
05:48
it can throw off that balance.
129
348260
4000
oreka hori hautsi dezaketela.
05:52
And instead of a balanced information diet,
130
352260
3000
Eta informazio dieta orekatu baten ordez,
05:55
you can end up surrounded
131
355260
2000
amaitu dezakegu
05:57
by information junk food.
132
357260
2000
zabor informazioz inguratuta.
05:59
What this suggests
133
359260
2000
Honek adierazten digu
06:01
is actually that we may have the story about the Internet wrong.
134
361260
3000
agian internetari buruzko historia oker ulertu dugula.
06:04
In a broadcast society --
135
364260
2000
Informazioaren Gizartean --
06:06
this is how the founding mythology goes --
136
366260
2000
hala dio sortu zeneko mitoak --
06:08
in a broadcast society,
137
368260
2000
informazioaren gizartean,
06:10
there were these gatekeepers, the editors,
138
370260
2000
ate zaindariak, editoreak ziren,
06:12
and they controlled the flows of information.
139
372260
3000
informazioaren fluxua kontrolatzen zuten.
06:15
And along came the Internet and it swept them out of the way,
140
375260
3000
Eta Interneta etorri zenean denak garbitu zituen,
06:18
and it allowed all of us to connect together,
141
378260
2000
posible eginez denok elkar konektatzea,
06:20
and it was awesome.
142
380260
2000
aparta benetan.
06:22
But that's not actually what's happening right now.
143
382260
3000
Baina ez da hori oraintxe gertatzen ari zaiguna.
06:26
What we're seeing is more of a passing of the torch
144
386260
3000
Ikusten duguna gehiago dirudi testigua pasatzea
06:29
from human gatekeepers
145
389260
2000
giza atezainetatik
06:31
to algorithmic ones.
146
391260
3000
argoritmoetara.
06:34
And the thing is that the algorithms
147
394260
3000
Eta gauza da algoritmoak oraindik
06:37
don't yet have the kind of embedded ethics
148
397260
3000
ez dutela halako berezko barne etikarik
06:40
that the editors did.
149
400260
3000
editoreek bezela.
06:43
So if algorithms are going to curate the world for us,
150
403260
3000
Beraz algoritmoak zuzenduko badute mundua guretzat,
06:46
if they're going to decide what we get to see and what we don't get to see,
151
406260
3000
eta erabaki zer ikusi behar dugun eta zer ez,
06:49
then we need to make sure
152
409260
2000
orduan zihurtatu behar dugu
06:51
that they're not just keyed to relevance.
153
411260
3000
gakoa ez dela niretzako garrantzizkoa dena bakarrik.
06:54
We need to make sure that they also show us things
154
414260
2000
Zihurtatu behar dugu erakusten dizkigutela
06:56
that are uncomfortable or challenging or important --
155
416260
3000
deseroso zaizkigun gauzak edo oro har garrantzitsuak direnak --
06:59
this is what TED does --
156
419260
2000
hori egiten du TED-ek --
07:01
other points of view.
157
421260
2000
beste ikuspuntuak.
07:03
And the thing is, we've actually been here before
158
423260
2000
Eta gauza da lehen izan ginela hemen
07:05
as a society.
159
425260
2000
gizarte bezela.
07:08
In 1915, it's not like newspapers were sweating a lot
160
428260
3000
1915.ean, kazetaritza ez zen gehiegi arduratzen
07:11
about their civic responsibilities.
161
431260
3000
gizarte erantzunkizunaren inguruan.
07:14
Then people noticed
162
434260
2000
Orduan jendea ohartu zen
07:16
that they were doing something really important.
163
436260
3000
zerbait garrantzitsua egiten ari zirela.
07:19
That, in fact, you couldn't have
164
439260
2000
Alegia, ezinezkoa dela
07:21
a functioning democracy
165
441260
2000
demokrazia batek funtzionatzea
07:23
if citizens didn't get a good flow of information,
166
443260
4000
herritarrak ez badute informazio fluxu egokirik.
07:28
that the newspapers were critical because they were acting as the filter,
167
448260
3000
Eta kazetaritzak kritikoa behar zuela, filtro gisa aritzen zelako,
07:31
and then journalistic ethics developed.
168
451260
2000
eta horrela kazetaritza etika bat garatu zen.
07:33
It wasn't perfect,
169
453260
2000
Ez zen perfektua,
07:35
but it got us through the last century.
170
455260
3000
baina lagundu zigun pasa den gizaldian zehar.
07:38
And so now,
171
458260
2000
Eta orain
07:40
we're kind of back in 1915 on the Web.
172
460260
3000
badirudi 1915.ean gaudela Web-ean.
07:44
And we need the new gatekeepers
173
464260
3000
Ate zaindari berriak behar ditugu
07:47
to encode that kind of responsibility
174
467260
2000
halako erantzunkizuna kudeatzeko
07:49
into the code that they're writing.
175
469260
2000
idazten ari diren kodearen baitan.
07:51
I know that there are a lot of people here from Facebook and from Google --
176
471260
3000
Badakit Facebook eta Google-eko jende asko dugula hemen --
07:54
Larry and Sergey --
177
474260
2000
Larry eta Sergey --
07:56
people who have helped build the Web as it is,
178
476260
2000
gaur egun dugun web-a eraikitzen lagundu dutenak,
07:58
and I'm grateful for that.
179
478260
2000
horregatik eskerrak eman nahi dizkiet.
08:00
But we really need you to make sure
180
480260
3000
Baina benetan behar dugu zuek zihurtatzea
08:03
that these algorithms have encoded in them
181
483260
3000
algoritmo horien kodean txertatua dagoela
08:06
a sense of the public life, a sense of civic responsibility.
182
486260
3000
bizitza publikoaren sentipena eta gizartearekiko erantzunkizuna.
08:09
We need you to make sure that they're transparent enough
183
489260
3000
Behar dugu zuek haren gardentasuna zihurtatzea
08:12
that we can see what the rules are
184
492260
2000
guk ikusi dezagun zein erregela diren
08:14
that determine what gets through our filters.
185
494260
3000
gure filtroak zeharkatu behar dituztenak.
08:17
And we need you to give us some control
186
497260
2000
Eta behar dugu kontrol esparru bat guk izatea,
08:19
so that we can decide
187
499260
2000
erabaki ahal izateko
08:21
what gets through and what doesn't.
188
501260
3000
handik zer pasatzen den eta zer ez.
08:24
Because I think
189
504260
2000
Uste dudalako
08:26
we really need the Internet to be that thing
190
506260
2000
guk behar dugun Internetak izan behar duela
08:28
that we all dreamed of it being.
191
508260
2000
denok amestu genuen hura.
08:30
We need it to connect us all together.
192
510260
3000
Denok elkarrekin konektatzen gaituena.
08:33
We need it to introduce us to new ideas
193
513260
3000
Ideia berriak aurkeztu diezazkigula,
08:36
and new people and different perspectives.
194
516260
3000
jende berria eta ikusmolde desberdinak.
08:40
And it's not going to do that
195
520260
2000
Baina ez da hortan ariko
08:42
if it leaves us all isolated in a Web of one.
196
522260
3000
bakoitzaren Web isolatuan uzten bagaitu.
08:45
Thank you.
197
525260
2000
Eskerrik asko.
08:47
(Applause)
198
527260
11000
(Txaloak)
Webgune honi buruz

Gune honek ingelesa ikasteko erabilgarriak diren YouTube bideoak aurkeztuko dizkizu. Mundu osoko goi mailako irakasleek emandako ingeleseko ikasgaiak ikusiko dituzu. Egin klik bikoitza bideo-orri bakoitzean bistaratzen diren ingelesezko azpitituluetan, hortik bideoa erreproduzitzeko. Azpitituluak bideoaren erreprodukzioarekin sinkronizatuta mugitzen dira. Iruzkin edo eskaerarik baduzu, jar zaitez gurekin harremanetan harremanetarako formulario hau erabiliz.

https://forms.gle/WvT1wiN1qDtmnspy7