How we can find ourselves in data | Giorgia Lupi

112,053 views ・ 2017-05-04

TED


Please double-click on the English subtitles below to play the video.

00:12
This is what my last week looked like.
0
12580
2880
00:16
What I did,
1
16500
1696
00:18
who I was with,
2
18220
1816
00:20
the main sensations I had for every waking hour ...
3
20060
3120
00:23
If the feeling came as I thought of my dad
4
23900
2496
00:26
who recently passed away,
5
26420
1776
00:28
or if I could have just definitely avoided the worries and anxieties.
6
28220
4056
00:32
And if you think I'm a little obsessive,
7
32300
1936
00:34
you're probably right.
8
34260
1856
00:36
But clearly, from this visualization,
9
36140
2256
00:38
you can learn much more about me than from this other one,
10
38420
3296
00:41
which are images you're probably more familiar with
11
41740
2976
00:44
and which you possibly even have on your phone right now.
12
44740
2816
00:47
Bar charts for the steps you walked,
13
47580
2216
00:49
pie charts for the quality of your sleep --
14
49820
2376
00:52
the path of your morning runs.
15
52220
1720
00:55
In my day job, I work with data.
16
55100
2296
00:57
I run a data visualization design company,
17
57420
2496
00:59
and we design and develop ways to make information accessible
18
59940
3336
01:03
through visual representations.
19
63300
2096
01:05
What my job has taught me over the years
20
65420
3016
01:08
is that to really understand data and their true potential,
21
68460
4216
01:12
sometimes we actually have to forget about them
22
72700
3096
01:15
and see through them instead.
23
75820
1760
01:18
Because data are always just a tool we use to represent reality.
24
78260
3576
01:21
They're always used as a placeholder for something else,
25
81860
2856
01:24
but they are never the real thing.
26
84740
2176
01:26
But let me step back for a moment
27
86940
1936
01:28
to when I first understood this personally.
28
88900
2480
01:32
In 1994, I was 13 years old.
29
92100
3336
01:35
I was a teenager in Italy.
30
95460
1936
01:37
I was too young to be interested in politics,
31
97420
2616
01:40
but I knew that a businessman, Silvio Berlusconi,
32
100060
2496
01:42
was running for president for the moderate right.
33
102580
2560
01:45
We lived in a very liberal town,
34
105940
2016
01:47
and my father was a politician for the Democratic Party.
35
107980
3296
01:51
And I remember that no one thought that Berlusconi could get elected --
36
111300
4336
01:55
that was totally not an option.
37
115660
1760
01:58
But it happened.
38
118380
1216
01:59
And I remember the feeling very vividly.
39
119620
2536
02:02
It was a complete surprise,
40
122180
1856
02:04
as my dad promised that in my town he knew nobody who voted for him.
41
124060
5040
02:10
This was the first time
42
130540
1496
02:12
when the data I had gave me a completely distorted image of reality.
43
132060
4560
02:17
My data sample was actually pretty limited and skewed,
44
137100
3296
02:20
so probably it was because of that, I thought, I lived in a bubble,
45
140420
3736
02:24
and I didn't have enough chances to see outside of it.
46
144180
2600
02:27
Now, fast-forward to November 8, 2016
47
147900
3776
02:31
in the United States.
48
151700
1200
02:33
The internet polls,
49
153780
1296
02:35
statistical models,
50
155100
1376
02:36
all the pundits agreeing on a possible outcome for the presidential election.
51
156500
4816
02:41
It looked like we had enough information this time,
52
161340
2616
02:43
and many more chances to see outside the closed circle we lived in --
53
163980
4096
02:48
but we clearly didn't.
54
168100
1320
02:49
The feeling felt very familiar.
55
169860
2096
02:51
I had been there before.
56
171980
1480
02:54
I think it's fair to say the data failed us this time --
57
174180
2856
02:57
and pretty spectacularly.
58
177060
1856
02:58
We believed in data,
59
178940
1696
03:00
but what happened,
60
180660
1416
03:02
even with the most respected newspaper,
61
182100
2696
03:04
is that the obsession to reduce everything to two simple percentage numbers
62
184820
4696
03:09
to make a powerful headline
63
189540
1976
03:11
made us focus on these two digits
64
191540
2056
03:13
and them alone.
65
193620
1200
03:15
In an effort to simplify the message
66
195380
2056
03:17
and draw a beautiful, inevitable red and blue map,
67
197460
3416
03:20
we lost the point completely.
68
200900
1880
03:23
We somehow forgot that there were stories --
69
203260
2136
03:25
stories of human beings behind these numbers.
70
205420
2360
03:29
In a different context,
71
209060
1576
03:30
but to a very similar point,
72
210660
1656
03:32
a peculiar challenge was presented to my team by this woman.
73
212340
3896
03:36
She came to us with a lot of data,
74
216260
2376
03:38
but ultimately she wanted to tell one of the most humane stories possible.
75
218660
4416
03:43
She's Samantha Cristoforetti.
76
223100
1696
03:44
She has been the first Italian woman astronaut,
77
224820
2576
03:47
and she contacted us before being launched
78
227420
2496
03:49
on a six-month-long expedition to the International Space Station.
79
229940
3896
03:53
She told us, "I'm going to space,
80
233860
2216
03:56
and I want to do something meaningful with the data of my mission
81
236100
3096
03:59
to reach out to people."
82
239220
1240
04:01
A mission to the International Space Station
83
241420
2536
04:03
comes with terabytes of data
84
243980
2096
04:06
about anything you can possibly imagine --
85
246100
2376
04:08
the orbits around Earth,
86
248500
1496
04:10
the speed and position of the ISS
87
250020
2096
04:12
and all of the other thousands of live streams from its sensors.
88
252140
3680
04:16
We had all of the hard data we could think of --
89
256660
2896
04:19
just like the pundits before the election --
90
259580
2416
04:22
but what is the point of all these numbers?
91
262020
2976
04:25
People are not interested in data for the sake of it,
92
265020
2736
04:27
because numbers are never the point.
93
267780
1855
04:29
They're always the means to an end.
94
269659
1961
04:32
The story we needed to tell
95
272659
1777
04:34
is that there is a human being in a teeny box
96
274460
2496
04:36
flying in space above your head,
97
276980
2256
04:39
and that you can actually see her with your naked eye on a clear night.
98
279260
4096
04:43
So we decided to use data to create a connection
99
283380
3096
04:46
between Samantha and all of the people looking at her from below.
100
286500
4056
04:50
We designed and developed what we called "Friends in Space,"
101
290580
3176
04:53
a web application that simply lets you say "hello" to Samantha
102
293780
4656
04:58
from where you are,
103
298460
1256
04:59
and "hello" to all the people who are online at the same time
104
299740
3536
05:03
from all over the world.
105
303300
1520
05:05
And all of these "hellos" left visible marks on the map
106
305460
3456
05:08
as Samantha was flying by
107
308940
2016
05:10
and as she was actually waving back every day at us
108
310980
3376
05:14
using Twitter from the ISS.
109
314380
1680
05:16
This made people see the mission's data from a very different perspective.
110
316700
4976
05:21
It all suddenly became much more about our human nature and our curiosity,
111
321700
4696
05:26
rather than technology.
112
326420
1656
05:28
So data powered the experience,
113
328100
2336
05:30
but stories of human beings were the drive.
114
330460
2400
05:34
The very positive response of its thousands of users
115
334660
3336
05:38
taught me a very important lesson --
116
338020
1936
05:39
that working with data means designing ways
117
339980
2856
05:42
to transform the abstract and the uncountable
118
342860
2736
05:45
into something that can be seen, felt and directly reconnected
119
345620
4016
05:49
to our lives and to our behaviors,
120
349660
2296
05:51
something that is hard to achieve
121
351980
1856
05:53
if we let the obsession for the numbers and the technology around them
122
353860
3896
05:57
lead us in the process.
123
357780
1280
06:00
But we can do even more to connect data to the stories they represent.
124
360420
4896
06:05
We can remove technology completely.
125
365340
2656
06:08
A few years ago, I met this other woman,
126
368020
2256
06:10
Stefanie Posavec --
127
370300
1376
06:11
a London-based designer who shares with me the passion and obsession about data.
128
371700
5816
06:17
We didn't know each other,
129
377540
1336
06:18
but we decided to run a very radical experiment,
130
378900
3256
06:22
starting a communication using only data,
131
382180
2536
06:24
no other language,
132
384740
1336
06:26
and we opted for using no technology whatsoever to share our data.
133
386100
4616
06:30
In fact, our only means of communication
134
390740
2896
06:33
would be through the old-fashioned post office.
135
393660
2856
06:36
For "Dear Data," every week for one year,
136
396540
2456
06:39
we used our personal data to get to know each other --
137
399020
3456
06:42
personal data around weekly shared mundane topics,
138
402500
3656
06:46
from our feelings
139
406180
1216
06:47
to the interactions with our partners,
140
407420
1856
06:49
from the compliments we received to the sounds of our surroundings.
141
409300
3160
06:53
Personal information that we would then manually hand draw
142
413300
3536
06:56
on a postcard-size sheet of paper
143
416860
2496
06:59
that we would every week send from London to New York,
144
419380
2936
07:02
where I live,
145
422340
1256
07:03
and from New York to London, where she lives.
146
423620
2200
07:06
The front of the postcard is the data drawing,
147
426300
3696
07:10
and the back of the card
148
430020
1296
07:11
contains the address of the other person, of course,
149
431340
2429
07:13
and the legend for how to interpret our drawing.
150
433793
2640
07:17
The very first week into the project,
151
437460
2016
07:19
we actually chose a pretty cold and impersonal topic.
152
439500
3056
07:22
How many times do we check the time in a week?
153
442580
3200
07:26
So here is the front of my card,
154
446540
1936
07:28
and you can see that every little symbol
155
448500
1976
07:30
represents all of the times that I checked the time,
156
450500
3416
07:33
positioned for days and different hours chronologically --
157
453940
3376
07:37
nothing really complicated here.
158
457340
2040
07:40
But then you see in the legend
159
460020
1576
07:41
how I added anecdotal details about these moments.
160
461620
3456
07:45
In fact, the different types of symbols indicate why I was checking the time --
161
465100
4576
07:49
what was I doing?
162
469700
1216
07:50
Was I bored? Was I hungry?
163
470940
1696
07:52
Was I late?
164
472660
1216
07:53
Did I check it on purpose or just casually glance at the clock?
165
473900
3216
07:57
And this is the key part --
166
477140
2256
07:59
representing the details of my days and my personality
167
479420
3696
08:03
through my data collection.
168
483140
1936
08:05
Using data as a lens or a filter to discover and reveal, for example,
169
485100
4696
08:09
my never-ending anxiety for being late,
170
489820
2176
08:12
even though I'm absolutely always on time.
171
492020
2200
08:16
Stefanie and I spent one year collecting our data manually
172
496020
4096
08:20
to force us to focus on the nuances that computers cannot gather --
173
500140
4496
08:24
or at least not yet --
174
504660
1536
08:26
using data also to explore our minds and the words we use,
175
506220
3496
08:29
and not only our activities.
176
509740
1936
08:31
Like at week number three,
177
511700
1416
08:33
where we tracked the "thank yous" we said and were received,
178
513140
3816
08:36
and when I realized that I thank mostly people that I don't know.
179
516980
4656
08:41
Apparently I'm a compulsive thanker to waitresses and waiters,
180
521660
4336
08:46
but I definitely don't thank enough the people who are close to me.
181
526020
3160
08:50
Over one year,
182
530820
1256
08:52
the process of actively noticing and counting these types of actions
183
532100
4496
08:56
became a ritual.
184
536620
1296
08:57
It actually changed ourselves.
185
537940
2056
09:00
We became much more in tune with ourselves,
186
540020
2696
09:02
much more aware of our behaviors and our surroundings.
187
542740
3120
09:06
Over one year, Stefanie and I connected at a very deep level
188
546500
2976
09:09
through our shared data diary,
189
549500
2016
09:11
but we could do this only because we put ourselves in these numbers,
190
551540
4296
09:15
adding the contexts of our very personal stories to them.
191
555860
3976
09:19
It was the only way to make them truly meaningful
192
559860
2456
09:22
and representative of ourselves.
193
562340
2200
09:26
I am not asking you to start drawing your personal data,
194
566300
3096
09:29
or to find a pen pal across the ocean.
195
569420
2856
09:32
But I'm asking you to consider data --
196
572300
2576
09:34
all kind of data --
197
574900
1456
09:36
as the beginning of the conversation
198
576380
1776
09:38
and not the end.
199
578180
1200
09:39
Because data alone will never give us a solution.
200
579900
3176
09:43
And this is why data failed us so badly --
201
583100
2696
09:45
because we failed to include the right amount of context
202
585820
3376
09:49
to represent reality --
203
589220
1456
09:50
a nuanced, complicated and intricate reality.
204
590700
3200
09:54
We kept looking at these two numbers,
205
594780
2456
09:57
obsessing with them
206
597260
1496
09:58
and pretending that our world could be reduced
207
598780
2496
10:01
to a couple digits and a horse race,
208
601300
2336
10:03
while the real stories,
209
603660
1256
10:04
the ones that really mattered,
210
604940
1456
10:06
were somewhere else.
211
606420
1416
10:07
What we missed looking at these stories only through models and algorithms
212
607860
4416
10:12
is what I call "data humanism."
213
612300
2520
10:15
In the Renaissance humanism,
214
615380
2016
10:17
European intellectuals
215
617420
1616
10:19
placed the human nature instead of God at the center of their view of the world.
216
619060
4920
10:24
I believe something similar needs to happen
217
624620
2216
10:26
with the universe of data.
218
626860
1776
10:28
Now data are apparently treated like a God --
219
628660
2976
10:31
keeper of infallible truth for our present and our future.
220
631660
3280
10:35
The experiences that I shared with you today
221
635660
2896
10:38
taught me that to make data faithfully representative of our human nature
222
638580
5016
10:43
and to make sure they will not mislead us anymore,
223
643620
3416
10:47
we need to start designing ways to include empathy, imperfection
224
647060
3696
10:50
and human qualities
225
650780
1576
10:52
in how we collect, process, analyze and display them.
226
652380
3720
10:57
I do see a place where, ultimately,
227
657100
2976
11:00
instead of using data only to become more efficient,
228
660100
3336
11:03
we will all use data to become more humane.
229
663460
2800
11:06
Thank you.
230
666700
1216
11:07
(Applause)
231
667940
4441
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7