Kenneth Cukier: Big data is better data

517,498 views ・ 2014-09-23

TED


μ•„λž˜ μ˜λ¬Έμžλ§‰μ„ λ”λΈ”ν΄λ¦­ν•˜μ‹œλ©΄ μ˜μƒμ΄ μž¬μƒλ©λ‹ˆλ‹€.

λ²ˆμ—­: Kyo young Chu κ²€ν† : Jeong-Lan Kinser
00:12
America's favorite pie is?
0
12787
3845
미ꡭ인듀이 κ°€μž₯ μ’‹μ•„ν•˜λŠ” νŒŒμ΄λŠ”?
00:16
Audience: Apple. Kenneth Cukier: Apple. Of course it is.
1
16632
3506
청쀑: μ‚¬κ³Όμš”. μΌ€λ„€μŠ€ μΏ ν‚€μ–΄: λ‹Ήμ—°νžˆ 사과죠.
00:20
How do we know it?
2
20138
1231
μ–΄λ–»κ²Œ μ•Œμ•˜μ„κΉŒμš”?
00:21
Because of data.
3
21369
2753
데이터가 μžˆμœΌλ‹ˆκΉŒ μ•Œμ£ .
00:24
You look at supermarket sales.
4
24122
2066
μŠˆνΌλ§ˆμΌ“μ˜ λ§€μΆœμ„ 보면 λ©λ‹ˆλ‹€.
00:26
You look at supermarket sales of 30-centimeter pies
5
26188
2866
30 μ„ΌμΉ˜ 크기의 냉동 파이 λ§€μΆœμ„ 보면
00:29
that are frozen, and apple wins, no contest.
6
29054
4075
사과 νŒŒμ΄κ°€ μ„ λ‘μž…λ‹ˆλ‹€. 경쟁이 μ•ˆ 되죠.
00:33
The majority of the sales are apple.
7
33129
5180
λŒ€λΆ€λΆ„μ˜ 맀좜이 사과 νŒŒμ΄μž…λ‹ˆλ‹€.
00:38
But then supermarkets started selling
8
38309
2964
ν•˜μ§€λ§Œ μŠˆνΌλ§ˆμΌ“λ“€μ΄ 11μ„ΌμΉ˜μ˜ μž‘μ€νŒŒμ΄λ“€μ„ νŒλ§€ν•˜κΈ° μ‹œμž‘ν–ˆκ³ ,
00:41
smaller, 11-centimeter pies,
9
41273
2583
00:43
and suddenly, apple fell to fourth or fifth place.
10
43856
4174
그러자 사과 νŒŒμ΄λŠ” 4번째, 5번째둜 μΆ”λ½ν–ˆμ£ .
00:48
Why? What happened?
11
48030
2875
μ™œμš”? 무슨 일이 μΌμ–΄λ‚œ κ±ΈκΉŒμš”?
00:50
Okay, think about it.
12
50905
2818
μƒκ°ν•΄λ΄…μ‹œλ‹€.
00:53
When you buy a 30-centimeter pie,
13
53723
3848
30μ„ΌμΉ˜ 파이λ₯Ό μ‚΄ λ•ŒλŠ”
00:57
the whole family has to agree,
14
57571
2261
κ°€μ‘± 전원이 λ™μ˜λ₯Ό ν•΄μ•Ό ν•©λ‹ˆλ‹€.
00:59
and apple is everyone's second favorite.
15
59832
3791
그리고 μ‚¬κ³ΌλŠ” λͺ¨λ‘κ°€ 두 번째둜 μ’‹μ•„ν•˜λŠ” νŒŒμ΄μ˜€λ˜ κ²λ‹ˆλ‹€.
01:03
(Laughter)
16
63623
1935
(μ›ƒμŒ)
01:05
But when you buy an individual 11-centimeter pie,
17
65558
3615
ν•˜μ§€λ§Œ 개인이 먹을 11μ„ΌμΉ˜ 파이λ₯Ό μ‚΄ λ•ŒλŠ”
01:09
you can buy the one that you want.
18
69173
3745
μžκΈ°κ°€ μ›ν•˜λŠ” 것을 μ‚΄ 수 μžˆμ–΄μš”.
01:12
You can get your first choice.
19
72918
4015
μžκΈ°κ°€ κ°€μž₯ μ›ν•˜λŠ” 것을 μ‚΄ 수 μžˆλŠ” 것이죠.
01:16
You have more data.
20
76933
1641
μžλ£Œκ°€ 더 λ§Žμ•„μ§„ κ²λ‹ˆλ‹€.
01:18
You can see something
21
78574
1554
μ—¬λŸ¬λΆ„μ€ 적은 μ–‘μ˜ λ°μ΄ν„°μ˜€μ„ λ•ŒλŠ” λ³Ό 수 μ—†λŠ” λ­”κ°€λ₯Ό λ³Ό 수 있죠.
01:20
that you couldn't see
22
80128
1132
01:21
when you only had smaller amounts of it.
23
81260
3953
01:25
Now, the point here is that more data
24
85213
2475
μ—¬κΈ°μ„œμ˜ μš”μ μ€ 더 λ§Žμ€ 데이터가 단지
01:27
doesn't just let us see more,
25
87688
2283
μš°λ¦¬κ°€ λ³΄λŠ” 같은 κ²ƒμ—μ„œ 더 λ§Žμ€ κ²ƒλ§Œμ„ λ³΄μ—¬μ£ΌλŠ” 것이 μ•„λ‹ˆλΌλŠ” κ²λ‹ˆλ‹€.
01:29
more of the same thing we were looking at.
26
89971
1854
01:31
More data allows us to see new.
27
91825
3613
더 λ§Žμ€ λ°μ΄ν„°λŠ” μƒˆλ‘œμš΄ κ±Έ λ³Ό 수 있게 ν•΄μ£Όμ£ .
01:35
It allows us to see better.
28
95438
3094
μš°λ¦¬κ°€ 더 잘 보게 ν•΄μ£Όκ³ , μš°λ¦¬κ°€ λ‹€λ₯Έ 것을 보게 ν•΄μ£Όκ³ ,
01:38
It allows us to see different.
29
98532
3656
01:42
In this case, it allows us to see
30
102188
3173
이 κ²½μš°μ—λŠ”, 미ꡭ인듀이 κ°€μž₯ μ’‹μ•„ν•˜λŠ” νŒŒμ΄κ°€ 사과가 μ•„λ‹ˆλž€ κ±Έ μ•Œλ €μ€λ‹ˆλ‹€.
01:45
what America's favorite pie is:
31
105361
2913
01:48
not apple.
32
108274
2542
01:50
Now, you probably all have heard the term big data.
33
110816
3614
λ‹€λ“€ λΉ…λ°μ΄ν„°λΌλŠ” 말을 듀어보셨을 κ²λ‹ˆλ‹€.
01:54
In fact, you're probably sick of hearing the term
34
114430
2057
사싀, κ·Έ, λΉ…λ°μ΄ν„°λΌλŠ” 단어에 μ‹μƒν•˜μ…¨μ„κ±°μ˜ˆμš”.
01:56
big data.
35
116487
1630
01:58
It is true that there is a lot of hype around the term,
36
118117
3330
빅데이터가 κ΅¬μ„€μˆ˜μ— 였λ₯΄λŠ” 건 μ‚¬μ‹€μ΄μ§€λ§Œ,
02:01
and that is very unfortunate,
37
121447
2332
ꡉμž₯히 μ•ˆνƒ€κΉκ²Œ μƒκ°ν•©λ‹ˆλ‹€.
02:03
because big data is an extremely important tool
38
123779
3046
μ™œλƒν•˜λ©΄ λΉ…λ°μ΄ν„°λŠ” μ‚¬νšŒμ˜ λ°œμ „μ— μžˆμ–΄μ„œ ꡉμž₯히 μ€‘μš”ν•œ 도ꡬ이기 λ•Œλ¬Έμž…λ‹ˆλ‹€.
02:06
by which society is going to advance.
39
126825
3734
02:10
In the past, we used to look at small data
40
130559
3561
과거에, μš°λ¦¬λŠ” μ†ŒλŸ‰μ˜ 자료λ₯Ό 보며
02:14
and think about what it would mean
41
134120
1704
세상을 μ΄ν•΄ν•˜λ € λ…Έλ ₯ν•˜λŠ” 게 무엇을 μ˜λ―Έν•˜λŠ”μ§€ μƒκ°ν–ˆμŠ΅λ‹ˆλ‹€.
02:15
to try to understand the world,
42
135824
1496
02:17
and now we have a lot more of it,
43
137320
1991
μ§€κΈˆ μš°λ¦¬λŠ” κ·Έ μ–΄λŠ λ•Œλ³΄λ‹€λ„
02:19
more than we ever could before.
44
139311
2722
λ§Žμ€ 자료λ₯Ό 가지고 μžˆμŠ΅λ‹ˆλ‹€.
02:22
What we find is that when we have
45
142033
1877
λ°©λŒ€ν•œ μ–‘μ˜ μžλ£Œκ°€ 있으면
02:23
a large body of data, we can fundamentally do things
46
143910
2724
근본적으둜 더 적은 μžλ£Œλ‘œλŠ”
02:26
that we couldn't do when we only had smaller amounts.
47
146634
3276
ν•  수 μ—†μ—ˆλ˜ 것듀을 ν•  수 μžˆλ‹€λŠ” 것을 κΉ¨λ‹¬μ•˜μŠ΅λ‹ˆλ‹€.
02:29
Big data is important, and big data is new,
48
149910
2641
λΉ…λ°μ΄ν„°λŠ” μ€‘μš”ν•˜κ³ , λΉ…λ°μ΄ν„°λŠ” μƒˆλ‘­μŠ΅λ‹ˆλ‹€.
02:32
and when you think about it,
49
152551
1777
μ—¬λŸ¬λΆ„μ΄ 생각해 λ³΄μ‹œλ©΄, 기아와 의료 지원, μ „λ ₯ 곡급,
02:34
the only way this planet is going to deal
50
154328
2216
02:36
with its global challenges β€”
51
156544
1789
02:38
to feed people, supply them with medical care,
52
158333
3537
02:41
supply them with energy, electricity,
53
161870
2810
02:44
and to make sure they're not burnt to a crisp
54
164680
1789
κΈ°μƒλ³€ν™”λ‘œκ·Έλ“€μ΄ κ³ ν†΅λ°›λŠ” λ¬Έμ œμ™€ 같은 세계적인 λ‚œκ΄€λ“€μ„ ν•΄μ³λ‚˜κ°ˆ 수 μžˆμ—ˆλ˜
02:46
because of global warming β€”
55
166469
1238
02:47
is because of the effective use of data.
56
167707
4195
μœ μΌν•œ 방법은 데이터λ₯Ό 효과적으둜 μ‚¬μš©ν–ˆκΈ° λ•Œλ¬Έμ΄μ£ .
02:51
So what is new about big data? What is the big deal?
57
171902
3870
그러면 λΉ… 데이터가 μƒˆλ‘­κ³  μ€‘μš”ν•œ μ΄μœ λŠ” λ­˜κΉŒμš”?
02:55
Well, to answer that question, let's think about
58
175772
2517
κΈ€μŽ„μš”, κ·Έ μ§ˆλ¬Έμ— λŒ€λ‹΅ ν•˜κΈ° μœ„ν•΄,
02:58
what information looked like,
59
178289
1896
κ³Όκ±°μ—λŠ” μ •λ³΄λΌλŠ” 것이 물리적으둜 μ–΄λ–€ ν˜•νƒœμ˜€λŠ”μ§€ μ•Œμ•„λ³΄μ£ .
03:00
physically looked like in the past.
60
180185
3034
03:03
In 1908, on the island of Crete,
61
183219
3611
1908년에, ν¬λ ˆν…Œ μ„¬μ—μ„œ,
03:06
archaeologists discovered a clay disc.
62
186830
4735
κ³ κ³ ν•™μžλ“€μ΄ 진흙 μ ‘μ‹œλ₯Ό ν•˜λ‚˜ λ°œκ²¬ν–ˆμŠ΅λ‹ˆλ‹€.
03:11
They dated it from 2000 B.C., so it's 4,000 years old.
63
191565
4059
기원전2000λ…„μ˜ μ ‘μ‹œλ‘œ μΆ”μ •ν–ˆμœΌλ‹ˆ 4000λ…„ 정도 λκ² λ„€μš”.
03:15
Now, there's inscriptions on this disc,
64
195624
2004
κ·Έ μ ‘μ‹œμ— λ­”κ°€ μƒˆκ²¨μ§„ 게 μžˆμ—ˆλŠ”λ°μš”,
03:17
but we actually don't know what it means.
65
197628
1327
뭘 μ˜λ―Έν•˜λŠ” μ§€λŠ” λͺ¨λ¦…λ‹ˆλ‹€. μ™„μ „νžˆ λ―Έμ§€μ˜ κ²ƒμ΄μ§€λ§Œ,
03:18
It's a complete mystery, but the point is that
66
198955
2098
μš”μ μ€, 4000λ…„ μ „μ—λŠ” 정보가 이런 ν˜•νƒœμ˜€λ‹€λŠ” μ‚¬μ‹€μž…λ‹ˆλ‹€.
03:21
this is what information used to look like
67
201053
1928
03:22
4,000 years ago.
68
202981
2089
03:25
This is how society stored
69
205070
2548
κ·Έ μ‚¬νšŒλŠ” 정보λ₯Ό 이런 ν˜•νƒœλ‘œ μ €μž₯ν•˜κ³  μ „νŒŒν•œ κ²ƒμž…λ‹ˆλ‹€.
03:27
and transmitted information.
70
207618
3524
03:31
Now, society hasn't advanced all that much.
71
211142
4160
아직 μ‚¬νšŒλŠ” 그리 많이 λ°œλ‹¬ν•˜μ§„ μ•Šμ•˜μ–΄μš”.
03:35
We still store information on discs,
72
215302
3474
μš°λ¦¬λŠ” μ§€κΈˆλ„ μ—¬μ „νžˆ λ””μŠ€ν¬μ— 정보λ₯Ό μ €μž₯ν•΄μš”.
03:38
but now we can store a lot more information,
73
218776
3184
ν•˜μ§€λ§Œ μ΄μ „μ˜ κ·Έ μ–΄λŠλ•Œ 보닀 훨씬 더 λ§Žμ€ 양을 μ €μž₯ν•  수 있죠.
03:41
more than ever before.
74
221960
1260
03:43
Searching it is easier. Copying it easier.
75
223220
3093
μˆ˜μƒ‰ν•˜λŠ” 것이 더 쉽고, λ³΅μ‚¬ν•˜λŠ” 것도 더 쉽고,
03:46
Sharing it is easier. Processing it is easier.
76
226313
3500
κ³΅μœ ν•˜κΈ°λ„ 더 쉽고, μ²˜λ¦¬λ„ 더 μ‰¬μ›Œμ‘ŒμŠ΅λ‹ˆλ‹€.
03:49
And what we can do is we can reuse this information
77
229813
2766
μš°λ¦¬κ°€ 이제 ν•  수 μžˆλŠ” 것은
데이터λ₯Ό 처음 μˆ˜μ§‘ν–ˆμ„ λ•ŒλŠ” μƒμƒν•˜μ§€ λͺ»ν•œ λ°©λ²•μœΌλ‘œ
03:52
for uses that we never even imagined
78
232579
1834
03:54
when we first collected the data.
79
234413
3195
이 정보λ₯Ό λ‹€μ‹œ μ‚¬μš©ν•  수 μžˆλ‹€λŠ” κ²λ‹ˆλ‹€.
03:57
In this respect, the data has gone
80
237608
2252
μ΄λŸ¬ν•œ κ΄€μ μ—μ„œ,
03:59
from a stock to a flow,
81
239860
3532
λ°μ΄ν„°λŠ” μ €μž₯된 κ²ƒμ—μ„œ νλ¦„μœΌλ‘œ,
04:03
from something that is stationary and static
82
243392
3938
κ³ μ •λ˜κ³  λ³€ν•˜μ§€ μ•ŠλŠ” κ²ƒμ—μ„œ
04:07
to something that is fluid and dynamic.
83
247330
3609
움직이고 λ³€ν™”ν•˜λŠ” μœ λ™μ μΈ κ²ƒμœΌλ‘œ λ°”λ€Œμ—ˆμŠ΅λ‹ˆλ‹€.
04:10
There is, if you will, a liquidity to information.
84
250939
4023
여기에 μ •λ³΄μ˜ μœ λ™μ„±μ΄ μžˆλŠ” κ²ƒμž…λ‹ˆλ‹€.
04:14
The disc that was discovered off of Crete
85
254962
3474
ν¬λ ˆν…Œ μ„¬μ—μ„œ 발견된 μ ‘μ‹œλŠ”
04:18
that's 4,000 years old, is heavy,
86
258436
3764
4,000λ…„μ΄λ‚˜ 지났고, λ¬΄κ±°μ› μœΌλ©°,
04:22
it doesn't store a lot of information,
87
262200
1962
μ •λ³΄μ˜ 양도 λ§Žμ§€ μ•Šμ•˜κ³ ,
04:24
and that information is unchangeable.
88
264162
3116
이λ₯Ό λ°”κΏ€ μˆ˜λ„ μ—†μ—ˆμ£ .
04:27
By contrast, all of the files
89
267278
4011
μ΄μ™€λŠ” λŒ€μ‘°μ μœΌλ‘œ
λ―Έκ΅­ κ΅­κ°€μ•ˆλ³΄κ΅­μ—μ„œ μ—λ“œμ›Œλ“œ μŠ€λ…Έλ“ μ΄ λΉΌλ‚Έ λͺ¨λ“  νŒŒμΌμ€
04:31
that Edward Snowden took
90
271289
1861
04:33
from the National Security Agency in the United States
91
273150
2621
04:35
fits on a memory stick
92
275771
2419
μ†ν†±λ§Œν•œ λ©”λͺ¨λ¦¬ μŠ€ν‹±μ—
04:38
the size of a fingernail,
93
278190
3010
λ‹€ λ“€μ–΄κ°”κ³ ,
04:41
and it can be shared at the speed of light.
94
281200
4745
λΉ›μ˜ μ†λ„λ‘œ κ³΅μœ ν•  수 μžˆμ—ˆμŠ΅λ‹ˆλ‹€.
04:45
More data. More.
95
285945
5255
더 λ§Žμ€ μžλ£ŒλŠ” 계속 λŠ˜μ–΄κ°‘λ‹ˆλ‹€.
04:51
Now, one reason why we have so much data in the world today
96
291200
1974
μ˜€λŠ˜λ‚  세계에 데이터가 λ„˜μ³λ‚˜λŠ” 이유 쀑 ν•˜λ‚˜λŠ”
04:53
is we are collecting things
97
293174
1432
정보λ₯Ό 제곡 받은 μžλ£Œλ“€μ„ 항상, μ–Έμ œλ‹€ λͺ¨μœΌκΈ° λ•Œλ¬Έμ΄κ³ μš”,
04:54
that we've always collected information on,
98
294606
3280
04:57
but another reason why is we're taking things
99
297886
2656
또 λ‹€λ₯Έ μ΄μœ λŠ”
μ •λ³΄λ‘œμ¨ κ°€μΉ˜λŠ” μžˆμ§€λ§Œ 자료의 ν˜•νƒœλ‘œ κ°€κ³΅λ˜μ§€ λͺ»ν–ˆλ˜ 것을 μ΄μ œλŠ”
05:00
that have always been informational
100
300542
2812
05:03
but have never been rendered into a data format
101
303354
2486
05:05
and we are putting it into data.
102
305840
2419
자료의 ν˜•νƒœλ‘œ 가곡할 수 있게 λ˜μ—ˆκΈ° λ•Œλ¬Έμž…λ‹ˆλ‹€.
05:08
Think, for example, the question of location.
103
308259
3308
에λ₯Ό λ“€μ–΄, μœ„μΉ˜μ— λŒ€ν•œ μ§ˆλ¬Έμ„ 생각해보죠.
05:11
Take, for example, Martin Luther.
104
311567
2249
λ§ˆν‹΄ 루터λ₯Ό 예둜 λ“€μ–΄ λ³Όκ²Œμš”.
05:13
If we wanted to know in the 1500s
105
313816
1597
1500λ…„ 경에 λ§ˆν‹΄ λ£¨ν„°μ˜ μœ„μΉ˜λ₯Ό νŒŒμ•…ν•˜λ €λ©΄,
05:15
where Martin Luther was,
106
315413
2667
05:18
we would have to follow him at all times,
107
318080
2092
항상 그의 κ½λ¬΄λ‹ˆλ₯Ό μ«“μ•„λ‹€λ‹ˆλ©΄μ„œ
05:20
maybe with a feathery quill and an inkwell,
108
320172
2137
펜과 μž‰ν¬λ³‘μ„ λ“€κ³ , 기둝을 ν•΄μ•Όλ§Œν–ˆμ£ .
05:22
and record it,
109
322309
1676
05:23
but now think about what it looks like today.
110
323985
2183
ν•˜μ§€λ§Œ μ§€κΈˆμ˜ λͺ¨μŠ΅μ€ 어떀지 생각해 λ³΄μ„Έμš”.
05:26
You know that somewhere,
111
326168
2122
μ—¬λŸ¬λΆ„λ“€μ˜ μœ„μΉ˜λ₯Ό μ‹€μ‹œκ°„μœΌλ‘œ κΈ°λ‘ν•˜λŠ” μŠ€ν”„λ ˆλ“œμ‹œνŠΈλ‚˜ μ•„λ‹ˆλ©΄
05:28
probably in a telecommunications carrier's database,
112
328290
2446
05:30
there is a spreadsheet or at least a database entry
113
330736
3036
μ΅œμ†Œν•œ λ°μ΄ν„°λ² μ΄μŠ€ μž…λ ₯ ν˜•νƒœλ‘œ 톡신사 λ°μ΄ν„°λ² μ΄μŠ€λΌλ˜κ°€,
05:33
that records your information
114
333772
2088
05:35
of where you've been at all times.
115
335860
2063
μ–΄λ”˜κ°€μ—λŠ” μ•„λ§ˆ μžˆμ„ κ²λ‹ˆλ‹€. νœ΄λŒ€μ „ν™”λ₯Ό 가지고 있고,
05:37
If you have a cell phone,
116
337923
1360
05:39
and that cell phone has GPS, but even if it doesn't have GPS,
117
339283
2847
κ·Έ 전화에 GPSκ°€ μžˆλ‹€λ©΄, μ•„λ‹ˆ μ—†λ‹€κ³  ν•˜λ”λΌλ„,
05:42
it can record your information.
118
342130
2385
μ—¬λŸ¬λΆ„μ˜ 정보λ₯Ό 기둝할 수 μžˆμŠ΅λ‹ˆλ‹€.
05:44
In this respect, location has been datafied.
119
344515
4084
즉, 이제 μœ„μΉ˜λŠ” 데이터화가 λ˜μ—ˆλ‹€λŠ” 말이죠.
05:48
Now think, for example, of the issue of posture,
120
348599
4601
이제 μžμ„Έλ₯Ό 예둜 λ“€μ–΄ λ³΄κ² μŠ΅λ‹ˆλ‹€.
05:53
the way that you are all sitting right now,
121
353200
1285
μ—¬λŸ¬λΆ„λ“€μ΄ μ•‰μ•„μžˆλŠ” μžμ„Έμš”.
05:54
the way that you sit,
122
354485
2030
이 λΆ„μ˜ μžμ„Έμ™€, μ € λΆ„μ˜ μžμ„Έ, 또 μ €μͺ½ λΆ„μ˜ μžμ„ΈλŠ” λͺ¨λ‘ λ‹€λ¦…λ‹ˆλ‹€.
05:56
the way that you sit, the way that you sit.
123
356515
2771
05:59
It's all different, and it's a function of your leg length
124
359286
2077
μ΄λŠ” 닀리 길이와 λ“±, 또 λ“±μ˜ ꡽어진 정도에 λ”°λ₯Έ ν•¨μˆ˜μ™€ κ°™κ±°λ“ μš”.
06:01
and your back and the contours of your back,
125
361363
2093
06:03
and if I were to put sensors, maybe 100 sensors
126
363456
2531
λ§Œμ•½ μ—¬λŸ¬λΆ„μ΄ μ§€κΈˆ μ•‰μ•„μžˆλŠ” μ˜μžμ—
06:05
into all of your chairs right now,
127
365987
1766
μˆ˜λ§Žμ€ μ„Όμ„œλ₯Ό 단닀고 ν•œλ‹€λ©΄,
06:07
I could create an index that's fairly unique to you,
128
367753
3600
κ°œκ°œμΈμ— λ§žλŠ” μƒμˆ˜λ“€μ„ λ‹€ λ§Œλ“€ 수 μžˆμ—ˆμ„ κ²λ‹ˆλ‹€.
06:11
sort of like a fingerprint, but it's not your finger.
129
371353
4409
손가락은 μ•„λ‹ˆμ§€λ§Œ, 지문과 λΉ„μŠ·ν•˜κ² κ΅°μš”.
06:15
So what could we do with this?
130
375762
2969
이걸둜 뭘 ν•  수 μžˆμ„κΉŒμš”?
06:18
Researchers in Tokyo are using it
131
378731
2397
λ„μΏ„μ˜ μ—°κ΅¬μžλ“€μ€ 이λ₯Ό 톡해
06:21
as a potential anti-theft device in cars.
132
381128
4388
μžλ™μ°¨ λ„λ‚œ 방지기λ₯Ό λ§Œλ“œλŠ” 것을 λͺ¨μƒ‰ν•˜κ³  μžˆμŠ΅λ‹ˆλ‹€.
06:25
The idea is that the carjacker sits behind the wheel,
133
385516
2924
μ›λ¦¬λŠ” μ΄λ ‡μŠ΅λ‹ˆλ‹€. μžλ™μ°¨ 도둑이 μš΄μ „μ„μ— μ•‰μ•„μ„œ
06:28
tries to stream off, but the car recognizes
134
388440
2104
μ°¨λ₯Ό 움직이렀고 ν•˜μ§€λ§Œ μžλ™μ°¨κ°€ μš΄μ „μ„μ— 앉은 μ‚¬λžŒμ΄
06:30
that a non-approved driver is behind the wheel,
135
390544
2362
인증된 μš΄μ „μžκ°€ μ•„λ‹˜μ„ μΈμ‹ν•˜κ³ 
06:32
and maybe the engine just stops, unless you
136
392906
2164
엔진을 λ„κ±°λ‚˜ ν•  κ²λ‹ˆλ‹€.
06:35
type in a password into the dashboard
137
395070
3177
κ·Έ μ‚¬λžŒμ΄ λΉ„λ°€λ²ˆν˜Έλ₯Ό μž…λ ₯ν•˜κ±°λ‚˜ ν•΄μ„œ
06:38
to say, "Hey, I have authorization to drive." Great.
138
398247
4658
μžμ‹ μ΄ ν—ˆκ°€λ°›μ•˜λ‹€λŠ” 것을 μ•Œλ¦¬μ§€ μ•ŠλŠ” ν•œ 말이죠.
μ’‹μ•„μš”.
06:42
What if every single car in Europe
139
402905
2553
μœ λŸ½μ— μžˆλŠ” λͺ¨λ“  μ°¨λŸ‰μ— 이 기술이 μ μš©λ˜μ—ˆλ‹€λ©΄ μ–΄λ–¨κΉŒμš”?
06:45
had this technology in it?
140
405458
1457
06:46
What could we do then?
141
406915
3165
그러면 무엇을 ν•  수 μžˆμ„κΉŒμš”?
06:50
Maybe, if we aggregated the data,
142
410080
2240
데이터λ₯Ό λͺ¨μ•„ 5초 ν›„μ˜ μžλ™μ°¨μ‚¬κ³ λ₯Ό μ˜ˆμΈ‘ν•  수 μžˆλŠ” 쑰짐을 μ•Œμ•„λ‚Ό 수 있죠.
06:52
maybe we could identify telltale signs
143
412320
3814
06:56
that best predict that a car accident
144
416134
2709
06:58
is going to take place in the next five seconds.
145
418843
5893
07:04
And then what we will have datafied
146
424736
2557
또 μš΄μ „μžμ˜ ν”Όλ‘œλ₯Ό λ°μ΄ν„°λ‘œ λ§Œλ“€μ–΄μ„œ μš΄μ „μžκ°€ κ·Έ μƒνƒœμ— μ ‘μ–΄λ“€κ²Œ 되면
07:07
is driver fatigue,
147
427293
1783
07:09
and the service would be when the car senses
148
429076
2334
07:11
that the person slumps into that position,
149
431410
3437
μ°¨λŸ‰μ΄ 이λ₯Ό κ°μ§€ν•˜κ³  μžλ™μ μœΌλ‘œ λ‚΄λΆ€ μ•ŒλžŒμ„ μž‘λ™μ‹œν‚€λŠ” κ±°μ£ .
07:14
automatically knows, hey, set an internal alarm
150
434847
3994
07:18
that would vibrate the steering wheel, honk inside
151
438841
2025
μš΄μ „λŒ€κ°€ μ§„λ™ν•œλ‹€λ˜μ§€, λ‚΄λΆ€μ—μ„œ, "μ •μ‹ μ°¨λ €! λ„λ‘œλ₯Ό 봐야지!"ν•˜λŠ”
07:20
to say, "Hey, wake up,
152
440866
1721
07:22
pay more attention to the road."
153
442587
1904
경보가 울리게 ν•˜λŠ” κ²ƒμ²˜λŸΌμš”.
07:24
These are the sorts of things we can do
154
444491
1853
우리의 삢을 더 뢄석할 수 μžˆλ‹€λ©΄ 이런 일듀이 κ°€λŠ₯ν•΄μ§‘λ‹ˆλ‹€.
07:26
when we datafy more aspects of our lives.
155
446344
2821
07:29
So what is the value of big data?
156
449165
3675
κ·Έλ ‡λ‹€λ©΄ λΉ…λ°μ΄ν„°λŠ” μ–΄λ–€ κ°€μΉ˜λ₯Ό 가지고 μžˆμ„κΉŒμš”?
07:32
Well, think about it.
157
452840
2190
자, 생각해 λ³΄μ„Έμš”.
07:35
You have more information.
158
455030
2412
더 λ§Žμ€ 정보가 있으면 μ „μ—” ν•  수 μ—†μ—ˆλ˜ 일듀을 ν•  수 μžˆμ–΄μš”.
07:37
You can do things that you couldn't do before.
159
457442
3341
07:40
One of the most impressive areas
160
460783
1676
이런 일이 μΌμ–΄λ‚˜λŠ” κ°€μž₯ 인상적인 λΆ„μ•Ό 쀑 ν•˜λ‚˜λŠ” λ°”λ‘œ 기계 ν•™μŠ΅μ˜ μ˜μ—­μž…λ‹ˆλ‹€.
07:42
where this concept is taking place
161
462459
1729
07:44
is in the area of machine learning.
162
464188
3307
07:47
Machine learning is a branch of artificial intelligence,
163
467495
3077
κΈ°κ³„ν•™μŠ΅μ€ 인곡지λŠ₯의 ν•œ 뢄야인데,
07:50
which itself is a branch of computer science.
164
470572
3378
인곡지λŠ₯은 컴퓨터 κ³΅ν•™μ˜ μΌλΆ€μž…λ‹ˆλ‹€.
07:53
The general idea is that instead of
165
473950
1543
μ‰½κ²Œ λ§ν•˜μžλ©΄
07:55
instructing a computer what do do,
166
475493
2117
μ»΄ν“¨ν„°μ—κ²Œ μ§€μ‹œλ₯Ό λ‚΄λ¦¬λŠ” λŒ€μ‹ μ—
07:57
we are going to simply throw data at the problem
167
477610
2620
데이터와 문제λ₯Ό μ£Όκ³ 
08:00
and tell the computer to figure it out for itself.
168
480230
3206
슀슀둜 μ•Œμ•„λ‚΄κ²Œ λ§Œλ“œλŠ” κ²λ‹ˆλ‹€.
08:03
And it will help you understand it
169
483436
1777
κ·Έ κΈ°μ›μœΌλ‘œ 거슬러 μ˜¬λΌκ°€λ³΄λ©΄
08:05
by seeing its origins.
170
485213
3552
이해에 도움이 될 κ±°μ˜ˆμš”.
08:08
In the 1950s, a computer scientist
171
488765
2388
1950년경에,
IBM의 컴퓨터 ν•™μžμΈ μ•„μ„œ μ‚¬λ¬΄μ—˜μ€ 첡컀 κ²Œμž„μ„ μ¦κ²Όμ–΄μš”.
08:11
at IBM named Arthur Samuel liked to play checkers,
172
491153
3592
08:14
so he wrote a computer program
173
494745
1402
κ·Έλž˜μ„œ, 컴퓨터 ν”„λ‘œκ·Έλž¨μ„ μ§œμ„œ
08:16
so he could play against the computer.
174
496147
2813
컴퓨터λ₯Ό μƒλŒ€λ‘œ κ²Œμž„μ„ ν–ˆμ£ .
08:18
He played. He won.
175
498960
2711
κ²Œμž„μ„ ν–ˆλ”λ‹ˆ 이기고,
08:21
He played. He won.
176
501671
2103
κ²Œμž„μ„ ν–ˆλ”λ‹ˆ 또 이기고,
08:23
He played. He won,
177
503774
3015
κ³„μ†ν•΄μ„œ 이겼죠.
08:26
because the computer only knew
178
506789
1778
μ™œλƒν•˜λ©΄ μ»΄ν“¨ν„°λŠ” 단지
08:28
what a legal move was.
179
508567
2227
κ²Œμž„κ·œμΉ™λ§Œ μ•Œμ•˜μ§€λ§Œ
08:30
Arthur Samuel knew something else.
180
510794
2087
μ•„μ„œ μ‚¬λ¬΄μ—˜μ€ κ·Έ 이상을 μ•Œμ•˜κΈ° λ•Œλ¬Έμ΄μ£ .
08:32
Arthur Samuel knew strategy.
181
512881
4629
μ‚¬λ¬΄μ—˜μ€ μ „λž΅μ΄λΌλŠ” κ±Έ μ•Œκ³  μžˆμ—ˆλ˜ κ²λ‹ˆλ‹€.
08:37
So he wrote a small sub-program alongside it
182
517510
2396
κ·Έλž˜μ„œ κ·ΈλŠ” 보쑰 ν”„λ‘œκ·Έλž¨μ„ κ°œλ°œν–ˆμ–΄μš”.
08:39
operating in the background, and all it did
183
519906
1974
κ·Έ ν”„λ‘œκ·Έλž¨μ€ μ•ˆμ—μ„œ μž‘λ™ν•˜λ©΄μ„œ 주어진 κ²Œμž„μ˜ ν˜•κ΅­μ—μ„œ 움직일 λ•Œλ§ˆλ‹€
08:41
was score the probability
184
521880
1817
08:43
that a given board configuration would likely lead
185
523697
2563
08:46
to a winning board versus a losing board
186
526260
2910
이길 ν˜•κ΅­κ³Ό 질 ν˜•κ΅­μœΌλ‘œ λ‚˜μ•„κ°ˆ ν™•λ₯ λ§Œμ„ κ³„μ‚°ν–ˆμŠ΅λ‹ˆλ‹€.
08:49
after every move.
187
529170
2508
08:51
He plays the computer. He wins.
188
531678
3150
κ·Έλž˜λ„ μ»΄ν“¨ν„°λž‘ κ²Œμž„μ„ ν•˜λ©΄ 이기고,
08:54
He plays the computer. He wins.
189
534828
2508
μ»΄ν“¨ν„°λž‘ κ²Œμž„μ„ ν•˜λ©΄, 또 이기고
08:57
He plays the computer. He wins.
190
537336
3731
μ»΄ν“¨ν„°λž‘ κ²Œμž„μ„ ν•˜λ©΄, κ³„μ†ν•΄μ„œ 이겼죠.
09:01
And then Arthur Samuel leaves the computer
191
541067
2277
그러고 λ‚˜μ„œ μ•„μ„œ μ‚¬λ¬΄μ—˜μ€ 컴퓨터가
09:03
to play itself.
192
543344
2227
혼자 κ²Œμž„μ„ ν•˜λ„λ‘ λ†”λ’€μŠ΅λ‹ˆλ‹€.
09:05
It plays itself. It collects more data.
193
545571
3509
슀슀둜 κ²Œμž„μ„ ν•˜λ©΄μ„œ 더 λ§Žμ€ 데이터λ₯Ό λͺ¨μ•˜μ£ .
09:09
It collects more data. It increases the accuracy of its prediction.
194
549080
4309
μžλ£Œκ°€ λ§Žμ•„μ§ˆμˆ˜λ‘ 예츑의 μ •ν™•λ„λŠ” λ†’μ•„μ‘ŒμŠ΅λ‹ˆλ‹€.
09:13
And then Arthur Samuel goes back to the computer
195
553389
2104
그리고 λ‚œ λ‹€μŒμ—, μ‚¬λ¬΄μ—˜μ΄ 컴퓨터와 λ‹€μ‹œ κ²Œμž„μ„ ν–ˆκ³ , κ·ΈλŠ” μ‘ŒμŠ΅λ‹ˆλ‹€.
09:15
and he plays it, and he loses,
196
555493
2318
09:17
and he plays it, and he loses,
197
557811
2069
κ·ΈλŠ” κ²Œμž„μ„ ν•˜κ³ , 지고
09:19
and he plays it, and he loses,
198
559880
2047
κ²Œμž„μ„ν•˜κ³  또 μ‘ŒμŠ΅λ‹ˆλ‹€.
09:21
and Arthur Samuel has created a machine
199
561927
2599
λ§ˆμΉ¨λ‚΄ μžμ‹ μ΄ κ°€λ₯΄μΉœ 일에 λŒ€ν•΄μ„œ
09:24
that surpasses his ability in a task that he taught it.
200
564526
6288
본인의 λŠ₯λ ₯을 μ΄ˆκ³Όν•˜λŠ” 기계λ₯Ό λ§Œλ“€μ–΄ λ‚Έκ²λ‹ˆλ‹€.
09:30
And this idea of machine learning
201
570814
2498
그리고 이 κΈ°κ³„ν•™μŠ΅μ˜ κ°œλ…μ€
09:33
is going everywhere.
202
573312
3927
μ—¬λŸ¬ 곳으둜 μ „νŒŒ 되죠.
09:37
How do you think we have self-driving cars?
203
577239
3149
μ–΄λ–»κ²Œ λ¬΄μΈμžλ™μ°¨κ°€ λ‚˜μ™”λ‹€κ³  μƒκ°ν•˜μ„Έμš”?
09:40
Are we any better off as a society
204
580388
2137
μ†Œν”„νŠΈμ›¨μ–΄μ— 길에 λŒ€ν•œ λͺ¨λ“  κ·œμΉ™μ„ λͺ¨μ…”놓은 것이 더 λ‚˜μ€
09:42
enshrining all the rules of the road into software?
205
582525
3285
μ‚¬νšŒλ‘œ κ°€λŠ” κΈΈμΌκΉŒμš”?
09:45
No. Memory is cheaper. No.
206
585810
2598
μ•„λ‹™λ‹ˆλ‹€. κΈ°μ–΅μž₯μΉ˜κ°€ 더 μ‹ΈκΈ° λ•Œλ¬Έλ„ μ•„λ‹™λ‹ˆλ‹€.
09:48
Algorithms are faster. No. Processors are better. No.
207
588408
3994
μ•Œκ³ λ¦¬μ¦˜μ΄ 더 λΉ¨λΌμ Έμ„œλ„ μ•„λ‹ˆκ³ . ν”„λ‘œμ„Έμ„œκ°€ 더 λ‚˜μ•„μ„œλ„ μ•„λ‹™λ‹ˆλ‹€.
09:52
All of those things matter, but that's not why.
208
592402
2772
이 λͺ¨λ“  것듀이 μ€‘μš”ν•˜κΈ΄ ν•˜μ§€λ§Œ 근본적인 μ΄μœ λŠ” μ•„λ‹™λ‹ˆλ‹€.
09:55
It's because we changed the nature of the problem.
209
595174
3141
μ΄μœ λŠ” μš°λ¦¬κ°€ 문제의 성격을 λ°”κΎΈμ—ˆκΈ° λ•Œλ¬Έμž…λ‹ˆλ‹€.
09:58
We changed the nature of the problem from one
210
598315
1530
μ»΄ν“¨ν„°μ—κ²Œ μš΄μ „ν•˜λŠ” 법을 κ³Όλ„ν•˜κ³  μžμ„Έν•˜κ²Œ μ„€λͺ…ν•˜λ €ν–ˆλ˜ 것을 λ‹€μŒμ²˜λŸΌ λ°”κΎΈμ—ˆμ£ :
09:59
in which we tried to overtly and explicitly
211
599845
2245
10:02
explain to the computer how to drive
212
602090
2581
10:04
to one in which we say,
213
604671
1316
10:05
"Here's a lot of data around the vehicle.
214
605987
1876
"μ—¬κΈ° μ°¨λŸ‰μ— λŒ€ν•œ λ§Žμ€ μžλ£Œκ°€ μžˆμ–΄. λ„€κ°€ μ•Œμ•„μ„œ 잘 해봐.
10:07
You figure it out.
215
607863
1533
10:09
You figure it out that that is a traffic light,
216
609396
1867
μ €κ²Œ μ‹ ν˜Έλ“±μΈμ§€λ„ μ•Œμ•„λ‚΄κ³ ,
10:11
that that traffic light is red and not green,
217
611263
2081
μ € μ‹ ν˜Έλ“±μ΄ μ΄ˆλ‘μƒ‰μ΄ μ•„λ‹ˆλΌ 빨간색인 것도 직접 μ•Œμ•„λ‚΄μ•Όν•΄.
10:13
that that means that you need to stop
218
613344
2014
그건 μ•žμœΌλ‘œ κ°€λΌλŠ” 게 μ•„λ‹ˆλΌ λ©ˆμΆ”λΌλŠ” 말인 것도 말이야."
10:15
and not go forward."
219
615358
3083
10:18
Machine learning is at the basis
220
618441
1518
κΈ°κ³„ν•™μŠ΅μ€ μΈν„°λ„·μ—μ„œ λ§Žμ€ κ²ƒλ“€μ˜ κΈ°λ°˜μž…λ‹ˆλ‹€.
10:19
of many of the things that we do online:
221
619959
1991
10:21
search engines,
222
621950
1857
κ·Έ μ€‘μ—λŠ” κ²€μƒ‰μ—”μ§„μ΄λ‚˜ μ•„λ§ˆμ‘΄μ˜ κ°œμΈν™” μ•Œκ³ λ¦¬μ¦˜, 컴퓨터 λ²ˆμ—­,
10:23
Amazon's personalization algorithm,
223
623807
3801
10:27
computer translation,
224
627608
2212
10:29
voice recognition systems.
225
629820
4290
그리고 μŒμ„± 인식 μ‹œμŠ€ν…œλ“€μ΄ 있죠.
10:34
Researchers recently have looked at
226
634110
2835
μ—°κ΅¬μžλ“€μ€ μ΅œκ·Όμ— 암세포 쑰직 검사λ₯Ό μ‚΄νŽ΄λ³΄κ³  μžˆμŠ΅λ‹ˆλ‹€.
10:36
the question of biopsies,
227
636945
3195
10:40
cancerous biopsies,
228
640140
2767
10:42
and they've asked the computer to identify
229
642907
2315
컴퓨터가 데이터와 μƒμ‘΄μœ¨λ‘œ 세포듀이 μ‹€μ œλ‘œ 암세포인지λ₯Ό νŒλ³„ν–ˆμ£ .
10:45
by looking at the data and survival rates
230
645222
2471
10:47
to determine whether cells are actually
231
647693
4667
10:52
cancerous or not,
232
652360
2544
10:54
and sure enough, when you throw the data at it,
233
654904
1778
λ‹Ήμ—°νžˆ 데이터λ₯Ό μ£Όλ©΄ κΈ°κ³„ν•™μŠ΅ μ•Œκ³ λ¦¬μ¦˜μ„ 톡해
10:56
through a machine-learning algorithm,
234
656682
2047
10:58
the machine was able to identify
235
658729
1877
μœ λ°©μ•” 쑰직의 검사가 암인지λ₯Ό μ˜ˆμΈ‘ν•  수 μžˆλŠ” 12개의 쑰짐을 νŒλ³„ν•  수 μžˆμ—ˆμ£ .
11:00
the 12 telltale signs that best predict
236
660606
2262
11:02
that this biopsy of the breast cancer cells
237
662868
3299
11:06
are indeed cancerous.
238
666167
3218
11:09
The problem: The medical literature
239
669385
2498
λ¬Έμ œλŠ” μ˜ν•™ μ €μ„œμ—λŠ” 단지 9개만이 μ•Œλ €μ € μžˆλ‹€λŠ” κ²λ‹ˆλ‹€.
11:11
only knew nine of them.
240
671883
2789
11:14
Three of the traits were ones
241
674672
1800
μ‚¬λžŒμ€ μ‚΄νŽ΄λ³Ό ν•„μš”κ°€ μ—†μ—ˆλ˜ 3가지 νŠΉμ„±μ„ 기계가 ν¬μ°©ν•œ κ²λ‹ˆλ‹€.
11:16
that people didn't need to look for,
242
676472
2975
11:19
but that the machine spotted.
243
679447
5531
11:24
Now, there are dark sides to big data as well.
244
684978
5925
자, λΉ…λ°μ΄ν„°μ˜ μ–΄λ‘μš΄ 면도 μžˆμŠ΅λ‹ˆλ‹€.
11:30
It will improve our lives, but there are problems
245
690903
2074
λΉ…λ°μ΄ν„°λŠ” 우리의 삢을 λ‚˜μ•„μ§€κ²Œ λ§Œλ“€ κ²λ‹ˆλ‹€.
11:32
that we need to be conscious of,
246
692977
2640
ν•˜μ§€λ§Œ 문제점이 μžˆλ‹€λŠ” 사싀도 μš°λ¦¬λŠ” μ•Œκ³  μžˆμ–΄μ•Ό ν•©λ‹ˆλ‹€.
11:35
and the first one is the idea
247
695617
2623
첫 λ²ˆμ§ΈλŠ”
μš°λ¦¬κ°€ μ˜ˆμΈ‘μ— μ˜ν•΄ μ²˜λ²Œμ„ 받을 μˆ˜λ„ μžˆλ‹€λŠ” κ°œλ…μž…λ‹ˆλ‹€.
11:38
that we may be punished for predictions,
248
698240
2686
11:40
that the police may use big data for their purposes,
249
700926
3870
경찰이 μ˜ν™”, "λ§ˆμ΄λ„ˆλ¦¬ν‹° 리포트"처럼
μžμ‹ λ“€μ˜ λͺ©μ μ„ μœ„ν•΄ 빅데이터λ₯Ό μ‚¬μš©ν•  μˆ˜λ„ μžˆλ‹€λŠ” κ²λ‹ˆλ‹€.
11:44
a little bit like "Minority Report."
250
704796
2351
11:47
Now, it's a term called predictive policing,
251
707147
2441
이λ₯Ό 예츑 μΉ˜μ•ˆ, λ˜λŠ” 논리 범죄학 이라고 ν•©λ‹ˆλ‹€.
11:49
or algorithmic criminology,
252
709588
2363
11:51
and the idea is that if we take a lot of data,
253
711951
2036
과거에 범죄가 μΌμ–΄λ‚œ 지역에 데이터λ₯Ό 많이 λͺ¨μ€λ‹€λ©΄
11:53
for example where past crimes have been,
254
713987
2159
11:56
we know where to send the patrols.
255
716146
2543
μ–΄λ””λ‘œ μˆœμ°°μ„ 보내야할 지 μ•Œ 수 μžˆλ‹€λŠ” κ²λ‹ˆλ‹€.
11:58
That makes sense, but the problem, of course,
256
718689
2115
λ§žλŠ” 말이죠. ν•˜μ§€λ§Œ λ¬Έμ œλŠ”,
12:00
is that it's not simply going to stop on location data,
257
720804
4544
λ‹¨μˆœνžˆ μœ„μΉ˜ μžλ£Œμ—λ§Œ λ¨Έλ¬΄λŠ” 것이 μ•„λ‹ˆλΌ
12:05
it's going to go down to the level of the individual.
258
725348
2959
개인 μˆ˜μ€€μ—κΉŒμ§€ 이λ₯Ό κ²ƒμ΄λΌλŠ” κ²λ‹ˆλ‹€.
12:08
Why don't we use data about the person's
259
728307
2250
개인의 고등학ꡐ μ„±μ ν‘œμ— λŒ€ν•œ 자료λ₯Ό μ‚¬μš©ν•˜λŠ” 건 μ–΄λ–¨κΉŒμš”?
12:10
high school transcript?
260
730557
2228
12:12
Maybe we should use the fact that
261
732785
1561
μ‚¬λžŒλ“€μ˜ 고용 μƒνƒœλ‚˜, μ‹ μš©μ μˆ˜, 인터넷 이용 행적, λ˜λŠ”
12:14
they're unemployed or not, their credit score,
262
734346
2028
12:16
their web-surfing behavior,
263
736374
1552
12:17
whether they're up late at night.
264
737926
1878
밀에 μžμ§€μ•Šκ³  κΉ¨μ–΄μžˆλŠ” μ§€μ˜ μƒνƒœ 등을 써야할 지도 λͺ¨λ₯΄μ£ .
12:19
Their Fitbit, when it's able to identify biochemistries,
265
739804
3161
Fitbit이 신체 생리λ₯Ό μ•Œ 수 μžˆλ‹€λ©΄
12:22
will show that they have aggressive thoughts.
266
742965
4236
μ‚¬λžŒλ“€μ΄ 곡격적인 생각을 ν•˜κ³  μžˆλŠ”μ§€ 보여쀄 κ²λ‹ˆλ‹€.
12:27
We may have algorithms that are likely to predict
267
747201
2221
μš°λ¦¬κ°€ μ·¨ν•  행동을 μ˜ˆμƒν•˜λŠ” μ•Œκ³ λ¦¬μ¦˜μ„
12:29
what we are about to do,
268
749422
1633
κ°€μ§ˆ μˆ˜λ„ μžˆμ„ 것이고,
12:31
and we may be held accountable
269
751055
1244
행동을 μ·¨ν•˜κΈ°λ„ 전에 그것에 λŒ€ν•΄ μ±…μž„μ„ μ Έμ•Όν•  지도 λͺ¨λ¦…λ‹ˆλ‹€.
12:32
before we've actually acted.
270
752299
2590
12:34
Privacy was the central challenge
271
754889
1732
μžλ£Œκ°€ 적을 λ‹Ήμ‹œμ—λŠ” μ‚¬μƒν™œ λ³΄ν˜Έκ°€ μŸμ μ΄μ—ˆμŠ΅λ‹ˆλ‹€.
12:36
in a small data era.
272
756621
2880
12:39
In the big data age,
273
759501
2149
빅데이터 μ‹œλŒ€μ—λŠ”
12:41
the challenge will be safeguarding free will,
274
761650
4523
μžμœ μ˜μ§€, 도덕적 선택, μΈκ°„μ˜ μ˜μ§€, 또 선택을 λ³΄ν˜Έν•˜λŠ” 게 쟁점이 λ©λ‹ˆλ‹€.
12:46
moral choice, human volition,
275
766173
3779
12:49
human agency.
276
769952
3068
12:54
There is another problem:
277
774540
2225
또 λ‹€λ₯Έ λ¬Έμ œκ°€ μžˆμ–΄μš”.
12:56
Big data is going to steal our jobs.
278
776765
3556
빅데이터 λ•Œλ¬Έμ— μΌμžλ¦¬λŠ” 쀄어듀 κ²λ‹ˆλ‹€.
13:00
Big data and algorithms are going to challenge
279
780321
3512
빅데이터와 μ•Œκ³ λ¦¬μ¦˜μ€
21μ„ΈκΈ°μ˜ μ „λ¬Έ μ§€μ‹λΆ„μ•Όμ˜ 일에 μ’…μ‚¬ν•˜κ³  μžˆλŠ”
13:03
white collar, professional knowledge work
280
783833
3061
13:06
in the 21st century
281
786894
1653
ν™”μ΄νŠΈ 칼라에 도전μž₯을 λ‚΄λ°€ κ²ƒμž…λ‹ˆλ‹€.
13:08
in the same way that factory automation
282
788547
2434
20세기에 곡μž₯ μžλ™ν™”μ™€ 생산 라인이
13:10
and the assembly line
283
790981
2189
블루 칼라 λ…Έλ™μžλ“€μ—κ²Œ λ„μ „ν•œ κ²ƒμ²˜λŸΌ 말이죠.
13:13
challenged blue collar labor in the 20th century.
284
793170
3026
13:16
Think about a lab technician
285
796196
2092
ν˜„λ―Έκ²½μœΌλ‘œ μ•” 쑰직검사λ₯Ό μ‚΄νŽ΄λ³΄κ³  암인지λ₯Ό νŒλ³„ν•˜λŠ” μ‹€ν—˜μ‹€μ˜ μ—°κ΅¬μžλ₯Ό λ– μ˜¬λ €λ³΄μ„Έμš”.
13:18
who is looking through a microscope
286
798288
1409
13:19
at a cancer biopsy
287
799697
1624
13:21
and determining whether it's cancerous or not.
288
801321
2637
13:23
The person went to university.
289
803958
1972
λŒ€ν•™λ„ λ‚˜μ™”κ³ , 뢀동산도 사고, νˆ¬ν‘œλ„ ν•˜κ³ , 또 μ‚¬νšŒμ˜ ꡬ성원이죠.
13:25
The person buys property.
290
805930
1430
13:27
He or she votes.
291
807360
1741
13:29
He or she is a stakeholder in society.
292
809101
3666
13:32
And that person's job,
293
812767
1394
κ·Έ 직업ꡰ 뿐만 μ•„λ‹ˆλΌ, λΉ„μŠ·ν•œ 직쒅에 μžˆλŠ” λͺ¨λ“  μ‚¬λžŒλ“€μ€,
13:34
as well as an entire fleet
294
814161
1609
13:35
of professionals like that person,
295
815770
1969
13:37
is going to find that their jobs are radically changed
296
817739
3150
μžμ‹ λ“€μ˜ 직업이 μ—„μ²­λ‚˜κ²Œ λ³€ν•˜κ±°λ‚˜, μ™„μ „νžˆ μ‚¬λΌμ§€λŠ” 것을 보게 될 κ²λ‹ˆλ‹€.
13:40
or actually completely eliminated.
297
820889
2357
13:43
Now, we like to think
298
823246
1284
μš°λ¦¬λŠ” λ‹¨κΈ°μ˜ 손싀 ν›„, μž₯기간에 걸쳐 κΈ°μˆ μ€ 일자리λ₯Ό λ§Œλ“ λ‹€κ³  μ•Œκ³  있죠.
13:44
that technology creates jobs over a period of time
299
824530
3187
13:47
after a short, temporary period of dislocation,
300
827717
3465
13:51
and that is true for the frame of reference
301
831182
1941
그리고 μš°λ¦¬κ°€ κ²ͺ은 μ‚°μ—…ν˜λͺ…κΈ°κ°„ λ™μ•ˆμ΄ μ΄λž¬κΈ°μ— λ§žλ‹€κ³  생각이라고 ν•  수 있죠.
13:53
with which we all live, the Industrial Revolution,
302
833123
2142
13:55
because that's precisely what happened.
303
835265
2328
13:57
But we forget something in that analysis:
304
837593
2333
ν•˜μ§€λ§Œ κ·Έ λΆ„μ„μ—μ„œ, μš°λ¦¬κ°€ μžŠμ€ 게 μžˆμŠ΅λ‹ˆλ‹€.
13:59
There are some categories of jobs
305
839926
1830
μ–΄λ–€ 직업듀은 μ—†μ–΄μ§€κ³ λ‚˜μ„œ λ‹€μ‹œ 생기진 μ•Šμ•˜μŠ΅λ‹ˆλ‹€.
14:01
that simply get eliminated and never come back.
306
841756
3420
14:05
The Industrial Revolution wasn't very good
307
845176
2004
μ—¬λŸ¬λΆ„μ΄ λ§μ΄μ—ˆλ‹€λ©΄
14:07
if you were a horse.
308
847180
4002
μ‚°μ—…ν˜λͺ…은 κ²°μ½” 쒋은 게 μ•„λ‹ˆμ—ˆμ„ κ²λ‹ˆλ‹€.
14:11
So we're going to need to be careful
309
851182
2055
μš°λ¦¬λŠ” 주의λ₯Ό 많이 κΈ°μšΈμ—¬μ•Ό ν•˜κ³ ,
14:13
and take big data and adjust it for our needs,
310
853237
3514
빅데이터λ₯Ό 우리 μΈκ°„μ˜ ν•„μš”μ— 맞게 μ‘°μ •ν•΄μ•Ό ν•  κ²ƒμž…λ‹ˆλ‹€.
14:16
our very human needs.
311
856751
3185
14:19
We have to be the master of this technology,
312
859936
1954
μš°λ¦¬λŠ” 이 기술의 λ…Έμ˜ˆκ°€ μ•„λ‹ˆλΌ 주인이 λ˜μ–΄μ•Όλ§Œ ν•©λ‹ˆλ‹€.
14:21
not its servant.
313
861890
1656
14:23
We are just at the outset of the big data era,
314
863546
2958
빅데이터 μ‹œλŒ€λŠ” 이제 막 μ‹œμž‘λκ³ , μ†”μ§νžˆ μš°λ¦¬λŠ” ν˜„μž¬μ˜ 자료λ₯Ό λ‹€λ£¨λŠ” 것도
14:26
and honestly, we are not very good
315
866504
3150
14:29
at handling all the data that we can now collect.
316
869654
4207
μ œλŒ€λ‘œ λͺ»ν•˜κ³  μžˆμŠ΅λ‹ˆλ‹€.
14:33
It's not just a problem for the National Security Agency.
317
873861
3330
κ΅­κ°€μ•ˆλ³΄κ΅­λ§Œμ˜ λ¬Έμ œκ°€ μ•„λ‹™λ‹ˆλ‹€.
14:37
Businesses collect lots of data, and they misuse it too,
318
877191
3038
기업듀도 λ§Žμ€ 자료λ₯Ό λͺ¨μ•„μ„œ μ˜€μš©ν•˜κ³  μžˆμŠ΅λ‹ˆλ‹€.
14:40
and we need to get better at this, and this will take time.
319
880229
3667
이λ₯Ό 더 μž˜ν•΄μ•Ό ν•˜λŠ”λ° μ‹œκ°„μ΄ μ’€ 걸리긴 ν•  κ²λ‹ˆλ‹€.
14:43
It's a little bit like the challenge that was faced
320
883896
1822
μ›μ‹œμΈμ΄ λΆˆμ„ 처음 봀을 λ•Œμ˜ 어렀움과 λΉ„μŠ·ν•˜λ‹€κ³  ν•  수 μžˆκ² λ„€μš”.
14:45
by primitive man and fire.
321
885718
2407
14:48
This is a tool, but this is a tool that,
322
888125
1885
λΉ…λ°μ΄ν„°λŠ” λ„κ΅¬μ΄μ§€λ§Œ
14:50
unless we're careful, will burn us.
323
890010
3559
μ‘°μ‹¬ν•˜μ§€ μ•ŠμœΌλ©΄ 데일 μœ„ν—˜μ΄ μžˆλŠ” λ„κ΅¬μž…λ‹ˆλ‹€.
14:56
Big data is going to transform how we live,
324
896008
3120
λΉ…λ°μ΄ν„°λŠ” μš°λ¦¬κ°€ μ‚΄κ³ , μΌν•˜κ³ ,
14:59
how we work and how we think.
325
899128
2801
μƒκ°ν•˜λŠ” 방식을 바꿔놓을 κ²λ‹ˆλ‹€.
15:01
It is going to help us manage our careers
326
901929
1889
λΉ…λ°μ΄ν„°λŠ” 우리의 κ²½λ ₯ 관리에 도움을 μ£Όκ³ , 만쑱슀럽고 희망찬,
15:03
and lead lives of satisfaction and hope
327
903818
3634
그리고 ν–‰λ³΅ν•˜κ³  κ±΄κ°•ν•œ μ‚ΆμœΌλ‘œ 우리λ₯Ό μ΄λŒμ–΄ 쀄 κ²λ‹ˆλ‹€.
15:07
and happiness and health,
328
907452
2992
15:10
but in the past, we've often looked at information technology
329
910444
3306
ν•˜μ§€λ§Œ 과거에 μš°λ¦¬λŠ” 정보 κΈ°μˆ μ„ μ’…μ’… λ– μ˜¬λ ΈμŠ΅λ‹ˆλ‹€.
15:13
and our eyes have only seen the T,
330
913750
2208
우리의 λˆˆμ€ 기술과 μ œν’ˆμ—λ§Œ λˆˆμ„ λ’€μŠ΅λ‹ˆλ‹€.
15:15
the technology, the hardware,
331
915958
1686
15:17
because that's what was physical.
332
917644
2262
물질적인 κ²ƒμ΄λ‹ˆκΉŒμš”.
15:19
We now need to recast our gaze at the I,
333
919906
2924
μ΄μ œλŠ” μ •λ³΄λ‘œ λˆˆμ„ λŒλ €μ•Ό ν•  λ•Œμ£ . 덜 κ°€μ‹œμ μ΄μ§€λ§Œ
15:22
the information,
334
922830
1380
15:24
which is less apparent,
335
924210
1373
15:25
but in some ways a lot more important.
336
925583
4109
μ–΄λ–€ μ˜λ―Έμ—μ„œλŠ” λ”μš± μ€‘μš”ν•œ 것이죠.
15:29
Humanity can finally learn from the information
337
929692
3465
세상과 κ·Έ μ•ˆμ—μ„œμ˜ 우리의 μœ„μΉ˜λ₯Ό μ΄ν•΄ν•˜λŠ”
15:33
that it can collect,
338
933157
2418
λ¬΄ν•œν•œ νƒν—˜μ˜ μΌλΆ€λ‘œμ¨
15:35
as part of our timeless quest
339
935575
2115
인λ₯˜λŠ” λ§ˆμΉ¨λ‚΄ μš°λ¦¬κ°€ μˆ˜μ§‘ν•  수 μžˆλŠ” μ •λ³΄λ‘œλΆ€ν„°
15:37
to understand the world and our place in it,
340
937690
3159
λ§Žμ€ 것을 배울 수 μžˆμŠ΅λ‹ˆλ‹€.
15:40
and that's why big data is a big deal.
341
940849
5631
그리고 이게 빅데이터가 μ€‘μš”ν•œ 관건인 μ΄μœ μž…λ‹ˆλ‹€.
15:46
(Applause)
342
946480
3568
(λ°•μˆ˜)
이 μ›Ήμ‚¬μ΄νŠΈ 정보

이 μ‚¬μ΄νŠΈλŠ” μ˜μ–΄ ν•™μŠ΅μ— μœ μš©ν•œ YouTube λ™μ˜μƒμ„ μ†Œκ°œν•©λ‹ˆλ‹€. μ „ 세계 졜고의 μ„ μƒλ‹˜λ“€μ΄ κ°€λ₯΄μΉ˜λŠ” μ˜μ–΄ μˆ˜μ—…μ„ 보게 될 κ²ƒμž…λ‹ˆλ‹€. 각 λ™μ˜μƒ νŽ˜μ΄μ§€μ— ν‘œμ‹œλ˜λŠ” μ˜μ–΄ μžλ§‰μ„ 더블 ν΄λ¦­ν•˜λ©΄ κ·Έκ³³μ—μ„œ λ™μ˜μƒμ΄ μž¬μƒλ©λ‹ˆλ‹€. λΉ„λ””μ˜€ μž¬μƒμ— 맞좰 μžλ§‰μ΄ μŠ€ν¬λ‘€λ©λ‹ˆλ‹€. μ˜κ²¬μ΄λ‚˜ μš”μ²­μ΄ μžˆλŠ” 경우 이 문의 양식을 μ‚¬μš©ν•˜μ—¬ λ¬Έμ˜ν•˜μ‹­μ‹œμ˜€.

https://forms.gle/WvT1wiN1qDtmnspy7