Leadership in the Age of AI | Paul Hudson and Lindsay Levin | TED

78,777 views ・ 2024-02-01

TED


Please double-click on the English subtitles below to play the video.

00:03
Lindsay Levin: So we're living in an era with multiple overlapping disruptions
0
3540
3840
00:07
that business is facing, and I want to dive straight in
1
7380
2600
00:09
and talk about one of the biggest of those, which is AI.
2
9980
2640
00:12
How are you approaching AI?
3
12660
1480
00:14
Paul Hudson: You know, AI at scale, it's a whole big subject, of course,
4
14180
4120
00:18
but for us, at Sanofi,
5
18340
1320
00:19
we aim to be the world's leading pharmaceutical company using AI at scale.
6
19700
5680
00:25
Why and how are we going to do that?
7
25860
1800
00:27
It's pretty straightforward.
8
27660
1360
00:29
We have 23,000 people using AI as often as every month,
9
29060
4280
00:33
9,000 people in the company using AI as often as every day.
10
33380
4920
00:38
We're boldly taking on the opportunity to completely disrupt the business.
11
38700
4280
00:43
We don't have a choice.
12
43020
1160
00:44
It's the fourth industrial revolution.
13
44180
1840
00:46
It's here whether we like it or not.
14
46060
2120
00:48
And it's amazing how resistant people can be across organizations
15
48180
4760
00:52
and across industries.
16
52980
1760
00:54
But we're all in and have been quite public about that.
17
54780
3640
00:58
Our aim is to provide daily decision intelligence,
18
58460
2880
01:01
to give people a sort of nudge in the right direction,
19
61380
3120
01:04
to give them deeper insights,
20
64540
1760
01:06
to allow them to be more effective at what they do.
21
66340
3400
01:09
And it's real.
22
69780
1160
01:10
And it's such a privilege to be involved in it.
23
70980
2200
01:13
LL: I mean, you're taking a very aggressive, active stance.
24
73180
3680
01:16
What surprised you?
25
76900
1280
01:19
PH: Well, a lot of things surprise you about AI.
26
79180
2320
01:21
I mean, for some people it's Skynet and Terminator.
27
81500
3040
01:24
For some people, they confuse AI with cyber.
28
84540
2760
01:27
I'm not saying everything is perfect,
29
87780
1800
01:29
but I'm surprised at the number of CEOs or executives who --
30
89620
4480
01:34
Their first response to an AI conversation is
31
94420
2560
01:37
"Governance, controls, rules, principles."
32
97020
2240
01:39
Of course, everything has its place,
33
99300
3200
01:42
but I think we have to be honest with ourselves.
34
102540
2280
01:44
If it is the fourth industrial revolution, which we believe it is,
35
104860
3200
01:48
then hesitating,
36
108100
2640
01:50
this fear that can take over,
37
110780
2400
01:53
can deprive you of so much opportunity.
38
113180
2760
01:55
And you have to go for it.
39
115940
1520
01:57
I find that when you talk to lots of CEOs, they really are very hesitant.
40
117500
4120
02:02
Some would say even frightened.
41
122060
1600
02:04
I look internally,
42
124020
1440
02:05
people are frightened that you get this radical data transparency
43
125460
3600
02:09
You can see their data real-time.
44
129060
1760
02:11
LL: And you're experiencing that.
45
131340
1600
02:12
PH: I've experienced that and still do.
46
132980
1920
02:14
You know, people are often shocked
47
134900
2080
02:17
that you may get the insight at the same time
48
137020
3440
02:20
as somebody lower down the organization.
49
140500
2040
02:23
And then there's the lost opportunity to polish a slide deck
50
143540
4440
02:28
and re-present it in the way that I'm supposed to be informed.
51
148020
4600
02:32
It's not a deliberate, sort of, misleading approach.
52
152620
3200
02:35
It's what people know.
53
155820
1920
02:37
They get the insight, they craft the story,
54
157740
2840
02:40
they push it upwards.
55
160620
1160
02:41
And that's life in many corporations.
56
161780
2320
02:44
For us, we get the data,
57
164500
1200
02:45
I get the same data every level of the organization does.
58
165700
2680
02:48
I get the insight exactly the same time.
59
168420
2320
02:50
And then people say, "Paul, don't look, the data is not 100 percent correct."
60
170740
5200
02:56
Well, make it correct because the data is live.
61
176660
3600
03:00
So if you really jump in and make it correct,
62
180260
2760
03:03
it'll better reflect what you're doing, right?
63
183060
2280
03:05
But if we wait for perfection it's simply not going to happen.
64
185380
3520
03:09
LL: So we're seeing fear and some of that, I guess, is not unreasonable.
65
189260
4880
03:14
You know, we read reports about the impact on job losses,
66
194180
3760
03:17
for example,
67
197980
1160
03:19
to come from AI.
68
199180
1200
03:20
I wonder what mindset you believe people need to adopt
69
200380
4400
03:24
in the workforce of all generations,
70
204820
2480
03:27
as they approach or we all approach this new future?
71
207340
3480
03:30
PH: You know, the adoption of AI in particular is not about jobs.
72
210860
5040
03:36
And I know people will think that
73
216260
1600
03:37
and inevitably, more meaningful work is created.
74
217860
2880
03:40
And of course, some roles change or some skills don't match.
75
220780
3680
03:44
And therefore, you know, with the help of many of the people in the room,
76
224500
3520
03:48
you get to reshape organizations.
77
228020
2200
03:50
But in the end,
78
230260
1480
03:51
it's really about using artificial intelligence
79
231780
3120
03:54
to create this real momentum of decision making
80
234940
4520
03:59
and to be able to take such an advantage over the competition.
81
239500
4120
04:04
And we believe, I believe,
82
244060
2040
04:06
that if you create more meaningful work
83
246100
1880
04:08
and people focus on insights and action
84
248020
1880
04:09
and less on Excel and PowerPoint and Word,
85
249940
3160
04:13
then there is a chance that they enjoy their work more.
86
253140
2800
04:15
Now it may lead to productivity gains, it may lead to efficiency.
87
255980
3080
04:19
It may lead to all those things, nobody's sort of denying that.
88
259060
3400
04:22
But what I've discovered so far is when it does,
89
262500
2280
04:24
people see it quite quickly
90
264820
1920
04:26
and they put their hand up to do something else,
91
266780
2400
04:29
or to focus more on insights
92
269180
2040
04:31
than ... data crunching
93
271260
4400
04:35
and aggregation.
94
275700
1480
04:37
You know, I'm old enough to remember when the internet was launched,
95
277220
4200
04:41
you know, commercially.
96
281460
1480
04:43
And it's sort of similar arguments, even when the cell phone was launched.
97
283380
4640
04:48
"Be careful."
98
288060
1400
04:49
But the truth is, they made all our lives better.
99
289860
2320
04:52
The use cases are coming.
100
292220
2040
04:54
But I think we're starting to understand now
101
294260
2560
04:56
how much this is going to change everybody's lives.
102
296860
2520
04:59
LL: So who is leading this in the organization?
103
299380
2360
05:01
Is it a new generation?
104
301780
1680
05:03
Is it specially appointed people?
105
303500
2160
05:05
Where's the leadership coming from?
106
305700
2200
05:07
PH: It's an excellent question
107
307940
1440
05:09
because maybe I'm the last of the great dinosaur CEOs
108
309420
4400
05:13
who got to the top by doing sort of every job.
109
313820
4520
05:19
I ran Japan, I ran North America,
110
319100
2760
05:21
you know, I was in global marketing.
111
321900
2120
05:24
You know, I've done all the tasks, to get to the top,
112
324020
2920
05:26
and then I've sort of seen everything.
113
326940
3320
05:30
And so I can be involved in every discussion.
114
330300
2640
05:33
And now the younger talent are saying,
115
333700
4320
05:38
"You didn't see AI, old man."
116
338060
1720
05:40
So, you know, "I have a better insight than you do."
117
340220
3880
05:44
And "Oh, and by the way, I'm not just going to push it up to my boss,
118
344420
3840
05:48
I'd like to tell you myself."
119
348260
1560
05:49
LL: Right.
120
349860
1200
05:51
PH: So the younger talent, justifiably is saying,
121
351060
2600
05:53
"Hey," you know, "I don't need to have my work shared upwards
122
353700
3760
05:57
by a bunch of guys
123
357460
2760
06:00
who are all sitting there going, 'What do you think?'
124
360260
2560
06:02
and none of them actually know."
125
362820
1600
06:04
And so we invert the pyramid.
126
364460
1920
06:06
We have to have different people with two, three,
127
366420
2680
06:09
four years experience in the room.
128
369140
2360
06:11
Because what do we know?
129
371540
1680
06:13
And that's sort of exciting, I think, really exciting.
130
373220
2720
06:16
LL: So is AI a unique disruptor?
131
376460
2680
06:19
If we think about some of the other big challenges, you know,
132
379180
2880
06:22
we've got to shift the entire global economy to be sustainable, for example.
133
382100
3640
06:25
Is that comparable in terms of complexity?
134
385740
2960
06:28
Maybe more so?
135
388740
1320
06:30
How do you tackle that kind of an issue?
136
390100
2560
06:32
PH: I think these are the big transformational moments for society.
137
392700
4720
06:37
And, you know, sustainability is,
138
397460
3240
06:40
you know, for many, it was carbon neutrality,
139
400700
2600
06:43
then it was net zero,
140
403300
1200
06:44
it was go to COP 28,
141
404500
1920
06:46
it was put a poster by the elevator
142
406460
2440
06:48
with the meaningful work you're doing to show your commitment.
143
408940
3720
06:52
But it's really different now.
144
412700
1800
06:54
I think there's a collective realization, certainly in healthcare,
145
414540
5320
06:59
that we didn't do well enough.
146
419860
2320
07:02
And we're a very purpose-driven organization --
147
422180
2600
07:04
an industry, in fact. We do health,
148
424780
2400
07:07
we transform the practice of medicine, we invent miracles often.
149
427180
3960
07:11
And so it's very easy to say, look, we're very purpose-driven.
150
431700
3920
07:16
But it doesn't abdicate the responsibility
151
436820
3600
07:20
of removing plastic from packaging of vaccines and medicines.
152
440420
4240
07:24
It's ridiculous to even think you wouldn't have to.
153
444660
2960
07:27
Often it's harder with the regulator, by the way, to get that done
154
447660
3120
07:30
than with your own people.
155
450780
1720
07:32
LL: So you've got those kinds of projects going on, have you?
156
452540
2880
07:35
Can you give us an example?
157
455420
1360
07:36
PH: We have to do it.
158
456820
1160
07:38
We have to do it because, you know, we have this sort of approach
159
458020
3640
07:41
of what can we do
160
461660
2920
07:44
that if we don't do, it won't happen?
161
464620
2160
07:47
And that's sort of our philosophy.
162
467180
2280
07:49
You know, in healthcare, it's a good example,
163
469500
2120
07:51
you know, delivery of healthcare creates more carbon
164
471620
3480
07:55
than the airline industry.
165
475100
2000
07:57
And that's half of that, let's say five percent.
166
477100
2240
07:59
Half of that is making drugs, shipping them, doing different things.
167
479380
3760
08:03
The other half is people driving to hospitals for an appointment
168
483180
3240
08:06
in an overheated, overcooled healthcare practice, too often,
169
486420
4200
08:10
without the use of a virtual hybrid delivery of healthcare.
170
490620
3240
08:14
And it's the same as the airline industry.
171
494180
2880
08:17
And, you know, we feel, because we're in healthcare,
172
497100
2840
08:19
we have this unique opportunity.
173
499980
1880
08:21
If somebody is pre-diabetic, for example,
174
501860
2520
08:24
and you coach them and they change their lifestyle
175
504420
2800
08:27
and don't become a diabetic,
176
507220
1880
08:29
that's a difference of them creating 16 tons of carbon
177
509140
4600
08:33
as a normal adult, healthy,
178
513780
2080
08:35
and 48 tons of carbon in their adult life if they become diabetic.
179
515900
4280
08:40
That's a 3X.
180
520740
2000
08:43
That's really meaningful.
181
523660
1640
08:45
And if we don't step in and help,
182
525300
1640
08:46
we just simply never get there and we're doing a lot of work,
183
526980
2880
08:49
I'm doing a lot of work with King Charles
184
529860
2000
08:51
and the Sustainable Markets Initiative
185
531860
2080
08:53
to get people to decarbonize the delivery of healthcare,
186
533940
3480
08:57
because it's such a massive opportunity.
187
537420
2120
08:59
LL: And presumably you've got to collaborate in very different ways
188
539540
3200
09:02
than in the past to do, you know,
189
542740
1680
09:04
you're talking about supply chains to deal with something like plastic.
190
544460
3480
09:08
Are you seeing different kinds of skills from people
191
548300
2520
09:10
in order to make those collaborations?
192
550860
2840
09:13
PH: Well, I think these functions and the sustainability groups, as I said,
193
553700
4160
09:17
have come from a poster by the elevator
194
557900
1920
09:19
to being very actively involved in a lot of work to do these things properly.
195
559860
3720
09:23
And, you know, you, it's not about a competitive advantage,
196
563620
4480
09:28
in Sanofi being better at wastewater management
197
568140
4000
09:32
or renewable energy than Pfizer or AstraZeneca.
198
572180
4520
09:36
That's not a competition.
199
576700
1920
09:38
The competition is us versus, you know, destroying the planet.
200
578660
3800
09:42
So we work a lot together to do the right thing.
201
582900
3840
09:46
I work with Novo Nordisk, AstraZeneca, GSK
202
586740
4120
09:50
to try and work out ways to be kinder to the environment
203
590900
3560
09:54
in the delivery of health, and it's the right thing to do, right?
204
594460
3080
09:57
It's a shared responsibility, a collective responsibility.
205
597540
3680
10:01
LL: And we're talking here about big social challenges
206
601260
2960
10:04
beyond any one business
207
604260
1880
10:06
or industry or even country,
208
606140
2600
10:08
with an expectation that business needs to step forward.
209
608740
3280
10:12
I think partly driven by the fact that policy doesn't always work,
210
612020
3120
10:15
and we're disappointed with leadership and with government.
211
615180
2760
10:17
So a lot of finger pointing as to who's responsible
212
617980
3080
10:21
and who can act on these big shifts.
213
621100
2520
10:23
I wonder how ambitious and bold
214
623620
3960
10:27
do you feel CEOs should be about stepping forward
215
627620
4600
10:32
and helping society through some of these mega transitions
216
632260
4400
10:36
that we are faced with?
217
636700
1560
10:38
PH: You know, it's clear that companies are being pulled more
218
638300
4880
10:43
into the conversation about individual's values.
219
643220
4000
10:47
And I think people who work in our company,
220
647660
3640
10:51
and all companies,
221
651300
1360
10:52
start to try and identify themselves, perhaps rightly,
222
652660
2960
10:55
with the values of the company.
223
655660
2000
10:57
And they're starting to have much higher expectations
224
657660
2920
11:00
about the company they work for.
225
660620
2520
11:03
And it can be on all the major social issues.
226
663180
4000
11:07
You know, there's been so many difficult moments and tragic moments.
227
667180
3720
11:11
I'm often written to by people from all over the world.
228
671340
2760
11:14
"You haven't declared which side you're on on this important subject.
229
674100
3880
11:18
Why not?"
230
678020
1160
11:20
And people want to know that you are fully vested
231
680020
3280
11:23
and the company is behind them.
232
683300
2160
11:25
To be clear, it's almost impossible to get everything right.
233
685500
3320
11:28
The world is almost, you know, perfectly divided.
234
688860
2840
11:32
You can pick an issue and half your employees will tell you
235
692220
3480
11:35
"We don't agree," and the other half will say, "Well done."
236
695740
3120
11:39
And we're not used to that as CEOs,
237
699260
1720
11:41
we're used to trying to find the right sort of moment
238
701020
3800
11:44
to get the majority to say, "I'm proud of my company."
239
704820
3360
11:48
So you have to retreat a little bit and say,
240
708740
2600
11:51
I'll spend my energy on making sure whatever the issue, that the people,
241
711380
4520
11:55
90,000 people in our case,
242
715940
2320
11:58
get the very best chance to live their best version of themselves.
243
718300
3600
12:01
Could be inclusion, it could be race,
244
721940
3080
12:05
it could be many different things.
245
725060
1920
12:07
It could be LGBTQ-plus.
246
727420
1880
12:09
But whatever those issues that are being debated strongly
247
729860
3400
12:13
or less strongly in different parts of the world,
248
733300
2440
12:15
it matters to us that our people feel
249
735780
2400
12:18
they can be the best version of themselves.
250
738180
2040
12:20
So when people ask, "What do you think about this?"
251
740260
2400
12:22
"What do you think about that?"
252
742700
1480
12:24
I can have a personal opinion, but our organization,
253
744220
2520
12:26
if you're looking to match your values with ours,
254
746780
2320
12:29
is really about, we're in healthcare,
255
749140
2160
12:31
we want you to have the best life possible.
256
751340
2320
12:33
How can the company facilitate that in this sort of, maelstrom of subjects?
257
753700
6080
12:39
And we focus then -- I'm not saying we're perfect,
258
759780
2400
12:42
I don't think any company really is perfect on this,
259
762180
2440
12:44
but the expectation of our own employees to be able to set a standard on an issue
260
764660
6280
12:50
and to see it through is real.
261
770940
2120
12:53
And leading in these times is more complicated, I think,
262
773060
2640
12:55
than perhaps it has been previously.
263
775740
1720
12:57
LL: So you're describing a world
264
777500
1840
12:59
where the work of the CEO is changing very fast.
265
779380
3600
13:03
Could you just sum up for us the role of a leader in this new era?
266
783740
5000
13:09
PH: Well, I took this role at the end of 2019 and thought,
267
789820
3200
13:13
I will roll through my 100-day plan
268
793060
1960
13:15
and I will amble through getting around town halls across the world
269
795060
4880
13:19
and shake a lot of hands and do a listening tour.
270
799980
2960
13:22
And a pandemic dropped on us within a few weeks after that
271
802980
3840
13:26
and worked out of the kitchen not far from here.
272
806860
4440
13:31
And the war, Russia-Ukraine,
273
811820
2800
13:34
current war, Israel and Hamas,
274
814620
5120
13:39
and mentioned the pandemic.
275
819780
2480
13:42
China-US, the social issues.
276
822300
3520
13:46
And I think what we realize is leading
277
826860
2240
13:49
is I think somebody described it as the perma crisis, you know,
278
829140
4240
13:53
a sort of, you know, a cadence of issues
279
833380
2560
13:55
that just is relentless.
280
835980
1320
13:57
And you really have to have some resiliency leading now, I think,
281
837940
5440
14:03
and you have to recognize that there are a series of sprints in the role.
282
843420
4440
14:07
There's the fundamentals of the business that must be continued.
283
847900
3120
14:11
Then there's a metronome for us.
284
851060
1720
14:12
We feel a responsibility to bring medicines
285
852820
2480
14:15
and transformational medicines forward is non-negotiable.
286
855340
3880
14:19
At the same time,
287
859220
1400
14:20
parts of your organization is in a very difficult situation
288
860660
3760
14:24
somewhere in the world.
289
864420
1360
14:25
And we have to make sure we have the right energy, experts,
290
865820
3800
14:29
support, crisis teams, often, more recently
291
869620
3960
14:33
to protect our people
292
873620
2560
14:36
and to continue the work we do.
293
876220
1560
14:37
We ship drugs all over the world, irrespective of the stance of politics
294
877820
5360
14:43
or anything else.
295
883220
1160
14:44
People are people and if they're sick we'll help.
296
884380
2320
14:46
But it really has got very complicated.
297
886740
1920
14:48
So resiliency, agility and being open-minded,
298
888660
3240
14:51
recognizing you're not the expert many times,
299
891940
3120
14:55
trusting the advice you get from your own people,
300
895060
2320
14:57
particularly those on the ground,
301
897420
1600
14:59
protecting your people where it's necessary,
302
899020
2080
15:01
moving your people where it's necessary.
303
901100
2720
15:03
That's the sort of new expectation, really.
304
903820
2600
15:06
LL: I mean, I think we're all experiencing this sense of perma crisis.
305
906820
3680
15:10
Just to finish,
306
910900
1160
15:12
it would be great to just get a sense
307
912060
2120
15:14
of what are you really excited about right now?
308
914180
2400
15:17
PH: Well, I'm incredibly excited,
309
917460
2720
15:20
I touched on it at the beginning about the use of artificial intelligence,
310
920180
3520
15:23
particularly large language models.
311
923700
1760
15:26
Because I think it changes everything.
312
926340
1920
15:28
It's got me questioning whether I can go back
313
928260
2360
15:30
and look at medicines that didn't quite make it,
314
930660
2600
15:33
and wonder if we just didn't know enough
315
933300
2320
15:35
with the data that we had to look deep enough.
316
935660
2880
15:39
It can be, you know, recently we just did our --
317
939140
2960
15:42
this is a small example, but it's just fun,
318
942140
2000
15:44
our engagement survey,
319
944180
1720
15:45
we had a 409,000 comments,
320
945900
3200
15:49
9 million words.
321
949100
1840
15:50
And normally, somebody would make a nice slide deck
322
950940
3000
15:53
to tell me, "The organization poll is very engaged.
323
953940
3040
15:57
Never been more engaged."
324
957020
1880
15:58
"How much more engaged?" "0.1."
325
958940
1600
16:00
"OK, good, thank you."
326
960540
2120
16:02
Definitely improvement.
327
962660
1760
16:04
And so I asked them to run the 409,000 comments
328
964460
3440
16:07
through a large language model.
329
967940
1520
16:09
Forty minutes later, it told me the three things
330
969780
2240
16:12
that people love about the company
331
972060
1640
16:13
and three things that people hate about it.
332
973740
2080
16:15
I didn't need a lot of external support, didn't need teams of people.
333
975820
4640
16:20
And it was clear, it was no hallucination,
334
980500
2160
16:22
because it was there right in front of me.
335
982700
2120
16:24
And it made sense.
336
984860
1160
16:26
And I shared them with people and they're like, yeah, that's us.
337
986060
3240
16:29
And I think that's the difference between meaningful work
338
989340
2880
16:32
using -- let's talk about what it tells us
339
992260
2520
16:34
about whether our people really like it here
340
994820
2680
16:37
and bring their best or not.
341
997500
1760
16:39
I think AI for me, a relatively new CEO,
342
999300
3520
16:42
I have a chance to disrupt structural biology, electron microscopy,
343
1002820
4520
16:47
I have a chance to invent medicines
344
1007380
1680
16:49
and druggable targets that were never touched before.
345
1009100
2520
16:51
I have a chance to take away the sort of,
346
1011620
2960
16:54
I should put it, the heavy lifting,
347
1014580
2280
16:56
unglamorous work that people don't like doing.
348
1016900
3520
17:01
I've a chance to reinvent everything,
349
1021020
1760
17:02
to do it more efficiently, reinvest in R and D.
350
1022820
2200
17:05
And I have a chance to get ahead of the competition
351
1025060
2920
17:08
while they're all worrying,
352
1028020
2680
17:10
we have governance,
353
1030740
1840
17:12
but every step forward by us
354
1032620
2120
17:14
is a step of leaving behind those that are overly sensitive,
355
1034740
3280
17:18
and we're happy to share.
356
1038060
1560
17:19
But we can do incredible things for patients
357
1039620
2080
17:21
and for the people in the company
358
1041700
1600
17:23
by being more bold about stepping into the new world.
359
1043340
2680
17:26
LL: Your passion is infectious.
360
1046060
1640
17:27
Thank you.
361
1047700
1200
17:28
Thank you very much, Paul.
362
1048900
1640
17:30
PH: Thank you.
363
1050580
1200
17:31
(Applause)
364
1051820
1600
About this website

This site will introduce you to YouTube videos that are useful for learning English. You will see English lessons taught by top-notch teachers from around the world. Double-click on the English subtitles displayed on each video page to play the video from there. The subtitles scroll in sync with the video playback. If you have any comments or requests, please contact us using this contact form.

https://forms.gle/WvT1wiN1qDtmnspy7