lexicap / vtt /episode_008_small.vtt
Shubham Gupta
Add files to lfs
323a418
WEBVTT
00:00.000 --> 00:02.800
The following is a conversation with Eric Schmidt.
00:02.800 --> 00:07.640
He was the CEO of Google for ten years and a chairman for six more, guiding the company
00:07.640 --> 00:13.080
through an incredible period of growth and a series of world changing innovations.
00:13.080 --> 00:19.280
He is one of the most impactful leaders in the era of the internet and the powerful voice
00:19.280 --> 00:22.400
for the promise of technology in our society.
00:22.400 --> 00:27.760
It was truly an honor to speak with him as part of the MIT course on artificial general
00:27.760 --> 00:32.040
intelligence and the artificial intelligence podcast.
00:32.040 --> 00:37.120
And now here's my conversation with Eric Schmidt.
00:37.120 --> 00:40.840
What was the first moment when you fell in love with technology?
00:40.840 --> 00:47.080
I grew up in the 1960s as a boy where every boy wanted to be an astronaut and part of
00:47.080 --> 00:49.000
the space program.
00:49.000 --> 00:54.400
So like everyone else of my age, we would go out to the cow pasture behind my house,
00:54.400 --> 00:58.640
which was literally a cow pasture, and we would shoot model rockets off.
00:58.640 --> 01:00.200
And that I think is the beginning.
01:00.200 --> 01:05.680
And of course, generationally, today, it would be video games and all the amazing things
01:05.680 --> 01:09.280
that you can do online with computers.
01:09.280 --> 01:15.080
There's a transformative, inspiring aspect of science and math that maybe rockets would
01:15.080 --> 01:17.560
bring, would instill in individuals.
01:17.560 --> 01:21.720
You've mentioned yesterday that eighth grade math is where the journey through Mathematical
01:21.720 --> 01:24.520
University diverges from many people.
01:24.520 --> 01:27.080
It's this fork in the roadway.
01:27.080 --> 01:31.160
There's a professor of math at Berkeley, Edward Franco.
01:31.160 --> 01:32.720
I'm not sure if you're familiar with him.
01:32.720 --> 01:33.720
I am.
01:33.720 --> 01:35.400
He has written this amazing book.
01:35.400 --> 01:41.960
I recommend to everybody called Love and Math, two of my favorite words.
01:41.960 --> 01:49.680
He says that if painting was taught like math, then students would be asked to paint a fence,
01:49.680 --> 01:54.520
which is his analogy of essentially how math is taught, and you never get a chance to discover
01:54.520 --> 01:59.400
the beauty of the art of painting or the beauty of the art of math.
01:59.400 --> 02:05.240
So how, when, and where did you discover that beauty?
02:05.240 --> 02:12.040
I think what happens with people like myself is that your math enabled pretty early, and
02:12.040 --> 02:16.640
all of a sudden you discover that you can use that to discover new insights.
02:16.640 --> 02:22.120
The great scientists will all tell a story, the men and women who are fantastic today,
02:22.120 --> 02:25.560
that somewhere when they were in high school or in college, they discovered that they could
02:25.560 --> 02:30.760
discover something themselves, and that sense of building something, of having an impact
02:30.760 --> 02:35.520
that you own drives knowledge, acquisition, and learning.
02:35.520 --> 02:41.000
In my case, it was programming, and the notion that I could build things that had not existed
02:41.000 --> 02:46.560
that I had built, that it had my name on it, and this was before open source, but you could
02:46.560 --> 02:49.160
think of it as open source contributions.
02:49.160 --> 02:53.760
So today, if I were a 16 or 17 year old boy, I'm sure that I would aspire as a computer
02:53.760 --> 02:59.000
scientist to make a contribution like the open source heroes of the world today.
02:59.000 --> 03:03.760
That would be what would be driving me, and I'd be trying and learning and making mistakes,
03:03.760 --> 03:06.720
and so forth, in the ways that it works.
03:06.720 --> 03:12.360
The repository that GitHub represents and that open source libraries represent is an
03:12.360 --> 03:17.840
enormous bank of knowledge of all of the people who are doing that, and one of the lessons
03:17.840 --> 03:22.240
that I learned at Google was that the world is a very big place, and there's an awful
03:22.240 --> 03:26.360
lot of smart people, and an awful lot of them are underutilized.
03:26.360 --> 03:32.240
So here's an opportunity, for example, building parts of programs, building new ideas to contribute
03:32.240 --> 03:36.640
to the greater of society.
03:36.640 --> 03:41.000
So in that moment in the 70s, the inspiring moment where there was nothing, and then you
03:41.000 --> 03:44.800
created something through programming, that magical moment.
03:44.800 --> 03:50.360
So in 1975, I think you've created a program called Lex, which I especially like because
03:50.360 --> 03:51.560
my name is Lex.
03:51.560 --> 03:52.560
So thank you.
03:52.560 --> 03:58.240
Thank you for creating a brand that established a reputation that's long lasting reliable
03:58.240 --> 04:01.240
and has a big impact on the world and still used today.
04:01.240 --> 04:03.000
So thank you for that.
04:03.000 --> 04:11.880
But more seriously, in that time, in the 70s, as an engineer, personal computers were being
04:11.880 --> 04:12.880
born.
04:12.880 --> 04:17.800
Do you think you'd be able to predict the 80s, 90s, and the aughts of where computers
04:17.800 --> 04:18.800
would go?
04:18.800 --> 04:23.160
I'm sure I could not and would not have gotten it right.
04:23.160 --> 04:27.960
I was the beneficiary of the great work of many, many people who saw it clearer than I
04:27.960 --> 04:29.160
did.
04:29.160 --> 04:34.760
With Lex, I worked with a fellow named Michael Lesk, who was my supervisor, and he essentially
04:34.760 --> 04:39.400
helped me architect and deliver a system that's still in use today.
04:39.400 --> 04:43.800
After that, I worked at Xerox Palo Alto Research Center, where the Alto was invented, and the
04:43.800 --> 04:50.600
Alto is the predecessor of the modern personal computer, or Macintosh, and so forth.
04:50.600 --> 04:55.360
And the altos were very rare, and I had to drive an hour from Berkeley to go use them,
04:55.360 --> 05:01.160
but I made a point of skipping classes and doing whatever it took to have access to this
05:01.160 --> 05:02.480
extraordinary achievement.
05:02.480 --> 05:05.000
I knew that they were consequential.
05:05.000 --> 05:08.240
What I did not understand was scaling.
05:08.240 --> 05:12.960
I did not understand what would happen when you had 100 million as opposed to 100.
05:12.960 --> 05:17.360
And so since then, and I have learned the benefit of scale, I always look for things
05:17.360 --> 05:23.160
which are going to scale to platforms, so mobile phones, Android, all those things.
05:23.160 --> 05:27.560
There are many, many people in the world.
05:27.560 --> 05:28.560
People really have needs.
05:28.560 --> 05:32.600
They really will use these platforms, and you can build big businesses on top of them.
05:32.600 --> 05:33.600
So it's interesting.
05:33.600 --> 05:36.880
So when you see a piece of technology, now you think, what will this technology look
05:36.880 --> 05:39.080
like when it's in the hands of a billion people?
05:39.080 --> 05:40.160
That's right.
05:40.160 --> 05:46.520
So an example would be that the market is so competitive now that if you can't figure
05:46.520 --> 05:51.880
out a way for something to have a million users or a billion users, it probably is not
05:51.880 --> 05:57.280
going to be successful because something else will become the general platform, and your
05:57.280 --> 06:04.360
idea will become a lost idea or a specialized service with relatively few users.
06:04.360 --> 06:06.000
So it's a path to generality.
06:06.000 --> 06:07.720
It's a path to general platform use.
06:07.720 --> 06:09.400
It's a path to broad applicability.
06:09.400 --> 06:15.000
Now, there are plenty of good businesses that are tiny, so luxury goods, for example.
06:15.000 --> 06:20.360
But if you want to have an impact at scale, you have to look for things which are of
06:20.360 --> 06:25.200
common value, common pricing, common distribution, and solve common problems, the problems that
06:25.200 --> 06:26.200
everyone has.
06:26.200 --> 06:30.440
And by the way, people have lots of problems, information, medicine, health, education,
06:30.440 --> 06:31.440
and so forth.
06:31.440 --> 06:32.440
Work on those problems.
06:32.440 --> 06:40.240
Like you said, you're a big fan of the middle class because there's so many of them by definition.
06:40.240 --> 06:46.600
So any product, any thing that has a huge impact, it improves their lives is a great
06:46.600 --> 06:48.960
business decision and it's just good for society.
06:48.960 --> 06:53.520
And there's nothing wrong with starting off in the high end as long as you have a plan
06:53.520 --> 06:55.520
to get to the middle class.
06:55.520 --> 06:59.280
There's nothing wrong with starting with a specialized market in order to learn and to
06:59.280 --> 07:01.080
build and to fund things.
07:01.080 --> 07:04.520
So you start a luxury market to build a general purpose market.
07:04.520 --> 07:09.640
But if you define yourself as only a narrow market, someone else can come along with a
07:09.640 --> 07:14.320
general purpose market that can push you to the corner, can restrict the scale of operation,
07:14.320 --> 07:17.320
can force you to be a lesser impact than you might be.
07:17.320 --> 07:22.800
So it's very important to think in terms of broad businesses and broad impact, even if
07:22.800 --> 07:26.360
you start in a little corner somewhere.
07:26.360 --> 07:33.200
So as you look to the 70s, but also in the decades to come, and you saw computers, did
07:33.200 --> 07:40.240
you see them as tools or was there a little element of another entity?
07:40.240 --> 07:46.240
I remember a quote saying AI began with our dream to create the gods.
07:46.240 --> 07:51.520
Is there a feeling when you wrote that program that you were creating another entity, giving
07:51.520 --> 07:52.800
life to something?
07:52.800 --> 07:58.880
I wish I could say otherwise, but I simply found the technology platforms so exciting.
07:58.880 --> 08:00.400
That's what I was focused on.
08:00.400 --> 08:04.640
I think the majority of the people that I've worked with, and there are a few exceptions,
08:04.640 --> 08:09.960
Steve Jobs being an example, really saw this as a great technological play.
08:09.960 --> 08:15.520
I think relatively few of the technical people understood the scale of its impact.
08:15.520 --> 08:19.680
So I used NCP, which is a predecessor to TCPIP.
08:19.680 --> 08:21.240
It just made sense to connect things.
08:21.240 --> 08:26.240
We didn't think of it in terms of the internet, and then companies, and then Facebook, and
08:26.240 --> 08:29.200
then Twitter, and then politics, and so forth.
08:29.200 --> 08:30.800
We never did that build.
08:30.800 --> 08:32.920
We didn't have that vision.
08:32.920 --> 08:38.200
And I think most people, it's a rare person who can see compounding at scale.
08:38.200 --> 08:41.520
Most people can see, if you ask people to predict the future, they'll say, they'll give
08:41.520 --> 08:44.080
you an answer of six to nine months or 12 months.
08:44.080 --> 08:47.560
Because that's about as far as people can imagine.
08:47.560 --> 08:51.020
But there's an old saying, which actually was attributed to a professor at MIT a long
08:51.020 --> 08:58.120
time ago, that we overestimate what can be done in one year, and we underestimate what
08:58.120 --> 09:00.280
can be done in a decade.
09:00.280 --> 09:05.560
And there's a great deal of evidence that these core platforms at hardware and software
09:05.560 --> 09:07.800
take a decade.
09:07.800 --> 09:09.600
So think about self driving cars.
09:09.600 --> 09:12.160
Self driving cars were thought about in the 90s.
09:12.160 --> 09:17.160
Over projects around them, the first DARPA Durand Challenge was roughly 2004.
09:17.160 --> 09:19.760
So that's roughly 15 years ago.
09:19.760 --> 09:25.400
And today we have self driving cars operating in a city in Arizona, right, so 15 years.
09:25.400 --> 09:31.720
And we still have a ways to go before they're more generally available.
09:31.720 --> 09:33.840
So you've spoken about the importance.
09:33.840 --> 09:37.080
You just talked about predicting into the future.
09:37.080 --> 09:41.640
You've spoken about the importance of thinking five years ahead and having a plan for those
09:41.640 --> 09:42.640
five years.
09:42.640 --> 09:47.840
And the way to say it is that almost everybody has a one year plan.
09:47.840 --> 09:50.960
Almost no one has a proper five year plan.
09:50.960 --> 09:55.160
And the key thing to having a five year plan is having a model for what's going to happen
09:55.160 --> 09:57.000
under the underlying platforms.
09:57.000 --> 09:59.840
So here's an example.
09:59.840 --> 10:05.120
Computer Moore's Law, as we know it, the thing that powered improvements in CPUs has largely
10:05.120 --> 10:10.400
halted in its traditional shrinking mechanism, because the costs have just gotten so high.
10:10.400 --> 10:12.200
It's getting harder and harder.
10:12.200 --> 10:16.600
But there's plenty of algorithmic improvements and specialized hardware improvements.
10:16.600 --> 10:21.240
So you need to understand the nature of those improvements and where they'll go in order
10:21.240 --> 10:24.360
to understand how it will change the platform.
10:24.360 --> 10:28.000
In the area of network connectivity, what are the gains that are going to be possible
10:28.000 --> 10:29.480
in wireless?
10:29.480 --> 10:35.720
It looks like there's an enormous expansion of wireless connectivity at many different
10:35.720 --> 10:36.960
bands, right?
10:36.960 --> 10:40.520
And that we will primarily, historically, I've always thought that we were primarily
10:40.520 --> 10:45.040
going to be using fiber, but now it looks like we're going to be using fiber plus very
10:45.040 --> 10:51.560
powerful high bandwidth sort of short distance connectivity to bridge the last mile, right?
10:51.560 --> 10:53.100
That's an amazing achievement.
10:53.100 --> 10:56.880
If you know that, then you're going to build your systems differently.
10:56.880 --> 10:59.800
By the way, those networks have different latency properties, right?
10:59.800 --> 11:05.040
Because they're more symmetric, the algorithms feel faster for that reason.
11:05.040 --> 11:09.920
And so when you think about whether it's a fiber or just technologies in general.
11:09.920 --> 11:15.920
So there's this barber, wooden poem or quote that I really like.
11:15.920 --> 11:20.400
It's from the champions of the impossible rather than the slaves of the possible that
11:20.400 --> 11:23.280
evolution draws its creative force.
11:23.280 --> 11:27.840
So in predicting the next five years, I'd like to talk about the impossible and the
11:27.840 --> 11:28.840
possible.
11:28.840 --> 11:34.720
Well, and again, one of the great things about humanity is that we produce dreamers, right?
11:34.720 --> 11:37.760
We literally have people who have a vision and a dream.
11:37.760 --> 11:43.400
They are, if you will, disagreeable in the sense that they disagree with the, they disagree
11:43.400 --> 11:46.240
with what the sort of zeitgeist is.
11:46.240 --> 11:48.040
They say there is another way.
11:48.040 --> 11:49.040
They have a belief.
11:49.040 --> 11:50.320
They have a vision.
11:50.320 --> 11:56.560
If you look at science, science is always marked by such people who, who went against
11:56.560 --> 12:01.360
some conventional wisdom, collected the knowledge at the time and assembled it in a way that
12:01.360 --> 12:03.760
produced a powerful platform.
12:03.760 --> 12:11.120
And you've been amazingly honest about in an inspiring way about things you've been wrong
12:11.120 --> 12:14.800
about predicting and you've obviously been right about a lot of things.
12:14.800 --> 12:23.880
But in this kind of tension, how do you balance as a company in predicting the next five years,
12:23.880 --> 12:26.520
the impossible, planning for the impossible.
12:26.520 --> 12:32.720
So listening to those crazy dreamers, letting them do, letting them run away and make the
12:32.720 --> 12:38.760
impossible real, make it happen and slow, you know, that's how programmers often think
12:38.760 --> 12:44.800
and slowing things down and saying, well, this is the rational, this is the possible,
12:44.800 --> 12:49.160
the pragmatic, the, the dreamer versus the pragmatist.
12:49.160 --> 12:56.680
So it's helpful to have a model which encourages a predictable revenue stream as well as the
12:56.680 --> 12:58.720
ability to do new things.
12:58.720 --> 13:03.120
So in Google's case, we're big enough and well enough managed and so forth that we have
13:03.120 --> 13:06.600
a pretty good sense of what our revenue will be for the next year or two, at least for
13:06.600 --> 13:07.960
a while.
13:07.960 --> 13:13.720
And so we have enough cash generation that we can make bets.
13:13.720 --> 13:18.760
And indeed, Google has become alphabet so the corporation is organized around these
13:18.760 --> 13:19.760
bets.
13:19.760 --> 13:25.560
And these bets are in areas of fundamental importance to, to the world, whether it's
13:25.560 --> 13:31.920
digital intelligence, medical technology, self driving cars, connectivity through balloons,
13:31.920 --> 13:33.440
on and on and on.
13:33.440 --> 13:36.080
And there's more coming and more coming.
13:36.080 --> 13:41.480
So one way you could express this is that the current business is successful enough
13:41.480 --> 13:44.720
that we have the luxury of making bets.
13:44.720 --> 13:48.960
And another one that you could say is that we have the, the wisdom of being able to see
13:48.960 --> 13:53.920
that a corporate structure needs to be created to enhance the likelihood of the success of
13:53.920 --> 13:55.320
those bets.
13:55.320 --> 13:59.760
So we essentially turned ourselves into a conglomerate of bets.
13:59.760 --> 14:04.360
And then this underlying corporation Google, which is itself innovative.
14:04.360 --> 14:08.160
So in order to pull this off, you have to have a bunch of belief systems.
14:08.160 --> 14:12.080
And one of them is that you have to have bottoms up and tops down the bottoms up.
14:12.080 --> 14:13.600
We call 20% time.
14:13.600 --> 14:17.040
And the idea is that people can spend 20% of the time on whatever they want.
14:17.040 --> 14:21.960
And the top down is that our founders in particular have a keen eye on technology and they're
14:21.960 --> 14:24.000
reviewing things constantly.
14:24.000 --> 14:27.520
So an example would be they'll, they'll hear about an idea or I'll hear about something
14:27.520 --> 14:28.880
and it sounds interesting.
14:28.880 --> 14:34.920
Let's go visit them and then let's begin to assemble the pieces to see if that's possible.
14:34.920 --> 14:39.920
And if you do this long enough, you get pretty good at predicting what's likely to work.
14:39.920 --> 14:42.120
So that's, that's a beautiful balance that struck.
14:42.120 --> 14:44.560
Is this something that applies at all scale?
14:44.560 --> 14:53.960
So seems seems to be that the Sergei, again, 15 years ago, came up with a concept that
14:53.960 --> 14:59.040
called 10% of the budget should be on things that are unrelated.
14:59.040 --> 15:05.040
It was called 70, 20, 10, 70% of our time on core business, 20% on adjacent business
15:05.040 --> 15:06.920
and 10% on other.
15:06.920 --> 15:11.200
And he proved mathematically, of course, he's a brilliant mathematician, that you needed
15:11.200 --> 15:18.800
that 10% right to make the sum of the growth work and it turns out he was right.
15:18.800 --> 15:24.320
So getting into the world of artificial intelligence, you've, you've talked quite extensively and
15:24.320 --> 15:32.160
effectively to the impact in the near term, the positive impact of artificial intelligence,
15:32.160 --> 15:38.720
whether it's machine, especially machine learning in medical applications and education
15:38.720 --> 15:44.160
and just making information more accessible, right in the AI community, there is a kind
15:44.160 --> 15:49.520
of debate, so there's this shroud of uncertainty as we face this new world with artificial
15:49.520 --> 15:50.800
intelligence in it.
15:50.800 --> 15:57.560
And there is some people like Elon Musk, you've disagreed on at least on the degree of emphasis
15:57.560 --> 16:00.800
he places on the existential threat of AI.
16:00.800 --> 16:07.120
So I've spoken with Stuart Russell, Max Tagmark, who share Elon Musk's view, and Yoshio Benjio,
16:07.120 --> 16:09.240
Steven Pinker, who do not.
16:09.240 --> 16:13.320
And so there's a, there's a, there's a lot of very smart people who are thinking about
16:13.320 --> 16:17.280
this stuff, disagreeing, which is really healthy, of course.
16:17.280 --> 16:22.800
So what do you think is the healthiest way for the AI community to, and really for the
16:22.800 --> 16:30.880
general public to think about AI and the concern of the technology being mismanaged
16:30.880 --> 16:33.000
in some, in some kind of way.
16:33.000 --> 16:37.560
So the source of education for the general public has been robot killer movies.
16:37.560 --> 16:38.760
Right.
16:38.760 --> 16:44.840
And Terminator, et cetera, and the one thing I can assure you we're not building are those
16:44.840 --> 16:46.200
kinds of solutions.
16:46.200 --> 16:51.240
Furthermore, if they were to show up, someone would notice and unplug them, right?
16:51.240 --> 16:57.760
So as exciting as those movies are, and they're great movies, where the killer robots to start,
16:57.760 --> 17:00.520
we would find a way to stop them, right?
17:00.520 --> 17:04.320
So I'm not concerned about that.
17:04.320 --> 17:08.680
And much of this has to do with the timeframe of conversation.
17:08.680 --> 17:16.120
So you can imagine a situation a hundred years from now, when the human brain is fully understood,
17:16.120 --> 17:19.880
and the next generation and next generation of brilliant MIT scientists have figured
17:19.880 --> 17:25.960
all this out, we're going to have a large number of ethics questions, right, around science
17:25.960 --> 17:29.760
and thinking and robots and computers and so forth and so on.
17:29.760 --> 17:32.360
So it depends on the question of the timeframe.
17:32.360 --> 17:37.280
In the next five to 10 years, we're not facing those questions.
17:37.280 --> 17:42.000
What we're facing in the next five to 10 years is how do we spread this disruptive technology
17:42.000 --> 17:46.520
as broadly as possible to gain the maximum benefit of it?
17:46.520 --> 17:51.880
The primary benefit should be in healthcare and in education, healthcare because it's
17:51.880 --> 17:52.880
obvious.
17:52.880 --> 17:55.840
We're all the same, even though we somehow believe we're not.
17:55.840 --> 18:00.400
As a medical matter, the fact that we have big data about our health will save lives,
18:00.400 --> 18:05.520
allow us to deal with skin cancer and other cancers, ophthalmological problems.
18:05.520 --> 18:10.080
There's people working on psychological diseases and so forth using these techniques.
18:10.080 --> 18:11.680
I go on and on.
18:11.680 --> 18:15.840
The promise of AI in medicine is extraordinary.
18:15.840 --> 18:20.360
There are many, many companies and startups and funds and solutions and we will all live
18:20.360 --> 18:22.120
much better for that.
18:22.120 --> 18:25.680
The same argument in education.
18:25.680 --> 18:31.760
Can you imagine that for each generation of child and even adult, you have a tutor educator
18:31.760 --> 18:37.320
that's AI based, that's not a human but is properly trained, that helps you get smarter,
18:37.320 --> 18:41.440
helps you address your language difficulties or your math difficulties or what have you.
18:41.440 --> 18:43.400
Why don't we focus on those two?
18:43.400 --> 18:49.240
The gains societally of making humans smarter and healthier are enormous.
18:49.240 --> 18:54.000
Those translate for decades and decades and will all benefit from them.
18:54.000 --> 18:58.560
There are people who are working on AI safety, which is the issue that you're describing.
18:58.560 --> 19:02.920
There are conversations in the community that should there be such problems, what should
19:02.920 --> 19:04.360
the rules be like?
19:04.360 --> 19:09.360
Google, for example, has announced its policies with respect to AI safety, which I certainly
19:09.360 --> 19:14.320
support and I think most everybody would support and they make sense.
19:14.320 --> 19:19.760
It helps guide the research but the killer robots are not arriving this year and they're
19:19.760 --> 19:23.840
not even being built.
19:23.840 --> 19:31.040
On that line of thinking, you said the timescale, in this topic or other topics, have you found
19:31.040 --> 19:37.920
a useful, on the business side or the intellectual side, to think beyond 5, 10 years, to think
19:37.920 --> 19:39.480
50 years out?
19:39.480 --> 19:42.000
Has it ever been useful or productive?
19:42.000 --> 19:49.040
In our industry, there are essentially no examples of 50 year predictions that have been correct.
19:49.040 --> 19:54.320
Let's review AI, which was largely invented here at MIT and a couple of other universities
19:54.320 --> 19:57.840
in 1956, 1957, 1958.
19:57.840 --> 20:01.800
The original claims were a decade or two.
20:01.800 --> 20:08.040
When I was a PhD student, I studied AI a bit and it entered during my looking at it a period
20:08.040 --> 20:13.880
which is known as AI winter, which went on for about 30 years, which is a whole generation
20:13.880 --> 20:18.800
of scientists and a whole group of people who didn't make a lot of progress because the
20:18.800 --> 20:22.160
algorithms had not improved and the computers had not improved.
20:22.160 --> 20:26.160
It took some brilliant mathematicians, starting with a fellow named Jeff Hinton at Toronto
20:26.160 --> 20:33.120
in Montreal, who basically invented this deep learning model which empowers us today.
20:33.120 --> 20:40.400
The seminal work there was 20 years ago and in the last 10 years, it's become popularized.
20:40.400 --> 20:43.520
Think about the time frames for that level of discovery.
20:43.520 --> 20:46.080
It's very hard to predict.
20:46.080 --> 20:50.240
Many people think that we'll be flying around in the equivalent of flying cars.
20:50.240 --> 20:51.240
Who knows?
20:51.240 --> 20:56.680
My own view, if I want to go out on a limb, is to say that we know a couple of things
20:56.680 --> 20:58.000
about 50 years from now.
20:58.000 --> 21:00.480
We know that there'll be more people alive.
21:00.480 --> 21:04.000
We know that we'll have to have platforms that are more sustainable because the earth
21:04.000 --> 21:09.440
is limited in the ways we all know and that the kind of platforms that are going to get
21:09.440 --> 21:13.000
billed will be consistent with the principles that I've described.
21:13.000 --> 21:17.560
They will be much more empowering of individuals, they'll be much more sensitive to the ecology
21:17.560 --> 21:20.440
because they have to be, they just have to be.
21:20.440 --> 21:24.160
I also think that humans are going to be a great deal smarter and I think they're going
21:24.160 --> 21:28.320
to be a lot smarter because of the tools that I've discussed with you and of course people
21:28.320 --> 21:29.320
will live longer.
21:29.320 --> 21:32.080
Life extension is continuing apace.
21:32.080 --> 21:36.840
A baby born today has a reasonable chance of living to 100, which is pretty exciting.
21:36.840 --> 21:40.760
It's well past the 21st century, so we better take care of them.
21:40.760 --> 21:46.160
You mentioned interesting statistic on some very large percentage, 60, 70% of people may
21:46.160 --> 21:48.360
live in cities.
21:48.360 --> 21:53.880
Today more than half the world lives in cities and one of the great stories of humanity in
21:53.880 --> 21:57.560
the last 20 years has been the rural to urban migration.
21:57.560 --> 22:02.720
This has occurred in the United States, it's occurred in Europe, it's occurring in Asia
22:02.720 --> 22:04.760
and it's occurring in Africa.
22:04.760 --> 22:09.280
When people move to cities, the cities get more crowded, but believe it or not their health
22:09.280 --> 22:15.560
gets better, their productivity gets better, their IQ and educational capabilities improve,
22:15.560 --> 22:19.840
so it's good news that people are moving to cities, but we have to make them livable
22:19.840 --> 22:22.800
and safe.
22:22.800 --> 22:28.240
So you, first of all, you are, but you've also worked with some of the greatest leaders
22:28.240 --> 22:30.180
in the history of tech.
22:30.180 --> 22:37.080
What insights do you draw from the difference in leadership styles of yourself, Steve Jobs,
22:37.080 --> 22:45.320
Elon Musk, Larry Page, now the new CEO, Sandra Pichai and others from the, I would say, calm
22:45.320 --> 22:49.600
sages to the mad geniuses?
22:49.600 --> 22:53.880
One of the things that I learned as a young executive is that there's no single formula
22:53.880 --> 22:56.200
for leadership.
22:56.200 --> 23:00.080
They try to teach one, but that's not how it really works.
23:00.080 --> 23:04.360
There are people who just understand what they need to do and they need to do it quickly.
23:04.360 --> 23:06.800
Those people are often entrepreneurs.
23:06.800 --> 23:09.080
They just know and they move fast.
23:09.080 --> 23:13.400
There are other people who are systems thinkers and planners, that's more who I am, somewhat
23:13.400 --> 23:18.760
more conservative, more thorough in execution, a little bit more risk averse.
23:18.760 --> 23:24.120
There's also people who are sort of slightly insane, right, in the sense that they are
23:24.120 --> 23:29.040
emphatic and charismatic and they feel it and they drive it and so forth.
23:29.040 --> 23:31.440
There's no single formula to success.
23:31.440 --> 23:35.320
There is one thing that unifies all of the people that you named, which is very high
23:35.320 --> 23:41.240
intelligence, at the end of the day, the thing that characterizes all of them is that they
23:41.240 --> 23:46.360
saw the world quicker, faster, they processed information faster, they didn't necessarily
23:46.360 --> 23:50.160
make the right decisions all the time, but they were on top of it.
23:50.160 --> 23:54.600
The other thing that's interesting about all those people is they all started young.
23:54.600 --> 23:58.560
Think about Steve Jobs starting Apple roughly at 18 or 19.
23:58.560 --> 24:01.840
Think about Bill Gates starting at roughly 2021.
24:01.840 --> 24:07.040
Think about by the time they were 30, Mark Zuckerberg, a more good example at 1920.
24:07.040 --> 24:13.720
By the time they were 30, they had 10 years, at 30 years old, they had 10 years of experience
24:13.720 --> 24:19.920
of dealing with people and products and shipments and the press and business and so forth.
24:19.920 --> 24:24.480
It's incredible how much experience they had compared to the rest of us who were busy getting
24:24.480 --> 24:25.480
our PhDs.
24:25.480 --> 24:26.480
Yes, exactly.
24:26.480 --> 24:32.760
We should celebrate these people because they've just had more life experience and that helps
24:32.760 --> 24:34.520
inform the judgment.
24:34.520 --> 24:41.360
At the end of the day, when you're at the top of these organizations, all the easy questions
24:41.360 --> 24:43.680
have been dealt with.
24:43.680 --> 24:45.840
How should we design the buildings?
24:45.840 --> 24:48.400
Where should we put the colors on our product?
24:48.400 --> 24:51.440
What should the box look like?
24:51.440 --> 24:55.520
The problems, that's why it's so interesting to be in these rooms, the problems that they
24:55.520 --> 25:00.200
face in terms of the way they operate, the way they deal with their employees, their
25:00.200 --> 25:04.160
customers, their innovation are profoundly challenging.
25:04.160 --> 25:09.360
Each of the companies is demonstrably different culturally.
25:09.360 --> 25:11.800
They are not, in fact, cut of the same.
25:11.800 --> 25:16.680
They behave differently based on input, their internal cultures are different, their compensation
25:16.680 --> 25:24.920
schemes are different, their values are different, so there's proof that diversity works.
25:24.920 --> 25:33.440
So when faced with a tough decision, in need of advice, it's been said that the best thing
25:33.440 --> 25:39.840
one can do is to find the best person in the world who can give that advice and find a
25:39.840 --> 25:44.880
way to be in a room with them, one on one and ask.
25:44.880 --> 25:51.920
So here we are, and let me ask in a long winded way, I wrote this down, in 1998 there were
25:51.920 --> 26:01.960
many good search engines, Lycos, Excite, Altavista, Infoseek, AskJeeves maybe, Yahoo even.
26:01.960 --> 26:07.040
So Google stepped in and disrupted everything, they disrupted the nature of search, the nature
26:07.040 --> 26:12.040
of our access to information, the way we discover new knowledge.
26:12.040 --> 26:19.120
So now it's 2018, actually 20 years later, there are many good personal AI assistants,
26:19.120 --> 26:22.360
including of course the best from Google.
26:22.360 --> 26:28.720
So you've spoken in medical and education, the impact of such an AI assistant could bring.
26:28.720 --> 26:34.920
So we arrive at this question, so it's a personal one for me, but I hope my situation represents
26:34.920 --> 26:41.200
that of many other, as we said, dreamers and the crazy engineers.
26:41.200 --> 26:46.680
So my whole life, I've dreamed of creating such an AI assistant.
26:46.680 --> 26:50.800
Every step I've taken has been towards that goal, now I'm a research scientist in Human
26:50.800 --> 26:58.920
Senate AI here at MIT, so the next step for me as I sit here, facing my passion, is to
26:58.920 --> 27:04.880
do what Larry and Sergey did in 1998, this simple start up.
27:04.880 --> 27:10.640
And so here's my simple question, given the low odds of success, the timing and luck required,
27:10.640 --> 27:14.280
the countless other factors that can't be controlled or predicted, which is all the
27:14.280 --> 27:16.560
things that Larry and Sergey faced.
27:16.560 --> 27:23.080
Is there some calculation, some strategy to follow in this step, or do you simply follow
27:23.080 --> 27:26.560
the passion just because there's no other choice?
27:26.560 --> 27:32.880
I think the people who are in universities are always trying to study the extraordinarily
27:32.880 --> 27:37.360
chaotic nature of innovation and entrepreneurship.
27:37.360 --> 27:42.880
My answer is that they didn't have that conversation, they just did it.
27:42.880 --> 27:48.840
They sensed a moment when, in the case of Google, there was all of this data that needed
27:48.840 --> 27:53.940
to be organized and they had a better algorithm, they had invented a better way.
27:53.940 --> 28:01.040
So today with Human Senate AI, which is your area of research, there must be new approaches.
28:01.040 --> 28:07.320
It's such a big field, there must be new approaches, different from what we and others are doing.
28:07.320 --> 28:12.320
There must be startups to fund, there must be research projects to try, there must be
28:12.320 --> 28:15.200
graduate students to work on new approaches.
28:15.200 --> 28:19.120
Here at MIT, there are people who are looking at learning from the standpoint of looking
28:19.120 --> 28:23.840
at child learning, how do children learn starting at each one?
28:23.840 --> 28:25.560
And the work is fantastic.
28:25.560 --> 28:30.120
Those approaches are different from the approach that most people are taking.
28:30.120 --> 28:33.980
Perhaps that's a bet that you should make, or perhaps there's another one.
28:33.980 --> 28:40.200
But at the end of the day, the successful entrepreneurs are not as crazy as they sound.
28:40.200 --> 28:43.200
They see an opportunity based on what's happened.
28:43.200 --> 28:45.400
Let's use Uber as an example.
28:45.400 --> 28:49.840
As Travis sells the story, he and his cofounder were sitting in Paris and they had this idea
28:49.840 --> 28:52.160
because they couldn't get a cab.
28:52.160 --> 28:56.800
And they said, we have smartphones and the rest is history.
28:56.800 --> 29:04.040
So what's the equivalent of that Travis Eiffel Tower, where is a cab moment that you could,
29:04.040 --> 29:08.800
as an entrepreneur, take advantage of, whether it's in Human Senate AI or something else?
29:08.800 --> 29:11.480
That's the next great start up.
29:11.480 --> 29:13.760
And the psychology of that moment.
29:13.760 --> 29:20.120
So when Sergey and Larry talk about, and listen to a few interviews, it's very nonchalant.
29:20.120 --> 29:25.280
Well, here's the very fascinating web data.
29:25.280 --> 29:29.080
And here's an algorithm we have for, you know, we just kind of want to play around with that
29:29.080 --> 29:30.080
data.
29:30.080 --> 29:32.520
And it seems like that's a really nice way to organize this data.
29:32.520 --> 29:38.000
Well, I should say what happened, remember, is that they were graduate students at Stanford
29:38.000 --> 29:41.320
and they thought this is interesting, so they built a search engine and they kept it in
29:41.320 --> 29:43.400
their room.
29:43.400 --> 29:47.520
And they had to get power from the room next door because they were using too much power
29:47.520 --> 29:48.520
in the room.
29:48.520 --> 29:51.640
So they ran an extension cord over.
29:51.640 --> 29:55.360
And then they went and they found a house and they had Google World headquarters of
29:55.360 --> 29:57.600
five people to start the company.
29:57.600 --> 30:02.560
And they raised $100,000 from Andy Bechtelstein, who was the sun founder to do this, and Dave
30:02.560 --> 30:04.520
Chariton and a few others.
30:04.520 --> 30:11.960
The point is their beginnings were very simple, but they were based on a powerful insight.
30:11.960 --> 30:14.320
That is a replicable model for any startup.
30:14.320 --> 30:16.520
It has to be a powerful insight.
30:16.520 --> 30:17.680
The beginnings are simple.
30:17.680 --> 30:19.960
And there has to be an innovation.
30:19.960 --> 30:24.280
In Larry and Sergey's case, it was PageRank, which was a brilliant idea, one of the most
30:24.280 --> 30:26.880
cited papers in the world today.
30:26.880 --> 30:29.880
What's the next one?
30:29.880 --> 30:37.280
So you're one of, if I may say, richest people in the world, and yet it seems that money
30:37.280 --> 30:43.200
is simply a side effect of your passions and not an inherent goal.
30:43.200 --> 30:48.360
But it's a, you're a fascinating person to ask.
30:48.360 --> 30:55.080
So much of our society at the individual level and at the company level and its nations is
30:55.080 --> 30:58.920
driven by the desire for wealth.
30:58.920 --> 31:01.280
What do you think about this drive?
31:01.280 --> 31:07.000
And what have you learned about, if I may romanticize the notion, the meaning of life,
31:07.000 --> 31:10.520
having achieved success on so many dimensions?
31:10.520 --> 31:16.960
There have been many studies of human happiness and above some threshold, which is typically
31:16.960 --> 31:23.600
relatively low for this conversation, there's no difference in happiness about money.
31:23.600 --> 31:30.120
The happiness is correlated with meaning and purpose, a sense of family, a sense of impact.
31:30.120 --> 31:34.440
So if you organize your life, assuming you have enough to get around and have a nice
31:34.440 --> 31:40.400
home and so forth, you'll be far happier if you figure out what you care about and work
31:40.400 --> 31:41.800
on that.
31:41.800 --> 31:44.120
It's often being in service to others.
31:44.120 --> 31:47.840
It's a great deal of evidence that people are happiest when they're serving others
31:47.840 --> 31:49.640
and not themselves.
31:49.640 --> 31:57.480
This goes directly against the sort of press induced excitement about powerful and wealthy
31:57.480 --> 32:01.840
leaders of one kind and indeed, these are consequential people.
32:01.840 --> 32:06.720
But if you are in a situation where you've been very fortunate as I have, you also have
32:06.720 --> 32:12.160
to take that as a responsibility and you have to basically work both to educate others and
32:12.160 --> 32:16.760
give them that opportunity, but also use that wealth to advance human society.
32:16.760 --> 32:20.440
In my case, I'm particularly interested in using the tools of artificial intelligence
32:20.440 --> 32:22.800
and machine learning to make society better.
32:22.800 --> 32:24.000
I've mentioned education.
32:24.000 --> 32:29.040
I've mentioned inequality and middle class and things like this, all of which are a passion
32:29.040 --> 32:30.160
of mine.
32:30.160 --> 32:31.920
It doesn't matter what you do.
32:31.920 --> 32:36.560
It matters that you believe in it, that it's important to you and that your life will be
32:36.560 --> 32:40.480
far more satisfying if you spend your life doing that.
32:40.480 --> 32:45.320
I think there's no better place to end than a discussion of the meaning of life.
32:45.320 --> 32:46.320
Eric, thank you so much.
32:46.320 --> 32:47.320
Thank you very much.
32:47.320 --> 33:16.320
Thank you.