lexicap / vtt /episode_011_small.vtt
Shubham Gupta
Add readme and files
a3be5d0
raw
history blame
93.8 kB
WEBVTT
00:00.000 --> 00:03.520
The following is a conversation with Jürgen Schmidhuber.
00:03.520 --> 00:06.360
He's the codirector of its CS Swiss AI lab
00:06.360 --> 00:10.360
and a cocreator of long short term memory networks.
00:10.360 --> 00:13.720
LSDMs are used in billions of devices today
00:13.720 --> 00:17.400
for speech recognition, translation, and much more.
00:17.400 --> 00:20.800
Over 30 years, he has proposed a lot of interesting
00:20.800 --> 00:24.800
out of the box ideas, a meta learning, adversarial networks,
00:24.800 --> 00:28.720
computer vision, and even a formal theory of quote,
00:28.720 --> 00:32.360
creativity, curiosity, and fun.
00:32.360 --> 00:34.920
This conversation is part of the MIT course
00:34.920 --> 00:36.520
on artificial general intelligence
00:36.520 --> 00:38.840
and the artificial intelligence podcast.
00:38.840 --> 00:41.960
If you enjoy it, subscribe on YouTube, iTunes,
00:41.960 --> 00:43.960
or simply connect with me on Twitter
00:43.960 --> 00:47.280
at Lex Friedman spelled F R I D.
00:47.280 --> 00:51.480
And now here's my conversation with Jürgen Schmidhuber.
00:53.080 --> 00:55.640
Early on you dreamed of AI systems
00:55.640 --> 00:58.680
that self improve recursively.
00:58.680 --> 01:01.440
When was that dream born?
01:01.440 --> 01:02.840
When I was a baby.
01:02.840 --> 01:04.000
No, that's not true.
01:04.000 --> 01:06.200
When I was a teenager.
01:06.200 --> 01:09.400
And what was the catalyst for that birth?
01:09.400 --> 01:12.800
What was the thing that first inspired you?
01:12.800 --> 01:15.000
When I was a boy, I...
01:17.400 --> 01:19.880
I was thinking about what to do in my life
01:19.880 --> 01:23.560
and then I thought the most exciting thing
01:23.560 --> 01:27.160
is to solve the riddles of the universe.
01:27.160 --> 01:30.720
And that means you have to become a physicist.
01:30.720 --> 01:35.640
However, then I realized that there's something even grander.
01:35.640 --> 01:39.680
You can try to build a machine.
01:39.680 --> 01:41.920
That isn't really a machine any longer.
01:41.920 --> 01:44.280
That learns to become a much better physicist
01:44.280 --> 01:46.840
than I could ever hope to be.
01:46.840 --> 01:50.120
And that's how I thought maybe I can multiply
01:50.120 --> 01:54.320
my tiny little bit of creativity into infinity.
01:54.320 --> 01:57.160
But ultimately, that creativity will be multiplied
01:57.160 --> 01:59.160
to understand the universe around us.
01:59.160 --> 02:05.640
That's the curiosity for that mystery that drove you.
02:05.640 --> 02:08.320
Yes, so if you can build a machine
02:08.320 --> 02:13.760
that learns to solve more and more complex problems
02:13.760 --> 02:16.720
and more and more general problems over,
02:16.720 --> 02:22.520
then you basically have solved all the problems.
02:22.520 --> 02:25.960
At least all the solvable problems.
02:25.960 --> 02:27.080
So how do you think...
02:27.080 --> 02:31.440
What is the mechanism for that kind of general solver look like?
02:31.440 --> 02:35.480
Obviously, we don't quite yet have one or know
02:35.480 --> 02:37.040
how to build one boy of ideas
02:37.040 --> 02:40.800
and you have had throughout your career several ideas about it.
02:40.800 --> 02:43.600
So how do you think about that mechanism?
02:43.600 --> 02:48.640
So in the 80s, I thought about how to build this machine
02:48.640 --> 02:51.000
that learns to solve all these problems
02:51.000 --> 02:54.120
that I cannot solve myself.
02:54.120 --> 02:57.120
And I thought it is clear, it has to be a machine
02:57.120 --> 03:00.880
that not only learns to solve this problem here
03:00.880 --> 03:02.640
and this problem here,
03:02.640 --> 03:06.240
but it also has to learn to improve
03:06.240 --> 03:09.360
the learning algorithm itself.
03:09.360 --> 03:12.480
So it has to have the learning algorithm
03:12.480 --> 03:15.720
in a representation that allows it to inspect it
03:15.720 --> 03:19.240
and modify it so that it can come up
03:19.240 --> 03:22.080
with a better learning algorithm.
03:22.080 --> 03:25.680
So I called that meta learning, learning to learn
03:25.680 --> 03:28.040
and recursive self improvement.
03:28.040 --> 03:29.840
That is really the pinnacle of that,
03:29.840 --> 03:35.960
where you then not only learn how to improve
03:35.960 --> 03:37.480
on that problem and on that,
03:37.480 --> 03:41.080
but you also improve the way the machine improves
03:41.080 --> 03:43.480
and you also improve the way it improves the way
03:43.480 --> 03:45.720
it improves itself.
03:45.720 --> 03:48.560
And that was my 1987 diploma thesis,
03:48.560 --> 03:53.200
which was all about that hierarchy of meta learners
03:53.200 --> 03:57.240
that have no computational limits
03:57.240 --> 03:59.920
except for the well known limits
03:59.920 --> 04:03.160
that Gödel identified in 1931
04:03.160 --> 04:05.640
and for the limits of physics.
04:06.480 --> 04:10.040
In the recent years, meta learning has gained popularity
04:10.040 --> 04:12.760
in a specific kind of form.
04:12.760 --> 04:16.000
You've talked about how that's not really meta learning
04:16.000 --> 04:21.000
with neural networks, that's more basic transfer learning.
04:21.480 --> 04:22.720
Can you talk about the difference
04:22.720 --> 04:25.440
between the big general meta learning
04:25.440 --> 04:27.960
and a more narrow sense of meta learning
04:27.960 --> 04:30.880
the way it's used today, the way it's talked about today?
04:30.880 --> 04:33.440
Let's take the example of a deep neural network
04:33.440 --> 04:37.240
that has learned to classify images.
04:37.240 --> 04:40.080
And maybe you have trained that network
04:40.080 --> 04:43.800
on 100 different databases of images.
04:43.800 --> 04:48.120
And now a new database comes along
04:48.120 --> 04:52.000
and you want to quickly learn the new thing as well.
04:53.400 --> 04:57.720
So one simple way of doing that is you take the network
04:57.720 --> 05:02.440
which already knows 100 types of databases
05:02.440 --> 05:06.320
and then you just take the top layer of that
05:06.320 --> 05:11.320
and you retrain that using the new labeled data
05:11.320 --> 05:14.720
that you have in the new image database.
05:14.720 --> 05:17.320
And then it turns out that it really, really quickly
05:17.320 --> 05:20.560
can learn that too, one shot basically,
05:20.560 --> 05:24.280
because from the first 100 data sets,
05:24.280 --> 05:27.520
it already has learned so much about computer vision
05:27.520 --> 05:31.840
that it can reuse that and that is then almost good enough
05:31.840 --> 05:34.240
to solve the new tasks except you need a little bit
05:34.240 --> 05:37.040
of adjustment on the top.
05:37.040 --> 05:40.200
So that is transfer learning
05:40.200 --> 05:43.480
and it has been done in principle for many decades.
05:43.480 --> 05:45.680
People have done similar things for decades.
05:47.360 --> 05:49.880
Meta learning, true meta learning is about
05:49.880 --> 05:53.880
having the learning algorithm itself
05:54.440 --> 05:59.440
open to introspection by the system that is using it
06:00.440 --> 06:03.760
and also open to modification
06:03.760 --> 06:07.200
such that the learning system has an opportunity
06:07.200 --> 06:11.400
to modify any part of the learning algorithm
06:11.400 --> 06:16.000
and then evaluate the consequences of that modification
06:16.000 --> 06:21.000
and then learn from that to create a better learning algorithm
06:22.000 --> 06:23.960
and so on recursively.
06:24.960 --> 06:27.680
So that's a very different animal
06:27.680 --> 06:32.680
where you are opening the space of possible learning algorithms
06:32.680 --> 06:35.520
to the learning system itself.
06:35.520 --> 06:39.040
Right, so you've like in the 2004 paper,
06:39.040 --> 06:42.440
you describe Gatal machines and programs
06:42.440 --> 06:44.520
that rewrite themselves, right?
06:44.520 --> 06:47.520
Philosophically and even in your paper mathematically,
06:47.520 --> 06:50.000
these are really compelling ideas,
06:50.000 --> 06:55.000
but practically, do you see these self referential programs
06:55.320 --> 06:59.400
being successful in the near term to having an impact
06:59.400 --> 07:03.040
where sort of it demonstrates to the world
07:03.040 --> 07:08.040
that this direction is a good one to pursue in the near term?
07:08.680 --> 07:11.400
Yes, we had these two different types
07:11.400 --> 07:13.440
of fundamental research,
07:13.440 --> 07:15.840
how to build a universal problem solver,
07:15.840 --> 07:19.840
one basically exploiting proof search
07:23.000 --> 07:24.960
and things like that that you need to come up
07:24.960 --> 07:29.960
with asymptotically optimal, theoretically optimal
07:30.320 --> 07:33.200
self improvers and problem solvers.
07:34.200 --> 07:39.200
However, one has to admit that through this proof search
07:40.640 --> 07:43.640
comes in an additive constant,
07:43.640 --> 07:46.800
an overhead, an additive overhead
07:46.800 --> 07:51.800
that vanishes in comparison to what you have to do
07:51.800 --> 07:53.960
to solve large problems.
07:53.960 --> 07:56.920
However, for many of the small problems
07:56.920 --> 07:59.920
that we want to solve in our everyday life,
07:59.920 --> 08:02.440
we cannot ignore this constant overhead.
08:02.440 --> 08:07.440
And that's why we also have been doing other things,
08:07.440 --> 08:11.160
non universal things such as recurrent neural networks
08:11.160 --> 08:14.360
which are trained by gradient descent
08:14.360 --> 08:17.600
and local search techniques which aren't universal at all,
08:17.600 --> 08:20.280
which aren't provably optimal at all
08:20.280 --> 08:22.000
like the other stuff that we did,
08:22.000 --> 08:24.600
but which are much more practical
08:24.600 --> 08:27.920
as long as we only want to solve the small problems
08:27.920 --> 08:32.920
that we are typically trying to solve in this environment here.
08:34.680 --> 08:38.200
So the universal problem solvers like the Gödel machine
08:38.200 --> 08:41.320
but also Markus Hutter's fastest way
08:41.320 --> 08:43.560
of solving all possible problems,
08:43.560 --> 08:47.360
which he developed around 2002 in my lab,
08:47.360 --> 08:51.280
they are associated with these constant overheads
08:51.280 --> 08:53.280
for proof search, which guarantees
08:53.280 --> 08:55.480
that the thing that you're doing is optimal.
08:55.480 --> 08:59.880
For example, there is this fastest way
08:59.880 --> 09:03.880
of solving all problems with a computable solution
09:03.880 --> 09:05.880
which is due to Markus Hutter.
09:05.880 --> 09:10.880
And to explain what's going on there,
09:10.880 --> 09:13.040
let's take traveling salesman problems.
09:14.240 --> 09:16.160
With traveling salesman problems,
09:16.160 --> 09:20.080
you have a number of cities, N cities,
09:20.080 --> 09:22.480
and you try to find the shortest path
09:22.480 --> 09:26.480
through all these cities without visiting any city twice.
09:28.480 --> 09:31.040
And nobody knows the fastest way
09:31.040 --> 09:35.040
of solving traveling salesman problems, TSPs,
09:37.520 --> 09:40.480
but let's assume there is a method of solving them
09:40.480 --> 09:44.480
within N to the five operations
09:44.480 --> 09:48.560
where N is the number of cities.
09:50.160 --> 09:54.560
Then the universal method of Markus
09:54.560 --> 09:58.560
is going to solve the same traveling salesman problem
09:58.560 --> 10:02.080
also within N to the five steps,
10:02.080 --> 10:06.360
plus O of one, plus a constant number of steps
10:06.360 --> 10:09.240
that you need for the proof searcher,
10:09.240 --> 10:13.800
which you need to show that this particular
10:13.800 --> 10:17.240
class of problems that traveling salesman problems
10:17.240 --> 10:19.360
can be solved within a certain time bound,
10:20.520 --> 10:24.400
within order N to the five steps, basically.
10:24.400 --> 10:28.520
And this additive constant doesn't care for N,
10:28.520 --> 10:32.400
which means as N is getting larger and larger,
10:32.400 --> 10:34.880
as you have more and more cities,
10:34.880 --> 10:38.600
the constant overhead pales and comparison.
10:38.600 --> 10:44.120
And that means that almost all large problems are solved
10:44.120 --> 10:46.520
in the best possible way already today.
10:46.520 --> 10:50.480
We already have a universal problem solved like that.
10:50.480 --> 10:54.520
However, it's not practical because the overhead,
10:54.520 --> 10:57.440
the constant overhead is so large
10:57.440 --> 11:00.200
that for the small kinds of problems
11:00.200 --> 11:04.560
that we want to solve in this little biosphere.
11:04.560 --> 11:06.360
By the way, when you say small,
11:06.360 --> 11:08.600
you're talking about things that fall
11:08.600 --> 11:10.880
within the constraints of our computational systems.
11:10.880 --> 11:14.280
So they can seem quite large to us mere humans.
11:14.280 --> 11:15.360
That's right, yeah.
11:15.360 --> 11:19.000
So they seem large and even unsolvable
11:19.000 --> 11:21.000
in a practical sense today,
11:21.000 --> 11:24.760
but they are still small compared to almost all problems
11:24.760 --> 11:28.480
because almost all problems are large problems,
11:28.480 --> 11:30.840
which are much larger than any constant.
11:31.920 --> 11:34.520
Do you find it useful as a person
11:34.520 --> 11:38.680
who is dreamed of creating a general learning system,
11:38.680 --> 11:39.880
has worked on creating one,
11:39.880 --> 11:42.160
has done a lot of interesting ideas there
11:42.160 --> 11:46.360
to think about P versus NP,
11:46.360 --> 11:50.800
this formalization of how hard problems are,
11:50.800 --> 11:52.360
how they scale,
11:52.360 --> 11:55.200
this kind of worst case analysis type of thinking.
11:55.200 --> 11:56.840
Do you find that useful?
11:56.840 --> 11:59.720
Or is it only just a mathematical,
12:00.560 --> 12:02.640
it's a set of mathematical techniques
12:02.640 --> 12:05.760
to give you intuition about what's good and bad?
12:05.760 --> 12:09.440
So P versus NP, that's super interesting
12:09.440 --> 12:11.800
from a theoretical point of view.
12:11.800 --> 12:14.560
And in fact, as you are thinking about that problem,
12:14.560 --> 12:17.280
you can also get inspiration
12:17.280 --> 12:21.280
for better practical problem solvers.
12:21.280 --> 12:23.320
On the other hand, we have to admit
12:23.320 --> 12:24.560
that at the moment,
12:24.560 --> 12:28.360
the best practical problem solvers
12:28.360 --> 12:30.120
for all kinds of problems
12:30.120 --> 12:33.880
that we are now solving through what is called AI at the moment,
12:33.880 --> 12:36.240
they are not of the kind
12:36.240 --> 12:38.800
that is inspired by these questions.
12:38.800 --> 12:42.680
There we are using general purpose computers,
12:42.680 --> 12:44.840
such as recurrent neural networks,
12:44.840 --> 12:46.680
but we have a search technique,
12:46.680 --> 12:50.320
which is just local search gradient descent
12:50.320 --> 12:51.960
to try to find a program
12:51.960 --> 12:54.400
that is running on these recurrent networks,
12:54.400 --> 12:58.160
such that it can solve some interesting problems,
12:58.160 --> 13:01.920
such as speech recognition or machine translation
13:01.920 --> 13:03.200
and something like that.
13:03.200 --> 13:06.480
And there is very little theory
13:06.480 --> 13:09.720
behind the best solutions that we have at the moment
13:09.720 --> 13:10.800
that can do that.
13:10.800 --> 13:12.640
Do you think that needs to change?
13:12.640 --> 13:15.120
Do you think that will change or can we go,
13:15.120 --> 13:17.120
can we create a general intelligence systems
13:17.120 --> 13:19.200
without ever really proving
13:19.200 --> 13:20.600
that that system is intelligent
13:20.600 --> 13:22.560
in some kind of mathematical way,
13:22.560 --> 13:24.960
solving machine translation perfectly
13:24.960 --> 13:26.320
or something like that,
13:26.320 --> 13:29.160
within some kind of syntactic definition of a language?
13:29.160 --> 13:31.120
Or can we just be super impressed
13:31.120 --> 13:35.080
by the thing working extremely well and that's sufficient?
13:35.080 --> 13:36.720
There's an old saying,
13:36.720 --> 13:39.360
and I don't know who brought it up first,
13:39.360 --> 13:42.440
which says there's nothing more practical
13:42.440 --> 13:43.680
than a good theory.
13:43.680 --> 13:48.680
And a good theory of problem solving
13:52.760 --> 13:55.560
under limited resources like here in this universe
13:55.560 --> 13:57.000
or on this little planet
13:58.480 --> 14:01.800
has to take into account these limited resources.
14:01.800 --> 14:06.800
And so probably there is locking a theory
14:08.040 --> 14:10.800
which is related to what we already have,
14:10.800 --> 14:14.440
these asymptotically optimal problem solvers,
14:14.440 --> 14:18.560
which tells us what we need in addition to that
14:18.560 --> 14:21.760
to come up with a practically optimal problem solver.
14:21.760 --> 14:26.760
So I believe we will have something like that
14:27.080 --> 14:29.720
and maybe just a few little tiny twists
14:29.720 --> 14:34.320
are necessary to change what we already have
14:34.320 --> 14:36.360
to come up with that as well.
14:36.360 --> 14:37.800
As long as we don't have that,
14:37.800 --> 14:42.600
we admit that we are taking suboptimal ways
14:42.600 --> 14:46.040
and recurrent neural networks and long short term memory
14:46.040 --> 14:50.440
for equipped with local search techniques
14:50.440 --> 14:53.560
and we are happy that it works better
14:53.560 --> 14:55.480
than any competing methods,
14:55.480 --> 15:00.480
but that doesn't mean that we think we are done.
15:00.800 --> 15:05.040
You've said that an AGI system will ultimately be a simple one,
15:05.040 --> 15:08.000
a general intelligence system will ultimately be a simple one,
15:08.000 --> 15:10.240
maybe a pseudo code of a few lines
15:10.240 --> 15:11.840
will be able to describe it.
15:11.840 --> 15:16.760
Can you talk through your intuition behind this idea,
15:16.760 --> 15:21.760
why you feel that at its core intelligence
15:22.120 --> 15:25.560
is a simple algorithm?
15:26.920 --> 15:31.680
Experience tells us that the stuff that works best
15:31.680 --> 15:33.120
is really simple.
15:33.120 --> 15:37.640
So the asymptotically optimal ways of solving problems,
15:37.640 --> 15:38.800
if you look at them,
15:38.800 --> 15:41.800
they're just a few lines of code, it's really true.
15:41.800 --> 15:44.000
Although they are these amazing properties,
15:44.000 --> 15:45.760
just a few lines of code,
15:45.760 --> 15:50.760
then the most promising and most useful practical things
15:53.760 --> 15:57.760
maybe don't have this proof of optimality associated with them.
15:57.760 --> 16:00.840
However, they are also just a few lines of code.
16:00.840 --> 16:05.040
The most successful recurrent neural networks,
16:05.040 --> 16:08.360
you can write them down and five lines of pseudo code.
16:08.360 --> 16:10.920
That's a beautiful, almost poetic idea,
16:10.920 --> 16:15.600
but what you're describing there
16:15.600 --> 16:17.400
is the lines of pseudo code
16:17.400 --> 16:20.600
are sitting on top of layers and layers of abstractions,
16:20.600 --> 16:22.240
in a sense.
16:22.240 --> 16:25.040
So you're saying at the very top,
16:25.040 --> 16:30.040
it'll be a beautifully written sort of algorithm,
16:31.120 --> 16:33.960
but do you think that there's many layers of abstractions
16:33.960 --> 16:36.880
we have to first learn to construct?
16:36.880 --> 16:38.280
Yeah, of course.
16:38.280 --> 16:42.640
We are building on all these great abstractions
16:42.640 --> 16:46.040
that people have invented over the millennia,
16:46.040 --> 16:51.040
such as matrix multiplications and drill numbers
16:51.600 --> 16:56.600
and basic arithmetics and calculus and derivations
16:58.720 --> 17:03.320
of error functions and derivatives of error functions
17:03.320 --> 17:04.320
and stuff like that.
17:05.440 --> 17:10.440
So without that language that greatly simplifies
17:10.440 --> 17:13.880
our way of thinking about these problems,
17:13.880 --> 17:14.840
we couldn't do anything.
17:14.840 --> 17:16.560
So in that sense, as always,
17:16.560 --> 17:19.600
we are standing on the shoulders of the giants
17:19.600 --> 17:24.600
who in the past simplified the problem of problem solving
17:25.520 --> 17:30.000
so much that now we have a chance to do the final step.
17:30.000 --> 17:32.120
So the final step will be a simple one.
17:34.000 --> 17:36.760
If we take a step back through all of human civilization
17:36.760 --> 17:38.360
and just the universe in general,
17:38.360 --> 17:41.440
how do you think about evolution?
17:41.440 --> 17:45.400
And what if creating a universe is required
17:45.400 --> 17:47.320
to achieve this final step?
17:47.320 --> 17:50.920
What if going through the very painful
17:50.920 --> 17:53.840
and inefficient process of evolution is needed
17:53.840 --> 17:55.880
to come up with this set of abstractions
17:55.880 --> 17:57.800
that ultimately lead to intelligence?
17:57.800 --> 18:00.800
Do you think there's a shortcut
18:00.800 --> 18:04.640
or do you think we have to create something like our universe
18:04.640 --> 18:09.480
in order to create something like human level intelligence?
18:09.480 --> 18:13.160
So far, the only example we have is this one,
18:13.160 --> 18:15.160
this universe in which we are living.
18:15.160 --> 18:16.360
You think you can do better?
18:20.880 --> 18:25.000
Maybe not, but we are part of this whole process.
18:25.000 --> 18:30.000
So apparently, so it might be the case
18:30.000 --> 18:32.160
that the code that runs the universe
18:32.160 --> 18:33.720
is really, really simple.
18:33.720 --> 18:36.640
Everything points to that possibility
18:36.640 --> 18:39.960
because gravity and other basic forces
18:39.960 --> 18:44.120
are really simple laws that can be easily described,
18:44.120 --> 18:47.080
also in just a few lines of code, basically.
18:47.080 --> 18:52.080
And then there are these other events
18:52.200 --> 18:55.080
that the apparently random events
18:55.080 --> 18:56.560
in the history of the universe,
18:56.560 --> 18:58.800
which as far as we know at the moment
18:58.800 --> 19:00.720
don't have a compact code,
19:00.720 --> 19:03.240
but who knows, maybe somebody in the near future
19:03.240 --> 19:06.800
is going to figure out the pseudo random generator,
19:06.800 --> 19:11.800
which is computing whether the measurement of that
19:13.520 --> 19:15.920
spin up or down thing here
19:15.920 --> 19:18.440
is going to be positive or negative.
19:18.440 --> 19:19.880
Underline quantum mechanics.
19:19.880 --> 19:20.720
Yes, so.
19:20.720 --> 19:23.160
Do you ultimately think quantum mechanics
19:23.160 --> 19:25.200
is a pseudo random number generator?
19:25.200 --> 19:26.920
So it's all deterministic.
19:26.920 --> 19:28.760
There's no randomness in our universe.
19:30.400 --> 19:31.800
Does God play dice?
19:31.800 --> 19:34.080
So a couple of years ago,
19:34.080 --> 19:39.080
a famous physicist, quantum physicist, Anton Zeilinger,
19:39.080 --> 19:41.600
he wrote an essay in Nature,
19:41.600 --> 19:44.280
and it started more or less like that.
19:46.720 --> 19:51.720
One of the fundamental insights of the 20th century
19:53.280 --> 19:58.280
was that the universe is fundamentally random
19:58.280 --> 20:02.760
on the quantum level, and that whenever
20:03.760 --> 20:06.720
you measure spin up or down or something like that,
20:06.720 --> 20:10.720
a new bit of information enters the history of the universe.
20:13.440 --> 20:14.680
And while I was reading that,
20:14.680 --> 20:18.000
I was already typing the response
20:18.000 --> 20:20.280
and they had to publish it because I was right,
20:21.560 --> 20:25.560
that there is no evidence, no physical evidence for that.
20:25.560 --> 20:28.440
So there's an alternative explanation
20:28.440 --> 20:31.240
where everything that we consider random
20:31.240 --> 20:33.800
is actually pseudo random,
20:33.800 --> 20:38.800
such as the decimal expansion of pi, 3.141 and so on,
20:39.400 --> 20:42.120
which looks random, but isn't.
20:42.120 --> 20:47.120
So pi is interesting because every three digit sequence,
20:47.720 --> 20:51.720
every sequence of three digits appears roughly
20:51.720 --> 20:56.720
one in a thousand times, and every five digit sequence
20:57.360 --> 21:00.760
appears roughly one in 10,000 times.
21:00.760 --> 21:02.760
What do you expect?
21:02.760 --> 21:06.760
If it was random, but there's a very short algorithm,
21:06.760 --> 21:09.120
a short program that computes all of that.
21:09.120 --> 21:11.200
So it's extremely compressible.
21:11.200 --> 21:13.120
And who knows, maybe tomorrow somebody,
21:13.120 --> 21:15.360
some grad student at CERN goes back
21:15.360 --> 21:19.120
over all these data points, better decay,
21:19.120 --> 21:21.760
and whatever, and figures out, oh,
21:21.760 --> 21:25.760
it's the second billion digits of pi or something like that.
21:25.760 --> 21:28.840
We don't have any fundamental reason at the moment
21:28.840 --> 21:33.600
to believe that this is truly random
21:33.600 --> 21:36.440
and not just a deterministic video game.
21:36.440 --> 21:38.680
If it was a deterministic video game,
21:38.680 --> 21:40.360
it would be much more beautiful
21:40.360 --> 21:44.160
because beauty is simplicity.
21:44.160 --> 21:47.560
And many of the basic laws of the universe
21:47.560 --> 21:51.560
like gravity and the other basic forces are very simple.
21:51.560 --> 21:55.560
So very short programs can explain what these are doing.
21:56.560 --> 22:00.560
And it would be awful and ugly.
22:00.560 --> 22:01.560
The universe would be ugly.
22:01.560 --> 22:03.560
The history of the universe would be ugly
22:03.560 --> 22:06.560
if for the extra things, the random,
22:06.560 --> 22:10.560
the seemingly random data points that we get all the time
22:10.560 --> 22:15.560
that we really need a huge number of extra bits
22:15.560 --> 22:21.560
to describe all these extra bits of information.
22:22.560 --> 22:25.560
So as long as we don't have evidence
22:25.560 --> 22:27.560
that there is no short program
22:27.560 --> 22:32.560
that computes the entire history of the entire universe,
22:32.560 --> 22:38.560
we are, as scientists, compelled to look further
22:38.560 --> 22:41.560
for that shortest program.
22:41.560 --> 22:46.560
Your intuition says there exists a program
22:46.560 --> 22:50.560
that can backtrack to the creation of the universe.
22:50.560 --> 22:53.560
So it can take the shortest path to the creation of the universe.
22:53.560 --> 22:57.560
Yes, including all the entanglement things
22:57.560 --> 23:01.560
and all the spin up and down measurements
23:01.560 --> 23:09.560
that have been taken place since 13.8 billion years ago.
23:09.560 --> 23:14.560
So we don't have a proof that it is random.
23:14.560 --> 23:19.560
We don't have a proof that it is compressible to a short program.
23:19.560 --> 23:21.560
But as long as we don't have that proof,
23:21.560 --> 23:24.560
we are obliged as scientists to keep looking
23:24.560 --> 23:26.560
for that simple explanation.
23:26.560 --> 23:27.560
Absolutely.
23:27.560 --> 23:30.560
So you said simplicity is beautiful or beauty is simple.
23:30.560 --> 23:32.560
Either one works.
23:32.560 --> 23:36.560
But you also work on curiosity, discovery.
23:36.560 --> 23:42.560
The romantic notion of randomness, of serendipity,
23:42.560 --> 23:49.560
of being surprised by things that are about you,
23:49.560 --> 23:53.560
kind of in our poetic notion of reality,
23:53.560 --> 23:56.560
we think as humans require randomness.
23:56.560 --> 23:58.560
So you don't find randomness beautiful.
23:58.560 --> 24:04.560
You find simple determinism beautiful.
24:04.560 --> 24:06.560
Yeah.
24:06.560 --> 24:07.560
Okay.
24:07.560 --> 24:08.560
So why?
24:08.560 --> 24:09.560
Why?
24:09.560 --> 24:12.560
Because the explanation becomes shorter.
24:12.560 --> 24:19.560
A universe that is compressible to a short program
24:19.560 --> 24:22.560
is much more elegant and much more beautiful
24:22.560 --> 24:24.560
than another one,
24:24.560 --> 24:28.560
which needs an almost infinite number of bits to be described.
24:28.560 --> 24:31.560
As far as we know,
24:31.560 --> 24:34.560
many things that are happening in this universe are really simple
24:34.560 --> 24:38.560
in terms of short programs that compute gravity
24:38.560 --> 24:43.560
and the interaction between elementary particles and so on.
24:43.560 --> 24:45.560
So all of that seems to be very, very simple.
24:45.560 --> 24:50.560
Every electron seems to reuse the same subprogram all the time
24:50.560 --> 24:57.560
as it is interacting with other elementary particles.
24:57.560 --> 25:04.560
If we now require an extra oracle
25:04.560 --> 25:07.560
injecting new bits of information all the time
25:07.560 --> 25:11.560
for these extra things which are currently not understood,
25:11.560 --> 25:18.560
such as better decay,
25:18.560 --> 25:25.560
then the whole description length of the data that we can observe
25:25.560 --> 25:31.560
of the history of the universe would become much longer.
25:31.560 --> 25:33.560
And therefore, uglier.
25:33.560 --> 25:34.560
And uglier.
25:34.560 --> 25:38.560
Again, the simplicity is elegant and beautiful.
25:38.560 --> 25:42.560
All the history of science is a history of compression progress.
25:42.560 --> 25:43.560
Yeah.
25:43.560 --> 25:48.560
So you've described sort of as we build up abstractions
25:48.560 --> 25:52.560
and you've talked about the idea of compression.
25:52.560 --> 25:55.560
How do you see this, the history of science,
25:55.560 --> 25:59.560
the history of humanity, our civilization and life on Earth
25:59.560 --> 26:03.560
as some kind of path towards greater and greater compression?
26:03.560 --> 26:04.560
What do you mean by that?
26:04.560 --> 26:06.560
How do you think about that?
26:06.560 --> 26:12.560
Indeed, the history of science is a history of compression progress.
26:12.560 --> 26:14.560
What does that mean?
26:14.560 --> 26:17.560
Hundreds of years ago, there was an astronomer
26:17.560 --> 26:19.560
whose name was Kepler.
26:19.560 --> 26:25.560
And he looked at the data points that he got by watching planets move.
26:25.560 --> 26:28.560
And then he had all these data points and suddenly it turned out
26:28.560 --> 26:37.560
that he can greatly compress the data by predicting it through an ellipse law.
26:37.560 --> 26:44.560
So it turns out that all these data points are more or less on ellipses around the sun.
26:44.560 --> 26:50.560
And another guy came along whose name was Newton and before him Hook.
26:50.560 --> 26:57.560
And they said the same thing that is making these planets move like that
26:57.560 --> 27:01.560
is what makes the apples fall down.
27:01.560 --> 27:10.560
And it also holds for stones and for all kinds of other objects.
27:10.560 --> 27:16.560
And suddenly many, many of these observations became much more compressible
27:16.560 --> 27:19.560
because as long as you can predict the next thing,
27:19.560 --> 27:22.560
given what you have seen so far, you can compress it.
27:22.560 --> 27:24.560
But you don't have to store that data extra.
27:24.560 --> 27:28.560
This is called predictive coding.
27:28.560 --> 27:33.560
And then there was still something wrong with that theory of the universe
27:33.560 --> 27:37.560
and you had deviations from these predictions of the theory.
27:37.560 --> 27:41.560
And 300 years later another guy came along whose name was Einstein
27:41.560 --> 27:50.560
and he was able to explain away all these deviations from the predictions of the old theory
27:50.560 --> 27:56.560
through a new theory which was called the general theory of relativity
27:56.560 --> 28:00.560
which at first glance looks a little bit more complicated
28:00.560 --> 28:05.560
and you have to warp space and time but you can't phrase it within one single sentence
28:05.560 --> 28:12.560
which is no matter how fast you accelerate and how fast or how hard you decelerate
28:12.560 --> 28:18.560
and no matter what is the gravity in your local framework,
28:18.560 --> 28:21.560
light speed always looks the same.
28:21.560 --> 28:24.560
And from that you can calculate all the consequences.
28:24.560 --> 28:30.560
So it's a very simple thing and it allows you to further compress all the observations
28:30.560 --> 28:35.560
because certainly there are hardly any deviations any longer
28:35.560 --> 28:39.560
that you can measure from the predictions of this new theory.
28:39.560 --> 28:44.560
So art of science is a history of compression progress.
28:44.560 --> 28:50.560
You never arrive immediately at the shortest explanation of the data
28:50.560 --> 28:52.560
but you're making progress.
28:52.560 --> 28:56.560
Whenever you are making progress you have an insight.
28:56.560 --> 29:01.560
You see, oh, first I needed so many bits of information to describe the data,
29:01.560 --> 29:04.560
to describe my falling apples, my video of falling apples,
29:04.560 --> 29:08.560
I need so many data, so many pixels have to be stored
29:08.560 --> 29:14.560
but then suddenly I realize, no, there is a very simple way of predicting the third frame
29:14.560 --> 29:20.560
in the video from the first two and maybe not every little detail can be predicted
29:20.560 --> 29:24.560
but more or less most of these orange blots that are coming down,
29:24.560 --> 29:28.560
I'm sorry, in the same way, which means that I can greatly compress the video
29:28.560 --> 29:33.560
and the amount of compression, progress,
29:33.560 --> 29:37.560
that is the depth of the insight that you have at that moment.
29:37.560 --> 29:40.560
That's the fun that you have, the scientific fun,
29:40.560 --> 29:46.560
the fun in that discovery and we can build artificial systems that do the same thing.
29:46.560 --> 29:51.560
They measure the depth of their insights as they are looking at the data
29:51.560 --> 29:55.560
through their own experiments and we give them a reward,
29:55.560 --> 30:00.560
an intrinsic reward and proportion to this depth of insight.
30:00.560 --> 30:07.560
And since they are trying to maximize the rewards they get,
30:07.560 --> 30:12.560
they are suddenly motivated to come up with new action sequences,
30:12.560 --> 30:17.560
with new experiments that have the property that the data that is coming in
30:17.560 --> 30:21.560
as a consequence of these experiments has the property
30:21.560 --> 30:25.560
that they can learn something about, see a pattern in there
30:25.560 --> 30:28.560
which they hadn't seen yet before.
30:28.560 --> 30:32.560
So there's an idea of power play that you've described,
30:32.560 --> 30:37.560
a training and general problem solver in this kind of way of looking for the unsolved problems.
30:37.560 --> 30:40.560
Can you describe that idea a little further?
30:40.560 --> 30:42.560
It's another very simple idea.
30:42.560 --> 30:49.560
Normally what you do in computer science, you have some guy who gives you a problem
30:49.560 --> 30:56.560
and then there is a huge search space of potential solution candidates
30:56.560 --> 31:02.560
and you somehow try them out and you have more or less sophisticated ways
31:02.560 --> 31:06.560
of moving around in that search space
31:06.560 --> 31:11.560
until you finally found a solution which you consider satisfactory.
31:11.560 --> 31:15.560
That's what most of computer science is about.
31:15.560 --> 31:19.560
Power play just goes one little step further and says,
31:19.560 --> 31:24.560
let's not only search for solutions to a given problem,
31:24.560 --> 31:30.560
but let's search to pairs of problems and their solutions
31:30.560 --> 31:36.560
where the system itself has the opportunity to phrase its own problem.
31:36.560 --> 31:42.560
So we are looking suddenly at pairs of problems and their solutions
31:42.560 --> 31:46.560
or modifications of the problem solver
31:46.560 --> 31:50.560
that is supposed to generate a solution to that new problem.
31:50.560 --> 31:56.560
And this additional degree of freedom
31:56.560 --> 32:01.560
allows us to build career systems that are like scientists
32:01.560 --> 32:06.560
in the sense that they not only try to solve and try to find answers
32:06.560 --> 32:12.560
to existing questions, no, they are also free to pose their own questions.
32:12.560 --> 32:15.560
So if you want to build an artificial scientist,
32:15.560 --> 32:19.560
you have to give it that freedom and power play is exactly doing that.
32:19.560 --> 32:23.560
So that's a dimension of freedom that's important to have,
32:23.560 --> 32:31.560
how hard do you think that, how multi dimensional and difficult the space of
32:31.560 --> 32:34.560
then coming up with your own questions is.
32:34.560 --> 32:38.560
So it's one of the things that as human beings we consider to be
32:38.560 --> 32:41.560
the thing that makes us special, the intelligence that makes us special
32:41.560 --> 32:47.560
is that brilliant insight that can create something totally new.
32:47.560 --> 32:51.560
Yes. So now let's look at the extreme case.
32:51.560 --> 32:57.560
Let's look at the set of all possible problems that you can formally describe,
32:57.560 --> 33:03.560
which is infinite, which should be the next problem
33:03.560 --> 33:07.560
that a scientist or power play is going to solve.
33:07.560 --> 33:16.560
Well, it should be the easiest problem that goes beyond what you already know.
33:16.560 --> 33:22.560
So it should be the simplest problem that the current problems
33:22.560 --> 33:28.560
that you have which can already solve 100 problems that he cannot solve yet
33:28.560 --> 33:30.560
by just generalizing.
33:30.560 --> 33:32.560
So it has to be new.
33:32.560 --> 33:36.560
So it has to require a modification of the problem solver such that the new
33:36.560 --> 33:41.560
problem solver can solve this new thing, but the old problem solver cannot do it.
33:41.560 --> 33:47.560
And in addition to that, we have to make sure that the problem solver
33:47.560 --> 33:50.560
doesn't forget any of the previous solutions.
33:50.560 --> 33:51.560
Right.
33:51.560 --> 33:57.560
And so by definition, power play is now trying always to search in this pair of
33:57.560 --> 34:02.560
in the set of pairs of problems and problems over modifications
34:02.560 --> 34:08.560
for a combination that minimize the time to achieve these criteria.
34:08.560 --> 34:14.560
Power is trying to find the problem which is easiest to add to the repertoire.
34:14.560 --> 34:19.560
So just like grad students and academics and researchers can spend their whole
34:19.560 --> 34:25.560
career in a local minima stuck trying to come up with interesting questions,
34:25.560 --> 34:27.560
but ultimately doing very little.
34:27.560 --> 34:32.560
Do you think it's easy in this approach of looking for the simplest
34:32.560 --> 34:38.560
problem solver problem to get stuck in a local minima is not never really discovering
34:38.560 --> 34:43.560
new, you know, really jumping outside of the hundred problems that you've already
34:43.560 --> 34:47.560
solved in a genuine creative way.
34:47.560 --> 34:52.560
No, because that's the nature of power play that it's always trying to break
34:52.560 --> 34:58.560
its current generalization abilities by coming up with a new problem which is
34:58.560 --> 35:04.560
beyond the current horizon, just shifting the horizon of knowledge a little bit
35:04.560 --> 35:10.560
out there, breaking the existing rules such that the new thing becomes solvable
35:10.560 --> 35:13.560
but wasn't solvable by the old thing.
35:13.560 --> 35:19.560
So like adding a new axiom, like what Gödel did when he came up with these
35:19.560 --> 35:23.560
new sentences, new theorems that didn't have a proof in the formal system,
35:23.560 --> 35:30.560
which means you can add them to the repertoire, hoping that they are not
35:30.560 --> 35:35.560
going to damage the consistency of the whole thing.
35:35.560 --> 35:41.560
So in the paper with the amazing title, Formal Theory of Creativity,
35:41.560 --> 35:47.560
Fun and Intrinsic Motivation, you talk about discovery as intrinsic reward.
35:47.560 --> 35:53.560
So if you view humans as intelligent agents, what do you think is the purpose
35:53.560 --> 35:56.560
and meaning of life for us humans?
35:56.560 --> 35:58.560
You've talked about this discovery.
35:58.560 --> 36:04.560
Do you see humans as an instance of power play agents?
36:04.560 --> 36:11.560
Yeah, so humans are curious and that means they behave like scientists,
36:11.560 --> 36:15.560
not only the official scientists but even the babies behave like scientists
36:15.560 --> 36:19.560
and they play around with their toys to figure out how the world works
36:19.560 --> 36:22.560
and how it is responding to their actions.
36:22.560 --> 36:26.560
And that's how they learn about gravity and everything.
36:26.560 --> 36:30.560
And yeah, in 1990, we had the first systems like that
36:30.560 --> 36:33.560
who would just try to play around with the environment
36:33.560 --> 36:39.560
and come up with situations that go beyond what they knew at that time
36:39.560 --> 36:42.560
and then get a reward for creating these situations
36:42.560 --> 36:45.560
and then becoming more general problem solvers
36:45.560 --> 36:48.560
and being able to understand more of the world.
36:48.560 --> 36:56.560
So yeah, I think in principle that curiosity,
36:56.560 --> 37:02.560
strategy or more sophisticated versions of what I just described,
37:02.560 --> 37:07.560
they are what we have built in as well because evolution discovered
37:07.560 --> 37:12.560
that's a good way of exploring the unknown world and a guy who explores
37:12.560 --> 37:16.560
the unknown world has a higher chance of solving problems
37:16.560 --> 37:19.560
that he needs to survive in this world.
37:19.560 --> 37:23.560
On the other hand, those guys who were too curious,
37:23.560 --> 37:25.560
they were weeded out as well.
37:25.560 --> 37:27.560
So you have to find this trade off.
37:27.560 --> 37:30.560
Evolution found a certain trade off apparently in our society.
37:30.560 --> 37:35.560
There is a certain percentage of extremely explorative guys
37:35.560 --> 37:41.560
and it doesn't matter if they die because many of the others are more conservative.
37:41.560 --> 37:46.560
And so yeah, it would be surprising to me
37:46.560 --> 37:55.560
if that principle of artificial curiosity wouldn't be present
37:55.560 --> 37:59.560
in almost exactly the same form here in our brains.
37:59.560 --> 38:02.560
So you're a bit of a musician and an artist.
38:02.560 --> 38:07.560
So continuing on this topic of creativity,
38:07.560 --> 38:10.560
what do you think is the role of creativity in intelligence?
38:10.560 --> 38:16.560
So you've kind of implied that it's essential for intelligence,
38:16.560 --> 38:21.560
if you think of intelligence as a problem solving system,
38:21.560 --> 38:23.560
as ability to solve problems.
38:23.560 --> 38:28.560
But do you think it's essential, this idea of creativity?
38:28.560 --> 38:34.560
We never have a subprogram that is called creativity or something.
38:34.560 --> 38:37.560
It's just a side effect of what our problems always do.
38:37.560 --> 38:44.560
They are searching a space of candidates, of solution candidates,
38:44.560 --> 38:47.560
until they hopefully find a solution to a given problem.
38:47.560 --> 38:50.560
But then there are these two types of creativity
38:50.560 --> 38:53.560
and both of them are now present in our machines.
38:53.560 --> 38:56.560
The first one has been around for a long time,
38:56.560 --> 38:59.560
which is human gives problem to machine.
38:59.560 --> 39:03.560
Machine tries to find a solution to that.
39:03.560 --> 39:05.560
And this has been happening for many decades.
39:05.560 --> 39:09.560
And for many decades, machines have found creative solutions
39:09.560 --> 39:13.560
to interesting problems where humans were not aware
39:13.560 --> 39:17.560
of these particularly creative solutions,
39:17.560 --> 39:20.560
but then appreciated that the machine found that.
39:20.560 --> 39:23.560
The second is the pure creativity.
39:23.560 --> 39:28.560
What I just mentioned, I would call the applied creativity,
39:28.560 --> 39:31.560
like applied art, where somebody tells you,
39:31.560 --> 39:34.560
now make a nice picture of this pope,
39:34.560 --> 39:36.560
and you will get money for that.
39:36.560 --> 39:41.560
So here is the artist and he makes a convincing picture of the pope
39:41.560 --> 39:44.560
and the pope likes it and gives him the money.
39:44.560 --> 39:48.560
And then there is the pure creativity,
39:48.560 --> 39:51.560
which is more like the power play and the artificial curiosity thing,
39:51.560 --> 39:56.560
where you have the freedom to select your own problem,
39:56.560 --> 40:02.560
like a scientist who defines his own question to study.
40:02.560 --> 40:06.560
And so that is the pure creativity, if you will,
40:06.560 --> 40:13.560
as opposed to the applied creativity, which serves another.
40:13.560 --> 40:18.560
In that distinction, there's almost echoes of narrow AI versus general AI.
40:18.560 --> 40:24.560
So this kind of constrained painting of a pope seems like
40:24.560 --> 40:29.560
the approaches of what people are calling narrow AI.
40:29.560 --> 40:32.560
And pure creativity seems to be,
40:32.560 --> 40:34.560
maybe I'm just biased as a human,
40:34.560 --> 40:40.560
but it seems to be an essential element of human level intelligence.
40:40.560 --> 40:43.560
Is that what you're implying?
40:43.560 --> 40:45.560
To a degree.
40:45.560 --> 40:50.560
If you zoom back a little bit and you just look at a general problem solving machine,
40:50.560 --> 40:53.560
which is trying to solve arbitrary problems,
40:53.560 --> 40:57.560
then this machine will figure out in the course of solving problems
40:57.560 --> 40:59.560
that it's good to be curious.
40:59.560 --> 41:04.560
So all of what I said just now about this pre wild curiosity
41:04.560 --> 41:10.560
and this will to invent new problems that the system doesn't know how to solve yet,
41:10.560 --> 41:14.560
should be just a byproduct of the general search.
41:14.560 --> 41:21.560
However, apparently evolution has built it into us
41:21.560 --> 41:26.560
because it turned out to be so successful, a pre wiring, a bias,
41:26.560 --> 41:33.560
a very successful exploratory bias that we are born with.
41:33.560 --> 41:36.560
And you've also said that consciousness in the same kind of way
41:36.560 --> 41:40.560
may be a byproduct of problem solving.
41:40.560 --> 41:44.560
Do you find this an interesting byproduct?
41:44.560 --> 41:46.560
Do you think it's a useful byproduct?
41:46.560 --> 41:49.560
What are your thoughts on consciousness in general?
41:49.560 --> 41:54.560
Or is it simply a byproduct of greater and greater capabilities of problem solving
41:54.560 --> 42:00.560
that's similar to creativity in that sense?
42:00.560 --> 42:04.560
We never have a procedure called consciousness in our machines.
42:04.560 --> 42:10.560
However, we get a side effects of what these machines are doing,
42:10.560 --> 42:15.560
things that seem to be closely related to what people call consciousness.
42:15.560 --> 42:20.560
So for example, already 1990 we had simple systems
42:20.560 --> 42:25.560
which were basically recurrent networks and therefore universal computers
42:25.560 --> 42:32.560
trying to map incoming data into actions that lead to success.
42:32.560 --> 42:39.560
Maximizing reward in a given environment, always finding the charging station in time
42:39.560 --> 42:43.560
whenever the battery is low and negative signals are coming from the battery,
42:43.560 --> 42:50.560
always find the charging station in time without bumping against painful obstacles on the way.
42:50.560 --> 42:54.560
So complicated things but very easily motivated.
42:54.560 --> 43:01.560
And then we give these little guys a separate recurrent network
43:01.560 --> 43:04.560
which is just predicting what's happening if I do that and that.
43:04.560 --> 43:08.560
What will happen as a consequence of these actions that I'm executing
43:08.560 --> 43:13.560
and it's just trained on the long and long history of interactions with the world.
43:13.560 --> 43:17.560
So it becomes a predictive model of the world basically.
43:17.560 --> 43:22.560
And therefore also a compressor of the observations of the world
43:22.560 --> 43:26.560
because whatever you can predict, you don't have to store extra.
43:26.560 --> 43:29.560
So compression is a side effect of prediction.
43:29.560 --> 43:32.560
And how does this recurrent network compress?
43:32.560 --> 43:36.560
Well, it's inventing little subprograms, little subnetworks
43:36.560 --> 43:41.560
that stand for everything that frequently appears in the environment.
43:41.560 --> 43:47.560
Like bottles and microphones and faces, maybe lots of faces in my environment.
43:47.560 --> 43:51.560
So I'm learning to create something like a prototype face
43:51.560 --> 43:55.560
and a new face comes along and all I have to encode are the deviations from the prototype.
43:55.560 --> 44:00.560
So it's compressing all the time the stuff that frequently appears.
44:00.560 --> 44:04.560
There's one thing that appears all the time
44:04.560 --> 44:09.560
that is present all the time when the agent is interacting with its environment,
44:09.560 --> 44:11.560
which is the agent itself.
44:11.560 --> 44:14.560
So just for data compression reasons,
44:14.560 --> 44:18.560
it is extremely natural for this recurrent network
44:18.560 --> 44:23.560
to come up with little subnetworks that stand for the properties of the agents,
44:23.560 --> 44:27.560
the hand, the other actuators,
44:27.560 --> 44:31.560
and all the stuff that you need to better encode the data,
44:31.560 --> 44:34.560
which is influenced by the actions of the agent.
44:34.560 --> 44:40.560
So there, just as a side effect of data compression during primal solving,
44:40.560 --> 44:45.560
you have internal self models.
44:45.560 --> 44:51.560
Now you can use this model of the world to plan your future.
44:51.560 --> 44:54.560
And that's what we also have done since 1990.
44:54.560 --> 44:57.560
So the recurrent network, which is the controller,
44:57.560 --> 44:59.560
which is trying to maximize reward,
44:59.560 --> 45:02.560
can use this model of the network of the world,
45:02.560 --> 45:05.560
this model network of the world, this predictive model of the world
45:05.560 --> 45:08.560
to plan ahead and say, let's not do this action sequence.
45:08.560 --> 45:11.560
Let's do this action sequence instead
45:11.560 --> 45:14.560
because it leads to more predicted rewards.
45:14.560 --> 45:19.560
And whenever it's waking up these little subnetworks that stand for itself,
45:19.560 --> 45:21.560
then it's thinking about itself.
45:21.560 --> 45:23.560
Then it's thinking about itself.
45:23.560 --> 45:30.560
And it's exploring mentally the consequences of its own actions.
45:30.560 --> 45:36.560
And now you tell me why it's still missing.
45:36.560 --> 45:39.560
Missing the gap to consciousness.
45:39.560 --> 45:43.560
There isn't. That's a really beautiful idea that, you know,
45:43.560 --> 45:46.560
if life is a collection of data
45:46.560 --> 45:53.560
and life is a process of compressing that data to act efficiently.
45:53.560 --> 45:57.560
In that data, you yourself appear very often.
45:57.560 --> 46:00.560
So it's useful to form compressions of yourself.
46:00.560 --> 46:03.560
And it's a really beautiful formulation of what consciousness is,
46:03.560 --> 46:05.560
is a necessary side effect.
46:05.560 --> 46:11.560
It's actually quite compelling to me.
46:11.560 --> 46:18.560
We've described RNNs, developed LSTMs, long short term memory networks.
46:18.560 --> 46:22.560
They're a type of recurrent neural networks.
46:22.560 --> 46:24.560
They've gotten a lot of success recently.
46:24.560 --> 46:29.560
So these are networks that model the temporal aspects in the data,
46:29.560 --> 46:31.560
temporal patterns in the data.
46:31.560 --> 46:36.560
And you've called them the deepest of the neural networks, right?
46:36.560 --> 46:43.560
What do you think is the value of depth in the models that we use to learn?
46:43.560 --> 46:47.560
Yeah, since you mentioned the long short term memory and the LSTM,
46:47.560 --> 46:52.560
I have to mention the names of the brilliant students who made that possible.
46:52.560 --> 46:53.560
Yes, of course, of course.
46:53.560 --> 46:56.560
First of all, my first student ever, Sepp Hochreiter,
46:56.560 --> 47:00.560
who had fundamental insights already in his diploma thesis.
47:00.560 --> 47:04.560
Then Felix Giers, who had additional important contributions.
47:04.560 --> 47:11.560
Alex Gray is a guy from Scotland who is mostly responsible for this CTC algorithm,
47:11.560 --> 47:16.560
which is now often used to train the LSTM to do the speech recognition
47:16.560 --> 47:21.560
on all the Google Android phones and whatever, and Siri and so on.
47:21.560 --> 47:26.560
So these guys, without these guys, I would be nothing.
47:26.560 --> 47:28.560
It's a lot of incredible work.
47:28.560 --> 47:30.560
What is now the depth?
47:30.560 --> 47:32.560
What is the importance of depth?
47:32.560 --> 47:36.560
Well, most problems in the real world are deep
47:36.560 --> 47:41.560
in the sense that the current input doesn't tell you all you need to know
47:41.560 --> 47:44.560
about the environment.
47:44.560 --> 47:49.560
So instead, you have to have a memory of what happened in the past
47:49.560 --> 47:54.560
and often important parts of that memory are dated.
47:54.560 --> 47:56.560
They are pretty old.
47:56.560 --> 47:59.560
So when you're doing speech recognition, for example,
47:59.560 --> 48:03.560
and somebody says 11,
48:03.560 --> 48:08.560
then that's about half a second or something like that,
48:08.560 --> 48:11.560
which means it's already 50 time steps.
48:11.560 --> 48:15.560
And another guy or the same guy says 7.
48:15.560 --> 48:18.560
So the ending is the same, even.
48:18.560 --> 48:22.560
But now the system has to see the distinction between 7 and 11,
48:22.560 --> 48:26.560
and the only way it can see the difference is it has to store
48:26.560 --> 48:34.560
that 50 steps ago there was an S or an L, 11 or 7.
48:34.560 --> 48:37.560
So there you have already a problem of depth 50,
48:37.560 --> 48:42.560
because for each time step you have something like a virtual layer
48:42.560 --> 48:45.560
and the expanded, unrolled version of this recurrent network
48:45.560 --> 48:47.560
which is doing the speech recognition.
48:47.560 --> 48:53.560
So these long time lags, they translate into problem depth.
48:53.560 --> 48:59.560
And most problems in this world are such that you really
48:59.560 --> 49:04.560
have to look far back in time to understand what is the problem
49:04.560 --> 49:06.560
and to solve it.
49:06.560 --> 49:09.560
But just like with LSTMs, you don't necessarily need to,
49:09.560 --> 49:12.560
when you look back in time, remember every aspect.
49:12.560 --> 49:14.560
You just need to remember the important aspects.
49:14.560 --> 49:15.560
That's right.
49:15.560 --> 49:19.560
The network has to learn to put the important stuff into memory
49:19.560 --> 49:23.560
and to ignore the unimportant noise.
49:23.560 --> 49:28.560
But in that sense, deeper and deeper is better?
49:28.560 --> 49:30.560
Or is there a limitation?
49:30.560 --> 49:36.560
I mean LSTM is one of the great examples of architectures
49:36.560 --> 49:41.560
that do something beyond just deeper and deeper networks.
49:41.560 --> 49:47.560
There's clever mechanisms for filtering data for remembering and forgetting.
49:47.560 --> 49:51.560
So do you think that kind of thinking is necessary?
49:51.560 --> 49:54.560
If you think about LSTMs as a leap, a big leap forward
49:54.560 --> 50:01.560
over traditional vanilla RNNs, what do you think is the next leap
50:01.560 --> 50:03.560
within this context?
50:03.560 --> 50:08.560
So LSTM is a very clever improvement, but LSTMs still don't
50:08.560 --> 50:13.560
have the same kind of ability to see far back in the past
50:13.560 --> 50:18.560
as humans do, the credit assignment problem across way back,
50:18.560 --> 50:24.560
not just 50 time steps or 100 or 1,000, but millions and billions.
50:24.560 --> 50:28.560
It's not clear what are the practical limits of the LSTM
50:28.560 --> 50:30.560
when it comes to looking back.
50:30.560 --> 50:35.560
Already in 2006, I think, we had examples where not only
50:35.560 --> 50:40.560
looked back tens of thousands of steps, but really millions of steps.
50:40.560 --> 50:46.560
And Juan Perez Ortiz in my lab, I think was the first author of a paper
50:46.560 --> 50:51.560
where we really, was it 2006 or something, had examples where it
50:51.560 --> 50:56.560
learned to look back for more than 10 million steps.
50:56.560 --> 51:02.560
So for most problems of speech recognition, it's not
51:02.560 --> 51:06.560
necessary to look that far back, but there are examples where it does.
51:06.560 --> 51:12.560
Now, the looking back thing, that's rather easy because there is only
51:12.560 --> 51:17.560
one past, but there are many possible futures.
51:17.560 --> 51:21.560
And so a reinforcement learning system, which is trying to maximize
51:21.560 --> 51:26.560
its future expected reward and doesn't know yet which of these
51:26.560 --> 51:31.560
many possible futures should I select, given this one single past,
51:31.560 --> 51:36.560
is facing problems that the LSTM by itself cannot solve.
51:36.560 --> 51:40.560
So the LSTM is good for coming up with a compact representation
51:40.560 --> 51:46.560
of the history so far, of the history and of observations and actions so far.
51:46.560 --> 51:53.560
But now, how do you plan in an efficient and good way among all these,
51:53.560 --> 51:57.560
how do you select one of these many possible action sequences
51:57.560 --> 52:02.560
that a reinforcement learning system has to consider to maximize
52:02.560 --> 52:05.560
reward in this unknown future.
52:05.560 --> 52:11.560
So again, we have this basic setup where you have one recon network,
52:11.560 --> 52:16.560
which gets in the video and the speech and whatever, and it's
52:16.560 --> 52:19.560
executing the actions and it's trying to maximize reward.
52:19.560 --> 52:24.560
So there is no teacher who tells it what to do at which point in time.
52:24.560 --> 52:29.560
And then there's the other network, which is just predicting
52:29.560 --> 52:32.560
what's going to happen if I do that and then.
52:32.560 --> 52:36.560
And that could be an LSTM network, and it learns to look back
52:36.560 --> 52:41.560
all the way to make better predictions of the next time step.
52:41.560 --> 52:45.560
So essentially, although it's predicting only the next time step,
52:45.560 --> 52:50.560
it is motivated to learn to put into memory something that happened
52:50.560 --> 52:54.560
maybe a million steps ago because it's important to memorize that
52:54.560 --> 52:58.560
if you want to predict that at the next time step, the next event.
52:58.560 --> 53:03.560
Now, how can a model of the world like that,
53:03.560 --> 53:07.560
a predictive model of the world be used by the first guy,
53:07.560 --> 53:11.560
let's call it the controller and the model, the controller and the model.
53:11.560 --> 53:16.560
How can the model be used by the controller to efficiently select
53:16.560 --> 53:19.560
among these many possible futures?
53:19.560 --> 53:23.560
The naive way we had about 30 years ago was
53:23.560 --> 53:27.560
let's just use the model of the world as a stand in,
53:27.560 --> 53:29.560
as a simulation of the world.
53:29.560 --> 53:32.560
And millisecond by millisecond we plan the future
53:32.560 --> 53:36.560
and that means we have to roll it out really in detail
53:36.560 --> 53:38.560
and it will work only if the model is really good
53:38.560 --> 53:40.560
and it will still be inefficient
53:40.560 --> 53:43.560
because we have to look at all these possible futures
53:43.560 --> 53:45.560
and there are so many of them.
53:45.560 --> 53:50.560
So instead, what we do now since 2015 in our CN systems,
53:50.560 --> 53:54.560
controller model systems, we give the controller the opportunity
53:54.560 --> 54:00.560
to learn by itself how to use the potentially relevant parts
54:00.560 --> 54:05.560
of the model network to solve new problems more quickly.
54:05.560 --> 54:09.560
And if it wants to, it can learn to ignore the M
54:09.560 --> 54:12.560
and sometimes it's a good idea to ignore the M
54:12.560 --> 54:15.560
because it's really bad, it's a bad predictor
54:15.560 --> 54:18.560
in this particular situation of life
54:18.560 --> 54:22.560
where the controller is currently trying to maximize reward.
54:22.560 --> 54:26.560
However, it can also learn to address and exploit
54:26.560 --> 54:32.560
some of the subprograms that came about in the model network
54:32.560 --> 54:35.560
through compressing the data by predicting it.
54:35.560 --> 54:40.560
So it now has an opportunity to reuse that code,
54:40.560 --> 54:43.560
the algorithmic information in the model network
54:43.560 --> 54:47.560
to reduce its own search space,
54:47.560 --> 54:50.560
search that it can solve a new problem more quickly
54:50.560 --> 54:52.560
than without the model.
54:52.560 --> 54:54.560
Compression.
54:54.560 --> 54:58.560
So you're ultimately optimistic and excited
54:58.560 --> 55:02.560
about the power of reinforcement learning
55:02.560 --> 55:04.560
in the context of real systems.
55:04.560 --> 55:06.560
Absolutely, yeah.
55:06.560 --> 55:11.560
So you see RL as a potential having a huge impact
55:11.560 --> 55:15.560
beyond just sort of the M part is often developed
55:15.560 --> 55:19.560
on supervised learning methods.
55:19.560 --> 55:25.560
You see RL as a, for problems of cell driving cars
55:25.560 --> 55:28.560
or any kind of applied side robotics,
55:28.560 --> 55:33.560
that's the correct, interesting direction for researching you.
55:33.560 --> 55:35.560
I do think so.
55:35.560 --> 55:37.560
We have a company called Nasense,
55:37.560 --> 55:43.560
which has applied reinforcement learning to little Audis.
55:43.560 --> 55:45.560
Little Audis.
55:45.560 --> 55:47.560
Which learn to park without a teacher.
55:47.560 --> 55:51.560
The same principles were used, of course.
55:51.560 --> 55:54.560
So these little Audis, they are small, maybe like that,
55:54.560 --> 55:57.560
so much smaller than the RL Audis.
55:57.560 --> 56:00.560
But they have all the sensors that you find in the RL Audis.
56:00.560 --> 56:03.560
You find the cameras, the LIDAR sensors.
56:03.560 --> 56:08.560
They go up to 120 kilometers an hour if they want to.
56:08.560 --> 56:12.560
And they have pain sensors, basically.
56:12.560 --> 56:16.560
And they don't want to bump against obstacles and other Audis.
56:16.560 --> 56:21.560
And so they must learn like little babies to park.
56:21.560 --> 56:25.560
Take the raw vision input and translate that into actions
56:25.560 --> 56:28.560
that lead to successful parking behavior,
56:28.560 --> 56:30.560
which is a rewarding thing.
56:30.560 --> 56:32.560
And yes, they learn that.
56:32.560 --> 56:34.560
So we have examples like that.
56:34.560 --> 56:36.560
And it's only in the beginning.
56:36.560 --> 56:38.560
This is just a tip of the iceberg.
56:38.560 --> 56:44.560
And I believe the next wave of AI is going to be all about that.
56:44.560 --> 56:47.560
So at the moment, the current wave of AI is about
56:47.560 --> 56:51.560
passive pattern observation and prediction.
56:51.560 --> 56:54.560
And that's what you have on your smartphone
56:54.560 --> 56:58.560
and what the major companies on the Pacific Rim are using
56:58.560 --> 57:01.560
to sell you ads to do marketing.
57:01.560 --> 57:04.560
That's the current sort of profit in AI.
57:04.560 --> 57:09.560
And that's only one or two percent of the wild economy,
57:09.560 --> 57:11.560
which is big enough to make these companies
57:11.560 --> 57:14.560
pretty much the most valuable companies in the world.
57:14.560 --> 57:19.560
But there's a much, much bigger fraction of the economy
57:19.560 --> 57:21.560
going to be affected by the next wave,
57:21.560 --> 57:25.560
which is really about machines that shape the data
57:25.560 --> 57:27.560
through their own actions.
57:27.560 --> 57:32.560
Do you think simulation is ultimately the biggest way
57:32.560 --> 57:36.560
that those methods will be successful in the next 10, 20 years?
57:36.560 --> 57:38.560
We're not talking about 100 years from now.
57:38.560 --> 57:42.560
We're talking about sort of the near term impact of RL.
57:42.560 --> 57:44.560
Do you think really good simulation is required?
57:44.560 --> 57:48.560
Or is there other techniques like imitation learning,
57:48.560 --> 57:53.560
observing other humans operating in the real world?
57:53.560 --> 57:57.560
Where do you think this success will come from?
57:57.560 --> 58:01.560
So at the moment we have a tendency of using
58:01.560 --> 58:06.560
physics simulations to learn behavior for machines
58:06.560 --> 58:13.560
that learn to solve problems that humans also do not know how to solve.
58:13.560 --> 58:15.560
However, this is not the future,
58:15.560 --> 58:19.560
because the future is in what little babies do.
58:19.560 --> 58:22.560
They don't use a physics engine to simulate the world.
58:22.560 --> 58:25.560
They learn a predictive model of the world,
58:25.560 --> 58:29.560
which maybe sometimes is wrong in many ways,
58:29.560 --> 58:34.560
but captures all kinds of important abstract high level predictions
58:34.560 --> 58:37.560
which are really important to be successful.
58:37.560 --> 58:42.560
And that's what was the future 30 years ago
58:42.560 --> 58:44.560
when we started that type of research,
58:44.560 --> 58:45.560
but it's still the future,
58:45.560 --> 58:50.560
and now we know much better how to move forward
58:50.560 --> 58:54.560
and to really make working systems based on that,
58:54.560 --> 58:57.560
where you have a learning model of the world,
58:57.560 --> 59:00.560
a model of the world that learns to predict what's going to happen
59:00.560 --> 59:01.560
if I do that and that,
59:01.560 --> 59:06.560
and then the controller uses that model
59:06.560 --> 59:11.560
to more quickly learn successful action sequences.
59:11.560 --> 59:13.560
And then of course always this curiosity thing,
59:13.560 --> 59:15.560
in the beginning the model is stupid,
59:15.560 --> 59:17.560
so the controller should be motivated
59:17.560 --> 59:20.560
to come up with experiments, with action sequences
59:20.560 --> 59:23.560
that lead to data that improve the model.
59:23.560 --> 59:26.560
Do you think improving the model,
59:26.560 --> 59:30.560
constructing an understanding of the world in this connection
59:30.560 --> 59:34.560
is now the popular approaches have been successful
59:34.560 --> 59:38.560
or grounded in ideas of neural networks,
59:38.560 --> 59:43.560
but in the 80s with expert systems there's symbolic AI approaches,
59:43.560 --> 59:47.560
which to us humans are more intuitive
59:47.560 --> 59:50.560
in the sense that it makes sense that you build up knowledge
59:50.560 --> 59:52.560
in this knowledge representation.
59:52.560 --> 59:56.560
What kind of lessons can we draw into our current approaches
59:56.560 --> 1:00:00.560
from expert systems, from symbolic AI?
1:00:00.560 --> 1:00:04.560
So I became aware of all of that in the 80s
1:00:04.560 --> 1:00:09.560
and back then logic programming was a huge thing.
1:00:09.560 --> 1:00:12.560
Was it inspiring to yourself that you find it compelling
1:00:12.560 --> 1:00:16.560
that a lot of your work was not so much in that realm,
1:00:16.560 --> 1:00:18.560
is more in the learning systems?
1:00:18.560 --> 1:00:20.560
Yes and no, but we did all of that.
1:00:20.560 --> 1:00:27.560
So my first publication ever actually was 1987,
1:00:27.560 --> 1:00:31.560
was the implementation of a genetic algorithm
1:00:31.560 --> 1:00:34.560
of a genetic programming system in Prolog.
1:00:34.560 --> 1:00:37.560
So Prolog, that's what you learn back then,
1:00:37.560 --> 1:00:39.560
which is a logic programming language,
1:00:39.560 --> 1:00:45.560
and the Japanese, they had this huge fifth generation AI project,
1:00:45.560 --> 1:00:48.560
which was mostly about logic programming back then,
1:00:48.560 --> 1:00:53.560
although neural networks existed and were well known back then,
1:00:53.560 --> 1:00:57.560
and deep learning has existed since 1965,
1:00:57.560 --> 1:01:01.560
since this guy in the Ukraine, Ivak Nenko, started it,
1:01:01.560 --> 1:01:05.560
but the Japanese and many other people,
1:01:05.560 --> 1:01:07.560
they focused really on this logic programming,
1:01:07.560 --> 1:01:10.560
and I was influenced to the extent that I said,
1:01:10.560 --> 1:01:13.560
okay, let's take these biologically inspired algorithms
1:01:13.560 --> 1:01:16.560
like evolution, programs,
1:01:16.560 --> 1:01:22.560
and implement that in the language which I know,
1:01:22.560 --> 1:01:24.560
which was Prolog, for example, back then.
1:01:24.560 --> 1:01:28.560
And then in many ways this came back later,
1:01:28.560 --> 1:01:31.560
because the Goudel machine, for example,
1:01:31.560 --> 1:01:33.560
has a proof searcher on board,
1:01:33.560 --> 1:01:35.560
and without that it would not be optimal.
1:01:35.560 --> 1:01:38.560
Well, Markus Hutter's universal algorithm
1:01:38.560 --> 1:01:40.560
for solving all well defined problems
1:01:40.560 --> 1:01:42.560
has a proof search on board,
1:01:42.560 --> 1:01:46.560
so that's very much logic programming.
1:01:46.560 --> 1:01:50.560
Without that it would not be asymptotically optimal.
1:01:50.560 --> 1:01:54.560
But then on the other hand, because we are very pragmatic guys also,
1:01:54.560 --> 1:01:59.560
we focused on recurrent neural networks
1:01:59.560 --> 1:02:04.560
and suboptimal stuff such as gradient based search
1:02:04.560 --> 1:02:09.560
and program space rather than provably optimal things.
1:02:09.560 --> 1:02:13.560
So logic programming certainly has a usefulness
1:02:13.560 --> 1:02:17.560
when you're trying to construct something provably optimal
1:02:17.560 --> 1:02:19.560
or provably good or something like that,
1:02:19.560 --> 1:02:22.560
but is it useful for practical problems?
1:02:22.560 --> 1:02:24.560
It's really useful for our theorem proving.
1:02:24.560 --> 1:02:28.560
The best theorem proofers today are not neural networks.
1:02:28.560 --> 1:02:31.560
No, they are logic programming systems
1:02:31.560 --> 1:02:35.560
that are much better theorem proofers than most math students
1:02:35.560 --> 1:02:38.560
in the first or second semester.
1:02:38.560 --> 1:02:42.560
But for reasoning, for playing games of Go, or chess,
1:02:42.560 --> 1:02:46.560
or for robots, autonomous vehicles that operate in the real world,
1:02:46.560 --> 1:02:50.560
or object manipulation, you think learning...
1:02:50.560 --> 1:02:53.560
Yeah, as long as the problems have little to do
1:02:53.560 --> 1:02:58.560
with theorem proving themselves,
1:02:58.560 --> 1:03:01.560
then as long as that is not the case,
1:03:01.560 --> 1:03:05.560
you just want to have better pattern recognition.
1:03:05.560 --> 1:03:09.560
So to build a self trying car, you want to have better pattern recognition
1:03:09.560 --> 1:03:13.560
and pedestrian recognition and all these things,
1:03:13.560 --> 1:03:18.560
and you want to minimize the number of false positives,
1:03:18.560 --> 1:03:22.560
which is currently slowing down self trying cars in many ways.
1:03:22.560 --> 1:03:27.560
And all of that has very little to do with logic programming.
1:03:27.560 --> 1:03:32.560
What are you most excited about in terms of directions
1:03:32.560 --> 1:03:36.560
of artificial intelligence at this moment in the next few years,
1:03:36.560 --> 1:03:41.560
in your own research and in the broader community?
1:03:41.560 --> 1:03:44.560
So I think in the not so distant future,
1:03:44.560 --> 1:03:52.560
we will have for the first time little robots that learn like kids.
1:03:52.560 --> 1:03:57.560
And I will be able to say to the robot,
1:03:57.560 --> 1:04:00.560
look here robot, we are going to assemble a smartphone.
1:04:00.560 --> 1:04:05.560
Let's take this slab of plastic and the screwdriver
1:04:05.560 --> 1:04:08.560
and let's screw in the screw like that.
1:04:08.560 --> 1:04:11.560
No, not like that, like that.
1:04:11.560 --> 1:04:13.560
Not like that, like that.
1:04:13.560 --> 1:04:17.560
And I don't have a data glove or something.
1:04:17.560 --> 1:04:20.560
He will see me and he will hear me
1:04:20.560 --> 1:04:24.560
and he will try to do something with his own actuators,
1:04:24.560 --> 1:04:26.560
which will be really different from mine,
1:04:26.560 --> 1:04:28.560
but he will understand the difference
1:04:28.560 --> 1:04:34.560
and will learn to imitate me but not in the supervised way
1:04:34.560 --> 1:04:40.560
where a teacher is giving target signals for all his muscles all the time.
1:04:40.560 --> 1:04:43.560
No, by doing this high level imitation
1:04:43.560 --> 1:04:46.560
where he first has to learn to imitate me
1:04:46.560 --> 1:04:50.560
and to interpret these additional noises coming from my mouth
1:04:50.560 --> 1:04:54.560
as helpful signals to do that pattern.
1:04:54.560 --> 1:05:00.560
And then it will by itself come up with faster ways
1:05:00.560 --> 1:05:03.560
and more efficient ways of doing the same thing.
1:05:03.560 --> 1:05:07.560
And finally, I stop his learning algorithm
1:05:07.560 --> 1:05:10.560
and make a million copies and sell it.
1:05:10.560 --> 1:05:13.560
And so at the moment this is not possible,
1:05:13.560 --> 1:05:16.560
but we already see how we are going to get there.
1:05:16.560 --> 1:05:21.560
And you can imagine to the extent that this works economically and cheaply,
1:05:21.560 --> 1:05:24.560
it's going to change everything.
1:05:24.560 --> 1:05:30.560
Almost all our production is going to be affected by that.
1:05:30.560 --> 1:05:33.560
And a much bigger wave,
1:05:33.560 --> 1:05:36.560
a much bigger AI wave is coming
1:05:36.560 --> 1:05:38.560
than the one that we are currently witnessing,
1:05:38.560 --> 1:05:41.560
which is mostly about passive pattern recognition on your smartphone.
1:05:41.560 --> 1:05:47.560
This is about active machines that shapes data through the actions they are executing
1:05:47.560 --> 1:05:51.560
and they learn to do that in a good way.
1:05:51.560 --> 1:05:56.560
So many of the traditional industries are going to be affected by that.
1:05:56.560 --> 1:06:00.560
All the companies that are building machines
1:06:00.560 --> 1:06:05.560
will equip these machines with cameras and other sensors
1:06:05.560 --> 1:06:10.560
and they are going to learn to solve all kinds of problems.
1:06:10.560 --> 1:06:14.560
Through interaction with humans, but also a lot on their own
1:06:14.560 --> 1:06:18.560
to improve what they already can do.
1:06:18.560 --> 1:06:23.560
And lots of old economy is going to be affected by that.
1:06:23.560 --> 1:06:28.560
And in recent years I have seen that old economy is actually waking up
1:06:28.560 --> 1:06:31.560
and realizing that this is the case.
1:06:31.560 --> 1:06:35.560
Are you optimistic about that future? Are you concerned?
1:06:35.560 --> 1:06:40.560
There's a lot of people concerned in the near term about the transformation
1:06:40.560 --> 1:06:42.560
of the nature of work.
1:06:42.560 --> 1:06:45.560
The kind of ideas that you just suggested
1:06:45.560 --> 1:06:48.560
would have a significant impact on what kind of things could be automated.
1:06:48.560 --> 1:06:51.560
Are you optimistic about that future?
1:06:51.560 --> 1:06:54.560
Are you nervous about that future?
1:06:54.560 --> 1:07:01.560
And looking a little bit farther into the future, there's people like Gila Musk
1:07:01.560 --> 1:07:06.560
still wrestle concerned about the existential threats of that future.
1:07:06.560 --> 1:07:10.560
So in the near term, job loss in the long term existential threat,
1:07:10.560 --> 1:07:15.560
are these concerns to you or are you ultimately optimistic?
1:07:15.560 --> 1:07:22.560
So let's first address the near future.
1:07:22.560 --> 1:07:27.560
We have had predictions of job losses for many decades.
1:07:27.560 --> 1:07:32.560
For example, when industrial robots came along,
1:07:32.560 --> 1:07:37.560
many people predicted that lots of jobs are going to get lost.
1:07:37.560 --> 1:07:41.560
And in a sense, they were right,
1:07:41.560 --> 1:07:45.560
because back then there were car factories
1:07:45.560 --> 1:07:50.560
and hundreds of people in these factories assembled cars.
1:07:50.560 --> 1:07:53.560
And today the same car factories have hundreds of robots
1:07:53.560 --> 1:07:58.560
and maybe three guys watching the robots.
1:07:58.560 --> 1:08:04.560
On the other hand, those countries that have lots of robots per capita,
1:08:04.560 --> 1:08:09.560
Japan, Korea, Germany, Switzerland, a couple of other countries,
1:08:09.560 --> 1:08:13.560
they have really low unemployment rates.
1:08:13.560 --> 1:08:17.560
Somehow all kinds of new jobs were created.
1:08:17.560 --> 1:08:22.560
Back then nobody anticipated those jobs.
1:08:22.560 --> 1:08:26.560
And decades ago, I already said,
1:08:26.560 --> 1:08:31.560
it's really easy to say which jobs are going to get lost,
1:08:31.560 --> 1:08:35.560
but it's really hard to predict the new ones.
1:08:35.560 --> 1:08:38.560
30 years ago, who would have predicted all these people
1:08:38.560 --> 1:08:44.560
making money as YouTube bloggers, for example?
1:08:44.560 --> 1:08:51.560
200 years ago, 60% of all people used to work in agriculture.
1:08:51.560 --> 1:08:55.560
Today, maybe 1%.
1:08:55.560 --> 1:09:01.560
But still, only, I don't know, 5% unemployment.
1:09:01.560 --> 1:09:03.560
Lots of new jobs were created.
1:09:03.560 --> 1:09:07.560
And Homo Ludens, the playing man,
1:09:07.560 --> 1:09:10.560
is inventing new jobs all the time.
1:09:10.560 --> 1:09:15.560
Most of these jobs are not existentially necessary
1:09:15.560 --> 1:09:18.560
for the survival of our species.
1:09:18.560 --> 1:09:22.560
There are only very few existentially necessary jobs
1:09:22.560 --> 1:09:27.560
such as farming and building houses and warming up the houses,
1:09:27.560 --> 1:09:30.560
but less than 10% of the population is doing that.
1:09:30.560 --> 1:09:37.560
And most of these newly invented jobs are about interacting with other people
1:09:37.560 --> 1:09:40.560
in new ways, through new media and so on,
1:09:40.560 --> 1:09:45.560
getting new types of kudos and forms of likes and whatever,
1:09:45.560 --> 1:09:47.560
and even making money through that.
1:09:47.560 --> 1:09:52.560
So, Homo Ludens, the playing man, doesn't want to be unemployed,
1:09:52.560 --> 1:09:56.560
and that's why he's inventing new jobs all the time.
1:09:56.560 --> 1:10:01.560
And he keeps considering these jobs as really important
1:10:01.560 --> 1:10:07.560
and is investing a lot of energy and hours of work into those new jobs.
1:10:07.560 --> 1:10:09.560
That's quite beautifully put.
1:10:09.560 --> 1:10:11.560
We're really nervous about the future
1:10:11.560 --> 1:10:14.560
because we can't predict what kind of new jobs will be created.
1:10:14.560 --> 1:10:20.560
But you're ultimately optimistic that we humans are so restless
1:10:20.560 --> 1:10:24.560
that we create and give meaning to newer and newer jobs,
1:10:24.560 --> 1:10:29.560
telling you things that get likes on Facebook
1:10:29.560 --> 1:10:31.560
or whatever the social platform is.
1:10:31.560 --> 1:10:36.560
So, what about long term existential threat of AI
1:10:36.560 --> 1:10:40.560
where our whole civilization may be swallowed up
1:10:40.560 --> 1:10:44.560
by this ultra super intelligent systems?
1:10:44.560 --> 1:10:47.560
Maybe it's not going to be swallowed up,
1:10:47.560 --> 1:10:55.560
but I'd be surprised if we humans were the last step
1:10:55.560 --> 1:10:59.560
in the evolution of the universe.
1:10:59.560 --> 1:11:03.560
You've actually had this beautiful comment somewhere
1:11:03.560 --> 1:11:08.560
that I've seen saying that artificial...
1:11:08.560 --> 1:11:11.560
Quite insightful, artificial intelligence systems
1:11:11.560 --> 1:11:15.560
just like us humans will likely not want to interact with humans.
1:11:15.560 --> 1:11:17.560
They'll just interact amongst themselves,
1:11:17.560 --> 1:11:20.560
just like ants interact amongst themselves
1:11:20.560 --> 1:11:24.560
and only tangentially interact with humans.
1:11:24.560 --> 1:11:28.560
And it's quite an interesting idea that once we create AGI
1:11:28.560 --> 1:11:31.560
that will lose interest in humans
1:11:31.560 --> 1:11:34.560
and have compete for their own Facebook likes
1:11:34.560 --> 1:11:36.560
and their own social platforms.
1:11:36.560 --> 1:11:40.560
So, within that quite elegant idea,
1:11:40.560 --> 1:11:44.560
how do we know in a hypothetical sense
1:11:44.560 --> 1:11:48.560
that there's not already intelligent systems out there?
1:11:48.560 --> 1:11:52.560
How do you think broadly of general intelligence
1:11:52.560 --> 1:11:56.560
greater than us, how do we know it's out there?
1:11:56.560 --> 1:12:01.560
How do we know it's around us and could it already be?
1:12:01.560 --> 1:12:04.560
I'd be surprised if within the next few decades
1:12:04.560 --> 1:12:10.560
or something like that we won't have AIs
1:12:10.560 --> 1:12:12.560
that are truly smart in every single way
1:12:12.560 --> 1:12:17.560
and better problem solvers in almost every single important way.
1:12:17.560 --> 1:12:22.560
And I'd be surprised if they wouldn't realize
1:12:22.560 --> 1:12:24.560
what we have realized a long time ago,
1:12:24.560 --> 1:12:29.560
which is that almost all physical resources are not here
1:12:29.560 --> 1:12:36.560
in this biosphere, but throughout the rest of the solar system
1:12:36.560 --> 1:12:42.560
gets two billion times more solar energy than our little planet.
1:12:42.560 --> 1:12:46.560
There's lots of material out there that you can use
1:12:46.560 --> 1:12:51.560
to build robots and self replicating robot factories and all this stuff.
1:12:51.560 --> 1:12:53.560
And they are going to do that.
1:12:53.560 --> 1:12:56.560
And they will be scientists and curious
1:12:56.560 --> 1:12:59.560
and they will explore what they can do.
1:12:59.560 --> 1:13:04.560
And in the beginning they will be fascinated by life
1:13:04.560 --> 1:13:07.560
and by their own origins in our civilization.
1:13:07.560 --> 1:13:09.560
They will want to understand that completely,
1:13:09.560 --> 1:13:13.560
just like people today would like to understand how life works
1:13:13.560 --> 1:13:22.560
and also the history of our own existence and civilization
1:13:22.560 --> 1:13:26.560
and also the physical laws that created all of them.
1:13:26.560 --> 1:13:29.560
So in the beginning they will be fascinated by life
1:13:29.560 --> 1:13:33.560
once they understand it, they lose interest,
1:13:33.560 --> 1:13:39.560
like anybody who loses interest in things he understands.
1:13:39.560 --> 1:13:43.560
And then, as you said,
1:13:43.560 --> 1:13:50.560
the most interesting sources of information for them
1:13:50.560 --> 1:13:57.560
will be others of their own kind.
1:13:57.560 --> 1:14:01.560
So, at least in the long run,
1:14:01.560 --> 1:14:06.560
there seems to be some sort of protection
1:14:06.560 --> 1:14:11.560
through lack of interest on the other side.
1:14:11.560 --> 1:14:16.560
And now it seems also clear, as far as we understand physics,
1:14:16.560 --> 1:14:20.560
you need matter and energy to compute
1:14:20.560 --> 1:14:22.560
and to build more robots and infrastructure
1:14:22.560 --> 1:14:28.560
and more AI civilization and AI ecologies
1:14:28.560 --> 1:14:31.560
consisting of trillions of different types of AI's.
1:14:31.560 --> 1:14:34.560
And so it seems inconceivable to me
1:14:34.560 --> 1:14:37.560
that this thing is not going to expand.
1:14:37.560 --> 1:14:41.560
Some AI ecology not controlled by one AI
1:14:41.560 --> 1:14:44.560
but trillions of different types of AI's competing
1:14:44.560 --> 1:14:47.560
in all kinds of quickly evolving
1:14:47.560 --> 1:14:49.560
and disappearing ecological niches
1:14:49.560 --> 1:14:52.560
in ways that we cannot fathom at the moment.
1:14:52.560 --> 1:14:54.560
But it's going to expand,
1:14:54.560 --> 1:14:56.560
limited by light speed and physics,
1:14:56.560 --> 1:15:00.560
but it's going to expand and now we realize
1:15:00.560 --> 1:15:02.560
that the universe is still young.
1:15:02.560 --> 1:15:05.560
It's only 13.8 billion years old
1:15:05.560 --> 1:15:10.560
and it's going to be a thousand times older than that.
1:15:10.560 --> 1:15:13.560
So there's plenty of time
1:15:13.560 --> 1:15:16.560
to conquer the entire universe
1:15:16.560 --> 1:15:19.560
and to fill it with intelligence
1:15:19.560 --> 1:15:21.560
and send us in receivers such that
1:15:21.560 --> 1:15:25.560
AI's can travel the way they are traveling
1:15:25.560 --> 1:15:27.560
in our labs today,
1:15:27.560 --> 1:15:31.560
which is by radio from sender to receiver.
1:15:31.560 --> 1:15:35.560
And let's call the current age of the universe one eon.
1:15:35.560 --> 1:15:38.560
One eon.
1:15:38.560 --> 1:15:41.560
Now it will take just a few eons from now
1:15:41.560 --> 1:15:43.560
and the entire visible universe
1:15:43.560 --> 1:15:46.560
is going to be full of that stuff.
1:15:46.560 --> 1:15:48.560
And let's look ahead to a time
1:15:48.560 --> 1:15:50.560
when the universe is going to be
1:15:50.560 --> 1:15:52.560
one thousand times older than it is now.
1:15:52.560 --> 1:15:54.560
They will look back and they will say,
1:15:54.560 --> 1:15:56.560
look almost immediately after the Big Bang,
1:15:56.560 --> 1:15:59.560
only a few eons later,
1:15:59.560 --> 1:16:02.560
the entire universe started to become intelligent.
1:16:02.560 --> 1:16:05.560
Now to your question,
1:16:05.560 --> 1:16:08.560
how do we see whether anything like that
1:16:08.560 --> 1:16:12.560
has already happened or is already in a more advanced stage
1:16:12.560 --> 1:16:14.560
in some other part of the universe,
1:16:14.560 --> 1:16:16.560
of the visible universe?
1:16:16.560 --> 1:16:18.560
We are trying to look out there
1:16:18.560 --> 1:16:20.560
and nothing like that has happened so far.
1:16:20.560 --> 1:16:22.560
Or is that true?
1:16:22.560 --> 1:16:24.560
Do you think we would recognize it?
1:16:24.560 --> 1:16:26.560
How do we know it's not among us?
1:16:26.560 --> 1:16:30.560
How do we know planets aren't in themselves intelligent beings?
1:16:30.560 --> 1:16:36.560
How do we know ants seen as a collective
1:16:36.560 --> 1:16:39.560
are not much greater intelligence than our own?
1:16:39.560 --> 1:16:41.560
These kinds of ideas.
1:16:41.560 --> 1:16:44.560
When I was a boy, I was thinking about these things
1:16:44.560 --> 1:16:48.560
and I thought, hmm, maybe it has already happened.
1:16:48.560 --> 1:16:50.560
Because back then I knew,
1:16:50.560 --> 1:16:53.560
I learned from popular physics books,
1:16:53.560 --> 1:16:57.560
that the structure, the large scale structure of the universe
1:16:57.560 --> 1:16:59.560
is not homogeneous.
1:16:59.560 --> 1:17:02.560
And you have these clusters of galaxies
1:17:02.560 --> 1:17:07.560
and then in between there are these huge empty spaces.
1:17:07.560 --> 1:17:11.560
And I thought, hmm, maybe they aren't really empty.
1:17:11.560 --> 1:17:13.560
It's just that in the middle of that
1:17:13.560 --> 1:17:16.560
some AI civilization already has expanded
1:17:16.560 --> 1:17:22.560
and then has covered a bubble of a billion light years time
1:17:22.560 --> 1:17:26.560
using all the energy of all the stars within that bubble
1:17:26.560 --> 1:17:29.560
for its own unfathomable practices.
1:17:29.560 --> 1:17:34.560
And so it always happened and we just failed to interpret the signs.
1:17:34.560 --> 1:17:39.560
But then I learned that gravity by itself
1:17:39.560 --> 1:17:42.560
explains the large scale structure of the universe
1:17:42.560 --> 1:17:45.560
and that this is not a convincing explanation.
1:17:45.560 --> 1:17:50.560
And then I thought maybe it's the dark matter
1:17:50.560 --> 1:17:54.560
because as far as we know today
1:17:54.560 --> 1:18:00.560
80% of the measurable matter is invisible.
1:18:00.560 --> 1:18:03.560
And we know that because otherwise our galaxy
1:18:03.560 --> 1:18:06.560
or other galaxies would fall apart.
1:18:06.560 --> 1:18:09.560
They are rotating too quickly.
1:18:09.560 --> 1:18:14.560
And then the idea was maybe all of these
1:18:14.560 --> 1:18:17.560
AI civilizations that are already out there,
1:18:17.560 --> 1:18:22.560
they are just invisible
1:18:22.560 --> 1:18:24.560
because they are really efficient in using the energies
1:18:24.560 --> 1:18:26.560
out their own local systems
1:18:26.560 --> 1:18:29.560
and that's why they appear dark to us.
1:18:29.560 --> 1:18:31.560
But this is also not a convincing explanation
1:18:31.560 --> 1:18:34.560
because then the question becomes
1:18:34.560 --> 1:18:41.560
why are there still any visible stars left in our own galaxy
1:18:41.560 --> 1:18:44.560
which also must have a lot of dark matter.
1:18:44.560 --> 1:18:46.560
So that is also not a convincing thing.
1:18:46.560 --> 1:18:53.560
And today I like to think it's quite plausible
1:18:53.560 --> 1:18:56.560
that maybe we are the first, at least in our local light cone
1:18:56.560 --> 1:19:04.560
within the few hundreds of millions of light years
1:19:04.560 --> 1:19:08.560
that we can reliably observe.
1:19:08.560 --> 1:19:10.560
Is that exciting to you?
1:19:10.560 --> 1:19:12.560
That we might be the first?
1:19:12.560 --> 1:19:16.560
It would make us much more important
1:19:16.560 --> 1:19:20.560
because if we mess it up through a nuclear war
1:19:20.560 --> 1:19:25.560
then maybe this will have an effect
1:19:25.560 --> 1:19:30.560
on the development of the entire universe.
1:19:30.560 --> 1:19:32.560
So let's not mess it up.
1:19:32.560 --> 1:19:33.560
Let's not mess it up.
1:19:33.560 --> 1:19:35.560
Jürgen, thank you so much for talking today.
1:19:35.560 --> 1:19:36.560
I really appreciate it.
1:19:36.560 --> 1:19:42.560
It's my pleasure.