Datasets:
Languages:
English
Multilinguality:
monolingual
Size Categories:
n<1K
Language Creators:
found
Source Datasets:
original
Tags:
karpathy,whisper,openai
WEBVTT | |
00:00.000 --> 00:04.200 | |
As part of MIT course 6S099, Artificial General Intelligence, | |
00:04.200 --> 00:06.600 | |
I've gotten the chance to sit down with Max Tegmark. | |
00:06.600 --> 00:08.680 | |
He is a professor here at MIT. | |
00:08.680 --> 00:11.920 | |
He's a physicist, spent a large part of his career | |
00:11.920 --> 00:16.960 | |
studying the mysteries of our cosmological universe. | |
00:16.960 --> 00:20.680 | |
But he's also studied and delved into the beneficial | |
00:20.680 --> 00:24.000 | |
possibilities and the existential risks | |
00:24.000 --> 00:25.800 | |
of artificial intelligence. | |
00:25.800 --> 00:29.040 | |
Amongst many other things, he is the cofounder | |
00:29.040 --> 00:33.080 | |
of the Future of Life Institute, author of two books, | |
00:33.080 --> 00:35.160 | |
both of which I highly recommend. | |
00:35.160 --> 00:37.260 | |
First, Our Mathematical Universe. | |
00:37.260 --> 00:40.160 | |
Second is Life 3.0. | |
00:40.160 --> 00:44.080 | |
He's truly an out of the box thinker and a fun personality, | |
00:44.080 --> 00:45.480 | |
so I really enjoy talking to him. | |
00:45.480 --> 00:47.980 | |
If you'd like to see more of these videos in the future, | |
00:47.980 --> 00:50.640 | |
please subscribe and also click the little bell icon | |
00:50.640 --> 00:52.720 | |
to make sure you don't miss any videos. | |
00:52.720 --> 00:56.840 | |
Also, Twitter, LinkedIn, agi.mit.edu | |
00:56.840 --> 00:59.600 | |
if you wanna watch other lectures | |
00:59.600 --> 01:01.080 | |
or conversations like this one. | |
01:01.080 --> 01:04.000 | |
Better yet, go read Max's book, Life 3.0. | |
01:04.000 --> 01:07.940 | |
Chapter seven on goals is my favorite. | |
01:07.940 --> 01:10.480 | |
It's really where philosophy and engineering come together | |
01:10.480 --> 01:13.440 | |
and it opens with a quote by Dostoevsky. | |
01:14.400 --> 01:17.940 | |
The mystery of human existence lies not in just staying alive | |
01:17.940 --> 01:20.520 | |
but in finding something to live for. | |
01:20.520 --> 01:23.920 | |
Lastly, I believe that every failure rewards us | |
01:23.920 --> 01:26.560 | |
with an opportunity to learn | |
01:26.560 --> 01:28.360 | |
and in that sense, I've been very fortunate | |
01:28.360 --> 01:30.960 | |
to fail in so many new and exciting ways | |
01:31.840 --> 01:34.020 | |
and this conversation was no different. | |
01:34.020 --> 01:36.160 | |
I've learned about something called | |
01:36.160 --> 01:40.840 | |
radio frequency interference, RFI, look it up. | |
01:40.840 --> 01:42.960 | |
Apparently, music and conversations | |
01:42.960 --> 01:45.480 | |
from local radio stations can bleed into the audio | |
01:45.480 --> 01:47.080 | |
that you're recording in such a way | |
01:47.080 --> 01:49.360 | |
that it almost completely ruins that audio. | |
01:49.360 --> 01:52.060 | |
It's an exceptionally difficult sound source to remove. | |
01:53.240 --> 01:55.520 | |
So, I've gotten the opportunity to learn | |
01:55.520 --> 02:00.200 | |
how to avoid RFI in the future during recording sessions. | |
02:00.200 --> 02:02.680 | |
I've also gotten the opportunity to learn | |
02:02.680 --> 02:06.240 | |
how to use Adobe Audition and iZotope RX 6 | |
02:06.240 --> 02:11.240 | |
to do some noise, some audio repair. | |
02:11.720 --> 02:14.380 | |
Of course, this is an exceptionally difficult noise | |
02:14.380 --> 02:15.220 | |
to remove. | |
02:15.220 --> 02:16.280 | |
I am an engineer. | |
02:16.280 --> 02:18.240 | |
I'm not an audio engineer. | |
02:18.240 --> 02:20.180 | |
Neither is anybody else in our group | |
02:20.180 --> 02:21.880 | |
but we did our best. | |
02:21.880 --> 02:25.040 | |
Nevertheless, I thank you for your patience | |
02:25.040 --> 02:27.960 | |
and I hope you're still able to enjoy this conversation. | |
02:27.960 --> 02:29.320 | |
Do you think there's intelligent life | |
02:29.320 --> 02:31.360 | |
out there in the universe? | |
02:31.360 --> 02:33.480 | |
Let's open up with an easy question. | |
02:33.480 --> 02:36.240 | |
I have a minority view here actually. | |
02:36.240 --> 02:39.440 | |
When I give public lectures, I often ask for a show of hands | |
02:39.440 --> 02:42.920 | |
who thinks there's intelligent life out there somewhere else | |
02:42.920 --> 02:45.440 | |
and almost everyone put their hands up | |
02:45.440 --> 02:47.360 | |
and when I ask why, they'll be like, | |
02:47.360 --> 02:50.900 | |
oh, there's so many galaxies out there, there's gotta be. | |
02:51.840 --> 02:54.560 | |
But I'm a numbers nerd, right? | |
02:54.560 --> 02:56.640 | |
So when you look more carefully at it, | |
02:56.640 --> 02:58.040 | |
it's not so clear at all. | |
02:59.080 --> 03:00.680 | |
When we talk about our universe, first of all, | |
03:00.680 --> 03:03.040 | |
we don't mean all of space. | |
03:03.040 --> 03:04.040 | |
We actually mean, I don't know, | |
03:04.040 --> 03:05.440 | |
you can throw me the universe if you want, | |
03:05.440 --> 03:07.280 | |
it's behind you there. | |
03:07.280 --> 03:11.440 | |
It's, we simply mean the spherical region of space | |
03:11.440 --> 03:15.360 | |
from which light has a time to reach us so far | |
03:15.360 --> 03:17.040 | |
during the 14.8 billion year, | |
03:17.040 --> 03:19.320 | |
13.8 billion years since our Big Bang. | |
03:19.320 --> 03:22.320 | |
There's more space here but this is what we call a universe | |
03:22.320 --> 03:24.040 | |
because that's all we have access to. | |
03:24.040 --> 03:25.960 | |
So is there intelligent life here | |
03:25.960 --> 03:28.920 | |
that's gotten to the point of building telescopes | |
03:28.920 --> 03:29.960 | |
and computers? | |
03:31.160 --> 03:34.540 | |
My guess is no, actually. | |
03:34.540 --> 03:37.800 | |
The probability of it happening on any given planet | |
03:39.240 --> 03:42.620 | |
is some number we don't know what it is. | |
03:42.620 --> 03:47.620 | |
And what we do know is that the number can't be super high | |
03:48.480 --> 03:50.300 | |
because there's over a billion Earth like planets | |
03:50.300 --> 03:52.880 | |
in the Milky Way galaxy alone, | |
03:52.880 --> 03:56.280 | |
many of which are billions of years older than Earth. | |
03:56.280 --> 04:00.600 | |
And aside from some UFO believers, | |
04:00.600 --> 04:01.880 | |
there isn't much evidence | |
04:01.880 --> 04:05.600 | |
that any superduran civilization has come here at all. | |
04:05.600 --> 04:08.440 | |
And so that's the famous Fermi paradox, right? | |
04:08.440 --> 04:10.180 | |
And then if you work the numbers, | |
04:10.180 --> 04:13.440 | |
what you find is that if you have no clue | |
04:13.440 --> 04:16.880 | |
what the probability is of getting life on a given planet, | |
04:16.880 --> 04:19.680 | |
so it could be 10 to the minus 10, 10 to the minus 20, | |
04:19.680 --> 04:22.960 | |
or 10 to the minus two, or any power of 10 | |
04:22.960 --> 04:23.800 | |
is sort of equally likely | |
04:23.800 --> 04:25.480 | |
if you wanna be really open minded, | |
04:25.480 --> 04:27.600 | |
that translates into it being equally likely | |
04:27.600 --> 04:31.800 | |
that our nearest neighbor is 10 to the 16 meters away, | |
04:31.800 --> 04:33.880 | |
10 to the 17 meters away, 10 to the 18. | |
04:35.400 --> 04:40.400 | |
By the time you get much less than 10 to the 16 already, | |
04:41.080 --> 04:45.960 | |
we pretty much know there is nothing else that close. | |
04:45.960 --> 04:47.280 | |
And when you get beyond 10. | |
04:47.280 --> 04:48.680 | |
Because they would have discovered us. | |
04:48.680 --> 04:50.360 | |
Yeah, they would have been discovered as long ago, | |
04:50.360 --> 04:51.440 | |
or if they're really close, | |
04:51.440 --> 04:53.560 | |
we would have probably noted some engineering projects | |
04:53.560 --> 04:54.640 | |
that they're doing. | |
04:54.640 --> 04:57.880 | |
And if it's beyond 10 to the 26 meters, | |
04:57.880 --> 05:00.000 | |
that's already outside of here. | |
05:00.000 --> 05:05.000 | |
So my guess is actually that we are the only life in here | |
05:05.800 --> 05:09.040 | |
that's gotten the point of building advanced tech, | |
05:09.040 --> 05:10.720 | |
which I think is very, | |
05:12.680 --> 05:15.360 | |
puts a lot of responsibility on our shoulders, not screw up. | |
05:15.360 --> 05:17.240 | |
I think people who take for granted | |
05:17.240 --> 05:20.120 | |
that it's okay for us to screw up, | |
05:20.120 --> 05:22.760 | |
have an accidental nuclear war or go extinct somehow | |
05:22.760 --> 05:25.960 | |
because there's a sort of Star Trek like situation out there | |
05:25.960 --> 05:28.360 | |
where some other life forms are gonna come and bail us out | |
05:28.360 --> 05:30.400 | |
and it doesn't matter as much. | |
05:30.400 --> 05:33.400 | |
I think they're leveling us into a false sense of security. | |
05:33.400 --> 05:35.200 | |
I think it's much more prudent to say, | |
05:35.200 --> 05:36.400 | |
let's be really grateful | |
05:36.400 --> 05:38.720 | |
for this amazing opportunity we've had | |
05:38.720 --> 05:43.720 | |
and make the best of it just in case it is down to us. | |
05:44.080 --> 05:45.680 | |
So from a physics perspective, | |
05:45.680 --> 05:48.800 | |
do you think intelligent life, | |
05:48.800 --> 05:51.360 | |
so it's unique from a sort of statistical view | |
05:51.360 --> 05:52.560 | |
of the size of the universe, | |
05:52.560 --> 05:55.840 | |
but from the basic matter of the universe, | |
05:55.840 --> 05:59.040 | |
how difficult is it for intelligent life to come about? | |
05:59.040 --> 06:01.280 | |
The kind of advanced tech building life | |
06:03.120 --> 06:05.720 | |
is implied in your statement that it's really difficult | |
06:05.720 --> 06:07.640 | |
to create something like a human species. | |
06:07.640 --> 06:11.560 | |
Well, I think what we know is that going from no life | |
06:11.560 --> 06:15.720 | |
to having life that can do a level of tech, | |
06:15.720 --> 06:18.720 | |
there's some sort of two going beyond that | |
06:18.720 --> 06:22.200 | |
than actually settling our whole universe with life. | |
06:22.200 --> 06:26.560 | |
There's some major roadblock there, | |
06:26.560 --> 06:30.880 | |
which is some great filter as it's sometimes called, | |
06:30.880 --> 06:33.520 | |
which is tough to get through. | |
06:33.520 --> 06:37.160 | |
It's either that roadblock is either behind us | |
06:37.160 --> 06:38.720 | |
or in front of us. | |
06:38.720 --> 06:41.080 | |
I'm hoping very much that it's behind us. | |
06:41.080 --> 06:45.960 | |
I'm super excited every time we get a new report from NASA | |
06:45.960 --> 06:48.480 | |
saying they failed to find any life on Mars. | |
06:48.480 --> 06:50.080 | |
I'm like, yes, awesome. | |
06:50.080 --> 06:51.680 | |
Because that suggests that the hard part, | |
06:51.680 --> 06:54.240 | |
maybe it was getting the first ribosome | |
06:54.240 --> 06:59.240 | |
or some very low level kind of stepping stone | |
06:59.520 --> 07:00.400 | |
so that we're home free. | |
07:00.400 --> 07:01.720 | |
Because if that's true, | |
07:01.720 --> 07:03.640 | |
then the future is really only limited | |
07:03.640 --> 07:05.200 | |
by our own imagination. | |
07:05.200 --> 07:07.360 | |
It would be much suckier if it turns out | |
07:07.360 --> 07:11.440 | |
that this level of life is kind of a dime a dozen, | |
07:11.440 --> 07:12.760 | |
but maybe there's some other problem. | |
07:12.760 --> 07:16.160 | |
Like as soon as a civilization gets advanced technology, | |
07:16.160 --> 07:17.000 | |
within a hundred years, | |
07:17.000 --> 07:20.320 | |
they get into some stupid fight with themselves and poof. | |
07:20.320 --> 07:21.760 | |
That would be a bummer. | |
07:21.760 --> 07:26.160 | |
Yeah, so you've explored the mysteries of the universe, | |
07:26.160 --> 07:29.000 | |
the cosmological universe, the one that's sitting | |
07:29.000 --> 07:30.000 | |
between us today. | |
07:31.080 --> 07:35.960 | |
I think you've also begun to explore the other universe, | |
07:35.960 --> 07:38.000 | |
which is sort of the mystery, | |
07:38.000 --> 07:40.960 | |
the mysterious universe of the mind of intelligence, | |
07:40.960 --> 07:42.840 | |
of intelligent life. | |
07:42.840 --> 07:45.280 | |
So is there a common thread between your interest | |
07:45.280 --> 07:48.760 | |
or the way you think about space and intelligence? | |
07:48.760 --> 07:51.040 | |
Oh yeah, when I was a teenager, | |
07:53.040 --> 07:57.280 | |
I was already very fascinated by the biggest questions. | |
07:57.280 --> 08:00.560 | |
And I felt that the two biggest mysteries of all in science | |
08:00.560 --> 08:05.000 | |
were our universe out there and our universe in here. | |
08:05.000 --> 08:08.120 | |
So it's quite natural after having spent | |
08:08.120 --> 08:11.040 | |
a quarter of a century on my career, | |
08:11.040 --> 08:12.680 | |
thinking a lot about this one, | |
08:12.680 --> 08:14.320 | |
that I'm now indulging in the luxury | |
08:14.320 --> 08:15.960 | |
of doing research on this one. | |
08:15.960 --> 08:17.720 | |
It's just so cool. | |
08:17.720 --> 08:20.120 | |
I feel the time is ripe now | |
08:20.120 --> 08:25.120 | |
for you trans greatly deepening our understanding of this. | |
08:25.120 --> 08:26.640 | |
Just start exploring this one. | |
08:26.640 --> 08:29.560 | |
Yeah, because I think a lot of people view intelligence | |
08:29.560 --> 08:33.520 | |
as something mysterious that can only exist | |
08:33.520 --> 08:36.120 | |
in biological organisms like us, | |
08:36.120 --> 08:37.680 | |
and therefore dismiss all talk | |
08:37.680 --> 08:41.160 | |
about artificial general intelligence as science fiction. | |
08:41.160 --> 08:43.200 | |
But from my perspective as a physicist, | |
08:43.200 --> 08:46.680 | |
I am a blob of quarks and electrons | |
08:46.680 --> 08:48.360 | |
moving around in a certain pattern | |
08:48.360 --> 08:50.080 | |
and processing information in certain ways. | |
08:50.080 --> 08:53.600 | |
And this is also a blob of quarks and electrons. | |
08:53.600 --> 08:55.360 | |
I'm not smarter than the water bottle | |
08:55.360 --> 08:57.880 | |
because I'm made of different kinds of quarks. | |
08:57.880 --> 08:59.640 | |
I'm made of up quarks and down quarks, | |
08:59.640 --> 09:01.400 | |
exact same kind as this. | |
09:01.400 --> 09:05.080 | |
There's no secret sauce, I think, in me. | |
09:05.080 --> 09:08.560 | |
It's all about the pattern of the information processing. | |
09:08.560 --> 09:12.240 | |
And this means that there's no law of physics | |
09:12.240 --> 09:15.600 | |
saying that we can't create technology, | |
09:15.600 --> 09:19.960 | |
which can help us by being incredibly intelligent | |
09:19.960 --> 09:21.680 | |
and help us crack mysteries that we couldn't. | |
09:21.680 --> 09:23.560 | |
In other words, I think we've really only seen | |
09:23.560 --> 09:26.480 | |
the tip of the intelligence iceberg so far. | |
09:26.480 --> 09:29.960 | |
Yeah, so the perceptronium. | |
09:29.960 --> 09:31.280 | |
Yeah. | |
09:31.280 --> 09:33.200 | |
So you coined this amazing term. | |
09:33.200 --> 09:35.760 | |
It's a hypothetical state of matter, | |
09:35.760 --> 09:38.360 | |
sort of thinking from a physics perspective, | |
09:38.360 --> 09:40.080 | |
what is the kind of matter that can help, | |
09:40.080 --> 09:42.920 | |
as you're saying, subjective experience emerge, | |
09:42.920 --> 09:44.280 | |
consciousness emerge. | |
09:44.280 --> 09:46.640 | |
So how do you think about consciousness | |
09:46.640 --> 09:48.160 | |
from this physics perspective? | |
09:49.960 --> 09:50.800 | |
Very good question. | |
09:50.800 --> 09:55.800 | |
So again, I think many people have underestimated | |
09:55.800 --> 09:59.120 | |
our ability to make progress on this | |
09:59.120 --> 10:01.320 | |
by convincing themselves it's hopeless | |
10:01.320 --> 10:05.840 | |
because somehow we're missing some ingredient that we need. | |
10:05.840 --> 10:09.560 | |
There's some new consciousness particle or whatever. | |
10:09.560 --> 10:12.720 | |
I happen to think that we're not missing anything | |
10:12.720 --> 10:16.320 | |
and that it's not the interesting thing | |
10:16.320 --> 10:18.560 | |
about consciousness that gives us | |
10:18.560 --> 10:21.400 | |
this amazing subjective experience of colors | |
10:21.400 --> 10:23.320 | |
and sounds and emotions. | |
10:23.320 --> 10:26.320 | |
It's rather something at the higher level | |
10:26.320 --> 10:28.800 | |
about the patterns of information processing. | |
10:28.800 --> 10:33.160 | |
And that's why I like to think about this idea | |
10:33.160 --> 10:34.480 | |
of perceptronium. | |
10:34.480 --> 10:36.920 | |
What does it mean for an arbitrary physical system | |
10:36.920 --> 10:41.920 | |
to be conscious in terms of what its particles are doing | |
10:41.920 --> 10:43.560 | |
or its information is doing? | |
10:43.560 --> 10:46.080 | |
I don't think, I hate carbon chauvinism, | |
10:46.080 --> 10:47.960 | |
this attitude you have to be made of carbon atoms | |
10:47.960 --> 10:50.160 | |
to be smart or conscious. | |
10:50.160 --> 10:53.520 | |
There's something about the information processing | |
10:53.520 --> 10:55.360 | |
that this kind of matter performs. | |
10:55.360 --> 10:57.840 | |
Yeah, and you can see I have my favorite equations here | |
10:57.840 --> 11:00.720 | |
describing various fundamental aspects of the world. | |
11:00.720 --> 11:02.560 | |
I feel that I think one day, | |
11:02.560 --> 11:04.360 | |
maybe someone who's watching this will come up | |
11:04.360 --> 11:07.280 | |
with the equations that information processing | |
11:07.280 --> 11:08.760 | |
has to satisfy to be conscious. | |
11:08.760 --> 11:11.800 | |
I'm quite convinced there is big discovery | |
11:11.800 --> 11:15.400 | |
to be made there because let's face it, | |
11:15.400 --> 11:18.720 | |
we know that so many things are made up of information. | |
11:18.720 --> 11:21.960 | |
We know that some information processing is conscious | |
11:21.960 --> 11:25.520 | |
because we are conscious. | |
11:25.520 --> 11:27.600 | |
But we also know that a lot of information processing | |
11:27.600 --> 11:28.440 | |
is not conscious. | |
11:28.440 --> 11:30.040 | |
Like most of the information processing happening | |
11:30.040 --> 11:32.680 | |
in your brain right now is not conscious. | |
11:32.680 --> 11:36.040 | |
There are like 10 megabytes per second coming in | |
11:36.040 --> 11:38.080 | |
even just through your visual system. | |
11:38.080 --> 11:40.480 | |
You're not conscious about your heartbeat regulation | |
11:40.480 --> 11:42.120 | |
or most things. | |
11:42.120 --> 11:45.680 | |
Even if I just ask you to like read what it says here, | |
11:45.680 --> 11:48.040 | |
you look at it and then, oh, now you know what it said. | |
11:48.040 --> 11:51.560 | |
But you're not aware of how the computation actually happened. | |
11:51.560 --> 11:53.680 | |
Your consciousness is like the CEO | |
11:53.680 --> 11:56.680 | |
that got an email at the end with the final answer. | |
11:56.680 --> 12:01.000 | |
So what is it that makes a difference? | |
12:01.000 --> 12:05.120 | |
I think that's both a great science mystery. | |
12:05.120 --> 12:07.080 | |
We're actually studying it a little bit in my lab here | |
12:07.080 --> 12:10.920 | |
at MIT, but I also think it's just a really urgent question | |
12:10.920 --> 12:12.080 | |
to answer. | |
12:12.080 --> 12:14.880 | |
For starters, I mean, if you're an emergency room doctor | |
12:14.880 --> 12:17.160 | |
and you have an unresponsive patient coming in, | |
12:17.160 --> 12:19.600 | |
wouldn't it be great if in addition to having | |
12:22.360 --> 12:25.320 | |
a CT scanner, you had a consciousness scanner | |
12:25.320 --> 12:27.920 | |
that could figure out whether this person | |
12:27.920 --> 12:30.960 | |
is actually having locked in syndrome | |
12:30.960 --> 12:32.440 | |
or is actually comatose. | |
12:33.360 --> 12:37.000 | |
And in the future, imagine if we build robots | |
12:37.000 --> 12:41.480 | |
or the machine that we can have really good conversations | |
12:41.480 --> 12:44.840 | |
with, which I think is very likely to happen. | |
12:44.840 --> 12:47.760 | |
Wouldn't you want to know if your home helper robot | |
12:47.760 --> 12:51.320 | |
is actually experiencing anything or just like a zombie, | |
12:51.320 --> 12:53.520 | |
I mean, would you prefer it? | |
12:53.520 --> 12:54.360 | |
What would you prefer? | |
12:54.360 --> 12:56.200 | |
Would you prefer that it's actually unconscious | |
12:56.200 --> 12:58.560 | |
so that you don't have to feel guilty about switching it off | |
12:58.560 --> 13:02.120 | |
or giving boring chores or what would you prefer? | |
13:02.120 --> 13:06.520 | |
Well, certainly we would prefer, | |
13:06.520 --> 13:08.960 | |
I would prefer the appearance of consciousness. | |
13:08.960 --> 13:11.720 | |
But the question is whether the appearance of consciousness | |
13:11.720 --> 13:15.040 | |
is different than consciousness itself. | |
13:15.040 --> 13:18.200 | |
And sort of to ask that as a question, | |
13:18.200 --> 13:21.760 | |
do you think we need to understand what consciousness is, | |
13:21.760 --> 13:23.520 | |
solve the hard problem of consciousness | |
13:23.520 --> 13:28.240 | |
in order to build something like an AGI system? | |
13:28.240 --> 13:30.440 | |
No, I don't think that. | |
13:30.440 --> 13:34.520 | |
And I think we will probably be able to build things | |
13:34.520 --> 13:36.080 | |
even if we don't answer that question. | |
13:36.080 --> 13:37.720 | |
But if we want to make sure that what happens | |
13:37.720 --> 13:40.960 | |
is a good thing, we better solve it first. | |
13:40.960 --> 13:44.960 | |
So it's a wonderful controversy you're raising there | |
13:44.960 --> 13:47.960 | |
where you have basically three points of view | |
13:47.960 --> 13:48.800 | |
about the hard problem. | |
13:48.800 --> 13:52.800 | |
So there are two different points of view. | |
13:52.800 --> 13:55.160 | |
They both conclude that the hard problem of consciousness | |
13:55.160 --> 13:56.840 | |
is BS. | |
13:56.840 --> 13:59.320 | |
On one hand, you have some people like Daniel Dennett | |
13:59.320 --> 14:01.480 | |
who say that consciousness is just BS | |
14:01.480 --> 14:05.000 | |
because consciousness is the same thing as intelligence. | |
14:05.000 --> 14:06.440 | |
There's no difference. | |
14:06.440 --> 14:11.080 | |
So anything which acts conscious is conscious, | |
14:11.080 --> 14:13.480 | |
just like we are. | |
14:13.480 --> 14:15.960 | |
And then there are also a lot of people, | |
14:15.960 --> 14:18.400 | |
including many top AI researchers I know, | |
14:18.400 --> 14:19.920 | |
who say, oh, consciousness is just bullshit | |
14:19.920 --> 14:22.760 | |
because, of course, machines can never be conscious. | |
14:22.760 --> 14:24.520 | |
They're always going to be zombies. | |
14:24.520 --> 14:27.880 | |
You never have to feel guilty about how you treat them. | |
14:27.880 --> 14:30.880 | |
And then there's a third group of people, | |
14:30.880 --> 14:34.920 | |
including Giulio Tononi, for example, | |
14:34.920 --> 14:37.440 | |
and Krzysztof Koch and a number of others. | |
14:37.440 --> 14:39.520 | |
I would put myself also in this middle camp | |
14:39.520 --> 14:41.880 | |
who say that actually some information processing | |
14:41.880 --> 14:44.160 | |
is conscious and some is not. | |
14:44.160 --> 14:46.960 | |
So let's find the equation which can be used | |
14:46.960 --> 14:49.080 | |
to determine which it is. | |
14:49.080 --> 14:52.040 | |
And I think we've just been a little bit lazy, | |
14:52.040 --> 14:54.960 | |
kind of running away from this problem for a long time. | |
14:54.960 --> 14:57.840 | |
It's been almost taboo to even mention the C word | |
14:57.840 --> 15:00.520 | |
in a lot of circles because, | |
15:00.520 --> 15:03.520 | |
but we should stop making excuses. | |
15:03.520 --> 15:07.920 | |
This is a science question and there are ways | |
15:07.920 --> 15:11.960 | |
we can even test any theory that makes predictions for this. | |
15:11.960 --> 15:13.640 | |
And coming back to this helper robot, | |
15:13.640 --> 15:16.080 | |
I mean, so you said you'd want your helper robot | |
15:16.080 --> 15:18.160 | |
to certainly act conscious and treat you, | |
15:18.160 --> 15:20.880 | |
like have conversations with you and stuff. | |
15:20.880 --> 15:21.720 | |
I think so. | |
15:21.720 --> 15:22.560 | |
But wouldn't you, would you feel, | |
15:22.560 --> 15:23.920 | |
would you feel a little bit creeped out | |
15:23.920 --> 15:27.680 | |
if you realized that it was just a glossed up tape recorder, | |
15:27.680 --> 15:31.560 | |
you know, that was just zombie and was a faking emotion? | |
15:31.560 --> 15:34.560 | |
Would you prefer that it actually had an experience | |
15:34.560 --> 15:37.000 | |
or would you prefer that it's actually | |
15:37.000 --> 15:39.120 | |
not experiencing anything so you feel, | |
15:39.120 --> 15:42.200 | |
you don't have to feel guilty about what you do to it? | |
15:42.200 --> 15:45.040 | |
It's such a difficult question because, you know, | |
15:45.040 --> 15:47.280 | |
it's like when you're in a relationship and you say, | |
15:47.280 --> 15:48.120 | |
well, I love you. | |
15:48.120 --> 15:49.760 | |
And the other person said, I love you back. | |
15:49.760 --> 15:52.640 | |
It's like asking, well, do they really love you back | |
15:52.640 --> 15:55.360 | |
or are they just saying they love you back? | |
15:55.360 --> 15:58.120 | |
Don't you really want them to actually love you? | |
15:58.120 --> 16:03.120 | |
It's hard to, it's hard to really know the difference | |
16:03.520 --> 16:08.520 | |
between everything seeming like there's consciousness | |
16:09.000 --> 16:10.640 | |
present, there's intelligence present, | |
16:10.640 --> 16:13.840 | |
there's affection, passion, love, | |
16:13.840 --> 16:16.200 | |
and it actually being there. | |
16:16.200 --> 16:17.720 | |
I'm not sure, do you have? | |
16:17.720 --> 16:19.400 | |
But like, can I ask you a question about this? | |
16:19.400 --> 16:20.760 | |
Like to make it a bit more pointed. | |
16:20.760 --> 16:22.920 | |
So Mass General Hospital is right across the river, right? | |
16:22.920 --> 16:23.760 | |
Yes. | |
16:23.760 --> 16:26.720 | |
Suppose you're going in for a medical procedure | |
16:26.720 --> 16:29.320 | |
and they're like, you know, for anesthesia, | |
16:29.320 --> 16:31.000 | |
what we're going to do is we're going to give you | |
16:31.000 --> 16:33.160 | |
muscle relaxants so you won't be able to move | |
16:33.160 --> 16:35.040 | |
and you're going to feel excruciating pain | |
16:35.040 --> 16:35.880 | |
during the whole surgery, | |
16:35.880 --> 16:37.600 | |
but you won't be able to do anything about it. | |
16:37.600 --> 16:39.200 | |
But then we're going to give you this drug | |
16:39.200 --> 16:40.760 | |
that erases your memory of it. | |
16:41.960 --> 16:43.440 | |
Would you be cool about that? | |
16:44.960 --> 16:47.600 | |
What's the difference that you're conscious about it | |
16:48.600 --> 16:51.640 | |
or not if there's no behavioral change, right? | |
16:51.640 --> 16:54.520 | |
Right, that's a really, that's a really clear way to put it. | |
16:54.520 --> 16:57.400 | |
That's, yeah, it feels like in that sense, | |
16:57.400 --> 17:01.080 | |
experiencing it is a valuable quality. | |
17:01.080 --> 17:04.800 | |
So actually being able to have subjective experiences, | |
17:05.840 --> 17:09.120 | |
at least in that case, is valuable. | |
17:09.120 --> 17:11.240 | |
And I think we humans have a little bit | |
17:11.240 --> 17:13.600 | |
of a bad track record also of making | |
17:13.600 --> 17:15.480 | |
these self serving arguments | |
17:15.480 --> 17:18.040 | |
that other entities aren't conscious. | |
17:18.040 --> 17:19.160 | |
You know, people often say, | |
17:19.160 --> 17:21.800 | |
oh, these animals can't feel pain. | |
17:21.800 --> 17:24.040 | |
It's okay to boil lobsters because we ask them | |
17:24.040 --> 17:25.960 | |
if it hurt and they didn't say anything. | |
17:25.960 --> 17:27.400 | |
And now there was just a paper out saying, | |
17:27.400 --> 17:29.320 | |
lobsters do feel pain when you boil them | |
17:29.320 --> 17:31.040 | |
and they're banning it in Switzerland. | |
17:31.040 --> 17:33.560 | |
And we did this with slaves too often and said, | |
17:33.560 --> 17:34.680 | |
oh, they don't mind. | |
17:36.240 --> 17:39.480 | |
They don't maybe aren't conscious | |
17:39.480 --> 17:41.160 | |
or women don't have souls or whatever. | |
17:41.160 --> 17:43.200 | |
So I'm a little bit nervous when I hear people | |
17:43.200 --> 17:46.360 | |
just take as an axiom that machines | |
17:46.360 --> 17:48.960 | |
can't have experience ever. | |
17:48.960 --> 17:51.560 | |
I think this is just a really fascinating science question | |
17:51.560 --> 17:52.400 | |
is what it is. | |
17:52.400 --> 17:54.720 | |
Let's research it and try to figure out | |
17:54.720 --> 17:56.000 | |
what it is that makes the difference | |
17:56.000 --> 17:58.880 | |
between unconscious intelligent behavior | |
17:58.880 --> 18:01.120 | |
and conscious intelligent behavior. | |
18:01.120 --> 18:04.680 | |
So in terms of, so if you think of a Boston Dynamics | |
18:04.680 --> 18:07.680 | |
human or robot being sort of with a broom | |
18:07.680 --> 18:11.920 | |
being pushed around, it starts pushing | |
18:11.920 --> 18:13.320 | |
on a consciousness question. | |
18:13.320 --> 18:17.040 | |
So let me ask, do you think an AGI system | |
18:17.040 --> 18:19.720 | |
like a few neuroscientists believe | |
18:19.720 --> 18:22.320 | |
needs to have a physical embodiment? | |
18:22.320 --> 18:25.720 | |
Needs to have a body or something like a body? | |
18:25.720 --> 18:28.280 | |
No, I don't think so. | |
18:28.280 --> 18:30.560 | |
You mean to have a conscious experience? | |
18:30.560 --> 18:31.640 | |
To have consciousness. | |
18:33.160 --> 18:36.080 | |
I do think it helps a lot to have a physical embodiment | |
18:36.080 --> 18:38.440 | |
to learn the kind of things about the world | |
18:38.440 --> 18:41.480 | |
that are important to us humans, for sure. | |
18:42.560 --> 18:45.600 | |
But I don't think the physical embodiment | |
18:45.600 --> 18:47.120 | |
is necessary after you've learned it | |
18:47.120 --> 18:48.760 | |
to just have the experience. | |
18:48.760 --> 18:51.400 | |
Think about when you're dreaming, right? | |
18:51.400 --> 18:52.600 | |
Your eyes are closed. | |
18:52.600 --> 18:54.240 | |
You're not getting any sensory input. | |
18:54.240 --> 18:55.960 | |
You're not behaving or moving in any way | |
18:55.960 --> 18:58.160 | |
but there's still an experience there, right? | |
18:59.720 --> 19:01.400 | |
And so clearly the experience that you have | |
19:01.400 --> 19:03.320 | |
when you see something cool in your dreams | |
19:03.320 --> 19:04.800 | |
isn't coming from your eyes. | |
19:04.800 --> 19:08.640 | |
It's just the information processing itself in your brain | |
19:08.640 --> 19:10.920 | |
which is that experience, right? | |
19:10.920 --> 19:13.640 | |
But if I put it another way, I'll say | |
19:13.640 --> 19:15.120 | |
because it comes from neuroscience | |
19:15.120 --> 19:18.280 | |
is the reason you want to have a body and a physical | |
19:18.280 --> 19:23.280 | |
something like a physical, you know, a physical system | |
19:23.920 --> 19:27.040 | |
is because you want to be able to preserve something. | |
19:27.040 --> 19:30.840 | |
In order to have a self, you could argue, | |
19:30.840 --> 19:35.840 | |
would you need to have some kind of embodiment of self | |
19:36.400 --> 19:37.960 | |
to want to preserve? | |
19:38.920 --> 19:42.400 | |
Well, now we're getting a little bit anthropomorphic | |
19:42.400 --> 19:45.200 | |
into anthropomorphizing things. | |
19:45.200 --> 19:47.280 | |
Maybe talking about self preservation instincts. | |
19:47.280 --> 19:50.560 | |
I mean, we are evolved organisms, right? | |
19:50.560 --> 19:53.520 | |
So Darwinian evolution endowed us | |
19:53.520 --> 19:57.120 | |
and other evolved organism with a self preservation instinct | |
19:57.120 --> 20:00.560 | |
because those that didn't have those self preservation genes | |
20:00.560 --> 20:02.960 | |
got cleaned out of the gene pool, right? | |
20:02.960 --> 20:06.880 | |
But if you build an artificial general intelligence | |
20:06.880 --> 20:10.040 | |
the mind space that you can design is much, much larger | |
20:10.040 --> 20:14.440 | |
than just a specific subset of minds that can evolve. | |
20:14.440 --> 20:17.280 | |
So an AGI mind doesn't necessarily have | |
20:17.280 --> 20:19.880 | |
to have any self preservation instinct. | |
20:19.880 --> 20:21.600 | |
It also doesn't necessarily have to be | |
20:21.600 --> 20:24.040 | |
so individualistic as us. | |
20:24.040 --> 20:26.080 | |
Like, imagine if you could just, first of all, | |
20:26.080 --> 20:27.960 | |
or we are also very afraid of death. | |
20:27.960 --> 20:29.920 | |
You know, I suppose you could back yourself up | |
20:29.920 --> 20:32.000 | |
every five minutes and then your airplane | |
20:32.000 --> 20:32.840 | |
is about to crash. | |
20:32.840 --> 20:36.680 | |
You're like, shucks, I'm gonna lose the last five minutes | |
20:36.680 --> 20:39.520 | |
of experiences since my last cloud backup, dang. | |
20:39.520 --> 20:41.520 | |
You know, it's not as big a deal. | |
20:41.520 --> 20:45.680 | |
Or if we could just copy experiences between our minds | |
20:45.680 --> 20:47.640 | |
easily like we, which we could easily do | |
20:47.640 --> 20:50.360 | |
if we were silicon based, right? | |
20:50.360 --> 20:54.040 | |
Then maybe we would feel a little bit more | |
20:54.040 --> 20:56.560 | |
like a hive mind actually, that maybe it's the, | |
20:56.560 --> 20:59.960 | |
so I don't think we should take for granted at all | |
20:59.960 --> 21:03.000 | |
that AGI will have to have any of those sort of | |
21:04.880 --> 21:07.360 | |
competitive as alpha male instincts. | |
21:07.360 --> 21:10.160 | |
On the other hand, you know, this is really interesting | |
21:10.160 --> 21:13.840 | |
because I think some people go too far and say, | |
21:13.840 --> 21:16.680 | |
of course we don't have to have any concerns either | |
21:16.680 --> 21:20.800 | |
that advanced AI will have those instincts | |
21:20.800 --> 21:22.680 | |
because we can build anything we want. | |
21:22.680 --> 21:26.280 | |
That there's a very nice set of arguments going back | |
21:26.280 --> 21:28.560 | |
to Steve Omohundro and Nick Bostrom and others | |
21:28.560 --> 21:32.280 | |
just pointing out that when we build machines, | |
21:32.280 --> 21:34.680 | |
we normally build them with some kind of goal, you know, | |
21:34.680 --> 21:38.520 | |
win this chess game, drive this car safely or whatever. | |
21:38.520 --> 21:40.960 | |
And as soon as you put in a goal into machine, | |
21:40.960 --> 21:42.760 | |
especially if it's kind of open ended goal | |
21:42.760 --> 21:44.640 | |
and the machine is very intelligent, | |
21:44.640 --> 21:47.000 | |
it'll break that down into a bunch of sub goals. | |
21:48.280 --> 21:51.280 | |
And one of those goals will almost always | |
21:51.280 --> 21:54.200 | |
be self preservation because if it breaks or dies | |
21:54.200 --> 21:56.120 | |
in the process, it's not gonna accomplish the goal, right? | |
21:56.120 --> 21:58.040 | |
Like suppose you just build a little, | |
21:58.040 --> 22:01.000 | |
you have a little robot and you tell it to go down | |
22:01.000 --> 22:04.040 | |
the store market here and get you some food, | |
22:04.040 --> 22:06.200 | |
make you cook an Italian dinner, you know, | |
22:06.200 --> 22:08.400 | |
and then someone mugs it and tries to break it | |
22:08.400 --> 22:09.480 | |
on the way. | |
22:09.480 --> 22:12.920 | |
That robot has an incentive to not get destroyed | |
22:12.920 --> 22:14.720 | |
and defend itself or run away, | |
22:14.720 --> 22:17.720 | |
because otherwise it's gonna fail in cooking your dinner. | |
22:17.720 --> 22:19.560 | |
It's not afraid of death, | |
22:19.560 --> 22:22.960 | |
but it really wants to complete the dinner cooking goal. | |
22:22.960 --> 22:25.040 | |
So it will have a self preservation instinct. | |
22:25.040 --> 22:27.920 | |
Continue being a functional agent somehow. | |
22:27.920 --> 22:32.920 | |
And similarly, if you give any kind of more ambitious goal | |
22:33.720 --> 22:37.000 | |
to an AGI, it's very likely they wanna acquire | |
22:37.000 --> 22:39.840 | |
more resources so it can do that better. | |
22:39.840 --> 22:42.720 | |
And it's exactly from those sort of sub goals | |
22:42.720 --> 22:43.800 | |
that we might not have intended | |
22:43.800 --> 22:47.160 | |
that some of the concerns about AGI safety come. | |
22:47.160 --> 22:50.600 | |
You give it some goal that seems completely harmless. | |
22:50.600 --> 22:53.360 | |
And then before you realize it, | |
22:53.360 --> 22:55.480 | |
it's also trying to do these other things | |
22:55.480 --> 22:56.920 | |
which you didn't want it to do. | |
22:56.920 --> 22:59.160 | |
And it's maybe smarter than us. | |
22:59.160 --> 23:01.000 | |
So it's fascinating. | |
23:01.000 --> 23:05.680 | |
And let me pause just because I am in a very kind | |
23:05.680 --> 23:08.720 | |
of human centric way, see fear of death | |
23:08.720 --> 23:11.840 | |
as a valuable motivator. | |
23:11.840 --> 23:16.440 | |
So you don't think, you think that's an artifact | |
23:16.440 --> 23:19.120 | |
of evolution, so that's the kind of mind space | |
23:19.120 --> 23:22.120 | |
evolution created that we're sort of almost obsessed | |
23:22.120 --> 23:24.400 | |
about self preservation, some kind of genetic flow. | |
23:24.400 --> 23:29.400 | |
You don't think that's necessary to be afraid of death. | |
23:29.480 --> 23:32.920 | |
So not just a kind of sub goal of self preservation | |
23:32.920 --> 23:34.920 | |
just so you can keep doing the thing, | |
23:34.920 --> 23:38.720 | |
but more fundamentally sort of have the finite thing | |
23:38.720 --> 23:43.080 | |
like this ends for you at some point. | |
23:43.080 --> 23:44.160 | |
Interesting. | |
23:44.160 --> 23:47.440 | |
Do I think it's necessary for what precisely? | |
23:47.440 --> 23:50.920 | |
For intelligence, but also for consciousness. | |
23:50.920 --> 23:55.040 | |
So for those, for both, do you think really | |
23:55.040 --> 23:59.120 | |
like a finite death and the fear of it is important? | |
23:59.120 --> 24:04.120 | |
So before I can answer, before we can agree | |
24:05.160 --> 24:06.960 | |
on whether it's necessary for intelligence | |
24:06.960 --> 24:08.360 | |
or for consciousness, we should be clear | |
24:08.360 --> 24:09.800 | |
on how we define those two words. | |
24:09.800 --> 24:11.960 | |
Cause a lot of really smart people define them | |
24:11.960 --> 24:13.320 | |
in very different ways. | |
24:13.320 --> 24:17.080 | |
I was on this panel with AI experts | |
24:17.080 --> 24:20.080 | |
and they couldn't agree on how to define intelligence even. | |
24:20.080 --> 24:22.000 | |
So I define intelligence simply | |
24:22.000 --> 24:24.760 | |
as the ability to accomplish complex goals. | |
24:25.640 --> 24:27.280 | |
I like your broad definition, because again | |
24:27.280 --> 24:29.040 | |
I don't want to be a carbon chauvinist. | |
24:29.040 --> 24:30.400 | |
Right. | |
24:30.400 --> 24:34.600 | |
And in that case, no, certainly | |
24:34.600 --> 24:36.480 | |
it doesn't require fear of death. | |
24:36.480 --> 24:40.120 | |
I would say alpha go, alpha zero is quite intelligent. | |
24:40.120 --> 24:43.080 | |
I don't think alpha zero has any fear of being turned off | |
24:43.080 --> 24:46.320 | |
because it doesn't understand the concept of it even. | |
24:46.320 --> 24:48.440 | |
And similarly consciousness. | |
24:48.440 --> 24:52.240 | |
I mean, you could certainly imagine very simple | |
24:52.240 --> 24:53.920 | |
kind of experience. | |
24:53.920 --> 24:57.200 | |
If certain plants have any kind of experience | |
24:57.200 --> 24:58.560 | |
I don't think they're very afraid of dying | |
24:58.560 --> 25:00.920 | |
or there's nothing they can do about it anyway much. | |
25:00.920 --> 25:04.560 | |
So there wasn't that much value in, but more seriously | |
25:04.560 --> 25:09.200 | |
I think if you ask, not just about being conscious | |
25:09.200 --> 25:14.200 | |
but maybe having what you would, we might call | |
25:14.320 --> 25:16.400 | |
an exciting life where you feel passion | |
25:16.400 --> 25:21.400 | |
and really appreciate the things. | |
25:21.480 --> 25:24.440 | |
Maybe there somehow, maybe there perhaps it does help | |
25:24.440 --> 25:27.880 | |
having a backdrop that, Hey, it's finite. | |
25:27.880 --> 25:31.200 | |
No, let's make the most of this, let's live to the fullest. | |
25:31.200 --> 25:33.800 | |
So if you knew you were going to live forever | |
25:34.880 --> 25:37.400 | |
do you think you would change your? | |
25:37.400 --> 25:39.560 | |
Yeah, I mean, in some perspective | |
25:39.560 --> 25:43.960 | |
it would be an incredibly boring life living forever. | |
25:43.960 --> 25:47.360 | |
So in the sort of loose subjective terms that you said | |
25:47.360 --> 25:50.480 | |
of something exciting and something in this | |
25:50.480 --> 25:53.240 | |
that other humans would understand, I think is, yeah | |
25:53.240 --> 25:57.120 | |
it seems that the finiteness of it is important. | |
25:57.120 --> 25:59.560 | |
Well, the good news I have for you then is | |
25:59.560 --> 26:02.120 | |
based on what we understand about cosmology | |
26:02.120 --> 26:05.120 | |
everything is in our universe is probably | |
26:05.120 --> 26:07.960 | |
ultimately probably finite, although. | |
26:07.960 --> 26:11.560 | |
Big crunch or big, what's the, the infinite expansion. | |
26:11.560 --> 26:13.840 | |
Yeah, we could have a big chill or a big crunch | |
26:13.840 --> 26:18.440 | |
or a big rip or that's the big snap or death bubbles. | |
26:18.440 --> 26:20.040 | |
All of them are more than a billion years away. | |
26:20.040 --> 26:24.600 | |
So we should, we certainly have vastly more time | |
26:24.600 --> 26:27.920 | |
than our ancestors thought, but there is still | |
26:29.160 --> 26:32.360 | |
it's still pretty hard to squeeze in an infinite number | |
26:32.360 --> 26:36.560 | |
of compute cycles, even though there are some loopholes | |
26:36.560 --> 26:37.720 | |
that just might be possible. | |
26:37.720 --> 26:41.960 | |
But I think, you know, some people like to say | |
26:41.960 --> 26:44.760 | |
that you should live as if you're about to | |
26:44.760 --> 26:46.720 | |
you're going to die in five years or so. | |
26:46.720 --> 26:47.960 | |
And that's sort of optimal. | |
26:47.960 --> 26:50.560 | |
Maybe it's a good assumption. | |
26:50.560 --> 26:54.680 | |
We should build our civilization as if it's all finite | |
26:54.680 --> 26:55.680 | |
to be on the safe side. | |
26:55.680 --> 26:56.960 | |
Right, exactly. | |
26:56.960 --> 26:59.720 | |
So you mentioned defining intelligence | |
26:59.720 --> 27:02.960 | |
as the ability to solve complex goals. | |
27:02.960 --> 27:05.440 | |
Where would you draw a line or how would you try | |
27:05.440 --> 27:08.200 | |
to define human level intelligence | |
27:08.200 --> 27:10.680 | |
and superhuman level intelligence? | |
27:10.680 --> 27:13.280 | |
Where is consciousness part of that definition? | |
27:13.280 --> 27:16.640 | |
No, consciousness does not come into this definition. | |
27:16.640 --> 27:20.280 | |
So, so I think of intelligence as it's a spectrum | |
27:20.280 --> 27:21.960 | |
but there are very many different kinds of goals | |
27:21.960 --> 27:22.800 | |
you can have. | |
27:22.800 --> 27:24.000 | |
You can have a goal to be a good chess player | |
27:24.000 --> 27:28.520 | |
a good goal player, a good car driver, a good investor | |
27:28.520 --> 27:31.160 | |
good poet, et cetera. | |
27:31.160 --> 27:34.320 | |
So intelligence that by its very nature | |
27:34.320 --> 27:36.680 | |
isn't something you can measure by this one number | |
27:36.680 --> 27:37.960 | |
or some overall goodness. | |
27:37.960 --> 27:38.800 | |
No, no. | |
27:38.800 --> 27:40.320 | |
There are some people who are more better at this. | |
27:40.320 --> 27:42.360 | |
Some people are better than that. | |
27:42.360 --> 27:45.440 | |
Right now we have machines that are much better than us | |
27:45.440 --> 27:49.040 | |
at some very narrow tasks like multiplying large numbers | |
27:49.040 --> 27:53.200 | |
fast, memorizing large databases, playing chess | |
27:53.200 --> 27:56.280 | |
playing go and soon driving cars. | |
27:57.480 --> 28:00.080 | |
But there's still no machine that can match | |
28:00.080 --> 28:02.720 | |
a human child in general intelligence | |
28:02.720 --> 28:05.720 | |
but artificial general intelligence, AGI | |
28:05.720 --> 28:07.880 | |
the name of your course, of course | |
28:07.880 --> 28:12.880 | |
that is by its very definition, the quest | |
28:13.400 --> 28:16.000 | |
to build a machine that can do everything | |
28:16.000 --> 28:17.800 | |
as well as we can. | |
28:17.800 --> 28:21.960 | |
So the old Holy grail of AI from back to its inception | |
28:21.960 --> 28:25.560 | |
in the sixties, if that ever happens, of course | |
28:25.560 --> 28:27.320 | |
I think it's going to be the biggest transition | |
28:27.320 --> 28:29.040 | |
in the history of life on earth | |
28:29.040 --> 28:33.200 | |
but it doesn't necessarily have to wait the big impact | |
28:33.200 --> 28:35.400 | |
until machines are better than us at knitting | |
28:35.400 --> 28:39.160 | |
that the really big change doesn't come exactly | |
28:39.160 --> 28:41.800 | |
at the moment they're better than us at everything. | |
28:41.800 --> 28:44.120 | |
The really big change comes first | |
28:44.120 --> 28:45.840 | |
there are big changes when they start becoming better | |
28:45.840 --> 28:48.800 | |
at us at doing most of the jobs that we do | |
28:48.800 --> 28:51.160 | |
because that takes away much of the demand | |
28:51.160 --> 28:53.200 | |
for human labor. | |
28:53.200 --> 28:55.640 | |
And then the really whopping change comes | |
28:55.640 --> 29:00.640 | |
when they become better than us at AI research, right? | |
29:01.040 --> 29:03.760 | |
Because right now the timescale of AI research | |
29:03.760 --> 29:08.400 | |
is limited by the human research and development cycle | |
29:08.400 --> 29:10.160 | |
of years typically, you know | |
29:10.160 --> 29:13.480 | |
how long does it take from one release of some software | |
29:13.480 --> 29:15.720 | |
or iPhone or whatever to the next? | |
29:15.720 --> 29:20.720 | |
But once Google can replace 40,000 engineers | |
29:20.920 --> 29:25.920 | |
by 40,000 equivalent pieces of software or whatever | |
29:26.400 --> 29:29.680 | |
but then there's no reason that has to be years | |
29:29.680 --> 29:31.840 | |
it can be in principle much faster | |
29:31.840 --> 29:36.040 | |
and the timescale of future progress in AI | |
29:36.040 --> 29:39.320 | |
and all of science and technology will be driven | |
29:39.320 --> 29:40.960 | |
by machines, not humans. | |
29:40.960 --> 29:45.960 | |
So it's this simple point which gives right | |
29:46.520 --> 29:48.720 | |
this incredibly fun controversy | |
29:48.720 --> 29:51.880 | |
about whether there can be intelligence explosion | |
29:51.880 --> 29:54.400 | |
so called singularity as Werner Vinge called it. | |
29:54.400 --> 29:57.040 | |
Now the idea is articulated by I.J. Good | |
29:57.040 --> 29:59.480 | |
is obviously way back fifties | |
29:59.480 --> 30:01.040 | |
but you can see Alan Turing | |
30:01.040 --> 30:03.640 | |
and others thought about it even earlier. | |
30:06.920 --> 30:10.080 | |
So you asked me what exactly would I define | |
30:10.080 --> 30:12.800 | |
human level intelligence, yeah. | |
30:12.800 --> 30:15.680 | |
So the glib answer is to say something | |
30:15.680 --> 30:18.520 | |
which is better than us at all cognitive tasks | |
30:18.520 --> 30:21.800 | |
with a better than any human at all cognitive tasks | |
30:21.800 --> 30:23.080 | |
but the really interesting bar | |
30:23.080 --> 30:25.760 | |
I think goes a little bit lower than that actually. | |
30:25.760 --> 30:27.920 | |
It's when they can, when they're better than us | |
30:27.920 --> 30:31.760 | |
at AI programming and general learning | |
30:31.760 --> 30:35.360 | |
so that they can if they want to get better | |
30:35.360 --> 30:37.240 | |
than us at anything by just studying. | |
30:37.240 --> 30:40.560 | |
So they're better is a key word and better is towards | |
30:40.560 --> 30:44.120 | |
this kind of spectrum of the complexity of goals | |
30:44.120 --> 30:45.680 | |
it's able to accomplish. | |
30:45.680 --> 30:50.360 | |
So another way to, and that's certainly | |
30:50.360 --> 30:53.040 | |
a very clear definition of human love. | |
30:53.040 --> 30:55.240 | |
So there's, it's almost like a sea that's rising | |
30:55.240 --> 30:56.800 | |
you can do more and more and more things | |
30:56.800 --> 30:58.640 | |
it's a geographic that you show | |
30:58.640 --> 30:59.880 | |
it's really nice way to put it. | |
30:59.880 --> 31:01.560 | |
So there's some peaks that | |
31:01.560 --> 31:03.280 | |
and there's an ocean level elevating | |
31:03.280 --> 31:04.800 | |
and you solve more and more problems | |
31:04.800 --> 31:07.720 | |
but just kind of to take a pause | |
31:07.720 --> 31:09.000 | |
and we took a bunch of questions | |
31:09.000 --> 31:10.240 | |
and a lot of social networks | |
31:10.240 --> 31:11.720 | |
and a bunch of people asked | |
31:11.720 --> 31:14.480 | |
a sort of a slightly different direction | |
31:14.480 --> 31:19.480 | |
on creativity and things that perhaps aren't a peak. | |
31:23.560 --> 31:24.720 | |
Human beings are flawed | |
31:24.720 --> 31:28.720 | |
and perhaps better means having contradiction | |
31:28.720 --> 31:30.200 | |
being flawed in some way. | |
31:30.200 --> 31:34.960 | |
So let me sort of start easy, first of all. | |
31:34.960 --> 31:36.600 | |
So you have a lot of cool equations. | |
31:36.600 --> 31:39.760 | |
Let me ask, what's your favorite equation, first of all? | |
31:39.760 --> 31:42.760 | |
I know they're all like your children, but like | |
31:42.760 --> 31:43.680 | |
which one is that? | |
31:43.680 --> 31:45.560 | |
This is the shirt in your equation. | |
31:45.560 --> 31:48.640 | |
It's the master key of quantum mechanics | |
31:48.640 --> 31:49.880 | |
of the micro world. | |
31:49.880 --> 31:52.800 | |
So this equation will protect everything | |
31:52.800 --> 31:55.840 | |
to do with atoms, molecules and all the way up. | |
31:55.840 --> 31:58.560 | |
Right? | |
31:58.560 --> 31:59.760 | |
Yeah, so, okay. | |
31:59.760 --> 32:02.080 | |
So quantum mechanics is certainly a beautiful | |
32:02.080 --> 32:05.160 | |
mysterious formulation of our world. | |
32:05.160 --> 32:08.760 | |
So I'd like to sort of ask you, just as an example | |
32:08.760 --> 32:12.160 | |
it perhaps doesn't have the same beauty as physics does | |
32:12.160 --> 32:16.960 | |
but in mathematics abstract, the Andrew Wiles | |
32:16.960 --> 32:19.360 | |
who proved the Fermat's last theorem. | |
32:19.360 --> 32:22.040 | |
So he just saw this recently | |
32:22.040 --> 32:24.160 | |
and it kind of caught my eye a little bit. | |
32:24.160 --> 32:27.960 | |
This is 358 years after it was conjectured. | |
32:27.960 --> 32:29.960 | |
So this is very simple formulation. | |
32:29.960 --> 32:32.640 | |
Everybody tried to prove it, everybody failed. | |
32:32.640 --> 32:34.800 | |
And so here's this guy comes along | |
32:34.800 --> 32:38.640 | |
and eventually proves it and then fails to prove it | |
32:38.640 --> 32:41.320 | |
and then proves it again in 94. | |
32:41.320 --> 32:43.480 | |
And he said like the moment when everything connected | |
32:43.480 --> 32:46.040 | |
into place in an interview said | |
32:46.040 --> 32:47.880 | |
it was so indescribably beautiful. | |
32:47.880 --> 32:51.040 | |
That moment when you finally realize the connecting piece | |
32:51.040 --> 32:52.800 | |
of two conjectures. | |
32:52.800 --> 32:55.280 | |
He said, it was so indescribably beautiful. | |
32:55.280 --> 32:57.040 | |
It was so simple and so elegant. | |
32:57.040 --> 32:58.760 | |
I couldn't understand how I'd missed it. | |
32:58.760 --> 33:02.080 | |
And I just stared at it in disbelief for 20 minutes. | |
33:02.080 --> 33:05.240 | |
Then during the day, I walked around the department | |
33:05.240 --> 33:07.880 | |
and I keep coming back to my desk | |
33:07.880 --> 33:09.840 | |
looking to see if it was still there. | |
33:09.840 --> 33:10.680 | |
It was still there. | |
33:10.680 --> 33:11.760 | |
I couldn't contain myself. | |
33:11.760 --> 33:12.880 | |
I was so excited. | |
33:12.880 --> 33:15.880 | |
It was the most important moment on my working life. | |
33:15.880 --> 33:18.960 | |
Nothing I ever do again will mean as much. | |
33:18.960 --> 33:20.800 | |
So that particular moment. | |
33:20.800 --> 33:24.640 | |
And it kind of made me think of what would it take? | |
33:24.640 --> 33:27.960 | |
And I think we have all been there at small levels. | |
33:29.480 --> 33:32.240 | |
Maybe let me ask, have you had a moment like that | |
33:32.240 --> 33:34.880 | |
in your life where you just had an idea? | |
33:34.880 --> 33:37.040 | |
It's like, wow, yes. | |
33:40.000 --> 33:42.480 | |
I wouldn't mention myself in the same breath | |
33:42.480 --> 33:44.760 | |
as Andrew Wiles, but I've certainly had a number | |
33:44.760 --> 33:52.200 | |
of aha moments when I realized something very cool | |
33:52.200 --> 33:56.000 | |
about physics, which has completely made my head explode. | |
33:56.000 --> 33:58.320 | |
In fact, some of my favorite discoveries I made later, | |
33:58.320 --> 34:01.080 | |
I later realized that they had been discovered earlier | |
34:01.080 --> 34:03.240 | |
by someone who sometimes got quite famous for it. | |
34:03.240 --> 34:05.480 | |
So it's too late for me to even publish it, | |
34:05.480 --> 34:07.440 | |
but that doesn't diminish in any way. | |
34:07.440 --> 34:09.760 | |
The emotional experience you have when you realize it, | |
34:09.760 --> 34:11.320 | |
like, wow. | |
34:11.320 --> 34:15.520 | |
Yeah, so what would it take in that moment, that wow, | |
34:15.520 --> 34:17.320 | |
that was yours in that moment? | |
34:17.320 --> 34:21.440 | |
So what do you think it takes for an intelligence system, | |
34:21.440 --> 34:24.520 | |
an AGI system, an AI system to have a moment like that? | |
34:25.640 --> 34:26.760 | |
That's a tricky question | |
34:26.760 --> 34:29.200 | |
because there are actually two parts to it, right? | |
34:29.200 --> 34:33.920 | |
One of them is, can it accomplish that proof? | |
34:33.920 --> 34:37.640 | |
Can it prove that you can never write A to the N | |
34:37.640 --> 34:42.760 | |
plus B to the N equals three to that equal Z to the N | |
34:42.760 --> 34:45.320 | |
for all integers, et cetera, et cetera, | |
34:45.320 --> 34:48.720 | |
when N is bigger than two? | |
34:48.720 --> 34:51.360 | |
That's simply a question about intelligence. | |
34:51.360 --> 34:54.120 | |
Can you build machines that are that intelligent? | |
34:54.120 --> 34:57.280 | |
And I think by the time we get a machine | |
34:57.280 --> 35:00.840 | |
that can independently come up with that level of proofs, | |
35:00.840 --> 35:03.360 | |
probably quite close to AGI. | |
35:03.360 --> 35:07.240 | |
The second question is a question about consciousness. | |
35:07.240 --> 35:11.760 | |
When will we, how likely is it that such a machine | |
35:11.760 --> 35:14.240 | |
will actually have any experience at all, | |
35:14.240 --> 35:16.160 | |
as opposed to just being like a zombie? | |
35:16.160 --> 35:20.560 | |
And would we expect it to have some sort of emotional response | |
35:20.560 --> 35:24.640 | |
to this or anything at all akin to human emotion | |
35:24.640 --> 35:28.320 | |
where when it accomplishes its machine goal, | |
35:28.320 --> 35:31.920 | |
it views it as somehow something very positive | |
35:31.920 --> 35:39.160 | |
and sublime and deeply meaningful? | |
35:39.160 --> 35:41.440 | |
I would certainly hope that if in the future | |
35:41.440 --> 35:45.120 | |
we do create machines that are our peers | |
35:45.120 --> 35:50.160 | |
or even our descendants, that I would certainly | |
35:50.160 --> 35:55.480 | |
hope that they do have this sublime appreciation of life. | |
35:55.480 --> 35:58.840 | |
In a way, my absolutely worst nightmare | |
35:58.840 --> 36:05.760 | |
would be that at some point in the future, | |
36:05.760 --> 36:07.400 | |
the distant future, maybe our cosmos | |
36:07.400 --> 36:10.600 | |
is teeming with all this post biological life doing | |
36:10.600 --> 36:12.880 | |
all the seemingly cool stuff. | |
36:12.880 --> 36:16.480 | |
And maybe the last humans, by the time | |
36:16.480 --> 36:20.120 | |
our species eventually fizzles out, | |
36:20.120 --> 36:21.920 | |
will be like, well, that's OK because we're | |
36:21.920 --> 36:23.600 | |
so proud of our descendants here. | |
36:23.600 --> 36:26.680 | |
And look what all the, my worst nightmare | |
36:26.680 --> 36:30.360 | |
is that we haven't solved the consciousness problem. | |
36:30.360 --> 36:32.880 | |
And we haven't realized that these are all the zombies. | |
36:32.880 --> 36:36.200 | |
They're not aware of anything any more than a tape recorder | |
36:36.200 --> 36:37.840 | |
has any kind of experience. | |
36:37.840 --> 36:40.040 | |
So the whole thing has just become | |
36:40.040 --> 36:41.520 | |
a play for empty benches. | |
36:41.520 --> 36:44.640 | |
That would be the ultimate zombie apocalypse. | |
36:44.640 --> 36:47.200 | |
So I would much rather, in that case, | |
36:47.200 --> 36:52.240 | |
that we have these beings which can really | |
36:52.240 --> 36:57.000 | |
appreciate how amazing it is. | |
36:57.000 --> 37:01.080 | |
And in that picture, what would be the role of creativity? | |
37:01.080 --> 37:04.960 | |
A few people ask about creativity. | |
37:04.960 --> 37:07.080 | |
When you think about intelligence, | |
37:07.080 --> 37:09.840 | |
certainly the story you told at the beginning of your book | |
37:09.840 --> 37:15.200 | |
involved creating movies and so on, making money. | |
37:15.200 --> 37:17.240 | |
You can make a lot of money in our modern world | |
37:17.240 --> 37:18.600 | |
with music and movies. | |
37:18.600 --> 37:20.880 | |
So if you are an intelligent system, | |
37:20.880 --> 37:22.960 | |
you may want to get good at that. | |
37:22.960 --> 37:26.280 | |
But that's not necessarily what I mean by creativity. | |
37:26.280 --> 37:29.640 | |
Is it important on that complex goals | |
37:29.640 --> 37:31.600 | |
where the sea is rising for there | |
37:31.600 --> 37:33.800 | |
to be something creative? | |
37:33.800 --> 37:37.400 | |
Or am I being very human centric and thinking creativity | |
37:37.400 --> 37:41.880 | |
somehow special relative to intelligence? | |
37:41.880 --> 37:47.240 | |
My hunch is that we should think of creativity simply | |
37:47.240 --> 37:50.760 | |
as an aspect of intelligence. | |
37:50.760 --> 37:57.840 | |
And we have to be very careful with human vanity. | |
37:57.840 --> 37:59.520 | |
We have this tendency to very often want | |
37:59.520 --> 38:01.560 | |
to say, as soon as machines can do something, | |
38:01.560 --> 38:03.560 | |
we try to diminish it and say, oh, but that's | |
38:03.560 --> 38:05.920 | |
not real intelligence. | |
38:05.920 --> 38:08.400 | |
Isn't it creative or this or that? | |
38:08.400 --> 38:12.200 | |
The other thing, if we ask ourselves | |
38:12.200 --> 38:14.320 | |
to write down a definition of what we actually mean | |
38:14.320 --> 38:18.840 | |
by being creative, what we mean by Andrew Wiles, what he did | |
38:18.840 --> 38:21.880 | |
there, for example, don't we often mean that someone takes | |
38:21.880 --> 38:26.000 | |
a very unexpected leap? | |
38:26.000 --> 38:29.680 | |
It's not like taking 573 and multiplying it | |
38:29.680 --> 38:33.840 | |
by 224 by just a step of straightforward cookbook | |
38:33.840 --> 38:36.520 | |
like rules, right? | |
38:36.520 --> 38:39.680 | |
You can maybe make a connection between two things | |
38:39.680 --> 38:42.640 | |
that people had never thought was connected or something | |
38:42.640 --> 38:44.480 | |
like that. | |
38:44.480 --> 38:47.720 | |
I think this is an aspect of intelligence. | |
38:47.720 --> 38:53.000 | |
And this is actually one of the most important aspects of it. | |
38:53.000 --> 38:55.520 | |
Maybe the reason we humans tend to be better at it | |
38:55.520 --> 38:57.840 | |
than traditional computers is because it's | |
38:57.840 --> 38:59.640 | |
something that comes more naturally if you're | |
38:59.640 --> 39:04.120 | |
a neural network than if you're a traditional logic gate | |
39:04.120 --> 39:05.720 | |
based computer machine. | |
39:05.720 --> 39:08.640 | |
We physically have all these connections. | |
39:08.640 --> 39:13.800 | |
And you activate here, activate here, activate here. | |
39:13.800 --> 39:16.560 | |
Bing. | |
39:16.560 --> 39:21.040 | |
My hunch is that if we ever build a machine where you could | |
39:21.040 --> 39:29.200 | |
just give it the task, hey, you say, hey, I just realized | |
39:29.200 --> 39:32.320 | |
I want to travel around the world instead this month. | |
39:32.320 --> 39:34.600 | |
Can you teach my AGI course for me? | |
39:34.600 --> 39:35.960 | |
And it's like, OK, I'll do it. | |
39:35.960 --> 39:37.920 | |
And it does everything that you would have done | |
39:37.920 --> 39:39.760 | |
and improvises and stuff. | |
39:39.760 --> 39:43.360 | |
That would, in my mind, involve a lot of creativity. | |
39:43.360 --> 39:45.680 | |
Yeah, so it's actually a beautiful way to put it. | |
39:45.680 --> 39:52.640 | |
I think we do try to grasp at the definition of intelligence | |
39:52.640 --> 39:56.360 | |
is everything we don't understand how to build. | |
39:56.360 --> 39:59.360 | |
So we as humans try to find things | |
39:59.360 --> 40:01.240 | |
that we have and machines don't have. | |
40:01.240 --> 40:03.800 | |
And maybe creativity is just one of the things, one | |
40:03.800 --> 40:05.480 | |
of the words we use to describe that. | |
40:05.480 --> 40:07.200 | |
That's a really interesting way to put it. | |
40:07.200 --> 40:09.520 | |
I don't think we need to be that defensive. | |
40:09.520 --> 40:11.560 | |
I don't think anything good comes out of saying, | |
40:11.560 --> 40:18.080 | |
well, we're somehow special, you know? | |
40:18.080 --> 40:21.040 | |
Contrary wise, there are many examples in history | |
40:21.040 --> 40:27.840 | |
of where trying to pretend that we're somehow superior | |
40:27.840 --> 40:33.120 | |
to all other intelligent beings has led to pretty bad results, | |
40:33.120 --> 40:35.960 | |
right? | |
40:35.960 --> 40:38.440 | |
Nazi Germany, they said that they were somehow superior | |
40:38.440 --> 40:40.080 | |
to other people. | |
40:40.080 --> 40:42.440 | |
Today, we still do a lot of cruelty to animals | |
40:42.440 --> 40:44.440 | |
by saying that we're so superior somehow, | |
40:44.440 --> 40:46.440 | |
and they can't feel pain. | |
40:46.440 --> 40:48.480 | |
Slavery was justified by the same kind | |
40:48.480 --> 40:52.200 | |
of just really weak arguments. | |
40:52.200 --> 40:57.120 | |
And I don't think if we actually go ahead and build | |
40:57.120 --> 40:59.440 | |
artificial general intelligence, it | |
40:59.440 --> 41:01.360 | |
can do things better than us, I don't | |
41:01.360 --> 41:04.080 | |
think we should try to found our self worth on some sort | |
41:04.080 --> 41:09.760 | |
of bogus claims of superiority in terms | |
41:09.760 --> 41:12.120 | |
of our intelligence. | |
41:12.120 --> 41:18.080 | |
I think we should instead find our calling | |
41:18.080 --> 41:23.360 | |
and the meaning of life from the experiences that we have. | |
41:23.360 --> 41:28.720 | |
I can have very meaningful experiences | |
41:28.720 --> 41:32.920 | |
even if there are other people who are smarter than me. | |
41:32.920 --> 41:34.400 | |
When I go to a faculty meeting here, | |
41:34.400 --> 41:36.520 | |
and we talk about something, and then I certainly realize, | |
41:36.520 --> 41:39.080 | |
oh, boy, he has an old prize, he has an old prize, | |
41:39.080 --> 41:40.800 | |
he has an old prize, I don't have one. | |
41:40.800 --> 41:43.760 | |
Does that make me enjoy life any less | |
41:43.760 --> 41:47.560 | |
or enjoy talking to those people less? | |
41:47.560 --> 41:49.560 | |
Of course not. | |
41:49.560 --> 41:54.160 | |
And the contrary, I feel very honored and privileged | |
41:54.160 --> 41:58.760 | |
to get to interact with other very intelligent beings that | |
41:58.760 --> 42:00.680 | |
are better than me at a lot of stuff. | |
42:00.680 --> 42:02.840 | |
So I don't think there's any reason why | |
42:02.840 --> 42:06.080 | |
we can't have the same approach with intelligent machines. | |
42:06.080 --> 42:07.320 | |
That's a really interesting. | |
42:07.320 --> 42:08.920 | |
So people don't often think about that. | |
42:08.920 --> 42:10.600 | |
They think about when there's going, | |
42:10.600 --> 42:13.320 | |
if there's machines that are more intelligent, | |
42:13.320 --> 42:15.080 | |
you naturally think that that's not | |
42:15.080 --> 42:19.080 | |
going to be a beneficial type of intelligence. | |
42:19.080 --> 42:23.000 | |
You don't realize it could be like peers with Nobel prizes | |
42:23.000 --> 42:25.120 | |
that would be just fun to talk with, | |
42:25.120 --> 42:27.560 | |
and they might be clever about certain topics, | |
42:27.560 --> 42:32.240 | |
and you can have fun having a few drinks with them. | |
42:32.240 --> 42:35.880 | |
Well, also, another example we can all | |
42:35.880 --> 42:39.320 | |
relate to of why it doesn't have to be a terrible thing | |
42:39.320 --> 42:42.560 | |
to be in the presence of people who are even smarter than us | |
42:42.560 --> 42:45.600 | |
all around is when you and I were both two years old, | |
42:45.600 --> 42:48.360 | |
I mean, our parents were much more intelligent than us, | |
42:48.360 --> 42:49.040 | |
right? | |
42:49.040 --> 42:51.960 | |
Worked out OK, because their goals | |
42:51.960 --> 42:53.960 | |
were aligned with our goals. | |
42:53.960 --> 42:58.680 | |
And that, I think, is really the number one key issue | |
42:58.680 --> 43:02.280 | |
we have to solve if we value align the value alignment | |
43:02.280 --> 43:03.080 | |
problem, exactly. | |
43:03.080 --> 43:06.520 | |
Because people who see too many Hollywood movies | |
43:06.520 --> 43:10.000 | |
with lousy science fiction plot lines, | |
43:10.000 --> 43:12.200 | |
they worry about the wrong thing, right? | |
43:12.200 --> 43:16.320 | |
They worry about some machine suddenly turning evil. | |
43:16.320 --> 43:21.480 | |
It's not malice that is the concern. | |
43:21.480 --> 43:22.880 | |
It's competence. | |
43:22.880 --> 43:27.440 | |
By definition, intelligent makes you very competent. | |
43:27.440 --> 43:31.920 | |
If you have a more intelligent goal playing, | |
43:31.920 --> 43:33.680 | |
computer playing is a less intelligent one. | |
43:33.680 --> 43:36.120 | |
And when we define intelligence as the ability | |
43:36.120 --> 43:38.600 | |
to accomplish goal winning, it's going | |
43:38.600 --> 43:40.560 | |
to be the more intelligent one that wins. | |
43:40.560 --> 43:43.560 | |
And if you have a human and then you | |
43:43.560 --> 43:47.720 | |
have an AGI that's more intelligent in all ways | |
43:47.720 --> 43:49.520 | |
and they have different goals, guess who's | |
43:49.520 --> 43:50.720 | |
going to get their way, right? | |
43:50.720 --> 43:57.120 | |
So I was just reading about this particular rhinoceros species | |
43:57.120 --> 43:59.200 | |
that was driven extinct just a few years ago. | |
43:59.200 --> 44:02.280 | |
Ellen Bummer is looking at this cute picture of a mommy | |
44:02.280 --> 44:05.080 | |
rhinoceros with its child. | |
44:05.080 --> 44:09.320 | |
And why did we humans drive it to extinction? | |
44:09.320 --> 44:12.800 | |
It wasn't because we were evil rhino haters as a whole. | |
44:12.800 --> 44:14.920 | |
It was just because our goals weren't aligned | |
44:14.920 --> 44:16.000 | |
with those of the rhinoceros. | |
44:16.000 --> 44:17.680 | |
And it didn't work out so well for the rhinoceros | |
44:17.680 --> 44:19.560 | |
because we were more intelligent, right? | |
44:19.560 --> 44:21.240 | |
So I think it's just so important | |
44:21.240 --> 44:27.120 | |
that if we ever do build AGI, before we unleash anything, | |
44:27.120 --> 44:31.840 | |
we have to make sure that it learns | |
44:31.840 --> 44:36.000 | |
to understand our goals, that it adopts our goals, | |
44:36.000 --> 44:37.920 | |
and that it retains those goals. | |
44:37.920 --> 44:40.520 | |
So the cool, interesting problem there | |
44:40.520 --> 44:47.040 | |
is us as human beings trying to formulate our values. | |
44:47.040 --> 44:51.360 | |
So you could think of the United States Constitution as a way | |
44:51.360 --> 44:56.680 | |
that people sat down, at the time a bunch of white men, | |
44:56.680 --> 44:59.680 | |
which is a good example, I should say. | |
44:59.680 --> 45:01.480 | |
They formulated the goals for this country. | |
45:01.480 --> 45:03.760 | |
And a lot of people agree that those goals actually | |
45:03.760 --> 45:05.360 | |
held up pretty well. | |
45:05.360 --> 45:07.160 | |
That's an interesting formulation of values | |
45:07.160 --> 45:09.440 | |
and failed miserably in other ways. | |
45:09.440 --> 45:13.320 | |
So for the value alignment problem and the solution to it, | |
45:13.320 --> 45:19.560 | |
we have to be able to put on paper or in a program | |
45:19.560 --> 45:20.400 | |
human values. | |
45:20.400 --> 45:22.400 | |
How difficult do you think that is? | |
45:22.400 --> 45:24.040 | |
Very. | |
45:24.040 --> 45:25.880 | |
But it's so important. | |
45:25.880 --> 45:28.000 | |
We really have to give it our best. | |
45:28.000 --> 45:30.120 | |
And it's difficult for two separate reasons. | |
45:30.120 --> 45:33.440 | |
There's the technical value alignment problem | |
45:33.440 --> 45:39.120 | |
of figuring out just how to make machines understand our goals, | |
45:39.120 --> 45:40.440 | |
adopt them, and retain them. | |
45:40.440 --> 45:43.200 | |
And then there's the separate part of it, | |
45:43.200 --> 45:44.200 | |
the philosophical part. | |
45:44.200 --> 45:45.920 | |
Whose values anyway? | |
45:45.920 --> 45:48.320 | |
And since it's not like we have any great consensus | |
45:48.320 --> 45:52.040 | |
on this planet on values, what mechanism should we | |
45:52.040 --> 45:54.120 | |
create then to aggregate and decide, OK, | |
45:54.120 --> 45:56.520 | |
what's a good compromise? | |
45:56.520 --> 45:58.440 | |
That second discussion can't just | |
45:58.440 --> 46:01.560 | |
be left to tech nerds like myself. | |
46:01.560 --> 46:05.720 | |
And if we refuse to talk about it and then AGI gets built, | |
46:05.720 --> 46:07.160 | |
who's going to be actually making | |
46:07.160 --> 46:08.480 | |
the decision about whose values? | |
46:08.480 --> 46:12.080 | |
It's going to be a bunch of dudes in some tech company. | |
46:12.080 --> 46:17.240 | |
And are they necessarily so representative of all | |
46:17.240 --> 46:19.400 | |
of humankind that we want to just entrust it to them? | |
46:19.400 --> 46:23.000 | |
Are they even uniquely qualified to speak | |
46:23.000 --> 46:25.240 | |
to future human happiness just because they're | |
46:25.240 --> 46:26.480 | |
good at programming AI? | |
46:26.480 --> 46:30.200 | |
I'd much rather have this be a really inclusive conversation. | |
46:30.200 --> 46:32.560 | |
But do you think it's possible? | |
46:32.560 --> 46:37.560 | |
So you create a beautiful vision that includes the diversity, | |
46:37.560 --> 46:40.960 | |
cultural diversity, and various perspectives on discussing | |
46:40.960 --> 46:43.600 | |
rights, freedoms, human dignity. | |
46:43.600 --> 46:46.520 | |
But how hard is it to come to that consensus? | |
46:46.520 --> 46:50.400 | |
Do you think it's certainly a really important thing | |
46:50.400 --> 46:51.880 | |
that we should all try to do? | |
46:51.880 --> 46:54.240 | |
But do you think it's feasible? | |
46:54.240 --> 47:00.160 | |
I think there's no better way to guarantee failure than to | |
47:00.160 --> 47:02.840 | |
refuse to talk about it or refuse to try. | |
47:02.840 --> 47:05.320 | |
And I also think it's a really bad strategy | |
47:05.320 --> 47:08.560 | |
to say, OK, let's first have a discussion for a long time. | |
47:08.560 --> 47:11.040 | |
And then once we reach complete consensus, | |
47:11.040 --> 47:13.360 | |
then we'll try to load it into some machine. | |
47:13.360 --> 47:16.560 | |
No, we shouldn't let perfect be the enemy of good. | |
47:16.560 --> 47:20.600 | |
Instead, we should start with the kindergarten ethics | |
47:20.600 --> 47:22.120 | |
that pretty much everybody agrees on | |
47:22.120 --> 47:24.360 | |
and put that into machines now. | |
47:24.360 --> 47:25.880 | |
We're not doing that even. | |
47:25.880 --> 47:31.000 | |
Look at anyone who builds this passenger aircraft, | |
47:31.000 --> 47:33.000 | |
wants it to never under any circumstances | |
47:33.000 --> 47:35.600 | |
fly into a building or a mountain. | |
47:35.600 --> 47:38.480 | |
Yet the September 11 hijackers were able to do that. | |
47:38.480 --> 47:41.800 | |
And even more embarrassingly, Andreas Lubitz, | |
47:41.800 --> 47:43.960 | |
this depressed Germanwings pilot, | |
47:43.960 --> 47:47.360 | |
when he flew his passenger jet into the Alps killing over 100 | |
47:47.360 --> 47:50.640 | |
people, he just told the autopilot to do it. | |
47:50.640 --> 47:53.200 | |
He told the freaking computer to change the altitude | |
47:53.200 --> 47:55.040 | |
to 100 meters. | |
47:55.040 --> 47:58.160 | |
And even though it had the GPS maps, everything, | |
47:58.160 --> 48:00.640 | |
the computer was like, OK. | |
48:00.640 --> 48:05.320 | |
So we should take those very basic values, | |
48:05.320 --> 48:08.400 | |
where the problem is not that we don't agree. | |
48:08.400 --> 48:10.120 | |
The problem is just we've been too lazy | |
48:10.120 --> 48:11.480 | |
to try to put it into our machines | |
48:11.480 --> 48:15.520 | |
and make sure that from now on, airplanes will just, | |
48:15.520 --> 48:16.920 | |
which all have computers in them, | |
48:16.920 --> 48:19.720 | |
but will just refuse to do something like that. | |
48:19.720 --> 48:22.160 | |
Go into safe mode, maybe lock the cockpit door, | |
48:22.160 --> 48:24.480 | |
go over to the nearest airport. | |
48:24.480 --> 48:28.080 | |
And there's so much other technology in our world | |
48:28.080 --> 48:31.320 | |
as well now, where it's really becoming quite timely | |
48:31.320 --> 48:34.120 | |
to put in some sort of very basic values like this. | |
48:34.120 --> 48:39.240 | |
Even in cars, we've had enough vehicle terrorism attacks | |
48:39.240 --> 48:42.040 | |
by now, where people have driven trucks and vans | |
48:42.040 --> 48:45.480 | |
into pedestrians, that it's not at all a crazy idea | |
48:45.480 --> 48:48.680 | |
to just have that hardwired into the car. | |
48:48.680 --> 48:50.280 | |
Because yeah, there are a lot of, | |
48:50.280 --> 48:52.240 | |
there's always going to be people who for some reason | |
48:52.240 --> 48:54.800 | |
want to harm others, but most of those people | |
48:54.800 --> 48:56.760 | |
don't have the technical expertise to figure out | |
48:56.760 --> 48:58.520 | |
how to work around something like that. | |
48:58.520 --> 49:01.760 | |
So if the car just won't do it, it helps. | |
49:01.760 --> 49:02.840 | |
So let's start there. | |
49:02.840 --> 49:04.960 | |
So there's a lot of, that's a great point. | |
49:04.960 --> 49:06.800 | |
So not chasing perfect. | |
49:06.800 --> 49:10.840 | |
There's a lot of things that most of the world agrees on. | |
49:10.840 --> 49:11.840 | |
Yeah, let's start there. | |
49:11.840 --> 49:12.680 | |
Let's start there. | |
49:12.680 --> 49:14.560 | |
And then once we start there, | |
49:14.560 --> 49:17.240 | |
we'll also get into the habit of having | |
49:17.240 --> 49:18.520 | |
these kind of conversations about, okay, | |
49:18.520 --> 49:21.760 | |
what else should we put in here and have these discussions? | |
49:21.760 --> 49:23.920 | |
This should be a gradual process then. | |
49:23.920 --> 49:28.600 | |
Great, so, but that also means describing these things | |
49:28.600 --> 49:31.240 | |
and describing it to a machine. | |
49:31.240 --> 49:34.200 | |
So one thing, we had a few conversations | |
49:34.200 --> 49:35.640 | |
with Stephen Wolfram. | |
49:35.640 --> 49:37.080 | |
I'm not sure if you're familiar with Stephen. | |
49:37.080 --> 49:38.360 | |
Oh yeah, I know him quite well. | |
49:38.360 --> 49:42.040 | |
So he is, he works with a bunch of things, | |
49:42.040 --> 49:46.560 | |
but cellular automata, these simple computable things, | |
49:46.560 --> 49:47.960 | |
these computation systems. | |
49:47.960 --> 49:49.880 | |
And he kind of mentioned that, | |
49:49.880 --> 49:52.480 | |
we probably have already within these systems | |
49:52.480 --> 49:54.680 | |
already something that's AGI, | |
49:56.120 --> 49:58.720 | |
meaning like we just don't know it | |
49:58.720 --> 50:00.400 | |
because we can't talk to it. | |
50:00.400 --> 50:04.800 | |
So if you give me this chance to try to at least | |
50:04.800 --> 50:06.720 | |
form a question out of this is, | |
50:07.600 --> 50:10.880 | |
I think it's an interesting idea to think | |
50:10.880 --> 50:12.680 | |
that we can have intelligent systems, | |
50:12.680 --> 50:15.600 | |
but we don't know how to describe something to them | |
50:15.600 --> 50:17.360 | |
and they can't communicate with us. | |
50:17.360 --> 50:19.840 | |
I know you're doing a little bit of work in explainable AI, | |
50:19.840 --> 50:22.040 | |
trying to get AI to explain itself. | |
50:22.040 --> 50:25.520 | |
So what are your thoughts of natural language processing | |
50:25.520 --> 50:27.640 | |
or some kind of other communication? | |
50:27.640 --> 50:30.120 | |
How does the AI explain something to us? | |
50:30.120 --> 50:33.640 | |
How do we explain something to it, to machines? | |
50:33.640 --> 50:35.320 | |
Or you think of it differently? | |
50:35.320 --> 50:39.960 | |
So there are two separate parts to your question there. | |
50:39.960 --> 50:42.440 | |
One of them has to do with communication, | |
50:42.440 --> 50:44.440 | |
which is super interesting, I'll get to that in a sec. | |
50:44.440 --> 50:47.280 | |
The other is whether we already have AGI | |
50:47.280 --> 50:49.240 | |
but we just haven't noticed it there. | |
50:49.240 --> 50:50.080 | |
Right. | |
50:51.800 --> 50:53.000 | |
There I beg to differ. | |
50:54.280 --> 50:56.480 | |
I don't think there's anything in any cellular automaton | |
50:56.480 --> 50:59.040 | |
or anything or the internet itself or whatever | |
50:59.040 --> 51:03.560 | |
that has artificial general intelligence | |
51:03.560 --> 51:05.520 | |
and that it can really do exactly everything | |
51:05.520 --> 51:07.000 | |
we humans can do better. | |
51:07.000 --> 51:11.600 | |
I think the day that happens, when that happens, | |
51:11.600 --> 51:15.600 | |
we will very soon notice, we'll probably notice even before | |
51:15.600 --> 51:17.440 | |
because in a very, very big way. | |
51:17.440 --> 51:18.840 | |
But for the second part, though. | |
51:18.840 --> 51:20.720 | |
Wait, can I ask, sorry. | |
51:20.720 --> 51:24.400 | |
So, because you have this beautiful way | |
51:24.400 --> 51:29.400 | |
to formulating consciousness as information processing, | |
51:30.360 --> 51:31.360 | |
and you can think of intelligence | |
51:31.360 --> 51:32.280 | |
as information processing, | |
51:32.280 --> 51:34.320 | |
and you can think of the entire universe | |
51:34.320 --> 51:38.720 | |
as these particles and these systems roaming around | |
51:38.720 --> 51:41.360 | |
that have this information processing power. | |
51:41.360 --> 51:44.840 | |
You don't think there is something with the power | |
51:44.840 --> 51:49.040 | |
to process information in the way that we human beings do | |
51:49.040 --> 51:54.040 | |
that's out there that needs to be sort of connected to. | |
51:55.400 --> 51:57.880 | |
It seems a little bit philosophical, perhaps, | |
51:57.880 --> 52:00.080 | |
but there's something compelling to the idea | |
52:00.080 --> 52:01.920 | |
that the power is already there, | |
52:01.920 --> 52:05.440 | |
which the focus should be more on being able | |
52:05.440 --> 52:07.360 | |
to communicate with it. | |
52:07.360 --> 52:11.960 | |
Well, I agree that in a certain sense, | |
52:11.960 --> 52:15.360 | |
the hardware processing power is already out there | |
52:15.360 --> 52:19.000 | |
because our universe itself can think of it | |
52:19.000 --> 52:21.000 | |
as being a computer already, right? | |
52:21.000 --> 52:23.800 | |
It's constantly computing what water waves, | |
52:23.800 --> 52:26.120 | |
how it devolved the water waves in the River Charles | |
52:26.120 --> 52:28.440 | |
and how to move the air molecules around. | |
52:28.440 --> 52:30.480 | |
Seth Lloyd has pointed out, my colleague here, | |
52:30.480 --> 52:32.920 | |
that you can even in a very rigorous way | |
52:32.920 --> 52:35.480 | |
think of our entire universe as being a quantum computer. | |
52:35.480 --> 52:37.680 | |
It's pretty clear that our universe | |
52:37.680 --> 52:40.320 | |
supports this amazing processing power | |
52:40.320 --> 52:42.160 | |
because you can even, | |
52:42.160 --> 52:44.920 | |
within this physics computer that we live in, right? | |
52:44.920 --> 52:47.040 | |
We can even build actual laptops and stuff, | |
52:47.040 --> 52:49.000 | |
so clearly the power is there. | |
52:49.000 --> 52:52.040 | |
It's just that most of the compute power that nature has, | |
52:52.040 --> 52:54.240 | |
it's, in my opinion, kind of wasting on boring stuff | |
52:54.240 --> 52:56.520 | |
like simulating yet another ocean wave somewhere | |
52:56.520 --> 52:58.040 | |
where no one is even looking, right? | |
52:58.040 --> 53:00.880 | |
So in a sense, what life does, what we are doing | |
53:00.880 --> 53:03.880 | |
when we build computers is we're rechanneling | |
53:03.880 --> 53:07.200 | |
all this compute that nature is doing anyway | |
53:07.200 --> 53:09.360 | |
into doing things that are more interesting | |
53:09.360 --> 53:11.440 | |
than just yet another ocean wave, | |
53:11.440 --> 53:13.200 | |
and let's do something cool here. | |
53:14.080 --> 53:17.080 | |
So the raw hardware power is there, for sure, | |
53:17.080 --> 53:21.080 | |
but then even just computing what's going to happen | |
53:21.080 --> 53:23.520 | |
for the next five seconds in this water bottle, | |
53:23.520 --> 53:26.000 | |
takes a ridiculous amount of compute | |
53:26.000 --> 53:27.920 | |
if you do it on a human computer. | |
53:27.920 --> 53:29.920 | |
This water bottle just did it. | |
53:29.920 --> 53:33.440 | |
But that does not mean that this water bottle has AGI | |
53:34.760 --> 53:37.040 | |
because AGI means it should also be able to, | |
53:37.040 --> 53:40.160 | |
like I've written my book, done this interview. | |
53:40.160 --> 53:42.080 | |
And I don't think it's just communication problems. | |
53:42.080 --> 53:46.760 | |
I don't really think it can do it. | |
53:46.760 --> 53:49.280 | |
Although Buddhists say when they watch the water | |
53:49.280 --> 53:51.240 | |
and that there is some beauty, | |
53:51.240 --> 53:53.720 | |
that there's some depth and beauty in nature | |
53:53.720 --> 53:54.840 | |
that they can communicate with. | |
53:54.840 --> 53:56.480 | |
Communication is also very important though | |
53:56.480 --> 54:01.200 | |
because I mean, look, part of my job is being a teacher. | |
54:01.200 --> 54:06.200 | |
And I know some very intelligent professors even | |
54:06.200 --> 54:09.800 | |
who just have a bit of hard time communicating. | |
54:09.800 --> 54:12.640 | |
They come up with all these brilliant ideas, | |
54:12.640 --> 54:14.520 | |
but to communicate with somebody else, | |
54:14.520 --> 54:16.920 | |
you have to also be able to simulate their own mind. | |
54:16.920 --> 54:18.360 | |
Yes, empathy. | |
54:18.360 --> 54:20.640 | |
Build well enough and understand model of their mind | |
54:20.640 --> 54:24.400 | |
that you can say things that they will understand. | |
54:24.400 --> 54:26.480 | |
And that's quite difficult. | |
54:26.480 --> 54:28.280 | |
And that's why today it's so frustrating | |
54:28.280 --> 54:32.600 | |
if you have a computer that makes some cancer diagnosis | |
54:32.600 --> 54:34.120 | |
and you ask it, well, why are you saying | |
54:34.120 --> 54:36.120 | |
I should have this surgery? | |
54:36.120 --> 54:37.960 | |
And if it can only reply, | |
54:37.960 --> 54:40.800 | |
I was trained on five terabytes of data | |
54:40.800 --> 54:45.080 | |
and this is my diagnosis, boop, boop, beep, beep. | |
54:45.080 --> 54:49.120 | |
It doesn't really instill a lot of confidence, right? | |
54:49.120 --> 54:51.120 | |
So I think we have a lot of work to do | |
54:51.120 --> 54:54.320 | |
on communication there. | |
54:54.320 --> 54:58.040 | |
So what kind of, I think you're doing a little bit of work | |
54:58.040 --> 54:59.320 | |
in explainable AI. | |
54:59.320 --> 55:01.320 | |
What do you think are the most promising avenues? | |
55:01.320 --> 55:05.240 | |
Is it mostly about sort of the Alexa problem | |
55:05.240 --> 55:07.200 | |
of natural language processing of being able | |
55:07.200 --> 55:11.600 | |
to actually use human interpretable methods | |
55:11.600 --> 55:13.160 | |
of communication? | |
55:13.160 --> 55:16.000 | |
So being able to talk to a system and it talk back to you, | |
55:16.000 --> 55:18.640 | |
or is there some more fundamental problems to be solved? | |
55:18.640 --> 55:21.160 | |
I think it's all of the above. | |
55:21.160 --> 55:23.520 | |
The natural language processing is obviously important, | |
55:23.520 --> 55:27.600 | |
but there are also more nerdy fundamental problems. | |
55:27.600 --> 55:31.640 | |
Like if you take, you play chess? | |
55:31.640 --> 55:33.040 | |
Of course, I'm Russian. | |
55:33.040 --> 55:33.880 | |
I have to. | |
55:33.880 --> 55:34.720 | |
You speak Russian? | |
55:34.720 --> 55:35.560 | |
Yes, I speak Russian. | |
55:35.560 --> 55:38.040 | |
Excellent, I didn't know. | |
55:38.040 --> 55:39.160 | |
When did you learn Russian? | |
55:39.160 --> 55:41.800 | |
I speak very bad Russian, I'm only an autodidact, | |
55:41.800 --> 55:44.560 | |
but I bought a book, Teach Yourself Russian, | |
55:44.560 --> 55:47.720 | |
read a lot, but it was very difficult. | |
55:47.720 --> 55:48.560 | |
Wow. | |
55:48.560 --> 55:49.960 | |
That's why I speak so bad. | |
55:49.960 --> 55:51.960 | |
How many languages do you know? | |
55:51.960 --> 55:53.840 | |
Wow, that's really impressive. | |
55:53.840 --> 55:56.320 | |
I don't know, my wife has some calculation, | |
55:56.320 --> 55:58.400 | |
but my point was, if you play chess, | |
55:58.400 --> 56:01.040 | |
have you looked at the AlphaZero games? | |
56:01.040 --> 56:02.600 | |
The actual games, no. | |
56:02.600 --> 56:05.000 | |
Check it out, some of them are just mind blowing, | |
56:06.320 --> 56:07.720 | |
really beautiful. | |
56:07.720 --> 56:12.400 | |
And if you ask, how did it do that? | |
56:13.760 --> 56:16.520 | |
You go talk to Demis Hassabis, | |
56:16.520 --> 56:18.240 | |
I know others from DeepMind, | |
56:19.120 --> 56:20.600 | |
all they'll ultimately be able to give you | |
56:20.600 --> 56:23.920 | |
is big tables of numbers, matrices, | |
56:23.920 --> 56:25.720 | |
that define the neural network. | |
56:25.720 --> 56:28.080 | |
And you can stare at these tables of numbers | |
56:28.080 --> 56:29.600 | |
till your face turn blue, | |
56:29.600 --> 56:32.520 | |
and you're not gonna understand much | |
56:32.520 --> 56:34.520 | |
about why it made that move. | |
56:34.520 --> 56:37.640 | |
And even if you have natural language processing | |
56:37.640 --> 56:40.280 | |
that can tell you in human language about, | |
56:40.280 --> 56:42.520 | |
oh, five, seven, points, two, eight, | |
56:42.520 --> 56:43.560 | |
still not gonna really help. | |
56:43.560 --> 56:47.480 | |
So I think there's a whole spectrum of fun challenges | |
56:47.480 --> 56:50.520 | |
that are involved in taking a computation | |
56:50.520 --> 56:52.240 | |
that does intelligent things | |
56:52.240 --> 56:56.240 | |
and transforming it into something equally good, | |
56:57.760 --> 57:01.840 | |
equally intelligent, but that's more understandable. | |
57:01.840 --> 57:03.240 | |
And I think that's really valuable | |
57:03.240 --> 57:07.440 | |
because I think as we put machines in charge | |
57:07.440 --> 57:09.760 | |
of ever more infrastructure in our world, | |
57:09.760 --> 57:12.680 | |
the power grid, the trading on the stock market, | |
57:12.680 --> 57:14.320 | |
weapon systems and so on, | |
57:14.320 --> 57:17.760 | |
it's absolutely crucial that we can trust | |
57:17.760 --> 57:19.400 | |
these AIs to do all we want. | |
57:19.400 --> 57:21.520 | |
And trust really comes from understanding | |
57:22.520 --> 57:24.400 | |
in a very fundamental way. | |
57:24.400 --> 57:27.560 | |
And that's why I'm working on this, | |
57:27.560 --> 57:29.160 | |
because I think the more, | |
57:29.160 --> 57:31.840 | |
if we're gonna have some hope of ensuring | |
57:31.840 --> 57:33.520 | |
that machines have adopted our goals | |
57:33.520 --> 57:35.800 | |
and that they're gonna retain them, | |
57:35.800 --> 57:38.800 | |
that kind of trust, I think, | |
57:38.800 --> 57:41.200 | |
needs to be based on things you can actually understand, | |
57:41.200 --> 57:44.240 | |
preferably even improve theorems on. | |
57:44.240 --> 57:46.080 | |
Even with a self driving car, right? | |
57:47.040 --> 57:48.680 | |
If someone just tells you it's been trained | |
57:48.680 --> 57:50.640 | |
on tons of data and it never crashed, | |
57:50.640 --> 57:54.200 | |
it's less reassuring than if someone actually has a proof. | |
57:54.200 --> 57:55.960 | |
Maybe it's a computer verified proof, | |
57:55.960 --> 57:58.800 | |
but still it says that under no circumstances | |
57:58.800 --> 58:02.320 | |
is this car just gonna swerve into oncoming traffic. | |
58:02.320 --> 58:04.640 | |
And that kind of information helps to build trust | |
58:04.640 --> 58:08.080 | |
and helps build the alignment of goals, | |
58:09.400 --> 58:12.200 | |
at least awareness that your goals, your values are aligned. | |
58:12.200 --> 58:13.840 | |
And I think even in the very short term, | |
58:13.840 --> 58:16.360 | |
if you look at how, you know, today, right? | |
58:16.360 --> 58:19.320 | |
This absolutely pathetic state of cybersecurity | |
58:19.320 --> 58:21.720 | |
that we have, where is it? | |
58:21.720 --> 58:25.960 | |
Three billion Yahoo accounts we can't pack, | |
58:27.200 --> 58:31.720 | |
almost every American's credit card and so on. | |
58:32.800 --> 58:34.120 | |
Why is this happening? | |
58:34.120 --> 58:37.960 | |
It's ultimately happening because we have software | |
58:37.960 --> 58:41.200 | |
that nobody fully understood how it worked. | |
58:41.200 --> 58:44.800 | |
That's why the bugs hadn't been found, right? | |
58:44.800 --> 58:47.480 | |
And I think AI can be used very effectively | |
58:47.480 --> 58:49.640 | |
for offense, for hacking, | |
58:49.640 --> 58:52.320 | |
but it can also be used for defense. | |
58:52.320 --> 58:55.360 | |
Hopefully automating verifiability | |
58:55.360 --> 59:00.360 | |
and creating systems that are built in different ways | |
59:00.680 --> 59:02.920 | |
so you can actually prove things about them. | |
59:02.920 --> 59:05.240 | |
And it's important. | |
59:05.240 --> 59:07.680 | |
So speaking of software that nobody understands | |
59:07.680 --> 59:10.640 | |
how it works, of course, a bunch of people ask | |
59:10.640 --> 59:12.160 | |
about your paper, about your thoughts | |
59:12.160 --> 59:14.680 | |
of why does deep and cheap learning work so well? | |
59:14.680 --> 59:15.520 | |
That's the paper. | |
59:15.520 --> 59:18.320 | |
But what are your thoughts on deep learning? | |
59:18.320 --> 59:21.880 | |
These kind of simplified models of our own brains | |
59:21.880 --> 59:26.440 | |
have been able to do some successful perception work, | |
59:26.440 --> 59:29.560 | |
pattern recognition work, and now with AlphaZero and so on, | |
59:29.560 --> 59:30.880 | |
do some clever things. | |
59:30.880 --> 59:33.880 | |
What are your thoughts about the promise limitations | |
59:33.880 --> 59:35.680 | |
of this piece? | |
59:35.680 --> 59:40.680 | |
Great, I think there are a number of very important insights, | |
59:43.080 --> 59:44.640 | |
very important lessons we can always draw | |
59:44.640 --> 59:47.120 | |
from these kinds of successes. | |
59:47.120 --> 59:48.960 | |
One of them is when you look at the human brain, | |
59:48.960 --> 59:51.480 | |
you see it's very complicated, 10th of 11 neurons, | |
59:51.480 --> 59:53.320 | |
and there are all these different kinds of neurons | |
59:53.320 --> 59:55.040 | |
and yada, yada, and there's been this long debate | |
59:55.040 --> 59:57.200 | |
about whether the fact that we have dozens | |
59:57.200 --> 1:00:00.160 | |
of different kinds is actually necessary for intelligence. | |
1:00:01.560 --> 1:00:03.360 | |
We can now, I think, quite convincingly answer | |
1:00:03.360 --> 1:00:07.640 | |
that question of no, it's enough to have just one kind. | |
1:00:07.640 --> 1:00:09.920 | |
If you look under the hood of AlphaZero, | |
1:00:09.920 --> 1:00:11.080 | |
there's only one kind of neuron | |
1:00:11.080 --> 1:00:15.000 | |
and it's ridiculously simple mathematical thing. | |
1:00:15.000 --> 1:00:17.280 | |
So it's just like in physics, | |
1:00:17.280 --> 1:00:20.320 | |
it's not, if you have a gas with waves in it, | |
1:00:20.320 --> 1:00:23.240 | |
it's not the detailed nature of the molecule that matter, | |
1:00:24.240 --> 1:00:26.040 | |
it's the collective behavior somehow. | |
1:00:26.040 --> 1:00:30.720 | |
Similarly, it's this higher level structure | |
1:00:30.720 --> 1:00:31.760 | |
of the network that matters, | |
1:00:31.760 --> 1:00:34.080 | |
not that you have 20 kinds of neurons. | |
1:00:34.080 --> 1:00:37.040 | |
I think our brain is such a complicated mess | |
1:00:37.040 --> 1:00:41.720 | |
because it wasn't evolved just to be intelligent, | |
1:00:41.720 --> 1:00:45.840 | |
it was involved to also be self assembling | |
1:00:47.000 --> 1:00:48.760 | |
and self repairing, right? | |
1:00:48.760 --> 1:00:51.920 | |
And evolutionarily attainable. | |
1:00:51.920 --> 1:00:53.560 | |
And so on and so on. | |
1:00:53.560 --> 1:00:54.720 | |
So I think it's pretty, | |
1:00:54.720 --> 1:00:57.040 | |
my hunch is that we're going to understand | |
1:00:57.040 --> 1:00:59.520 | |
how to build AGI before we fully understand | |
1:00:59.520 --> 1:01:02.600 | |
how our brains work, just like we understood | |
1:01:02.600 --> 1:01:05.560 | |
how to build flying machines long before | |
1:01:05.560 --> 1:01:07.800 | |
we were able to build a mechanical bird. | |
1:01:07.800 --> 1:01:08.640 | |
Yeah, that's right. | |
1:01:08.640 --> 1:01:13.280 | |
You've given the example exactly of mechanical birds | |
1:01:13.280 --> 1:01:15.680 | |
and airplanes and airplanes do a pretty good job | |
1:01:15.680 --> 1:01:18.560 | |
of flying without really mimicking bird flight. | |
1:01:18.560 --> 1:01:20.920 | |
And even now after 100 years later, | |
1:01:20.920 --> 1:01:23.880 | |
did you see the Ted talk with this German mechanical bird? | |
1:01:23.880 --> 1:01:25.040 | |
I heard you mention it. | |
1:01:25.040 --> 1:01:26.520 | |
Check it out, it's amazing. | |
1:01:26.520 --> 1:01:27.760 | |
But even after that, right, | |
1:01:27.760 --> 1:01:29.360 | |
we still don't fly in mechanical birds | |
1:01:29.360 --> 1:01:32.720 | |
because it turned out the way we came up with was simpler | |
1:01:32.720 --> 1:01:33.840 | |
and it's better for our purposes. | |
1:01:33.840 --> 1:01:35.280 | |
And I think it might be the same there. | |
1:01:35.280 --> 1:01:36.280 | |
That's one lesson. | |
1:01:37.520 --> 1:01:42.520 | |
And another lesson, it's more what our paper was about. | |
1:01:42.640 --> 1:01:45.800 | |
First, as a physicist thought it was fascinating | |
1:01:45.800 --> 1:01:48.240 | |
how there's a very close mathematical relationship | |
1:01:48.240 --> 1:01:50.800 | |
actually between our artificial neural networks | |
1:01:50.800 --> 1:01:54.560 | |
and a lot of things that we've studied for in physics | |
1:01:54.560 --> 1:01:57.520 | |
go by nerdy names like the renormalization group equation | |
1:01:57.520 --> 1:01:59.800 | |
and Hamiltonians and yada, yada, yada. | |
1:01:59.800 --> 1:02:04.360 | |
And when you look a little more closely at this, | |
1:02:05.720 --> 1:02:06.560 | |
you have, | |
1:02:10.320 --> 1:02:12.360 | |
at first I was like, well, there's something crazy here | |
1:02:12.360 --> 1:02:13.520 | |
that doesn't make sense. | |
1:02:13.520 --> 1:02:18.520 | |
Because we know that if you even want to build | |
1:02:19.200 --> 1:02:22.560 | |
a super simple neural network to tell apart cat pictures | |
1:02:22.560 --> 1:02:23.400 | |
and dog pictures, right, | |
1:02:23.400 --> 1:02:25.400 | |
that you can do that very, very well now. | |
1:02:25.400 --> 1:02:27.520 | |
But if you think about it a little bit, | |
1:02:27.520 --> 1:02:29.080 | |
you convince yourself it must be impossible | |
1:02:29.080 --> 1:02:31.920 | |
because if I have one megapixel, | |
1:02:31.920 --> 1:02:34.160 | |
even if each pixel is just black or white, | |
1:02:34.160 --> 1:02:36.960 | |
there's two to the power of 1 million possible images, | |
1:02:36.960 --> 1:02:38.960 | |
which is way more than there are atoms in our universe, | |
1:02:38.960 --> 1:02:41.000 | |
right, so in order to, | |
1:02:42.040 --> 1:02:43.200 | |
and then for each one of those, | |
1:02:43.200 --> 1:02:44.640 | |
I have to assign a number, | |
1:02:44.640 --> 1:02:47.080 | |
which is the probability that it's a dog. | |
1:02:47.080 --> 1:02:49.440 | |
So an arbitrary function of images | |
1:02:49.440 --> 1:02:54.440 | |
is a list of more numbers than there are atoms in our universe. | |
1:02:54.440 --> 1:02:57.360 | |
So clearly I can't store that under the hood of my GPU | |
1:02:57.360 --> 1:03:00.640 | |
or my computer, yet somehow it works. | |
1:03:00.640 --> 1:03:01.480 | |
So what does that mean? | |
1:03:01.480 --> 1:03:04.960 | |
Well, it means that out of all of the problems | |
1:03:04.960 --> 1:03:08.200 | |
that you could try to solve with a neural network, | |
1:03:10.120 --> 1:03:12.880 | |
almost all of them are impossible to solve | |
1:03:12.880 --> 1:03:14.560 | |
with a reasonably sized one. | |
1:03:15.480 --> 1:03:17.440 | |
But then what we showed in our paper | |
1:03:17.440 --> 1:03:22.360 | |
was that the fraction, the kind of problems, | |
1:03:22.360 --> 1:03:23.800 | |
the fraction of all the problems | |
1:03:23.800 --> 1:03:26.520 | |
that you could possibly pose, | |
1:03:26.520 --> 1:03:29.480 | |
that we actually care about given the laws of physics | |
1:03:29.480 --> 1:03:32.480 | |
is also an infinite testimony, tiny little part. | |
1:03:32.480 --> 1:03:35.440 | |
And amazingly, they're basically the same part. | |
1:03:35.440 --> 1:03:37.560 | |
Yeah, it's almost like our world was created for, | |
1:03:37.560 --> 1:03:39.000 | |
I mean, they kind of come together. | |
1:03:39.000 --> 1:03:42.800 | |
Yeah, well, you could say maybe where the world was created | |
1:03:42.800 --> 1:03:44.960 | |
for us, but I have a more modest interpretation, | |
1:03:44.960 --> 1:03:46.680 | |
which is that the world was created for us, | |
1:03:46.680 --> 1:03:48.040 | |
but I have a more modest interpretation, | |
1:03:48.040 --> 1:03:50.360 | |
which is that instead evolution endowed us | |
1:03:50.360 --> 1:03:53.120 | |
with neural networks precisely for that reason. | |
1:03:53.120 --> 1:03:54.640 | |
Because this particular architecture, | |
1:03:54.640 --> 1:03:56.040 | |
as opposed to the one in your laptop, | |
1:03:56.040 --> 1:04:01.040 | |
is very, very well adapted to solving the kind of problems | |
1:04:02.480 --> 1:04:05.560 | |
that nature kept presenting our ancestors with. | |
1:04:05.560 --> 1:04:08.120 | |
So it makes sense that why do we have a brain | |
1:04:08.120 --> 1:04:09.280 | |
in the first place? | |
1:04:09.280 --> 1:04:11.880 | |
It's to be able to make predictions about the future | |
1:04:11.880 --> 1:04:12.880 | |
and so on. | |
1:04:12.880 --> 1:04:16.440 | |
So if we had a sucky system, which could never solve it, | |
1:04:16.440 --> 1:04:18.280 | |
we wouldn't have a world. | |
1:04:18.280 --> 1:04:23.280 | |
So this is, I think, a very beautiful fact. | |
1:04:23.680 --> 1:04:24.520 | |
Yeah. | |
1:04:24.520 --> 1:04:29.000 | |
We also realize that there's been earlier work | |
1:04:29.000 --> 1:04:32.040 | |
on why deeper networks are good, | |
1:04:32.040 --> 1:04:34.680 | |
but we were able to show an additional cool fact there, | |
1:04:34.680 --> 1:04:38.360 | |
which is that even incredibly simple problems, | |
1:04:38.360 --> 1:04:41.080 | |
like suppose I give you a thousand numbers | |
1:04:41.080 --> 1:04:42.720 | |
and ask you to multiply them together, | |
1:04:42.720 --> 1:04:46.680 | |
and you can write a few lines of code, boom, done, trivial. | |
1:04:46.680 --> 1:04:49.520 | |
If you just try to do that with a neural network | |
1:04:49.520 --> 1:04:52.440 | |
that has only one single hidden layer in it, | |
1:04:52.440 --> 1:04:53.400 | |
you can do it, | |
1:04:54.320 --> 1:04:57.360 | |
but you're going to need two to the power of a thousand | |
1:04:57.360 --> 1:05:00.920 | |
neurons to multiply a thousand numbers, | |
1:05:00.920 --> 1:05:02.520 | |
which is, again, more neurons than there are atoms | |
1:05:02.520 --> 1:05:03.360 | |
in our universe. | |
1:05:04.600 --> 1:05:05.480 | |
That's fascinating. | |
1:05:05.480 --> 1:05:09.960 | |
But if you allow yourself to make it a deep network | |
1:05:09.960 --> 1:05:13.240 | |
with many layers, you only need 4,000 neurons. | |
1:05:13.240 --> 1:05:14.520 | |
It's perfectly feasible. | |
1:05:16.400 --> 1:05:17.960 | |
That's really interesting. | |
1:05:17.960 --> 1:05:18.800 | |
Yeah. | |
1:05:18.800 --> 1:05:21.040 | |
So on another architecture type, | |
1:05:21.040 --> 1:05:22.720 | |
I mean, you mentioned Schrodinger's equation, | |
1:05:22.720 --> 1:05:26.360 | |
and what are your thoughts about quantum computing | |
1:05:27.240 --> 1:05:32.240 | |
and the role of this kind of computational unit | |
1:05:32.400 --> 1:05:34.880 | |
in creating an intelligence system? | |
1:05:34.880 --> 1:05:39.520 | |
In some Hollywood movies that I will not mention by name | |
1:05:39.520 --> 1:05:41.040 | |
because I don't want to spoil them. | |
1:05:41.040 --> 1:05:44.240 | |
The way they get AGI is building a quantum computer. | |
1:05:45.480 --> 1:05:47.600 | |
Because the word quantum sounds cool and so on. | |
1:05:47.600 --> 1:05:48.440 | |
That's right. | |
1:05:50.040 --> 1:05:52.880 | |
First of all, I think we don't need quantum computers | |
1:05:52.880 --> 1:05:54.920 | |
to build AGI. | |
1:05:54.920 --> 1:05:59.240 | |
I suspect your brain is not a quantum computer | |
1:05:59.240 --> 1:06:00.640 | |
in any profound sense. | |
1:06:01.600 --> 1:06:03.200 | |
So you don't even wrote a paper about that | |
1:06:03.200 --> 1:06:04.560 | |
a lot many years ago. | |
1:06:04.560 --> 1:06:08.120 | |
I calculated the so called decoherence time, | |
1:06:08.120 --> 1:06:10.320 | |
how long it takes until the quantum computerness | |
1:06:10.320 --> 1:06:13.400 | |
of what your neurons are doing gets erased | |
1:06:15.320 --> 1:06:17.960 | |
by just random noise from the environment. | |
1:06:17.960 --> 1:06:21.320 | |
And it's about 10 to the minus 21 seconds. | |
1:06:21.320 --> 1:06:24.600 | |
So as cool as it would be to have a quantum computer | |
1:06:24.600 --> 1:06:27.320 | |
in my head, I don't think that fast. | |
1:06:27.320 --> 1:06:28.360 | |
On the other hand, | |
1:06:28.360 --> 1:06:33.040 | |
there are very cool things you could do | |
1:06:33.040 --> 1:06:34.200 | |
with quantum computers. | |
1:06:35.240 --> 1:06:37.480 | |
Or I think we'll be able to do soon | |
1:06:37.480 --> 1:06:39.360 | |
when we get bigger ones. | |
1:06:39.360 --> 1:06:40.960 | |
That might actually help machine learning | |
1:06:40.960 --> 1:06:43.160 | |
do even better than the brain. | |
1:06:43.160 --> 1:06:45.640 | |
So for example, | |
1:06:47.040 --> 1:06:50.760 | |
one, this is just a moonshot, | |
1:06:50.760 --> 1:06:55.760 | |
but learning is very much same thing as search. | |
1:07:01.800 --> 1:07:03.160 | |
If you're trying to train a neural network | |
1:07:03.160 --> 1:07:06.240 | |
to get really learned to do something really well, | |
1:07:06.240 --> 1:07:07.280 | |
you have some loss function, | |
1:07:07.280 --> 1:07:10.360 | |
you have a bunch of knobs you can turn, | |
1:07:10.360 --> 1:07:12.080 | |
represented by a bunch of numbers, | |
1:07:12.080 --> 1:07:12.920 | |
and you're trying to tweak them | |
1:07:12.920 --> 1:07:15.080 | |
so that it becomes as good as possible at this thing. | |
1:07:15.080 --> 1:07:19.680 | |
So if you think of a landscape with some valley, | |
1:07:20.720 --> 1:07:22.120 | |
where each dimension of the landscape | |
1:07:22.120 --> 1:07:24.120 | |
corresponds to some number you can change, | |
1:07:24.120 --> 1:07:25.640 | |
you're trying to find the minimum. | |
1:07:25.640 --> 1:07:26.760 | |
And it's well known that | |
1:07:26.760 --> 1:07:29.040 | |
if you have a very high dimensional landscape, | |
1:07:29.040 --> 1:07:31.840 | |
complicated things, it's super hard to find the minimum. | |
1:07:31.840 --> 1:07:35.840 | |
Quantum mechanics is amazingly good at this. | |
1:07:35.840 --> 1:07:38.240 | |
Like if I want to know what's the lowest energy state | |
1:07:38.240 --> 1:07:39.720 | |
this water can possibly have, | |
1:07:41.720 --> 1:07:42.560 | |
incredibly hard to compute, | |
1:07:42.560 --> 1:07:45.400 | |
but nature will happily figure this out for you | |
1:07:45.400 --> 1:07:48.000 | |
if you just cool it down, make it very, very cold. | |
1:07:49.800 --> 1:07:50.880 | |
If you put a ball somewhere, | |
1:07:50.880 --> 1:07:52.240 | |
it'll roll down to its minimum. | |
1:07:52.240 --> 1:07:54.280 | |
And this happens metaphorically | |
1:07:54.280 --> 1:07:56.320 | |
at the energy landscape too. | |
1:07:56.320 --> 1:07:59.280 | |
And quantum mechanics even uses some clever tricks, | |
1:07:59.280 --> 1:08:02.520 | |
which today's machine learning systems don't. | |
1:08:02.520 --> 1:08:04.160 | |
Like if you're trying to find the minimum | |
1:08:04.160 --> 1:08:06.960 | |
and you get stuck in the little local minimum here, | |
1:08:06.960 --> 1:08:08.760 | |
in quantum mechanics you can actually tunnel | |
1:08:08.760 --> 1:08:11.840 | |
through the barrier and get unstuck again. | |
1:08:13.480 --> 1:08:14.320 | |
That's really interesting. | |
1:08:14.320 --> 1:08:16.120 | |
Yeah, so it may be, for example, | |
1:08:16.120 --> 1:08:19.160 | |
that we'll one day use quantum computers | |
1:08:19.160 --> 1:08:22.840 | |
that help train neural networks better. | |
1:08:22.840 --> 1:08:23.680 | |
That's really interesting. | |
1:08:23.680 --> 1:08:27.040 | |
Okay, so as a component of kind of the learning process, | |
1:08:27.040 --> 1:08:27.880 | |
for example. | |
1:08:27.880 --> 1:08:29.440 | |
Yeah. | |
1:08:29.440 --> 1:08:33.080 | |
Let me ask sort of wrapping up here a little bit, | |
1:08:33.080 --> 1:08:36.880 | |
let me return to the questions of our human nature | |
1:08:36.880 --> 1:08:40.000 | |
and love, as I mentioned. | |
1:08:40.000 --> 1:08:41.640 | |
So do you think, | |
1:08:44.280 --> 1:08:46.000 | |
you mentioned sort of a helper robot, | |
1:08:46.000 --> 1:08:48.640 | |
but you could think of also personal robots. | |
1:08:48.640 --> 1:08:52.480 | |
Do you think the way we human beings fall in love | |
1:08:52.480 --> 1:08:54.680 | |
and get connected to each other | |
1:08:54.680 --> 1:08:58.040 | |
is possible to achieve in an AI system | |
1:08:58.040 --> 1:09:00.360 | |
and human level AI intelligence system? | |
1:09:00.360 --> 1:09:03.720 | |
Do you think we would ever see that kind of connection? | |
1:09:03.720 --> 1:09:06.160 | |
Or, you know, in all this discussion | |
1:09:06.160 --> 1:09:08.520 | |
about solving complex goals, | |
1:09:08.520 --> 1:09:10.760 | |
is this kind of human social connection, | |
1:09:10.760 --> 1:09:12.560 | |
do you think that's one of the goals | |
1:09:12.560 --> 1:09:16.280 | |
on the peaks and valleys with the raising sea levels | |
1:09:16.280 --> 1:09:17.360 | |
that we'll be able to achieve? | |
1:09:17.360 --> 1:09:20.040 | |
Or do you think that's something that's ultimately, | |
1:09:20.040 --> 1:09:21.760 | |
or at least in the short term, | |
1:09:21.760 --> 1:09:23.640 | |
relative to the other goals is not achievable? | |
1:09:23.640 --> 1:09:25.120 | |
I think it's all possible. | |
1:09:25.120 --> 1:09:27.600 | |
And I mean, in recent, | |
1:09:27.600 --> 1:09:30.840 | |
there's a very wide range of guesses, as you know, | |
1:09:30.840 --> 1:09:33.720 | |
among AI researchers, when we're going to get AGI. | |
1:09:35.120 --> 1:09:37.640 | |
Some people, you know, like our friend Rodney Brooks | |
1:09:37.640 --> 1:09:41.040 | |
says it's going to be hundreds of years at least. | |
1:09:41.040 --> 1:09:42.200 | |
And then there are many others | |
1:09:42.200 --> 1:09:44.040 | |
who think it's going to happen much sooner. | |
1:09:44.040 --> 1:09:45.520 | |
And recent polls, | |
1:09:46.840 --> 1:09:48.640 | |
maybe half or so of AI researchers | |
1:09:48.640 --> 1:09:50.920 | |
think we're going to get AGI within decades. | |
1:09:50.920 --> 1:09:52.720 | |
So if that happens, of course, | |
1:09:52.720 --> 1:09:55.040 | |
then I think these things are all possible. | |
1:09:55.040 --> 1:09:56.840 | |
But in terms of whether it will happen, | |
1:09:56.840 --> 1:10:00.600 | |
I think we shouldn't spend so much time asking | |
1:10:00.600 --> 1:10:03.240 | |
what do we think will happen in the future? | |
1:10:03.240 --> 1:10:05.160 | |
As if we are just some sort of pathetic, | |
1:10:05.160 --> 1:10:07.040 | |
your passive bystanders, you know, | |
1:10:07.040 --> 1:10:09.280 | |
waiting for the future to happen to us. | |
1:10:09.280 --> 1:10:11.640 | |
Hey, we're the ones creating this future, right? | |
1:10:11.640 --> 1:10:15.520 | |
So we should be proactive about it | |
1:10:15.520 --> 1:10:16.920 | |
and ask ourselves what sort of future | |
1:10:16.920 --> 1:10:18.240 | |
we would like to have happen. | |
1:10:18.240 --> 1:10:19.920 | |
We're going to make it like that. | |
1:10:19.920 --> 1:10:22.720 | |
Well, what I prefer is just some sort of incredibly boring, | |
1:10:22.720 --> 1:10:24.320 | |
zombie like future where there's all these | |
1:10:24.320 --> 1:10:26.040 | |
mechanical things happening and there's no passion, | |
1:10:26.040 --> 1:10:28.040 | |
no emotion, no experience, maybe even. | |
1:10:29.600 --> 1:10:32.040 | |
No, I would of course, much rather prefer it | |
1:10:32.040 --> 1:10:35.240 | |
if all the things that we find that we value the most | |
1:10:36.240 --> 1:10:40.680 | |
about humanity are our subjective experience, | |
1:10:40.680 --> 1:10:43.000 | |
passion, inspiration, love, you know. | |
1:10:43.000 --> 1:10:48.000 | |
If we can create a future where those things do happen, | |
1:10:48.000 --> 1:10:50.840 | |
where those things do exist, you know, | |
1:10:50.840 --> 1:10:54.560 | |
I think ultimately it's not our universe | |
1:10:54.560 --> 1:10:57.960 | |
giving meaning to us, it's us giving meaning to our universe. | |
1:10:57.960 --> 1:11:01.840 | |
And if we build more advanced intelligence, | |
1:11:01.840 --> 1:11:03.680 | |
let's make sure we build it in such a way | |
1:11:03.680 --> 1:11:08.680 | |
that meaning is part of it. | |
1:11:09.120 --> 1:11:11.400 | |
A lot of people that seriously study this problem | |
1:11:11.400 --> 1:11:13.600 | |
and think of it from different angles | |
1:11:13.600 --> 1:11:16.880 | |
have trouble in the majority of cases, | |
1:11:16.880 --> 1:11:19.160 | |
if they think through that happen, | |
1:11:19.160 --> 1:11:22.520 | |
are the ones that are not beneficial to humanity. | |
1:11:22.520 --> 1:11:25.560 | |
And so, yeah, so what are your thoughts? | |
1:11:25.560 --> 1:11:29.400 | |
What's should people, you know, | |
1:11:29.400 --> 1:11:32.040 | |
I really don't like people to be terrified. | |
1:11:33.440 --> 1:11:35.040 | |
What's a way for people to think about it | |
1:11:35.040 --> 1:11:39.600 | |
in a way we can solve it and we can make it better? | |
1:11:39.600 --> 1:11:42.960 | |
No, I don't think panicking is going to help in any way. | |
1:11:42.960 --> 1:11:44.840 | |
It's not going to increase chances | |
1:11:44.840 --> 1:11:45.880 | |
of things going well either. | |
1:11:45.880 --> 1:11:48.400 | |
Even if you are in a situation where there is a real threat, | |
1:11:48.400 --> 1:11:51.080 | |
does it help if everybody just freaks out? | |
1:11:51.080 --> 1:11:52.680 | |
No, of course, of course not. | |
1:11:53.640 --> 1:11:56.600 | |
I think, yeah, there are of course ways | |
1:11:56.600 --> 1:11:58.440 | |
in which things can go horribly wrong. | |
1:11:59.560 --> 1:12:03.680 | |
First of all, it's important when we think about this thing, | |
1:12:03.680 --> 1:12:05.280 | |
about the problems and risks, | |
1:12:05.280 --> 1:12:07.160 | |
to also remember how huge the upsides can be | |
1:12:07.160 --> 1:12:08.440 | |
if we get it right, right? | |
1:12:08.440 --> 1:12:12.360 | |
Everything we love about society and civilization | |
1:12:12.360 --> 1:12:13.400 | |
is a product of intelligence. | |
1:12:13.400 --> 1:12:15.320 | |
So if we can amplify our intelligence | |
1:12:15.320 --> 1:12:18.760 | |
with machine intelligence and not anymore lose our loved one | |
1:12:18.760 --> 1:12:21.080 | |
to what we're told is an incurable disease | |
1:12:21.080 --> 1:12:24.800 | |
and things like this, of course, we should aspire to that. | |
1:12:24.800 --> 1:12:26.680 | |
So that can be a motivator, I think, | |
1:12:26.680 --> 1:12:29.120 | |
reminding ourselves that the reason we try to solve problems | |
1:12:29.120 --> 1:12:33.520 | |
is not just because we're trying to avoid gloom, | |
1:12:33.520 --> 1:12:35.760 | |
but because we're trying to do something great. | |
1:12:35.760 --> 1:12:37.680 | |
But then in terms of the risks, | |
1:12:37.680 --> 1:12:42.680 | |
I think the really important question is to ask, | |
1:12:42.680 --> 1:12:45.480 | |
what can we do today that will actually help | |
1:12:45.480 --> 1:12:47.320 | |
make the outcome good, right? | |
1:12:47.320 --> 1:12:49.880 | |
And dismissing the risk is not one of them. | |
1:12:51.240 --> 1:12:54.800 | |
I find it quite funny often when I'm in discussion panels | |
1:12:54.800 --> 1:12:55.960 | |
about these things, | |
1:12:55.960 --> 1:13:00.960 | |
how the people who work for companies, | |
1:13:01.200 --> 1:13:03.120 | |
always be like, oh, nothing to worry about, | |
1:13:03.120 --> 1:13:04.760 | |
nothing to worry about, nothing to worry about. | |
1:13:04.760 --> 1:13:09.600 | |
And it's only academics sometimes express concerns. | |
1:13:09.600 --> 1:13:11.880 | |
That's not surprising at all if you think about it. | |
1:13:11.880 --> 1:13:12.880 | |
Right. | |
1:13:12.880 --> 1:13:15.200 | |
Upton Sinclair quipped, right, | |
1:13:15.200 --> 1:13:18.040 | |
that it's hard to make a man believe in something | |
1:13:18.040 --> 1:13:20.120 | |
when his income depends on not believing in it. | |
1:13:20.120 --> 1:13:24.080 | |
And frankly, we know a lot of these people in companies | |
1:13:24.080 --> 1:13:26.240 | |
that they're just as concerned as anyone else. | |
1:13:26.240 --> 1:13:28.480 | |
But if you're the CEO of a company, | |
1:13:28.480 --> 1:13:30.280 | |
that's not something you want to go on record saying | |
1:13:30.280 --> 1:13:33.440 | |
when you have silly journalists who are gonna put a picture | |
1:13:33.440 --> 1:13:35.720 | |
of a Terminator robot when they quote you. | |
1:13:35.720 --> 1:13:39.040 | |
So the issues are real. | |
1:13:39.040 --> 1:13:41.920 | |
And the way I think about what the issue is, | |
1:13:41.920 --> 1:13:46.920 | |
is basically the real choice we have is, | |
1:13:48.040 --> 1:13:50.840 | |
first of all, are we gonna just dismiss the risks | |
1:13:50.840 --> 1:13:54.480 | |
and say, well, let's just go ahead and build machines | |
1:13:54.480 --> 1:13:57.560 | |
that can do everything we can do better and cheaper. | |
1:13:57.560 --> 1:14:00.200 | |
Let's just make ourselves obsolete as fast as possible. | |
1:14:00.200 --> 1:14:01.720 | |
What could possibly go wrong? | |
1:14:01.720 --> 1:14:03.440 | |
That's one attitude. | |
1:14:03.440 --> 1:14:05.440 | |
The opposite attitude, I think, is to say, | |
1:14:06.400 --> 1:14:08.800 | |
here's this incredible potential, | |
1:14:08.800 --> 1:14:11.960 | |
let's think about what kind of future | |
1:14:11.960 --> 1:14:14.640 | |
we're really, really excited about. | |
1:14:14.640 --> 1:14:18.480 | |
What are the shared goals that we can really aspire towards? | |
1:14:18.480 --> 1:14:19.960 | |
And then let's think really hard | |
1:14:19.960 --> 1:14:22.000 | |
about how we can actually get there. | |
1:14:22.000 --> 1:14:24.160 | |
So start with, don't start thinking about the risks, | |
1:14:24.160 --> 1:14:26.720 | |
start thinking about the goals. | |
1:14:26.720 --> 1:14:28.200 | |
And then when you do that, | |
1:14:28.200 --> 1:14:30.480 | |
then you can think about the obstacles you want to avoid. | |
1:14:30.480 --> 1:14:32.840 | |
I often get students coming in right here into my office | |
1:14:32.840 --> 1:14:34.120 | |
for career advice. | |
1:14:34.120 --> 1:14:35.560 | |
I always ask them this very question, | |
1:14:35.560 --> 1:14:37.920 | |
where do you want to be in the future? | |
1:14:37.920 --> 1:14:40.640 | |
If all she can say is, oh, maybe I'll have cancer, | |
1:14:40.640 --> 1:14:42.480 | |
maybe I'll get run over by a truck. | |
1:14:42.480 --> 1:14:44.280 | |
Yeah, focus on the obstacles instead of the goals. | |
1:14:44.280 --> 1:14:46.880 | |
She's just going to end up a hypochondriac paranoid. | |
1:14:47.920 --> 1:14:49.920 | |
Whereas if she comes in and fire in her eyes | |
1:14:49.920 --> 1:14:51.840 | |
and is like, I want to be there. | |
1:14:51.840 --> 1:14:53.960 | |
And then we can talk about the obstacles | |
1:14:53.960 --> 1:14:55.760 | |
and see how we can circumvent them. | |
1:14:55.760 --> 1:14:58.880 | |
That's, I think, a much, much healthier attitude. | |
1:14:58.880 --> 1:15:03.880 | |
And I feel it's very challenging to come up with a vision | |
1:15:03.880 --> 1:15:08.120 | |
for the future, which we are unequivocally excited about. | |
1:15:08.120 --> 1:15:10.320 | |
I'm not just talking now in the vague terms, | |
1:15:10.320 --> 1:15:12.360 | |
like, yeah, let's cure cancer, fine. | |
1:15:12.360 --> 1:15:14.720 | |
I'm talking about what kind of society | |
1:15:14.720 --> 1:15:15.840 | |
do we want to create? | |
1:15:15.840 --> 1:15:20.360 | |
What do we want it to mean to be human in the age of AI, | |
1:15:20.360 --> 1:15:21.720 | |
in the age of AGI? | |
1:15:22.840 --> 1:15:25.360 | |
So if we can have this conversation, | |
1:15:25.360 --> 1:15:28.200 | |
broad, inclusive conversation, | |
1:15:28.200 --> 1:15:31.400 | |
and gradually start converging towards some, | |
1:15:31.400 --> 1:15:34.240 | |
some future that with some direction, at least, | |
1:15:34.240 --> 1:15:35.400 | |
that we want to steer towards, right, | |
1:15:35.400 --> 1:15:38.160 | |
then we'll be much more motivated | |
1:15:38.160 --> 1:15:39.960 | |
to constructively take on the obstacles. | |
1:15:39.960 --> 1:15:43.560 | |
And I think if I had, if I had to, | |
1:15:43.560 --> 1:15:46.640 | |
if I try to wrap this up in a more succinct way, | |
1:15:46.640 --> 1:15:51.480 | |
I think we can all agree already now | |
1:15:51.480 --> 1:15:56.160 | |
that we should aspire to build AGI | |
1:15:56.160 --> 1:16:05.160 | |
that doesn't overpower us, but that empowers us. | |
1:16:05.160 --> 1:16:08.560 | |
And think of the many various ways that can do that, | |
1:16:08.560 --> 1:16:11.000 | |
whether that's from my side of the world | |
1:16:11.000 --> 1:16:12.720 | |
of autonomous vehicles. | |
1:16:12.720 --> 1:16:14.720 | |
I'm personally actually from the camp | |
1:16:14.720 --> 1:16:16.800 | |
that believes this human level intelligence | |
1:16:16.800 --> 1:16:20.480 | |
is required to achieve something like vehicles | |
1:16:20.480 --> 1:16:23.880 | |
that would actually be something we would enjoy using | |
1:16:23.880 --> 1:16:25.120 | |
and being part of. | |
1:16:25.120 --> 1:16:27.040 | |
So that's one example, and certainly there's a lot | |
1:16:27.040 --> 1:16:30.920 | |
of other types of robots and medicine and so on. | |
1:16:30.920 --> 1:16:33.880 | |
So focusing on those and then coming up with the obstacles, | |
1:16:33.880 --> 1:16:35.920 | |
coming up with the ways that that can go wrong | |
1:16:35.920 --> 1:16:38.160 | |
and solving those one at a time. | |
1:16:38.160 --> 1:16:41.520 | |
And just because you can build an autonomous vehicle, | |
1:16:41.520 --> 1:16:42.800 | |
even if you could build one | |
1:16:42.800 --> 1:16:45.080 | |
that would drive just fine without you, | |
1:16:45.080 --> 1:16:46.720 | |
maybe there are some things in life | |
1:16:46.720 --> 1:16:48.400 | |
that we would actually want to do ourselves. | |
1:16:48.400 --> 1:16:49.240 | |
That's right. | |
1:16:49.240 --> 1:16:51.400 | |
Right, like, for example, | |
1:16:51.400 --> 1:16:53.040 | |
if you think of our society as a whole, | |
1:16:53.040 --> 1:16:56.320 | |
there are some things that we find very meaningful to do. | |
1:16:57.200 --> 1:16:59.640 | |
And that doesn't mean we have to stop doing them | |
1:16:59.640 --> 1:17:02.000 | |
just because machines can do them better. | |
1:17:02.000 --> 1:17:04.080 | |
I'm not gonna stop playing tennis | |
1:17:04.080 --> 1:17:07.360 | |
just the day someone builds a tennis robot and beat me. | |
1:17:07.360 --> 1:17:09.600 | |
People are still playing chess and even go. | |
1:17:09.600 --> 1:17:14.600 | |
Yeah, and in the very near term even, | |
1:17:14.600 --> 1:17:18.880 | |
some people are advocating basic income, replace jobs. | |
1:17:18.880 --> 1:17:20.840 | |
But if the government is gonna be willing | |
1:17:20.840 --> 1:17:24.040 | |
to just hand out cash to people for doing nothing, | |
1:17:24.040 --> 1:17:25.840 | |
then one should also seriously consider | |
1:17:25.840 --> 1:17:27.640 | |
whether the government should also hire | |
1:17:27.640 --> 1:17:29.480 | |
a lot more teachers and nurses | |
1:17:29.480 --> 1:17:32.160 | |
and the kind of jobs which people often | |
1:17:32.160 --> 1:17:34.440 | |
find great fulfillment in doing, right? | |
1:17:34.440 --> 1:17:36.320 | |
We get very tired of hearing politicians saying, | |
1:17:36.320 --> 1:17:39.320 | |
oh, we can't afford hiring more teachers, | |
1:17:39.320 --> 1:17:41.480 | |
but we're gonna maybe have basic income. | |
1:17:41.480 --> 1:17:44.000 | |
If we can have more serious research and thought | |
1:17:44.000 --> 1:17:46.200 | |
into what gives meaning to our lives, | |
1:17:46.200 --> 1:17:48.960 | |
the jobs give so much more than income, right? | |
1:17:48.960 --> 1:17:50.520 | |
Mm hmm. | |
1:17:50.520 --> 1:17:53.320 | |
And then think about in the future, | |
1:17:53.320 --> 1:17:58.320 | |
what are the roles that we wanna have people | |
1:18:00.000 --> 1:18:03.040 | |
continually feeling empowered by machines? | |
1:18:03.040 --> 1:18:06.120 | |
And I think sort of, I come from Russia, | |
1:18:06.120 --> 1:18:07.240 | |
from the Soviet Union. | |
1:18:07.240 --> 1:18:10.160 | |
And I think for a lot of people in the 20th century, | |
1:18:10.160 --> 1:18:14.080 | |
going to the moon, going to space was an inspiring thing. | |
1:18:14.080 --> 1:18:18.080 | |
I feel like the universe of the mind, | |
1:18:18.080 --> 1:18:20.880 | |
so AI, understanding, creating intelligence | |
1:18:20.880 --> 1:18:23.240 | |
is that for the 21st century. | |
1:18:23.240 --> 1:18:24.400 | |
So it's really surprising. | |
1:18:24.400 --> 1:18:25.640 | |
And I've heard you mention this. | |
1:18:25.640 --> 1:18:27.400 | |
It's really surprising to me, | |
1:18:27.400 --> 1:18:29.240 | |
both on the research funding side, | |
1:18:29.240 --> 1:18:31.760 | |
that it's not funded as greatly as it could be, | |
1:18:31.760 --> 1:18:34.760 | |
but most importantly, on the politician side, | |
1:18:34.760 --> 1:18:36.520 | |
that it's not part of the public discourse | |
1:18:36.520 --> 1:18:40.800 | |
except in the killer bots terminator kind of view, | |
1:18:40.800 --> 1:18:44.880 | |
that people are not yet, I think, perhaps excited | |
1:18:44.880 --> 1:18:46.680 | |
by the possible positive future | |
1:18:46.680 --> 1:18:48.120 | |
that we can build together. | |
1:18:48.120 --> 1:18:51.520 | |
So we should be, because politicians usually just focus | |
1:18:51.520 --> 1:18:53.320 | |
on the next election cycle, right? | |
1:18:54.480 --> 1:18:57.160 | |
The single most important thing I feel we humans have learned | |
1:18:57.160 --> 1:18:59.320 | |
in the entire history of science | |
1:18:59.320 --> 1:19:02.040 | |
is they were the masters of underestimation. | |
1:19:02.040 --> 1:19:07.040 | |
We underestimated the size of our cosmos again and again, | |
1:19:08.480 --> 1:19:10.200 | |
realizing that everything we thought existed | |
1:19:10.200 --> 1:19:12.240 | |
was just a small part of something grander, right? | |
1:19:12.240 --> 1:19:16.640 | |
Planet, solar system, the galaxy, clusters of galaxies. | |
1:19:16.640 --> 1:19:17.560 | |
The universe. | |
1:19:18.440 --> 1:19:23.120 | |
And we now know that the future has just | |
1:19:23.120 --> 1:19:25.160 | |
so much more potential | |
1:19:25.160 --> 1:19:27.640 | |
than our ancestors could ever have dreamt of. | |
1:19:27.640 --> 1:19:32.360 | |
This cosmos, imagine if all of Earth | |
1:19:33.600 --> 1:19:35.440 | |
was completely devoid of life, | |
1:19:36.640 --> 1:19:38.520 | |
except for Cambridge, Massachusetts. | |
1:19:39.560 --> 1:19:42.680 | |
Wouldn't it be kind of lame if all we ever aspired to | |
1:19:42.680 --> 1:19:45.560 | |
was to stay in Cambridge, Massachusetts forever | |
1:19:45.560 --> 1:19:47.160 | |
and then go extinct in one week, | |
1:19:47.160 --> 1:19:49.760 | |
even though Earth was gonna continue on for longer? | |
1:19:49.760 --> 1:19:52.800 | |
That sort of attitude I think we have now | |
1:19:54.200 --> 1:19:57.800 | |
on the cosmic scale, life can flourish on Earth, | |
1:19:57.800 --> 1:20:00.840 | |
not for four years, but for billions of years. | |
1:20:00.840 --> 1:20:02.920 | |
I can even tell you about how to move it out of harm's way | |
1:20:02.920 --> 1:20:04.840 | |
when the sun gets too hot. | |
1:20:04.840 --> 1:20:09.520 | |
And then we have so much more resources out here, | |
1:20:09.520 --> 1:20:12.480 | |
which today, maybe there are a lot of other planets | |
1:20:12.480 --> 1:20:14.960 | |
with bacteria or cow like life on them, | |
1:20:14.960 --> 1:20:19.880 | |
but most of this, all this opportunity seems, | |
1:20:19.880 --> 1:20:22.440 | |
as far as we can tell, to be largely dead, | |
1:20:22.440 --> 1:20:23.560 | |
like the Sahara Desert. | |
1:20:23.560 --> 1:20:28.480 | |
And yet we have the opportunity to help life flourish | |
1:20:28.480 --> 1:20:30.280 | |
around this for billions of years. | |
1:20:30.280 --> 1:20:32.680 | |
So let's quit squabbling about | |
1:20:34.080 --> 1:20:36.480 | |
whether some little border should be drawn | |
1:20:36.480 --> 1:20:38.440 | |
one mile to the left or right, | |
1:20:38.440 --> 1:20:41.080 | |
and look up into the skies and realize, | |
1:20:41.080 --> 1:20:44.040 | |
hey, we can do such incredible things. | |
1:20:44.040 --> 1:20:46.640 | |
Yeah, and that's, I think, why it's really exciting | |
1:20:46.640 --> 1:20:49.440 | |
that you and others are connected | |
1:20:49.440 --> 1:20:51.880 | |
with some of the work Elon Musk is doing, | |
1:20:51.880 --> 1:20:54.480 | |
because he's literally going out into that space, | |
1:20:54.480 --> 1:20:57.000 | |
really exploring our universe, and it's wonderful. | |
1:20:57.000 --> 1:21:02.000 | |
That is exactly why Elon Musk is so misunderstood, right? | |
1:21:02.000 --> 1:21:05.000 | |
Misconstrued him as some kind of pessimistic doomsayer. | |
1:21:05.000 --> 1:21:07.640 | |
The reason he cares so much about AI safety | |
1:21:07.640 --> 1:21:12.080 | |
is because he more than almost anyone else appreciates | |
1:21:12.080 --> 1:21:14.280 | |
these amazing opportunities that we'll squander | |
1:21:14.280 --> 1:21:16.640 | |
if we wipe out here on Earth. | |
1:21:16.640 --> 1:21:19.680 | |
We're not just going to wipe out the next generation, | |
1:21:19.680 --> 1:21:23.320 | |
all generations, and this incredible opportunity | |
1:21:23.320 --> 1:21:25.400 | |
that's out there, and that would really be a waste. | |
1:21:25.400 --> 1:21:30.080 | |
And AI, for people who think that it would be better | |
1:21:30.080 --> 1:21:33.600 | |
to do without technology, let me just mention that | |
1:21:34.680 --> 1:21:36.320 | |
if we don't improve our technology, | |
1:21:36.320 --> 1:21:39.320 | |
the question isn't whether humanity is going to go extinct. | |
1:21:39.320 --> 1:21:41.160 | |
The question is just whether we're going to get taken out | |
1:21:41.160 --> 1:21:44.800 | |
by the next big asteroid or the next super volcano | |
1:21:44.800 --> 1:21:48.280 | |
or something else dumb that we could easily prevent | |
1:21:48.280 --> 1:21:49.840 | |
with more tech, right? | |
1:21:49.840 --> 1:21:53.160 | |
And if we want life to flourish throughout the cosmos, | |
1:21:53.160 --> 1:21:54.760 | |
AI is the key to it. | |
1:21:56.120 --> 1:21:59.840 | |
As I mentioned in a lot of detail in my book right there, | |
1:21:59.840 --> 1:22:04.840 | |
even many of the most inspired sci fi writers, | |
1:22:04.880 --> 1:22:08.120 | |
I feel have totally underestimated the opportunities | |
1:22:08.120 --> 1:22:11.240 | |
for space travel, especially at the other galaxies, | |
1:22:11.240 --> 1:22:15.360 | |
because they weren't thinking about the possibility of AGI, | |
1:22:15.360 --> 1:22:17.520 | |
which just makes it so much easier. | |
1:22:17.520 --> 1:22:18.440 | |
Right, yeah. | |
1:22:18.440 --> 1:22:23.440 | |
So that goes to your view of AGI that enables our progress, | |
1:22:24.080 --> 1:22:25.760 | |
that enables a better life. | |
1:22:25.760 --> 1:22:28.320 | |
So that's a beautiful way to put it | |
1:22:28.320 --> 1:22:29.960 | |
and then something to strive for. | |
1:22:29.960 --> 1:22:31.440 | |
So Max, thank you so much. | |
1:22:31.440 --> 1:22:32.560 | |
Thank you for your time today. | |
1:22:32.560 --> 1:22:33.560 | |
It's been awesome. | |
1:22:33.560 --> 1:22:34.400 | |
Thank you so much. | |
1:22:34.400 --> 1:22:35.240 | |
Thanks. | |
1:22:35.240 --> 1:22:40.240 | |
Have a great day. | |