diff --git "a/vtt/episode_001_large.vtt" "b/vtt/episode_001_large.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_001_large.vtt" @@ -0,0 +1,5342 @@ +WEBVTT + +00:00.000 --> 00:04.200 + As part of MIT course 6S099, Artificial General Intelligence, + +00:04.200 --> 00:06.600 + I've gotten the chance to sit down with Max Tegmark. + +00:06.600 --> 00:08.680 + He is a professor here at MIT. + +00:08.680 --> 00:11.920 + He's a physicist, spent a large part of his career + +00:11.920 --> 00:16.960 + studying the mysteries of our cosmological universe. + +00:16.960 --> 00:20.680 + But he's also studied and delved into the beneficial + +00:20.680 --> 00:24.000 + possibilities and the existential risks + +00:24.000 --> 00:25.800 + of artificial intelligence. + +00:25.800 --> 00:29.040 + Amongst many other things, he is the cofounder + +00:29.040 --> 00:33.080 + of the Future of Life Institute, author of two books, + +00:33.080 --> 00:35.160 + both of which I highly recommend. + +00:35.160 --> 00:37.260 + First, Our Mathematical Universe. + +00:37.260 --> 00:40.160 + Second is Life 3.0. + +00:40.160 --> 00:44.080 + He's truly an out of the box thinker and a fun personality, + +00:44.080 --> 00:45.480 + so I really enjoy talking to him. + +00:45.480 --> 00:47.980 + If you'd like to see more of these videos in the future, + +00:47.980 --> 00:50.640 + please subscribe and also click the little bell icon + +00:50.640 --> 00:52.720 + to make sure you don't miss any videos. + +00:52.720 --> 00:56.840 + Also, Twitter, LinkedIn, agi.mit.edu + +00:56.840 --> 00:59.600 + if you wanna watch other lectures + +00:59.600 --> 01:01.080 + or conversations like this one. + +01:01.080 --> 01:04.000 + Better yet, go read Max's book, Life 3.0. + +01:04.000 --> 01:07.940 + Chapter seven on goals is my favorite. + +01:07.940 --> 01:10.480 + It's really where philosophy and engineering come together + +01:10.480 --> 01:13.440 + and it opens with a quote by Dostoevsky. + +01:14.400 --> 01:17.940 + The mystery of human existence lies not in just staying alive + +01:17.940 --> 01:20.520 + but in finding something to live for. + +01:20.520 --> 01:23.920 + Lastly, I believe that every failure rewards us + +01:23.920 --> 01:26.560 + with an opportunity to learn + +01:26.560 --> 01:28.360 + and in that sense, I've been very fortunate + +01:28.360 --> 01:30.960 + to fail in so many new and exciting ways + +01:31.840 --> 01:34.020 + and this conversation was no different. + +01:34.020 --> 01:36.160 + I've learned about something called + +01:36.160 --> 01:40.840 + radio frequency interference, RFI, look it up. + +01:40.840 --> 01:42.960 + Apparently, music and conversations + +01:42.960 --> 01:45.480 + from local radio stations can bleed into the audio + +01:45.480 --> 01:47.080 + that you're recording in such a way + +01:47.080 --> 01:49.360 + that it almost completely ruins that audio. + +01:49.360 --> 01:52.060 + It's an exceptionally difficult sound source to remove. + +01:53.240 --> 01:55.520 + So, I've gotten the opportunity to learn + +01:55.520 --> 02:00.200 + how to avoid RFI in the future during recording sessions. + +02:00.200 --> 02:02.680 + I've also gotten the opportunity to learn + +02:02.680 --> 02:06.240 + how to use Adobe Audition and iZotope RX 6 + +02:06.240 --> 02:11.240 + to do some noise, some audio repair. + +02:11.720 --> 02:14.380 + Of course, this is an exceptionally difficult noise + +02:14.380 --> 02:15.220 + to remove. + +02:15.220 --> 02:16.280 + I am an engineer. + +02:16.280 --> 02:18.240 + I'm not an audio engineer. + +02:18.240 --> 02:20.180 + Neither is anybody else in our group + +02:20.180 --> 02:21.880 + but we did our best. + +02:21.880 --> 02:25.040 + Nevertheless, I thank you for your patience + +02:25.040 --> 02:27.960 + and I hope you're still able to enjoy this conversation. + +02:27.960 --> 02:29.320 + Do you think there's intelligent life + +02:29.320 --> 02:31.360 + out there in the universe? + +02:31.360 --> 02:33.480 + Let's open up with an easy question. + +02:33.480 --> 02:36.240 + I have a minority view here actually. + +02:36.240 --> 02:39.440 + When I give public lectures, I often ask for a show of hands + +02:39.440 --> 02:42.920 + who thinks there's intelligent life out there somewhere else + +02:42.920 --> 02:45.440 + and almost everyone put their hands up + +02:45.440 --> 02:47.360 + and when I ask why, they'll be like, + +02:47.360 --> 02:50.900 + oh, there's so many galaxies out there, there's gotta be. + +02:51.840 --> 02:54.560 + But I'm a numbers nerd, right? + +02:54.560 --> 02:56.640 + So when you look more carefully at it, + +02:56.640 --> 02:58.040 + it's not so clear at all. + +02:59.080 --> 03:00.680 + When we talk about our universe, first of all, + +03:00.680 --> 03:03.040 + we don't mean all of space. + +03:03.040 --> 03:04.040 + We actually mean, I don't know, + +03:04.040 --> 03:05.440 + you can throw me the universe if you want, + +03:05.440 --> 03:07.280 + it's behind you there. + +03:07.280 --> 03:11.440 + It's, we simply mean the spherical region of space + +03:11.440 --> 03:15.360 + from which light has a time to reach us so far + +03:15.360 --> 03:17.040 + during the 14.8 billion year, + +03:17.040 --> 03:19.320 + 13.8 billion years since our Big Bang. + +03:19.320 --> 03:22.320 + There's more space here but this is what we call a universe + +03:22.320 --> 03:24.040 + because that's all we have access to. + +03:24.040 --> 03:25.960 + So is there intelligent life here + +03:25.960 --> 03:28.920 + that's gotten to the point of building telescopes + +03:28.920 --> 03:29.960 + and computers? + +03:31.160 --> 03:34.540 + My guess is no, actually. + +03:34.540 --> 03:37.800 + The probability of it happening on any given planet + +03:39.240 --> 03:42.620 + is some number we don't know what it is. + +03:42.620 --> 03:47.620 + And what we do know is that the number can't be super high + +03:48.480 --> 03:50.300 + because there's over a billion Earth like planets + +03:50.300 --> 03:52.880 + in the Milky Way galaxy alone, + +03:52.880 --> 03:56.280 + many of which are billions of years older than Earth. + +03:56.280 --> 04:00.600 + And aside from some UFO believers, + +04:00.600 --> 04:01.880 + there isn't much evidence + +04:01.880 --> 04:05.600 + that any superduran civilization has come here at all. + +04:05.600 --> 04:08.440 + And so that's the famous Fermi paradox, right? + +04:08.440 --> 04:10.180 + And then if you work the numbers, + +04:10.180 --> 04:13.440 + what you find is that if you have no clue + +04:13.440 --> 04:16.880 + what the probability is of getting life on a given planet, + +04:16.880 --> 04:19.680 + so it could be 10 to the minus 10, 10 to the minus 20, + +04:19.680 --> 04:22.960 + or 10 to the minus two, or any power of 10 + +04:22.960 --> 04:23.800 + is sort of equally likely + +04:23.800 --> 04:25.480 + if you wanna be really open minded, + +04:25.480 --> 04:27.600 + that translates into it being equally likely + +04:27.600 --> 04:31.800 + that our nearest neighbor is 10 to the 16 meters away, + +04:31.800 --> 04:33.880 + 10 to the 17 meters away, 10 to the 18. + +04:35.400 --> 04:40.400 + By the time you get much less than 10 to the 16 already, + +04:41.080 --> 04:45.960 + we pretty much know there is nothing else that close. + +04:45.960 --> 04:47.280 + And when you get beyond 10. + +04:47.280 --> 04:48.680 + Because they would have discovered us. + +04:48.680 --> 04:50.360 + Yeah, they would have been discovered as long ago, + +04:50.360 --> 04:51.440 + or if they're really close, + +04:51.440 --> 04:53.560 + we would have probably noted some engineering projects + +04:53.560 --> 04:54.640 + that they're doing. + +04:54.640 --> 04:57.880 + And if it's beyond 10 to the 26 meters, + +04:57.880 --> 05:00.000 + that's already outside of here. + +05:00.000 --> 05:05.000 + So my guess is actually that we are the only life in here + +05:05.800 --> 05:09.040 + that's gotten the point of building advanced tech, + +05:09.040 --> 05:10.720 + which I think is very, + +05:12.680 --> 05:15.360 + puts a lot of responsibility on our shoulders, not screw up. + +05:15.360 --> 05:17.240 + I think people who take for granted + +05:17.240 --> 05:20.120 + that it's okay for us to screw up, + +05:20.120 --> 05:22.760 + have an accidental nuclear war or go extinct somehow + +05:22.760 --> 05:25.960 + because there's a sort of Star Trek like situation out there + +05:25.960 --> 05:28.360 + where some other life forms are gonna come and bail us out + +05:28.360 --> 05:30.400 + and it doesn't matter as much. + +05:30.400 --> 05:33.400 + I think they're leveling us into a false sense of security. + +05:33.400 --> 05:35.200 + I think it's much more prudent to say, + +05:35.200 --> 05:36.400 + let's be really grateful + +05:36.400 --> 05:38.720 + for this amazing opportunity we've had + +05:38.720 --> 05:43.720 + and make the best of it just in case it is down to us. + +05:44.080 --> 05:45.680 + So from a physics perspective, + +05:45.680 --> 05:48.800 + do you think intelligent life, + +05:48.800 --> 05:51.360 + so it's unique from a sort of statistical view + +05:51.360 --> 05:52.560 + of the size of the universe, + +05:52.560 --> 05:55.840 + but from the basic matter of the universe, + +05:55.840 --> 05:59.040 + how difficult is it for intelligent life to come about? + +05:59.040 --> 06:01.280 + The kind of advanced tech building life + +06:03.120 --> 06:05.720 + is implied in your statement that it's really difficult + +06:05.720 --> 06:07.640 + to create something like a human species. + +06:07.640 --> 06:11.560 + Well, I think what we know is that going from no life + +06:11.560 --> 06:15.720 + to having life that can do a level of tech, + +06:15.720 --> 06:18.720 + there's some sort of two going beyond that + +06:18.720 --> 06:22.200 + than actually settling our whole universe with life. + +06:22.200 --> 06:26.560 + There's some major roadblock there, + +06:26.560 --> 06:30.880 + which is some great filter as it's sometimes called, + +06:30.880 --> 06:33.520 + which is tough to get through. + +06:33.520 --> 06:37.160 + It's either that roadblock is either behind us + +06:37.160 --> 06:38.720 + or in front of us. + +06:38.720 --> 06:41.080 + I'm hoping very much that it's behind us. + +06:41.080 --> 06:45.960 + I'm super excited every time we get a new report from NASA + +06:45.960 --> 06:48.480 + saying they failed to find any life on Mars. + +06:48.480 --> 06:50.080 + I'm like, yes, awesome. + +06:50.080 --> 06:51.680 + Because that suggests that the hard part, + +06:51.680 --> 06:54.240 + maybe it was getting the first ribosome + +06:54.240 --> 06:59.240 + or some very low level kind of stepping stone + +06:59.520 --> 07:00.400 + so that we're home free. + +07:00.400 --> 07:01.720 + Because if that's true, + +07:01.720 --> 07:03.640 + then the future is really only limited + +07:03.640 --> 07:05.200 + by our own imagination. + +07:05.200 --> 07:07.360 + It would be much suckier if it turns out + +07:07.360 --> 07:11.440 + that this level of life is kind of a dime a dozen, + +07:11.440 --> 07:12.760 + but maybe there's some other problem. + +07:12.760 --> 07:16.160 + Like as soon as a civilization gets advanced technology, + +07:16.160 --> 07:17.000 + within a hundred years, + +07:17.000 --> 07:20.320 + they get into some stupid fight with themselves and poof. + +07:20.320 --> 07:21.760 + That would be a bummer. + +07:21.760 --> 07:26.160 + Yeah, so you've explored the mysteries of the universe, + +07:26.160 --> 07:29.000 + the cosmological universe, the one that's sitting + +07:29.000 --> 07:30.000 + between us today. + +07:31.080 --> 07:35.960 + I think you've also begun to explore the other universe, + +07:35.960 --> 07:38.000 + which is sort of the mystery, + +07:38.000 --> 07:40.960 + the mysterious universe of the mind of intelligence, + +07:40.960 --> 07:42.840 + of intelligent life. + +07:42.840 --> 07:45.280 + So is there a common thread between your interest + +07:45.280 --> 07:48.760 + or the way you think about space and intelligence? + +07:48.760 --> 07:51.040 + Oh yeah, when I was a teenager, + +07:53.040 --> 07:57.280 + I was already very fascinated by the biggest questions. + +07:57.280 --> 08:00.560 + And I felt that the two biggest mysteries of all in science + +08:00.560 --> 08:05.000 + were our universe out there and our universe in here. + +08:05.000 --> 08:08.120 + So it's quite natural after having spent + +08:08.120 --> 08:11.040 + a quarter of a century on my career, + +08:11.040 --> 08:12.680 + thinking a lot about this one, + +08:12.680 --> 08:14.320 + that I'm now indulging in the luxury + +08:14.320 --> 08:15.960 + of doing research on this one. + +08:15.960 --> 08:17.720 + It's just so cool. + +08:17.720 --> 08:20.120 + I feel the time is ripe now + +08:20.120 --> 08:25.120 + for you trans greatly deepening our understanding of this. + +08:25.120 --> 08:26.640 + Just start exploring this one. + +08:26.640 --> 08:29.560 + Yeah, because I think a lot of people view intelligence + +08:29.560 --> 08:33.520 + as something mysterious that can only exist + +08:33.520 --> 08:36.120 + in biological organisms like us, + +08:36.120 --> 08:37.680 + and therefore dismiss all talk + +08:37.680 --> 08:41.160 + about artificial general intelligence as science fiction. + +08:41.160 --> 08:43.200 + But from my perspective as a physicist, + +08:43.200 --> 08:46.680 + I am a blob of quarks and electrons + +08:46.680 --> 08:48.360 + moving around in a certain pattern + +08:48.360 --> 08:50.080 + and processing information in certain ways. + +08:50.080 --> 08:53.600 + And this is also a blob of quarks and electrons. + +08:53.600 --> 08:55.360 + I'm not smarter than the water bottle + +08:55.360 --> 08:57.880 + because I'm made of different kinds of quarks. + +08:57.880 --> 08:59.640 + I'm made of up quarks and down quarks, + +08:59.640 --> 09:01.400 + exact same kind as this. + +09:01.400 --> 09:05.080 + There's no secret sauce, I think, in me. + +09:05.080 --> 09:08.560 + It's all about the pattern of the information processing. + +09:08.560 --> 09:12.240 + And this means that there's no law of physics + +09:12.240 --> 09:15.600 + saying that we can't create technology, + +09:15.600 --> 09:19.960 + which can help us by being incredibly intelligent + +09:19.960 --> 09:21.680 + and help us crack mysteries that we couldn't. + +09:21.680 --> 09:23.560 + In other words, I think we've really only seen + +09:23.560 --> 09:26.480 + the tip of the intelligence iceberg so far. + +09:26.480 --> 09:29.960 + Yeah, so the perceptronium. + +09:29.960 --> 09:31.280 + Yeah. + +09:31.280 --> 09:33.200 + So you coined this amazing term. + +09:33.200 --> 09:35.760 + It's a hypothetical state of matter, + +09:35.760 --> 09:38.360 + sort of thinking from a physics perspective, + +09:38.360 --> 09:40.080 + what is the kind of matter that can help, + +09:40.080 --> 09:42.920 + as you're saying, subjective experience emerge, + +09:42.920 --> 09:44.280 + consciousness emerge. + +09:44.280 --> 09:46.640 + So how do you think about consciousness + +09:46.640 --> 09:48.160 + from this physics perspective? + +09:49.960 --> 09:50.800 + Very good question. + +09:50.800 --> 09:55.800 + So again, I think many people have underestimated + +09:55.800 --> 09:59.120 + our ability to make progress on this + +09:59.120 --> 10:01.320 + by convincing themselves it's hopeless + +10:01.320 --> 10:05.840 + because somehow we're missing some ingredient that we need. + +10:05.840 --> 10:09.560 + There's some new consciousness particle or whatever. + +10:09.560 --> 10:12.720 + I happen to think that we're not missing anything + +10:12.720 --> 10:16.320 + and that it's not the interesting thing + +10:16.320 --> 10:18.560 + about consciousness that gives us + +10:18.560 --> 10:21.400 + this amazing subjective experience of colors + +10:21.400 --> 10:23.320 + and sounds and emotions. + +10:23.320 --> 10:26.320 + It's rather something at the higher level + +10:26.320 --> 10:28.800 + about the patterns of information processing. + +10:28.800 --> 10:33.160 + And that's why I like to think about this idea + +10:33.160 --> 10:34.480 + of perceptronium. + +10:34.480 --> 10:36.920 + What does it mean for an arbitrary physical system + +10:36.920 --> 10:41.920 + to be conscious in terms of what its particles are doing + +10:41.920 --> 10:43.560 + or its information is doing? + +10:43.560 --> 10:46.080 + I don't think, I hate carbon chauvinism, + +10:46.080 --> 10:47.960 + this attitude you have to be made of carbon atoms + +10:47.960 --> 10:50.160 + to be smart or conscious. + +10:50.160 --> 10:53.520 + There's something about the information processing + +10:53.520 --> 10:55.360 + that this kind of matter performs. + +10:55.360 --> 10:57.840 + Yeah, and you can see I have my favorite equations here + +10:57.840 --> 11:00.720 + describing various fundamental aspects of the world. + +11:00.720 --> 11:02.560 + I feel that I think one day, + +11:02.560 --> 11:04.360 + maybe someone who's watching this will come up + +11:04.360 --> 11:07.280 + with the equations that information processing + +11:07.280 --> 11:08.760 + has to satisfy to be conscious. + +11:08.760 --> 11:11.800 + I'm quite convinced there is big discovery + +11:11.800 --> 11:15.400 + to be made there because let's face it, + +11:15.400 --> 11:18.720 + we know that so many things are made up of information. + +11:18.720 --> 11:21.960 + We know that some information processing is conscious + +11:21.960 --> 11:25.520 + because we are conscious. + +11:25.520 --> 11:27.600 + But we also know that a lot of information processing + +11:27.600 --> 11:28.440 + is not conscious. + +11:28.440 --> 11:30.040 + Like most of the information processing happening + +11:30.040 --> 11:32.680 + in your brain right now is not conscious. + +11:32.680 --> 11:36.040 + There are like 10 megabytes per second coming in + +11:36.040 --> 11:38.080 + even just through your visual system. + +11:38.080 --> 11:40.480 + You're not conscious about your heartbeat regulation + +11:40.480 --> 11:42.120 + or most things. + +11:42.120 --> 11:45.680 + Even if I just ask you to like read what it says here, + +11:45.680 --> 11:48.040 + you look at it and then, oh, now you know what it said. + +11:48.040 --> 11:51.560 + But you're not aware of how the computation actually happened. + +11:51.560 --> 11:53.680 + Your consciousness is like the CEO + +11:53.680 --> 11:56.680 + that got an email at the end with the final answer. + +11:56.680 --> 12:01.000 + So what is it that makes a difference? + +12:01.000 --> 12:05.120 + I think that's both a great science mystery. + +12:05.120 --> 12:07.080 + We're actually studying it a little bit in my lab here + +12:07.080 --> 12:10.920 + at MIT, but I also think it's just a really urgent question + +12:10.920 --> 12:12.080 + to answer. + +12:12.080 --> 12:14.880 + For starters, I mean, if you're an emergency room doctor + +12:14.880 --> 12:17.160 + and you have an unresponsive patient coming in, + +12:17.160 --> 12:19.600 + wouldn't it be great if in addition to having + +12:22.360 --> 12:25.320 + a CT scanner, you had a consciousness scanner + +12:25.320 --> 12:27.920 + that could figure out whether this person + +12:27.920 --> 12:30.960 + is actually having locked in syndrome + +12:30.960 --> 12:32.440 + or is actually comatose. + +12:33.360 --> 12:37.000 + And in the future, imagine if we build robots + +12:37.000 --> 12:41.480 + or the machine that we can have really good conversations + +12:41.480 --> 12:44.840 + with, which I think is very likely to happen. + +12:44.840 --> 12:47.760 + Wouldn't you want to know if your home helper robot + +12:47.760 --> 12:51.320 + is actually experiencing anything or just like a zombie, + +12:51.320 --> 12:53.520 + I mean, would you prefer it? + +12:53.520 --> 12:54.360 + What would you prefer? + +12:54.360 --> 12:56.200 + Would you prefer that it's actually unconscious + +12:56.200 --> 12:58.560 + so that you don't have to feel guilty about switching it off + +12:58.560 --> 13:02.120 + or giving boring chores or what would you prefer? + +13:02.120 --> 13:06.520 + Well, certainly we would prefer, + +13:06.520 --> 13:08.960 + I would prefer the appearance of consciousness. + +13:08.960 --> 13:11.720 + But the question is whether the appearance of consciousness + +13:11.720 --> 13:15.040 + is different than consciousness itself. + +13:15.040 --> 13:18.200 + And sort of to ask that as a question, + +13:18.200 --> 13:21.760 + do you think we need to understand what consciousness is, + +13:21.760 --> 13:23.520 + solve the hard problem of consciousness + +13:23.520 --> 13:28.240 + in order to build something like an AGI system? + +13:28.240 --> 13:30.440 + No, I don't think that. + +13:30.440 --> 13:34.520 + And I think we will probably be able to build things + +13:34.520 --> 13:36.080 + even if we don't answer that question. + +13:36.080 --> 13:37.720 + But if we want to make sure that what happens + +13:37.720 --> 13:40.960 + is a good thing, we better solve it first. + +13:40.960 --> 13:44.960 + So it's a wonderful controversy you're raising there + +13:44.960 --> 13:47.960 + where you have basically three points of view + +13:47.960 --> 13:48.800 + about the hard problem. + +13:48.800 --> 13:52.800 + So there are two different points of view. + +13:52.800 --> 13:55.160 + They both conclude that the hard problem of consciousness + +13:55.160 --> 13:56.840 + is BS. + +13:56.840 --> 13:59.320 + On one hand, you have some people like Daniel Dennett + +13:59.320 --> 14:01.480 + who say that consciousness is just BS + +14:01.480 --> 14:05.000 + because consciousness is the same thing as intelligence. + +14:05.000 --> 14:06.440 + There's no difference. + +14:06.440 --> 14:11.080 + So anything which acts conscious is conscious, + +14:11.080 --> 14:13.480 + just like we are. + +14:13.480 --> 14:15.960 + And then there are also a lot of people, + +14:15.960 --> 14:18.400 + including many top AI researchers I know, + +14:18.400 --> 14:19.920 + who say, oh, consciousness is just bullshit + +14:19.920 --> 14:22.760 + because, of course, machines can never be conscious. + +14:22.760 --> 14:24.520 + They're always going to be zombies. + +14:24.520 --> 14:27.880 + You never have to feel guilty about how you treat them. + +14:27.880 --> 14:30.880 + And then there's a third group of people, + +14:30.880 --> 14:34.920 + including Giulio Tononi, for example, + +14:34.920 --> 14:37.440 + and Krzysztof Koch and a number of others. + +14:37.440 --> 14:39.520 + I would put myself also in this middle camp + +14:39.520 --> 14:41.880 + who say that actually some information processing + +14:41.880 --> 14:44.160 + is conscious and some is not. + +14:44.160 --> 14:46.960 + So let's find the equation which can be used + +14:46.960 --> 14:49.080 + to determine which it is. + +14:49.080 --> 14:52.040 + And I think we've just been a little bit lazy, + +14:52.040 --> 14:54.960 + kind of running away from this problem for a long time. + +14:54.960 --> 14:57.840 + It's been almost taboo to even mention the C word + +14:57.840 --> 15:00.520 + in a lot of circles because, + +15:00.520 --> 15:03.520 + but we should stop making excuses. + +15:03.520 --> 15:07.920 + This is a science question and there are ways + +15:07.920 --> 15:11.960 + we can even test any theory that makes predictions for this. + +15:11.960 --> 15:13.640 + And coming back to this helper robot, + +15:13.640 --> 15:16.080 + I mean, so you said you'd want your helper robot + +15:16.080 --> 15:18.160 + to certainly act conscious and treat you, + +15:18.160 --> 15:20.880 + like have conversations with you and stuff. + +15:20.880 --> 15:21.720 + I think so. + +15:21.720 --> 15:22.560 + But wouldn't you, would you feel, + +15:22.560 --> 15:23.920 + would you feel a little bit creeped out + +15:23.920 --> 15:27.680 + if you realized that it was just a glossed up tape recorder, + +15:27.680 --> 15:31.560 + you know, that was just zombie and was a faking emotion? + +15:31.560 --> 15:34.560 + Would you prefer that it actually had an experience + +15:34.560 --> 15:37.000 + or would you prefer that it's actually + +15:37.000 --> 15:39.120 + not experiencing anything so you feel, + +15:39.120 --> 15:42.200 + you don't have to feel guilty about what you do to it? + +15:42.200 --> 15:45.040 + It's such a difficult question because, you know, + +15:45.040 --> 15:47.280 + it's like when you're in a relationship and you say, + +15:47.280 --> 15:48.120 + well, I love you. + +15:48.120 --> 15:49.760 + And the other person said, I love you back. + +15:49.760 --> 15:52.640 + It's like asking, well, do they really love you back + +15:52.640 --> 15:55.360 + or are they just saying they love you back? + +15:55.360 --> 15:58.120 + Don't you really want them to actually love you? + +15:58.120 --> 16:03.120 + It's hard to, it's hard to really know the difference + +16:03.520 --> 16:08.520 + between everything seeming like there's consciousness + +16:09.000 --> 16:10.640 + present, there's intelligence present, + +16:10.640 --> 16:13.840 + there's affection, passion, love, + +16:13.840 --> 16:16.200 + and it actually being there. + +16:16.200 --> 16:17.720 + I'm not sure, do you have? + +16:17.720 --> 16:19.400 + But like, can I ask you a question about this? + +16:19.400 --> 16:20.760 + Like to make it a bit more pointed. + +16:20.760 --> 16:22.920 + So Mass General Hospital is right across the river, right? + +16:22.920 --> 16:23.760 + Yes. + +16:23.760 --> 16:26.720 + Suppose you're going in for a medical procedure + +16:26.720 --> 16:29.320 + and they're like, you know, for anesthesia, + +16:29.320 --> 16:31.000 + what we're going to do is we're going to give you + +16:31.000 --> 16:33.160 + muscle relaxants so you won't be able to move + +16:33.160 --> 16:35.040 + and you're going to feel excruciating pain + +16:35.040 --> 16:35.880 + during the whole surgery, + +16:35.880 --> 16:37.600 + but you won't be able to do anything about it. + +16:37.600 --> 16:39.200 + But then we're going to give you this drug + +16:39.200 --> 16:40.760 + that erases your memory of it. + +16:41.960 --> 16:43.440 + Would you be cool about that? + +16:44.960 --> 16:47.600 + What's the difference that you're conscious about it + +16:48.600 --> 16:51.640 + or not if there's no behavioral change, right? + +16:51.640 --> 16:54.520 + Right, that's a really, that's a really clear way to put it. + +16:54.520 --> 16:57.400 + That's, yeah, it feels like in that sense, + +16:57.400 --> 17:01.080 + experiencing it is a valuable quality. + +17:01.080 --> 17:04.800 + So actually being able to have subjective experiences, + +17:05.840 --> 17:09.120 + at least in that case, is valuable. + +17:09.120 --> 17:11.240 + And I think we humans have a little bit + +17:11.240 --> 17:13.600 + of a bad track record also of making + +17:13.600 --> 17:15.480 + these self serving arguments + +17:15.480 --> 17:18.040 + that other entities aren't conscious. + +17:18.040 --> 17:19.160 + You know, people often say, + +17:19.160 --> 17:21.800 + oh, these animals can't feel pain. + +17:21.800 --> 17:24.040 + It's okay to boil lobsters because we ask them + +17:24.040 --> 17:25.960 + if it hurt and they didn't say anything. + +17:25.960 --> 17:27.400 + And now there was just a paper out saying, + +17:27.400 --> 17:29.320 + lobsters do feel pain when you boil them + +17:29.320 --> 17:31.040 + and they're banning it in Switzerland. + +17:31.040 --> 17:33.560 + And we did this with slaves too often and said, + +17:33.560 --> 17:34.680 + oh, they don't mind. + +17:36.240 --> 17:39.480 + They don't maybe aren't conscious + +17:39.480 --> 17:41.160 + or women don't have souls or whatever. + +17:41.160 --> 17:43.200 + So I'm a little bit nervous when I hear people + +17:43.200 --> 17:46.360 + just take as an axiom that machines + +17:46.360 --> 17:48.960 + can't have experience ever. + +17:48.960 --> 17:51.560 + I think this is just a really fascinating science question + +17:51.560 --> 17:52.400 + is what it is. + +17:52.400 --> 17:54.720 + Let's research it and try to figure out + +17:54.720 --> 17:56.000 + what it is that makes the difference + +17:56.000 --> 17:58.880 + between unconscious intelligent behavior + +17:58.880 --> 18:01.120 + and conscious intelligent behavior. + +18:01.120 --> 18:04.680 + So in terms of, so if you think of a Boston Dynamics + +18:04.680 --> 18:07.680 + human or robot being sort of with a broom + +18:07.680 --> 18:11.920 + being pushed around, it starts pushing + +18:11.920 --> 18:13.320 + on a consciousness question. + +18:13.320 --> 18:17.040 + So let me ask, do you think an AGI system + +18:17.040 --> 18:19.720 + like a few neuroscientists believe + +18:19.720 --> 18:22.320 + needs to have a physical embodiment? + +18:22.320 --> 18:25.720 + Needs to have a body or something like a body? + +18:25.720 --> 18:28.280 + No, I don't think so. + +18:28.280 --> 18:30.560 + You mean to have a conscious experience? + +18:30.560 --> 18:31.640 + To have consciousness. + +18:33.160 --> 18:36.080 + I do think it helps a lot to have a physical embodiment + +18:36.080 --> 18:38.440 + to learn the kind of things about the world + +18:38.440 --> 18:41.480 + that are important to us humans, for sure. + +18:42.560 --> 18:45.600 + But I don't think the physical embodiment + +18:45.600 --> 18:47.120 + is necessary after you've learned it + +18:47.120 --> 18:48.760 + to just have the experience. + +18:48.760 --> 18:51.400 + Think about when you're dreaming, right? + +18:51.400 --> 18:52.600 + Your eyes are closed. + +18:52.600 --> 18:54.240 + You're not getting any sensory input. + +18:54.240 --> 18:55.960 + You're not behaving or moving in any way + +18:55.960 --> 18:58.160 + but there's still an experience there, right? + +18:59.720 --> 19:01.400 + And so clearly the experience that you have + +19:01.400 --> 19:03.320 + when you see something cool in your dreams + +19:03.320 --> 19:04.800 + isn't coming from your eyes. + +19:04.800 --> 19:08.640 + It's just the information processing itself in your brain + +19:08.640 --> 19:10.920 + which is that experience, right? + +19:10.920 --> 19:13.640 + But if I put it another way, I'll say + +19:13.640 --> 19:15.120 + because it comes from neuroscience + +19:15.120 --> 19:18.280 + is the reason you want to have a body and a physical + +19:18.280 --> 19:23.280 + something like a physical, you know, a physical system + +19:23.920 --> 19:27.040 + is because you want to be able to preserve something. + +19:27.040 --> 19:30.840 + In order to have a self, you could argue, + +19:30.840 --> 19:35.840 + would you need to have some kind of embodiment of self + +19:36.400 --> 19:37.960 + to want to preserve? + +19:38.920 --> 19:42.400 + Well, now we're getting a little bit anthropomorphic + +19:42.400 --> 19:45.200 + into anthropomorphizing things. + +19:45.200 --> 19:47.280 + Maybe talking about self preservation instincts. + +19:47.280 --> 19:50.560 + I mean, we are evolved organisms, right? + +19:50.560 --> 19:53.520 + So Darwinian evolution endowed us + +19:53.520 --> 19:57.120 + and other evolved organism with a self preservation instinct + +19:57.120 --> 20:00.560 + because those that didn't have those self preservation genes + +20:00.560 --> 20:02.960 + got cleaned out of the gene pool, right? + +20:02.960 --> 20:06.880 + But if you build an artificial general intelligence + +20:06.880 --> 20:10.040 + the mind space that you can design is much, much larger + +20:10.040 --> 20:14.440 + than just a specific subset of minds that can evolve. + +20:14.440 --> 20:17.280 + So an AGI mind doesn't necessarily have + +20:17.280 --> 20:19.880 + to have any self preservation instinct. + +20:19.880 --> 20:21.600 + It also doesn't necessarily have to be + +20:21.600 --> 20:24.040 + so individualistic as us. + +20:24.040 --> 20:26.080 + Like, imagine if you could just, first of all, + +20:26.080 --> 20:27.960 + or we are also very afraid of death. + +20:27.960 --> 20:29.920 + You know, I suppose you could back yourself up + +20:29.920 --> 20:32.000 + every five minutes and then your airplane + +20:32.000 --> 20:32.840 + is about to crash. + +20:32.840 --> 20:36.680 + You're like, shucks, I'm gonna lose the last five minutes + +20:36.680 --> 20:39.520 + of experiences since my last cloud backup, dang. + +20:39.520 --> 20:41.520 + You know, it's not as big a deal. + +20:41.520 --> 20:45.680 + Or if we could just copy experiences between our minds + +20:45.680 --> 20:47.640 + easily like we, which we could easily do + +20:47.640 --> 20:50.360 + if we were silicon based, right? + +20:50.360 --> 20:54.040 + Then maybe we would feel a little bit more + +20:54.040 --> 20:56.560 + like a hive mind actually, that maybe it's the, + +20:56.560 --> 20:59.960 + so I don't think we should take for granted at all + +20:59.960 --> 21:03.000 + that AGI will have to have any of those sort of + +21:04.880 --> 21:07.360 + competitive as alpha male instincts. + +21:07.360 --> 21:10.160 + On the other hand, you know, this is really interesting + +21:10.160 --> 21:13.840 + because I think some people go too far and say, + +21:13.840 --> 21:16.680 + of course we don't have to have any concerns either + +21:16.680 --> 21:20.800 + that advanced AI will have those instincts + +21:20.800 --> 21:22.680 + because we can build anything we want. + +21:22.680 --> 21:26.280 + That there's a very nice set of arguments going back + +21:26.280 --> 21:28.560 + to Steve Omohundro and Nick Bostrom and others + +21:28.560 --> 21:32.280 + just pointing out that when we build machines, + +21:32.280 --> 21:34.680 + we normally build them with some kind of goal, you know, + +21:34.680 --> 21:38.520 + win this chess game, drive this car safely or whatever. + +21:38.520 --> 21:40.960 + And as soon as you put in a goal into machine, + +21:40.960 --> 21:42.760 + especially if it's kind of open ended goal + +21:42.760 --> 21:44.640 + and the machine is very intelligent, + +21:44.640 --> 21:47.000 + it'll break that down into a bunch of sub goals. + +21:48.280 --> 21:51.280 + And one of those goals will almost always + +21:51.280 --> 21:54.200 + be self preservation because if it breaks or dies + +21:54.200 --> 21:56.120 + in the process, it's not gonna accomplish the goal, right? + +21:56.120 --> 21:58.040 + Like suppose you just build a little, + +21:58.040 --> 22:01.000 + you have a little robot and you tell it to go down + +22:01.000 --> 22:04.040 + the store market here and get you some food, + +22:04.040 --> 22:06.200 + make you cook an Italian dinner, you know, + +22:06.200 --> 22:08.400 + and then someone mugs it and tries to break it + +22:08.400 --> 22:09.480 + on the way. + +22:09.480 --> 22:12.920 + That robot has an incentive to not get destroyed + +22:12.920 --> 22:14.720 + and defend itself or run away, + +22:14.720 --> 22:17.720 + because otherwise it's gonna fail in cooking your dinner. + +22:17.720 --> 22:19.560 + It's not afraid of death, + +22:19.560 --> 22:22.960 + but it really wants to complete the dinner cooking goal. + +22:22.960 --> 22:25.040 + So it will have a self preservation instinct. + +22:25.040 --> 22:27.920 + Continue being a functional agent somehow. + +22:27.920 --> 22:32.920 + And similarly, if you give any kind of more ambitious goal + +22:33.720 --> 22:37.000 + to an AGI, it's very likely they wanna acquire + +22:37.000 --> 22:39.840 + more resources so it can do that better. + +22:39.840 --> 22:42.720 + And it's exactly from those sort of sub goals + +22:42.720 --> 22:43.800 + that we might not have intended + +22:43.800 --> 22:47.160 + that some of the concerns about AGI safety come. + +22:47.160 --> 22:50.600 + You give it some goal that seems completely harmless. + +22:50.600 --> 22:53.360 + And then before you realize it, + +22:53.360 --> 22:55.480 + it's also trying to do these other things + +22:55.480 --> 22:56.920 + which you didn't want it to do. + +22:56.920 --> 22:59.160 + And it's maybe smarter than us. + +22:59.160 --> 23:01.000 + So it's fascinating. + +23:01.000 --> 23:05.680 + And let me pause just because I am in a very kind + +23:05.680 --> 23:08.720 + of human centric way, see fear of death + +23:08.720 --> 23:11.840 + as a valuable motivator. + +23:11.840 --> 23:16.440 + So you don't think, you think that's an artifact + +23:16.440 --> 23:19.120 + of evolution, so that's the kind of mind space + +23:19.120 --> 23:22.120 + evolution created that we're sort of almost obsessed + +23:22.120 --> 23:24.400 + about self preservation, some kind of genetic flow. + +23:24.400 --> 23:29.400 + You don't think that's necessary to be afraid of death. + +23:29.480 --> 23:32.920 + So not just a kind of sub goal of self preservation + +23:32.920 --> 23:34.920 + just so you can keep doing the thing, + +23:34.920 --> 23:38.720 + but more fundamentally sort of have the finite thing + +23:38.720 --> 23:43.080 + like this ends for you at some point. + +23:43.080 --> 23:44.160 + Interesting. + +23:44.160 --> 23:47.440 + Do I think it's necessary for what precisely? + +23:47.440 --> 23:50.920 + For intelligence, but also for consciousness. + +23:50.920 --> 23:55.040 + So for those, for both, do you think really + +23:55.040 --> 23:59.120 + like a finite death and the fear of it is important? + +23:59.120 --> 24:04.120 + So before I can answer, before we can agree + +24:05.160 --> 24:06.960 + on whether it's necessary for intelligence + +24:06.960 --> 24:08.360 + or for consciousness, we should be clear + +24:08.360 --> 24:09.800 + on how we define those two words. + +24:09.800 --> 24:11.960 + Cause a lot of really smart people define them + +24:11.960 --> 24:13.320 + in very different ways. + +24:13.320 --> 24:17.080 + I was on this panel with AI experts + +24:17.080 --> 24:20.080 + and they couldn't agree on how to define intelligence even. + +24:20.080 --> 24:22.000 + So I define intelligence simply + +24:22.000 --> 24:24.760 + as the ability to accomplish complex goals. + +24:25.640 --> 24:27.280 + I like your broad definition, because again + +24:27.280 --> 24:29.040 + I don't want to be a carbon chauvinist. + +24:29.040 --> 24:30.400 + Right. + +24:30.400 --> 24:34.600 + And in that case, no, certainly + +24:34.600 --> 24:36.480 + it doesn't require fear of death. + +24:36.480 --> 24:40.120 + I would say alpha go, alpha zero is quite intelligent. + +24:40.120 --> 24:43.080 + I don't think alpha zero has any fear of being turned off + +24:43.080 --> 24:46.320 + because it doesn't understand the concept of it even. + +24:46.320 --> 24:48.440 + And similarly consciousness. + +24:48.440 --> 24:52.240 + I mean, you could certainly imagine very simple + +24:52.240 --> 24:53.920 + kind of experience. + +24:53.920 --> 24:57.200 + If certain plants have any kind of experience + +24:57.200 --> 24:58.560 + I don't think they're very afraid of dying + +24:58.560 --> 25:00.920 + or there's nothing they can do about it anyway much. + +25:00.920 --> 25:04.560 + So there wasn't that much value in, but more seriously + +25:04.560 --> 25:09.200 + I think if you ask, not just about being conscious + +25:09.200 --> 25:14.200 + but maybe having what you would, we might call + +25:14.320 --> 25:16.400 + an exciting life where you feel passion + +25:16.400 --> 25:21.400 + and really appreciate the things. + +25:21.480 --> 25:24.440 + Maybe there somehow, maybe there perhaps it does help + +25:24.440 --> 25:27.880 + having a backdrop that, Hey, it's finite. + +25:27.880 --> 25:31.200 + No, let's make the most of this, let's live to the fullest. + +25:31.200 --> 25:33.800 + So if you knew you were going to live forever + +25:34.880 --> 25:37.400 + do you think you would change your? + +25:37.400 --> 25:39.560 + Yeah, I mean, in some perspective + +25:39.560 --> 25:43.960 + it would be an incredibly boring life living forever. + +25:43.960 --> 25:47.360 + So in the sort of loose subjective terms that you said + +25:47.360 --> 25:50.480 + of something exciting and something in this + +25:50.480 --> 25:53.240 + that other humans would understand, I think is, yeah + +25:53.240 --> 25:57.120 + it seems that the finiteness of it is important. + +25:57.120 --> 25:59.560 + Well, the good news I have for you then is + +25:59.560 --> 26:02.120 + based on what we understand about cosmology + +26:02.120 --> 26:05.120 + everything is in our universe is probably + +26:05.120 --> 26:07.960 + ultimately probably finite, although. + +26:07.960 --> 26:11.560 + Big crunch or big, what's the, the infinite expansion. + +26:11.560 --> 26:13.840 + Yeah, we could have a big chill or a big crunch + +26:13.840 --> 26:18.440 + or a big rip or that's the big snap or death bubbles. + +26:18.440 --> 26:20.040 + All of them are more than a billion years away. + +26:20.040 --> 26:24.600 + So we should, we certainly have vastly more time + +26:24.600 --> 26:27.920 + than our ancestors thought, but there is still + +26:29.160 --> 26:32.360 + it's still pretty hard to squeeze in an infinite number + +26:32.360 --> 26:36.560 + of compute cycles, even though there are some loopholes + +26:36.560 --> 26:37.720 + that just might be possible. + +26:37.720 --> 26:41.960 + But I think, you know, some people like to say + +26:41.960 --> 26:44.760 + that you should live as if you're about to + +26:44.760 --> 26:46.720 + you're going to die in five years or so. + +26:46.720 --> 26:47.960 + And that's sort of optimal. + +26:47.960 --> 26:50.560 + Maybe it's a good assumption. + +26:50.560 --> 26:54.680 + We should build our civilization as if it's all finite + +26:54.680 --> 26:55.680 + to be on the safe side. + +26:55.680 --> 26:56.960 + Right, exactly. + +26:56.960 --> 26:59.720 + So you mentioned defining intelligence + +26:59.720 --> 27:02.960 + as the ability to solve complex goals. + +27:02.960 --> 27:05.440 + Where would you draw a line or how would you try + +27:05.440 --> 27:08.200 + to define human level intelligence + +27:08.200 --> 27:10.680 + and superhuman level intelligence? + +27:10.680 --> 27:13.280 + Where is consciousness part of that definition? + +27:13.280 --> 27:16.640 + No, consciousness does not come into this definition. + +27:16.640 --> 27:20.280 + So, so I think of intelligence as it's a spectrum + +27:20.280 --> 27:21.960 + but there are very many different kinds of goals + +27:21.960 --> 27:22.800 + you can have. + +27:22.800 --> 27:24.000 + You can have a goal to be a good chess player + +27:24.000 --> 27:28.520 + a good goal player, a good car driver, a good investor + +27:28.520 --> 27:31.160 + good poet, et cetera. + +27:31.160 --> 27:34.320 + So intelligence that by its very nature + +27:34.320 --> 27:36.680 + isn't something you can measure by this one number + +27:36.680 --> 27:37.960 + or some overall goodness. + +27:37.960 --> 27:38.800 + No, no. + +27:38.800 --> 27:40.320 + There are some people who are more better at this. + +27:40.320 --> 27:42.360 + Some people are better than that. + +27:42.360 --> 27:45.440 + Right now we have machines that are much better than us + +27:45.440 --> 27:49.040 + at some very narrow tasks like multiplying large numbers + +27:49.040 --> 27:53.200 + fast, memorizing large databases, playing chess + +27:53.200 --> 27:56.280 + playing go and soon driving cars. + +27:57.480 --> 28:00.080 + But there's still no machine that can match + +28:00.080 --> 28:02.720 + a human child in general intelligence + +28:02.720 --> 28:05.720 + but artificial general intelligence, AGI + +28:05.720 --> 28:07.880 + the name of your course, of course + +28:07.880 --> 28:12.880 + that is by its very definition, the quest + +28:13.400 --> 28:16.000 + to build a machine that can do everything + +28:16.000 --> 28:17.800 + as well as we can. + +28:17.800 --> 28:21.960 + So the old Holy grail of AI from back to its inception + +28:21.960 --> 28:25.560 + in the sixties, if that ever happens, of course + +28:25.560 --> 28:27.320 + I think it's going to be the biggest transition + +28:27.320 --> 28:29.040 + in the history of life on earth + +28:29.040 --> 28:33.200 + but it doesn't necessarily have to wait the big impact + +28:33.200 --> 28:35.400 + until machines are better than us at knitting + +28:35.400 --> 28:39.160 + that the really big change doesn't come exactly + +28:39.160 --> 28:41.800 + at the moment they're better than us at everything. + +28:41.800 --> 28:44.120 + The really big change comes first + +28:44.120 --> 28:45.840 + there are big changes when they start becoming better + +28:45.840 --> 28:48.800 + at us at doing most of the jobs that we do + +28:48.800 --> 28:51.160 + because that takes away much of the demand + +28:51.160 --> 28:53.200 + for human labor. + +28:53.200 --> 28:55.640 + And then the really whopping change comes + +28:55.640 --> 29:00.640 + when they become better than us at AI research, right? + +29:01.040 --> 29:03.760 + Because right now the timescale of AI research + +29:03.760 --> 29:08.400 + is limited by the human research and development cycle + +29:08.400 --> 29:10.160 + of years typically, you know + +29:10.160 --> 29:13.480 + how long does it take from one release of some software + +29:13.480 --> 29:15.720 + or iPhone or whatever to the next? + +29:15.720 --> 29:20.720 + But once Google can replace 40,000 engineers + +29:20.920 --> 29:25.920 + by 40,000 equivalent pieces of software or whatever + +29:26.400 --> 29:29.680 + but then there's no reason that has to be years + +29:29.680 --> 29:31.840 + it can be in principle much faster + +29:31.840 --> 29:36.040 + and the timescale of future progress in AI + +29:36.040 --> 29:39.320 + and all of science and technology will be driven + +29:39.320 --> 29:40.960 + by machines, not humans. + +29:40.960 --> 29:45.960 + So it's this simple point which gives right + +29:46.520 --> 29:48.720 + this incredibly fun controversy + +29:48.720 --> 29:51.880 + about whether there can be intelligence explosion + +29:51.880 --> 29:54.400 + so called singularity as Werner Vinge called it. + +29:54.400 --> 29:57.040 + Now the idea is articulated by I.J. Good + +29:57.040 --> 29:59.480 + is obviously way back fifties + +29:59.480 --> 30:01.040 + but you can see Alan Turing + +30:01.040 --> 30:03.640 + and others thought about it even earlier. + +30:06.920 --> 30:10.080 + So you asked me what exactly would I define + +30:10.080 --> 30:12.800 + human level intelligence, yeah. + +30:12.800 --> 30:15.680 + So the glib answer is to say something + +30:15.680 --> 30:18.520 + which is better than us at all cognitive tasks + +30:18.520 --> 30:21.800 + with a better than any human at all cognitive tasks + +30:21.800 --> 30:23.080 + but the really interesting bar + +30:23.080 --> 30:25.760 + I think goes a little bit lower than that actually. + +30:25.760 --> 30:27.920 + It's when they can, when they're better than us + +30:27.920 --> 30:31.760 + at AI programming and general learning + +30:31.760 --> 30:35.360 + so that they can if they want to get better + +30:35.360 --> 30:37.240 + than us at anything by just studying. + +30:37.240 --> 30:40.560 + So they're better is a key word and better is towards + +30:40.560 --> 30:44.120 + this kind of spectrum of the complexity of goals + +30:44.120 --> 30:45.680 + it's able to accomplish. + +30:45.680 --> 30:50.360 + So another way to, and that's certainly + +30:50.360 --> 30:53.040 + a very clear definition of human love. + +30:53.040 --> 30:55.240 + So there's, it's almost like a sea that's rising + +30:55.240 --> 30:56.800 + you can do more and more and more things + +30:56.800 --> 30:58.640 + it's a geographic that you show + +30:58.640 --> 30:59.880 + it's really nice way to put it. + +30:59.880 --> 31:01.560 + So there's some peaks that + +31:01.560 --> 31:03.280 + and there's an ocean level elevating + +31:03.280 --> 31:04.800 + and you solve more and more problems + +31:04.800 --> 31:07.720 + but just kind of to take a pause + +31:07.720 --> 31:09.000 + and we took a bunch of questions + +31:09.000 --> 31:10.240 + and a lot of social networks + +31:10.240 --> 31:11.720 + and a bunch of people asked + +31:11.720 --> 31:14.480 + a sort of a slightly different direction + +31:14.480 --> 31:19.480 + on creativity and things that perhaps aren't a peak. + +31:23.560 --> 31:24.720 + Human beings are flawed + +31:24.720 --> 31:28.720 + and perhaps better means having contradiction + +31:28.720 --> 31:30.200 + being flawed in some way. + +31:30.200 --> 31:34.960 + So let me sort of start easy, first of all. + +31:34.960 --> 31:36.600 + So you have a lot of cool equations. + +31:36.600 --> 31:39.760 + Let me ask, what's your favorite equation, first of all? + +31:39.760 --> 31:42.760 + I know they're all like your children, but like + +31:42.760 --> 31:43.680 + which one is that? + +31:43.680 --> 31:45.560 + This is the shirt in your equation. + +31:45.560 --> 31:48.640 + It's the master key of quantum mechanics + +31:48.640 --> 31:49.880 + of the micro world. + +31:49.880 --> 31:52.800 + So this equation will protect everything + +31:52.800 --> 31:55.840 + to do with atoms, molecules and all the way up. + +31:55.840 --> 31:58.560 + Right? + +31:58.560 --> 31:59.760 + Yeah, so, okay. + +31:59.760 --> 32:02.080 + So quantum mechanics is certainly a beautiful + +32:02.080 --> 32:05.160 + mysterious formulation of our world. + +32:05.160 --> 32:08.760 + So I'd like to sort of ask you, just as an example + +32:08.760 --> 32:12.160 + it perhaps doesn't have the same beauty as physics does + +32:12.160 --> 32:16.960 + but in mathematics abstract, the Andrew Wiles + +32:16.960 --> 32:19.360 + who proved the Fermat's last theorem. + +32:19.360 --> 32:22.040 + So he just saw this recently + +32:22.040 --> 32:24.160 + and it kind of caught my eye a little bit. + +32:24.160 --> 32:27.960 + This is 358 years after it was conjectured. + +32:27.960 --> 32:29.960 + So this is very simple formulation. + +32:29.960 --> 32:32.640 + Everybody tried to prove it, everybody failed. + +32:32.640 --> 32:34.800 + And so here's this guy comes along + +32:34.800 --> 32:38.640 + and eventually proves it and then fails to prove it + +32:38.640 --> 32:41.320 + and then proves it again in 94. + +32:41.320 --> 32:43.480 + And he said like the moment when everything connected + +32:43.480 --> 32:46.040 + into place in an interview said + +32:46.040 --> 32:47.880 + it was so indescribably beautiful. + +32:47.880 --> 32:51.040 + That moment when you finally realize the connecting piece + +32:51.040 --> 32:52.800 + of two conjectures. + +32:52.800 --> 32:55.280 + He said, it was so indescribably beautiful. + +32:55.280 --> 32:57.040 + It was so simple and so elegant. + +32:57.040 --> 32:58.760 + I couldn't understand how I'd missed it. + +32:58.760 --> 33:02.080 + And I just stared at it in disbelief for 20 minutes. + +33:02.080 --> 33:05.240 + Then during the day, I walked around the department + +33:05.240 --> 33:07.880 + and I keep coming back to my desk + +33:07.880 --> 33:09.840 + looking to see if it was still there. + +33:09.840 --> 33:10.680 + It was still there. + +33:10.680 --> 33:11.760 + I couldn't contain myself. + +33:11.760 --> 33:12.880 + I was so excited. + +33:12.880 --> 33:15.880 + It was the most important moment on my working life. + +33:15.880 --> 33:18.960 + Nothing I ever do again will mean as much. + +33:18.960 --> 33:20.800 + So that particular moment. + +33:20.800 --> 33:24.640 + And it kind of made me think of what would it take? + +33:24.640 --> 33:27.960 + And I think we have all been there at small levels. + +33:29.480 --> 33:32.240 + Maybe let me ask, have you had a moment like that + +33:32.240 --> 33:34.880 + in your life where you just had an idea? + +33:34.880 --> 33:37.040 + It's like, wow, yes. + +33:40.000 --> 33:42.480 + I wouldn't mention myself in the same breath + +33:42.480 --> 33:44.760 + as Andrew Wiles, but I've certainly had a number + +33:44.760 --> 33:52.200 + of aha moments when I realized something very cool + +33:52.200 --> 33:56.000 + about physics, which has completely made my head explode. + +33:56.000 --> 33:58.320 + In fact, some of my favorite discoveries I made later, + +33:58.320 --> 34:01.080 + I later realized that they had been discovered earlier + +34:01.080 --> 34:03.240 + by someone who sometimes got quite famous for it. + +34:03.240 --> 34:05.480 + So it's too late for me to even publish it, + +34:05.480 --> 34:07.440 + but that doesn't diminish in any way. + +34:07.440 --> 34:09.760 + The emotional experience you have when you realize it, + +34:09.760 --> 34:11.320 + like, wow. + +34:11.320 --> 34:15.520 + Yeah, so what would it take in that moment, that wow, + +34:15.520 --> 34:17.320 + that was yours in that moment? + +34:17.320 --> 34:21.440 + So what do you think it takes for an intelligence system, + +34:21.440 --> 34:24.520 + an AGI system, an AI system to have a moment like that? + +34:25.640 --> 34:26.760 + That's a tricky question + +34:26.760 --> 34:29.200 + because there are actually two parts to it, right? + +34:29.200 --> 34:33.920 + One of them is, can it accomplish that proof? + +34:33.920 --> 34:37.640 + Can it prove that you can never write A to the N + +34:37.640 --> 34:42.760 + plus B to the N equals three to that equal Z to the N + +34:42.760 --> 34:45.320 + for all integers, et cetera, et cetera, + +34:45.320 --> 34:48.720 + when N is bigger than two? + +34:48.720 --> 34:51.360 + That's simply a question about intelligence. + +34:51.360 --> 34:54.120 + Can you build machines that are that intelligent? + +34:54.120 --> 34:57.280 + And I think by the time we get a machine + +34:57.280 --> 35:00.840 + that can independently come up with that level of proofs, + +35:00.840 --> 35:03.360 + probably quite close to AGI. + +35:03.360 --> 35:07.240 + The second question is a question about consciousness. + +35:07.240 --> 35:11.760 + When will we, how likely is it that such a machine + +35:11.760 --> 35:14.240 + will actually have any experience at all, + +35:14.240 --> 35:16.160 + as opposed to just being like a zombie? + +35:16.160 --> 35:20.560 + And would we expect it to have some sort of emotional response + +35:20.560 --> 35:24.640 + to this or anything at all akin to human emotion + +35:24.640 --> 35:28.320 + where when it accomplishes its machine goal, + +35:28.320 --> 35:31.920 + it views it as somehow something very positive + +35:31.920 --> 35:39.160 + and sublime and deeply meaningful? + +35:39.160 --> 35:41.440 + I would certainly hope that if in the future + +35:41.440 --> 35:45.120 + we do create machines that are our peers + +35:45.120 --> 35:50.160 + or even our descendants, that I would certainly + +35:50.160 --> 35:55.480 + hope that they do have this sublime appreciation of life. + +35:55.480 --> 35:58.840 + In a way, my absolutely worst nightmare + +35:58.840 --> 36:05.760 + would be that at some point in the future, + +36:05.760 --> 36:07.400 + the distant future, maybe our cosmos + +36:07.400 --> 36:10.600 + is teeming with all this post biological life doing + +36:10.600 --> 36:12.880 + all the seemingly cool stuff. + +36:12.880 --> 36:16.480 + And maybe the last humans, by the time + +36:16.480 --> 36:20.120 + our species eventually fizzles out, + +36:20.120 --> 36:21.920 + will be like, well, that's OK because we're + +36:21.920 --> 36:23.600 + so proud of our descendants here. + +36:23.600 --> 36:26.680 + And look what all the, my worst nightmare + +36:26.680 --> 36:30.360 + is that we haven't solved the consciousness problem. + +36:30.360 --> 36:32.880 + And we haven't realized that these are all the zombies. + +36:32.880 --> 36:36.200 + They're not aware of anything any more than a tape recorder + +36:36.200 --> 36:37.840 + has any kind of experience. + +36:37.840 --> 36:40.040 + So the whole thing has just become + +36:40.040 --> 36:41.520 + a play for empty benches. + +36:41.520 --> 36:44.640 + That would be the ultimate zombie apocalypse. + +36:44.640 --> 36:47.200 + So I would much rather, in that case, + +36:47.200 --> 36:52.240 + that we have these beings which can really + +36:52.240 --> 36:57.000 + appreciate how amazing it is. + +36:57.000 --> 37:01.080 + And in that picture, what would be the role of creativity? + +37:01.080 --> 37:04.960 + A few people ask about creativity. + +37:04.960 --> 37:07.080 + When you think about intelligence, + +37:07.080 --> 37:09.840 + certainly the story you told at the beginning of your book + +37:09.840 --> 37:15.200 + involved creating movies and so on, making money. + +37:15.200 --> 37:17.240 + You can make a lot of money in our modern world + +37:17.240 --> 37:18.600 + with music and movies. + +37:18.600 --> 37:20.880 + So if you are an intelligent system, + +37:20.880 --> 37:22.960 + you may want to get good at that. + +37:22.960 --> 37:26.280 + But that's not necessarily what I mean by creativity. + +37:26.280 --> 37:29.640 + Is it important on that complex goals + +37:29.640 --> 37:31.600 + where the sea is rising for there + +37:31.600 --> 37:33.800 + to be something creative? + +37:33.800 --> 37:37.400 + Or am I being very human centric and thinking creativity + +37:37.400 --> 37:41.880 + somehow special relative to intelligence? + +37:41.880 --> 37:47.240 + My hunch is that we should think of creativity simply + +37:47.240 --> 37:50.760 + as an aspect of intelligence. + +37:50.760 --> 37:57.840 + And we have to be very careful with human vanity. + +37:57.840 --> 37:59.520 + We have this tendency to very often want + +37:59.520 --> 38:01.560 + to say, as soon as machines can do something, + +38:01.560 --> 38:03.560 + we try to diminish it and say, oh, but that's + +38:03.560 --> 38:05.920 + not real intelligence. + +38:05.920 --> 38:08.400 + Isn't it creative or this or that? + +38:08.400 --> 38:12.200 + The other thing, if we ask ourselves + +38:12.200 --> 38:14.320 + to write down a definition of what we actually mean + +38:14.320 --> 38:18.840 + by being creative, what we mean by Andrew Wiles, what he did + +38:18.840 --> 38:21.880 + there, for example, don't we often mean that someone takes + +38:21.880 --> 38:26.000 + a very unexpected leap? + +38:26.000 --> 38:29.680 + It's not like taking 573 and multiplying it + +38:29.680 --> 38:33.840 + by 224 by just a step of straightforward cookbook + +38:33.840 --> 38:36.520 + like rules, right? + +38:36.520 --> 38:39.680 + You can maybe make a connection between two things + +38:39.680 --> 38:42.640 + that people had never thought was connected or something + +38:42.640 --> 38:44.480 + like that. + +38:44.480 --> 38:47.720 + I think this is an aspect of intelligence. + +38:47.720 --> 38:53.000 + And this is actually one of the most important aspects of it. + +38:53.000 --> 38:55.520 + Maybe the reason we humans tend to be better at it + +38:55.520 --> 38:57.840 + than traditional computers is because it's + +38:57.840 --> 38:59.640 + something that comes more naturally if you're + +38:59.640 --> 39:04.120 + a neural network than if you're a traditional logic gate + +39:04.120 --> 39:05.720 + based computer machine. + +39:05.720 --> 39:08.640 + We physically have all these connections. + +39:08.640 --> 39:13.800 + And you activate here, activate here, activate here. + +39:13.800 --> 39:16.560 + Bing. + +39:16.560 --> 39:21.040 + My hunch is that if we ever build a machine where you could + +39:21.040 --> 39:29.200 + just give it the task, hey, you say, hey, I just realized + +39:29.200 --> 39:32.320 + I want to travel around the world instead this month. + +39:32.320 --> 39:34.600 + Can you teach my AGI course for me? + +39:34.600 --> 39:35.960 + And it's like, OK, I'll do it. + +39:35.960 --> 39:37.920 + And it does everything that you would have done + +39:37.920 --> 39:39.760 + and improvises and stuff. + +39:39.760 --> 39:43.360 + That would, in my mind, involve a lot of creativity. + +39:43.360 --> 39:45.680 + Yeah, so it's actually a beautiful way to put it. + +39:45.680 --> 39:52.640 + I think we do try to grasp at the definition of intelligence + +39:52.640 --> 39:56.360 + is everything we don't understand how to build. + +39:56.360 --> 39:59.360 + So we as humans try to find things + +39:59.360 --> 40:01.240 + that we have and machines don't have. + +40:01.240 --> 40:03.800 + And maybe creativity is just one of the things, one + +40:03.800 --> 40:05.480 + of the words we use to describe that. + +40:05.480 --> 40:07.200 + That's a really interesting way to put it. + +40:07.200 --> 40:09.520 + I don't think we need to be that defensive. + +40:09.520 --> 40:11.560 + I don't think anything good comes out of saying, + +40:11.560 --> 40:18.080 + well, we're somehow special, you know? + +40:18.080 --> 40:21.040 + Contrary wise, there are many examples in history + +40:21.040 --> 40:27.840 + of where trying to pretend that we're somehow superior + +40:27.840 --> 40:33.120 + to all other intelligent beings has led to pretty bad results, + +40:33.120 --> 40:35.960 + right? + +40:35.960 --> 40:38.440 + Nazi Germany, they said that they were somehow superior + +40:38.440 --> 40:40.080 + to other people. + +40:40.080 --> 40:42.440 + Today, we still do a lot of cruelty to animals + +40:42.440 --> 40:44.440 + by saying that we're so superior somehow, + +40:44.440 --> 40:46.440 + and they can't feel pain. + +40:46.440 --> 40:48.480 + Slavery was justified by the same kind + +40:48.480 --> 40:52.200 + of just really weak arguments. + +40:52.200 --> 40:57.120 + And I don't think if we actually go ahead and build + +40:57.120 --> 40:59.440 + artificial general intelligence, it + +40:59.440 --> 41:01.360 + can do things better than us, I don't + +41:01.360 --> 41:04.080 + think we should try to found our self worth on some sort + +41:04.080 --> 41:09.760 + of bogus claims of superiority in terms + +41:09.760 --> 41:12.120 + of our intelligence. + +41:12.120 --> 41:18.080 + I think we should instead find our calling + +41:18.080 --> 41:23.360 + and the meaning of life from the experiences that we have. + +41:23.360 --> 41:28.720 + I can have very meaningful experiences + +41:28.720 --> 41:32.920 + even if there are other people who are smarter than me. + +41:32.920 --> 41:34.400 + When I go to a faculty meeting here, + +41:34.400 --> 41:36.520 + and we talk about something, and then I certainly realize, + +41:36.520 --> 41:39.080 + oh, boy, he has an old prize, he has an old prize, + +41:39.080 --> 41:40.800 + he has an old prize, I don't have one. + +41:40.800 --> 41:43.760 + Does that make me enjoy life any less + +41:43.760 --> 41:47.560 + or enjoy talking to those people less? + +41:47.560 --> 41:49.560 + Of course not. + +41:49.560 --> 41:54.160 + And the contrary, I feel very honored and privileged + +41:54.160 --> 41:58.760 + to get to interact with other very intelligent beings that + +41:58.760 --> 42:00.680 + are better than me at a lot of stuff. + +42:00.680 --> 42:02.840 + So I don't think there's any reason why + +42:02.840 --> 42:06.080 + we can't have the same approach with intelligent machines. + +42:06.080 --> 42:07.320 + That's a really interesting. + +42:07.320 --> 42:08.920 + So people don't often think about that. + +42:08.920 --> 42:10.600 + They think about when there's going, + +42:10.600 --> 42:13.320 + if there's machines that are more intelligent, + +42:13.320 --> 42:15.080 + you naturally think that that's not + +42:15.080 --> 42:19.080 + going to be a beneficial type of intelligence. + +42:19.080 --> 42:23.000 + You don't realize it could be like peers with Nobel prizes + +42:23.000 --> 42:25.120 + that would be just fun to talk with, + +42:25.120 --> 42:27.560 + and they might be clever about certain topics, + +42:27.560 --> 42:32.240 + and you can have fun having a few drinks with them. + +42:32.240 --> 42:35.880 + Well, also, another example we can all + +42:35.880 --> 42:39.320 + relate to of why it doesn't have to be a terrible thing + +42:39.320 --> 42:42.560 + to be in the presence of people who are even smarter than us + +42:42.560 --> 42:45.600 + all around is when you and I were both two years old, + +42:45.600 --> 42:48.360 + I mean, our parents were much more intelligent than us, + +42:48.360 --> 42:49.040 + right? + +42:49.040 --> 42:51.960 + Worked out OK, because their goals + +42:51.960 --> 42:53.960 + were aligned with our goals. + +42:53.960 --> 42:58.680 + And that, I think, is really the number one key issue + +42:58.680 --> 43:02.280 + we have to solve if we value align the value alignment + +43:02.280 --> 43:03.080 + problem, exactly. + +43:03.080 --> 43:06.520 + Because people who see too many Hollywood movies + +43:06.520 --> 43:10.000 + with lousy science fiction plot lines, + +43:10.000 --> 43:12.200 + they worry about the wrong thing, right? + +43:12.200 --> 43:16.320 + They worry about some machine suddenly turning evil. + +43:16.320 --> 43:21.480 + It's not malice that is the concern. + +43:21.480 --> 43:22.880 + It's competence. + +43:22.880 --> 43:27.440 + By definition, intelligent makes you very competent. + +43:27.440 --> 43:31.920 + If you have a more intelligent goal playing, + +43:31.920 --> 43:33.680 + computer playing is a less intelligent one. + +43:33.680 --> 43:36.120 + And when we define intelligence as the ability + +43:36.120 --> 43:38.600 + to accomplish goal winning, it's going + +43:38.600 --> 43:40.560 + to be the more intelligent one that wins. + +43:40.560 --> 43:43.560 + And if you have a human and then you + +43:43.560 --> 43:47.720 + have an AGI that's more intelligent in all ways + +43:47.720 --> 43:49.520 + and they have different goals, guess who's + +43:49.520 --> 43:50.720 + going to get their way, right? + +43:50.720 --> 43:57.120 + So I was just reading about this particular rhinoceros species + +43:57.120 --> 43:59.200 + that was driven extinct just a few years ago. + +43:59.200 --> 44:02.280 + Ellen Bummer is looking at this cute picture of a mommy + +44:02.280 --> 44:05.080 + rhinoceros with its child. + +44:05.080 --> 44:09.320 + And why did we humans drive it to extinction? + +44:09.320 --> 44:12.800 + It wasn't because we were evil rhino haters as a whole. + +44:12.800 --> 44:14.920 + It was just because our goals weren't aligned + +44:14.920 --> 44:16.000 + with those of the rhinoceros. + +44:16.000 --> 44:17.680 + And it didn't work out so well for the rhinoceros + +44:17.680 --> 44:19.560 + because we were more intelligent, right? + +44:19.560 --> 44:21.240 + So I think it's just so important + +44:21.240 --> 44:27.120 + that if we ever do build AGI, before we unleash anything, + +44:27.120 --> 44:31.840 + we have to make sure that it learns + +44:31.840 --> 44:36.000 + to understand our goals, that it adopts our goals, + +44:36.000 --> 44:37.920 + and that it retains those goals. + +44:37.920 --> 44:40.520 + So the cool, interesting problem there + +44:40.520 --> 44:47.040 + is us as human beings trying to formulate our values. + +44:47.040 --> 44:51.360 + So you could think of the United States Constitution as a way + +44:51.360 --> 44:56.680 + that people sat down, at the time a bunch of white men, + +44:56.680 --> 44:59.680 + which is a good example, I should say. + +44:59.680 --> 45:01.480 + They formulated the goals for this country. + +45:01.480 --> 45:03.760 + And a lot of people agree that those goals actually + +45:03.760 --> 45:05.360 + held up pretty well. + +45:05.360 --> 45:07.160 + That's an interesting formulation of values + +45:07.160 --> 45:09.440 + and failed miserably in other ways. + +45:09.440 --> 45:13.320 + So for the value alignment problem and the solution to it, + +45:13.320 --> 45:19.560 + we have to be able to put on paper or in a program + +45:19.560 --> 45:20.400 + human values. + +45:20.400 --> 45:22.400 + How difficult do you think that is? + +45:22.400 --> 45:24.040 + Very. + +45:24.040 --> 45:25.880 + But it's so important. + +45:25.880 --> 45:28.000 + We really have to give it our best. + +45:28.000 --> 45:30.120 + And it's difficult for two separate reasons. + +45:30.120 --> 45:33.440 + There's the technical value alignment problem + +45:33.440 --> 45:39.120 + of figuring out just how to make machines understand our goals, + +45:39.120 --> 45:40.440 + adopt them, and retain them. + +45:40.440 --> 45:43.200 + And then there's the separate part of it, + +45:43.200 --> 45:44.200 + the philosophical part. + +45:44.200 --> 45:45.920 + Whose values anyway? + +45:45.920 --> 45:48.320 + And since it's not like we have any great consensus + +45:48.320 --> 45:52.040 + on this planet on values, what mechanism should we + +45:52.040 --> 45:54.120 + create then to aggregate and decide, OK, + +45:54.120 --> 45:56.520 + what's a good compromise? + +45:56.520 --> 45:58.440 + That second discussion can't just + +45:58.440 --> 46:01.560 + be left to tech nerds like myself. + +46:01.560 --> 46:05.720 + And if we refuse to talk about it and then AGI gets built, + +46:05.720 --> 46:07.160 + who's going to be actually making + +46:07.160 --> 46:08.480 + the decision about whose values? + +46:08.480 --> 46:12.080 + It's going to be a bunch of dudes in some tech company. + +46:12.080 --> 46:17.240 + And are they necessarily so representative of all + +46:17.240 --> 46:19.400 + of humankind that we want to just entrust it to them? + +46:19.400 --> 46:23.000 + Are they even uniquely qualified to speak + +46:23.000 --> 46:25.240 + to future human happiness just because they're + +46:25.240 --> 46:26.480 + good at programming AI? + +46:26.480 --> 46:30.200 + I'd much rather have this be a really inclusive conversation. + +46:30.200 --> 46:32.560 + But do you think it's possible? + +46:32.560 --> 46:37.560 + So you create a beautiful vision that includes the diversity, + +46:37.560 --> 46:40.960 + cultural diversity, and various perspectives on discussing + +46:40.960 --> 46:43.600 + rights, freedoms, human dignity. + +46:43.600 --> 46:46.520 + But how hard is it to come to that consensus? + +46:46.520 --> 46:50.400 + Do you think it's certainly a really important thing + +46:50.400 --> 46:51.880 + that we should all try to do? + +46:51.880 --> 46:54.240 + But do you think it's feasible? + +46:54.240 --> 47:00.160 + I think there's no better way to guarantee failure than to + +47:00.160 --> 47:02.840 + refuse to talk about it or refuse to try. + +47:02.840 --> 47:05.320 + And I also think it's a really bad strategy + +47:05.320 --> 47:08.560 + to say, OK, let's first have a discussion for a long time. + +47:08.560 --> 47:11.040 + And then once we reach complete consensus, + +47:11.040 --> 47:13.360 + then we'll try to load it into some machine. + +47:13.360 --> 47:16.560 + No, we shouldn't let perfect be the enemy of good. + +47:16.560 --> 47:20.600 + Instead, we should start with the kindergarten ethics + +47:20.600 --> 47:22.120 + that pretty much everybody agrees on + +47:22.120 --> 47:24.360 + and put that into machines now. + +47:24.360 --> 47:25.880 + We're not doing that even. + +47:25.880 --> 47:31.000 + Look at anyone who builds this passenger aircraft, + +47:31.000 --> 47:33.000 + wants it to never under any circumstances + +47:33.000 --> 47:35.600 + fly into a building or a mountain. + +47:35.600 --> 47:38.480 + Yet the September 11 hijackers were able to do that. + +47:38.480 --> 47:41.800 + And even more embarrassingly, Andreas Lubitz, + +47:41.800 --> 47:43.960 + this depressed Germanwings pilot, + +47:43.960 --> 47:47.360 + when he flew his passenger jet into the Alps killing over 100 + +47:47.360 --> 47:50.640 + people, he just told the autopilot to do it. + +47:50.640 --> 47:53.200 + He told the freaking computer to change the altitude + +47:53.200 --> 47:55.040 + to 100 meters. + +47:55.040 --> 47:58.160 + And even though it had the GPS maps, everything, + +47:58.160 --> 48:00.640 + the computer was like, OK. + +48:00.640 --> 48:05.320 + So we should take those very basic values, + +48:05.320 --> 48:08.400 + where the problem is not that we don't agree. + +48:08.400 --> 48:10.120 + The problem is just we've been too lazy + +48:10.120 --> 48:11.480 + to try to put it into our machines + +48:11.480 --> 48:15.520 + and make sure that from now on, airplanes will just, + +48:15.520 --> 48:16.920 + which all have computers in them, + +48:16.920 --> 48:19.720 + but will just refuse to do something like that. + +48:19.720 --> 48:22.160 + Go into safe mode, maybe lock the cockpit door, + +48:22.160 --> 48:24.480 + go over to the nearest airport. + +48:24.480 --> 48:28.080 + And there's so much other technology in our world + +48:28.080 --> 48:31.320 + as well now, where it's really becoming quite timely + +48:31.320 --> 48:34.120 + to put in some sort of very basic values like this. + +48:34.120 --> 48:39.240 + Even in cars, we've had enough vehicle terrorism attacks + +48:39.240 --> 48:42.040 + by now, where people have driven trucks and vans + +48:42.040 --> 48:45.480 + into pedestrians, that it's not at all a crazy idea + +48:45.480 --> 48:48.680 + to just have that hardwired into the car. + +48:48.680 --> 48:50.280 + Because yeah, there are a lot of, + +48:50.280 --> 48:52.240 + there's always going to be people who for some reason + +48:52.240 --> 48:54.800 + want to harm others, but most of those people + +48:54.800 --> 48:56.760 + don't have the technical expertise to figure out + +48:56.760 --> 48:58.520 + how to work around something like that. + +48:58.520 --> 49:01.760 + So if the car just won't do it, it helps. + +49:01.760 --> 49:02.840 + So let's start there. + +49:02.840 --> 49:04.960 + So there's a lot of, that's a great point. + +49:04.960 --> 49:06.800 + So not chasing perfect. + +49:06.800 --> 49:10.840 + There's a lot of things that most of the world agrees on. + +49:10.840 --> 49:11.840 + Yeah, let's start there. + +49:11.840 --> 49:12.680 + Let's start there. + +49:12.680 --> 49:14.560 + And then once we start there, + +49:14.560 --> 49:17.240 + we'll also get into the habit of having + +49:17.240 --> 49:18.520 + these kind of conversations about, okay, + +49:18.520 --> 49:21.760 + what else should we put in here and have these discussions? + +49:21.760 --> 49:23.920 + This should be a gradual process then. + +49:23.920 --> 49:28.600 + Great, so, but that also means describing these things + +49:28.600 --> 49:31.240 + and describing it to a machine. + +49:31.240 --> 49:34.200 + So one thing, we had a few conversations + +49:34.200 --> 49:35.640 + with Stephen Wolfram. + +49:35.640 --> 49:37.080 + I'm not sure if you're familiar with Stephen. + +49:37.080 --> 49:38.360 + Oh yeah, I know him quite well. + +49:38.360 --> 49:42.040 + So he is, he works with a bunch of things, + +49:42.040 --> 49:46.560 + but cellular automata, these simple computable things, + +49:46.560 --> 49:47.960 + these computation systems. + +49:47.960 --> 49:49.880 + And he kind of mentioned that, + +49:49.880 --> 49:52.480 + we probably have already within these systems + +49:52.480 --> 49:54.680 + already something that's AGI, + +49:56.120 --> 49:58.720 + meaning like we just don't know it + +49:58.720 --> 50:00.400 + because we can't talk to it. + +50:00.400 --> 50:04.800 + So if you give me this chance to try to at least + +50:04.800 --> 50:06.720 + form a question out of this is, + +50:07.600 --> 50:10.880 + I think it's an interesting idea to think + +50:10.880 --> 50:12.680 + that we can have intelligent systems, + +50:12.680 --> 50:15.600 + but we don't know how to describe something to them + +50:15.600 --> 50:17.360 + and they can't communicate with us. + +50:17.360 --> 50:19.840 + I know you're doing a little bit of work in explainable AI, + +50:19.840 --> 50:22.040 + trying to get AI to explain itself. + +50:22.040 --> 50:25.520 + So what are your thoughts of natural language processing + +50:25.520 --> 50:27.640 + or some kind of other communication? + +50:27.640 --> 50:30.120 + How does the AI explain something to us? + +50:30.120 --> 50:33.640 + How do we explain something to it, to machines? + +50:33.640 --> 50:35.320 + Or you think of it differently? + +50:35.320 --> 50:39.960 + So there are two separate parts to your question there. + +50:39.960 --> 50:42.440 + One of them has to do with communication, + +50:42.440 --> 50:44.440 + which is super interesting, I'll get to that in a sec. + +50:44.440 --> 50:47.280 + The other is whether we already have AGI + +50:47.280 --> 50:49.240 + but we just haven't noticed it there. + +50:49.240 --> 50:50.080 + Right. + +50:51.800 --> 50:53.000 + There I beg to differ. + +50:54.280 --> 50:56.480 + I don't think there's anything in any cellular automaton + +50:56.480 --> 50:59.040 + or anything or the internet itself or whatever + +50:59.040 --> 51:03.560 + that has artificial general intelligence + +51:03.560 --> 51:05.520 + and that it can really do exactly everything + +51:05.520 --> 51:07.000 + we humans can do better. + +51:07.000 --> 51:11.600 + I think the day that happens, when that happens, + +51:11.600 --> 51:15.600 + we will very soon notice, we'll probably notice even before + +51:15.600 --> 51:17.440 + because in a very, very big way. + +51:17.440 --> 51:18.840 + But for the second part, though. + +51:18.840 --> 51:20.720 + Wait, can I ask, sorry. + +51:20.720 --> 51:24.400 + So, because you have this beautiful way + +51:24.400 --> 51:29.400 + to formulating consciousness as information processing, + +51:30.360 --> 51:31.360 + and you can think of intelligence + +51:31.360 --> 51:32.280 + as information processing, + +51:32.280 --> 51:34.320 + and you can think of the entire universe + +51:34.320 --> 51:38.720 + as these particles and these systems roaming around + +51:38.720 --> 51:41.360 + that have this information processing power. + +51:41.360 --> 51:44.840 + You don't think there is something with the power + +51:44.840 --> 51:49.040 + to process information in the way that we human beings do + +51:49.040 --> 51:54.040 + that's out there that needs to be sort of connected to. + +51:55.400 --> 51:57.880 + It seems a little bit philosophical, perhaps, + +51:57.880 --> 52:00.080 + but there's something compelling to the idea + +52:00.080 --> 52:01.920 + that the power is already there, + +52:01.920 --> 52:05.440 + which the focus should be more on being able + +52:05.440 --> 52:07.360 + to communicate with it. + +52:07.360 --> 52:11.960 + Well, I agree that in a certain sense, + +52:11.960 --> 52:15.360 + the hardware processing power is already out there + +52:15.360 --> 52:19.000 + because our universe itself can think of it + +52:19.000 --> 52:21.000 + as being a computer already, right? + +52:21.000 --> 52:23.800 + It's constantly computing what water waves, + +52:23.800 --> 52:26.120 + how it devolved the water waves in the River Charles + +52:26.120 --> 52:28.440 + and how to move the air molecules around. + +52:28.440 --> 52:30.480 + Seth Lloyd has pointed out, my colleague here, + +52:30.480 --> 52:32.920 + that you can even in a very rigorous way + +52:32.920 --> 52:35.480 + think of our entire universe as being a quantum computer. + +52:35.480 --> 52:37.680 + It's pretty clear that our universe + +52:37.680 --> 52:40.320 + supports this amazing processing power + +52:40.320 --> 52:42.160 + because you can even, + +52:42.160 --> 52:44.920 + within this physics computer that we live in, right? + +52:44.920 --> 52:47.040 + We can even build actual laptops and stuff, + +52:47.040 --> 52:49.000 + so clearly the power is there. + +52:49.000 --> 52:52.040 + It's just that most of the compute power that nature has, + +52:52.040 --> 52:54.240 + it's, in my opinion, kind of wasting on boring stuff + +52:54.240 --> 52:56.520 + like simulating yet another ocean wave somewhere + +52:56.520 --> 52:58.040 + where no one is even looking, right? + +52:58.040 --> 53:00.880 + So in a sense, what life does, what we are doing + +53:00.880 --> 53:03.880 + when we build computers is we're rechanneling + +53:03.880 --> 53:07.200 + all this compute that nature is doing anyway + +53:07.200 --> 53:09.360 + into doing things that are more interesting + +53:09.360 --> 53:11.440 + than just yet another ocean wave, + +53:11.440 --> 53:13.200 + and let's do something cool here. + +53:14.080 --> 53:17.080 + So the raw hardware power is there, for sure, + +53:17.080 --> 53:21.080 + but then even just computing what's going to happen + +53:21.080 --> 53:23.520 + for the next five seconds in this water bottle, + +53:23.520 --> 53:26.000 + takes a ridiculous amount of compute + +53:26.000 --> 53:27.920 + if you do it on a human computer. + +53:27.920 --> 53:29.920 + This water bottle just did it. + +53:29.920 --> 53:33.440 + But that does not mean that this water bottle has AGI + +53:34.760 --> 53:37.040 + because AGI means it should also be able to, + +53:37.040 --> 53:40.160 + like I've written my book, done this interview. + +53:40.160 --> 53:42.080 + And I don't think it's just communication problems. + +53:42.080 --> 53:46.760 + I don't really think it can do it. + +53:46.760 --> 53:49.280 + Although Buddhists say when they watch the water + +53:49.280 --> 53:51.240 + and that there is some beauty, + +53:51.240 --> 53:53.720 + that there's some depth and beauty in nature + +53:53.720 --> 53:54.840 + that they can communicate with. + +53:54.840 --> 53:56.480 + Communication is also very important though + +53:56.480 --> 54:01.200 + because I mean, look, part of my job is being a teacher. + +54:01.200 --> 54:06.200 + And I know some very intelligent professors even + +54:06.200 --> 54:09.800 + who just have a bit of hard time communicating. + +54:09.800 --> 54:12.640 + They come up with all these brilliant ideas, + +54:12.640 --> 54:14.520 + but to communicate with somebody else, + +54:14.520 --> 54:16.920 + you have to also be able to simulate their own mind. + +54:16.920 --> 54:18.360 + Yes, empathy. + +54:18.360 --> 54:20.640 + Build well enough and understand model of their mind + +54:20.640 --> 54:24.400 + that you can say things that they will understand. + +54:24.400 --> 54:26.480 + And that's quite difficult. + +54:26.480 --> 54:28.280 + And that's why today it's so frustrating + +54:28.280 --> 54:32.600 + if you have a computer that makes some cancer diagnosis + +54:32.600 --> 54:34.120 + and you ask it, well, why are you saying + +54:34.120 --> 54:36.120 + I should have this surgery? + +54:36.120 --> 54:37.960 + And if it can only reply, + +54:37.960 --> 54:40.800 + I was trained on five terabytes of data + +54:40.800 --> 54:45.080 + and this is my diagnosis, boop, boop, beep, beep. + +54:45.080 --> 54:49.120 + It doesn't really instill a lot of confidence, right? + +54:49.120 --> 54:51.120 + So I think we have a lot of work to do + +54:51.120 --> 54:54.320 + on communication there. + +54:54.320 --> 54:58.040 + So what kind of, I think you're doing a little bit of work + +54:58.040 --> 54:59.320 + in explainable AI. + +54:59.320 --> 55:01.320 + What do you think are the most promising avenues? + +55:01.320 --> 55:05.240 + Is it mostly about sort of the Alexa problem + +55:05.240 --> 55:07.200 + of natural language processing of being able + +55:07.200 --> 55:11.600 + to actually use human interpretable methods + +55:11.600 --> 55:13.160 + of communication? + +55:13.160 --> 55:16.000 + So being able to talk to a system and it talk back to you, + +55:16.000 --> 55:18.640 + or is there some more fundamental problems to be solved? + +55:18.640 --> 55:21.160 + I think it's all of the above. + +55:21.160 --> 55:23.520 + The natural language processing is obviously important, + +55:23.520 --> 55:27.600 + but there are also more nerdy fundamental problems. + +55:27.600 --> 55:31.640 + Like if you take, you play chess? + +55:31.640 --> 55:33.040 + Of course, I'm Russian. + +55:33.040 --> 55:33.880 + I have to. + +55:33.880 --> 55:34.720 + You speak Russian? + +55:34.720 --> 55:35.560 + Yes, I speak Russian. + +55:35.560 --> 55:38.040 + Excellent, I didn't know. + +55:38.040 --> 55:39.160 + When did you learn Russian? + +55:39.160 --> 55:41.800 + I speak very bad Russian, I'm only an autodidact, + +55:41.800 --> 55:44.560 + but I bought a book, Teach Yourself Russian, + +55:44.560 --> 55:47.720 + read a lot, but it was very difficult. + +55:47.720 --> 55:48.560 + Wow. + +55:48.560 --> 55:49.960 + That's why I speak so bad. + +55:49.960 --> 55:51.960 + How many languages do you know? + +55:51.960 --> 55:53.840 + Wow, that's really impressive. + +55:53.840 --> 55:56.320 + I don't know, my wife has some calculation, + +55:56.320 --> 55:58.400 + but my point was, if you play chess, + +55:58.400 --> 56:01.040 + have you looked at the AlphaZero games? + +56:01.040 --> 56:02.600 + The actual games, no. + +56:02.600 --> 56:05.000 + Check it out, some of them are just mind blowing, + +56:06.320 --> 56:07.720 + really beautiful. + +56:07.720 --> 56:12.400 + And if you ask, how did it do that? + +56:13.760 --> 56:16.520 + You go talk to Demis Hassabis, + +56:16.520 --> 56:18.240 + I know others from DeepMind, + +56:19.120 --> 56:20.600 + all they'll ultimately be able to give you + +56:20.600 --> 56:23.920 + is big tables of numbers, matrices, + +56:23.920 --> 56:25.720 + that define the neural network. + +56:25.720 --> 56:28.080 + And you can stare at these tables of numbers + +56:28.080 --> 56:29.600 + till your face turn blue, + +56:29.600 --> 56:32.520 + and you're not gonna understand much + +56:32.520 --> 56:34.520 + about why it made that move. + +56:34.520 --> 56:37.640 + And even if you have natural language processing + +56:37.640 --> 56:40.280 + that can tell you in human language about, + +56:40.280 --> 56:42.520 + oh, five, seven, points, two, eight, + +56:42.520 --> 56:43.560 + still not gonna really help. + +56:43.560 --> 56:47.480 + So I think there's a whole spectrum of fun challenges + +56:47.480 --> 56:50.520 + that are involved in taking a computation + +56:50.520 --> 56:52.240 + that does intelligent things + +56:52.240 --> 56:56.240 + and transforming it into something equally good, + +56:57.760 --> 57:01.840 + equally intelligent, but that's more understandable. + +57:01.840 --> 57:03.240 + And I think that's really valuable + +57:03.240 --> 57:07.440 + because I think as we put machines in charge + +57:07.440 --> 57:09.760 + of ever more infrastructure in our world, + +57:09.760 --> 57:12.680 + the power grid, the trading on the stock market, + +57:12.680 --> 57:14.320 + weapon systems and so on, + +57:14.320 --> 57:17.760 + it's absolutely crucial that we can trust + +57:17.760 --> 57:19.400 + these AIs to do all we want. + +57:19.400 --> 57:21.520 + And trust really comes from understanding + +57:22.520 --> 57:24.400 + in a very fundamental way. + +57:24.400 --> 57:27.560 + And that's why I'm working on this, + +57:27.560 --> 57:29.160 + because I think the more, + +57:29.160 --> 57:31.840 + if we're gonna have some hope of ensuring + +57:31.840 --> 57:33.520 + that machines have adopted our goals + +57:33.520 --> 57:35.800 + and that they're gonna retain them, + +57:35.800 --> 57:38.800 + that kind of trust, I think, + +57:38.800 --> 57:41.200 + needs to be based on things you can actually understand, + +57:41.200 --> 57:44.240 + preferably even improve theorems on. + +57:44.240 --> 57:46.080 + Even with a self driving car, right? + +57:47.040 --> 57:48.680 + If someone just tells you it's been trained + +57:48.680 --> 57:50.640 + on tons of data and it never crashed, + +57:50.640 --> 57:54.200 + it's less reassuring than if someone actually has a proof. + +57:54.200 --> 57:55.960 + Maybe it's a computer verified proof, + +57:55.960 --> 57:58.800 + but still it says that under no circumstances + +57:58.800 --> 58:02.320 + is this car just gonna swerve into oncoming traffic. + +58:02.320 --> 58:04.640 + And that kind of information helps to build trust + +58:04.640 --> 58:08.080 + and helps build the alignment of goals, + +58:09.400 --> 58:12.200 + at least awareness that your goals, your values are aligned. + +58:12.200 --> 58:13.840 + And I think even in the very short term, + +58:13.840 --> 58:16.360 + if you look at how, you know, today, right? + +58:16.360 --> 58:19.320 + This absolutely pathetic state of cybersecurity + +58:19.320 --> 58:21.720 + that we have, where is it? + +58:21.720 --> 58:25.960 + Three billion Yahoo accounts we can't pack, + +58:27.200 --> 58:31.720 + almost every American's credit card and so on. + +58:32.800 --> 58:34.120 + Why is this happening? + +58:34.120 --> 58:37.960 + It's ultimately happening because we have software + +58:37.960 --> 58:41.200 + that nobody fully understood how it worked. + +58:41.200 --> 58:44.800 + That's why the bugs hadn't been found, right? + +58:44.800 --> 58:47.480 + And I think AI can be used very effectively + +58:47.480 --> 58:49.640 + for offense, for hacking, + +58:49.640 --> 58:52.320 + but it can also be used for defense. + +58:52.320 --> 58:55.360 + Hopefully automating verifiability + +58:55.360 --> 59:00.360 + and creating systems that are built in different ways + +59:00.680 --> 59:02.920 + so you can actually prove things about them. + +59:02.920 --> 59:05.240 + And it's important. + +59:05.240 --> 59:07.680 + So speaking of software that nobody understands + +59:07.680 --> 59:10.640 + how it works, of course, a bunch of people ask + +59:10.640 --> 59:12.160 + about your paper, about your thoughts + +59:12.160 --> 59:14.680 + of why does deep and cheap learning work so well? + +59:14.680 --> 59:15.520 + That's the paper. + +59:15.520 --> 59:18.320 + But what are your thoughts on deep learning? + +59:18.320 --> 59:21.880 + These kind of simplified models of our own brains + +59:21.880 --> 59:26.440 + have been able to do some successful perception work, + +59:26.440 --> 59:29.560 + pattern recognition work, and now with AlphaZero and so on, + +59:29.560 --> 59:30.880 + do some clever things. + +59:30.880 --> 59:33.880 + What are your thoughts about the promise limitations + +59:33.880 --> 59:35.680 + of this piece? + +59:35.680 --> 59:40.680 + Great, I think there are a number of very important insights, + +59:43.080 --> 59:44.640 + very important lessons we can always draw + +59:44.640 --> 59:47.120 + from these kinds of successes. + +59:47.120 --> 59:48.960 + One of them is when you look at the human brain, + +59:48.960 --> 59:51.480 + you see it's very complicated, 10th of 11 neurons, + +59:51.480 --> 59:53.320 + and there are all these different kinds of neurons + +59:53.320 --> 59:55.040 + and yada, yada, and there's been this long debate + +59:55.040 --> 59:57.200 + about whether the fact that we have dozens + +59:57.200 --> 1:00:00.160 + of different kinds is actually necessary for intelligence. + +1:00:01.560 --> 1:00:03.360 + We can now, I think, quite convincingly answer + +1:00:03.360 --> 1:00:07.640 + that question of no, it's enough to have just one kind. + +1:00:07.640 --> 1:00:09.920 + If you look under the hood of AlphaZero, + +1:00:09.920 --> 1:00:11.080 + there's only one kind of neuron + +1:00:11.080 --> 1:00:15.000 + and it's ridiculously simple mathematical thing. + +1:00:15.000 --> 1:00:17.280 + So it's just like in physics, + +1:00:17.280 --> 1:00:20.320 + it's not, if you have a gas with waves in it, + +1:00:20.320 --> 1:00:23.240 + it's not the detailed nature of the molecule that matter, + +1:00:24.240 --> 1:00:26.040 + it's the collective behavior somehow. + +1:00:26.040 --> 1:00:30.720 + Similarly, it's this higher level structure + +1:00:30.720 --> 1:00:31.760 + of the network that matters, + +1:00:31.760 --> 1:00:34.080 + not that you have 20 kinds of neurons. + +1:00:34.080 --> 1:00:37.040 + I think our brain is such a complicated mess + +1:00:37.040 --> 1:00:41.720 + because it wasn't evolved just to be intelligent, + +1:00:41.720 --> 1:00:45.840 + it was involved to also be self assembling + +1:00:47.000 --> 1:00:48.760 + and self repairing, right? + +1:00:48.760 --> 1:00:51.920 + And evolutionarily attainable. + +1:00:51.920 --> 1:00:53.560 + And so on and so on. + +1:00:53.560 --> 1:00:54.720 + So I think it's pretty, + +1:00:54.720 --> 1:00:57.040 + my hunch is that we're going to understand + +1:00:57.040 --> 1:00:59.520 + how to build AGI before we fully understand + +1:00:59.520 --> 1:01:02.600 + how our brains work, just like we understood + +1:01:02.600 --> 1:01:05.560 + how to build flying machines long before + +1:01:05.560 --> 1:01:07.800 + we were able to build a mechanical bird. + +1:01:07.800 --> 1:01:08.640 + Yeah, that's right. + +1:01:08.640 --> 1:01:13.280 + You've given the example exactly of mechanical birds + +1:01:13.280 --> 1:01:15.680 + and airplanes and airplanes do a pretty good job + +1:01:15.680 --> 1:01:18.560 + of flying without really mimicking bird flight. + +1:01:18.560 --> 1:01:20.920 + And even now after 100 years later, + +1:01:20.920 --> 1:01:23.880 + did you see the Ted talk with this German mechanical bird? + +1:01:23.880 --> 1:01:25.040 + I heard you mention it. + +1:01:25.040 --> 1:01:26.520 + Check it out, it's amazing. + +1:01:26.520 --> 1:01:27.760 + But even after that, right, + +1:01:27.760 --> 1:01:29.360 + we still don't fly in mechanical birds + +1:01:29.360 --> 1:01:32.720 + because it turned out the way we came up with was simpler + +1:01:32.720 --> 1:01:33.840 + and it's better for our purposes. + +1:01:33.840 --> 1:01:35.280 + And I think it might be the same there. + +1:01:35.280 --> 1:01:36.280 + That's one lesson. + +1:01:37.520 --> 1:01:42.520 + And another lesson, it's more what our paper was about. + +1:01:42.640 --> 1:01:45.800 + First, as a physicist thought it was fascinating + +1:01:45.800 --> 1:01:48.240 + how there's a very close mathematical relationship + +1:01:48.240 --> 1:01:50.800 + actually between our artificial neural networks + +1:01:50.800 --> 1:01:54.560 + and a lot of things that we've studied for in physics + +1:01:54.560 --> 1:01:57.520 + go by nerdy names like the renormalization group equation + +1:01:57.520 --> 1:01:59.800 + and Hamiltonians and yada, yada, yada. + +1:01:59.800 --> 1:02:04.360 + And when you look a little more closely at this, + +1:02:05.720 --> 1:02:06.560 + you have, + +1:02:10.320 --> 1:02:12.360 + at first I was like, well, there's something crazy here + +1:02:12.360 --> 1:02:13.520 + that doesn't make sense. + +1:02:13.520 --> 1:02:18.520 + Because we know that if you even want to build + +1:02:19.200 --> 1:02:22.560 + a super simple neural network to tell apart cat pictures + +1:02:22.560 --> 1:02:23.400 + and dog pictures, right, + +1:02:23.400 --> 1:02:25.400 + that you can do that very, very well now. + +1:02:25.400 --> 1:02:27.520 + But if you think about it a little bit, + +1:02:27.520 --> 1:02:29.080 + you convince yourself it must be impossible + +1:02:29.080 --> 1:02:31.920 + because if I have one megapixel, + +1:02:31.920 --> 1:02:34.160 + even if each pixel is just black or white, + +1:02:34.160 --> 1:02:36.960 + there's two to the power of 1 million possible images, + +1:02:36.960 --> 1:02:38.960 + which is way more than there are atoms in our universe, + +1:02:38.960 --> 1:02:41.000 + right, so in order to, + +1:02:42.040 --> 1:02:43.200 + and then for each one of those, + +1:02:43.200 --> 1:02:44.640 + I have to assign a number, + +1:02:44.640 --> 1:02:47.080 + which is the probability that it's a dog. + +1:02:47.080 --> 1:02:49.440 + So an arbitrary function of images + +1:02:49.440 --> 1:02:54.440 + is a list of more numbers than there are atoms in our universe. + +1:02:54.440 --> 1:02:57.360 + So clearly I can't store that under the hood of my GPU + +1:02:57.360 --> 1:03:00.640 + or my computer, yet somehow it works. + +1:03:00.640 --> 1:03:01.480 + So what does that mean? + +1:03:01.480 --> 1:03:04.960 + Well, it means that out of all of the problems + +1:03:04.960 --> 1:03:08.200 + that you could try to solve with a neural network, + +1:03:10.120 --> 1:03:12.880 + almost all of them are impossible to solve + +1:03:12.880 --> 1:03:14.560 + with a reasonably sized one. + +1:03:15.480 --> 1:03:17.440 + But then what we showed in our paper + +1:03:17.440 --> 1:03:22.360 + was that the fraction, the kind of problems, + +1:03:22.360 --> 1:03:23.800 + the fraction of all the problems + +1:03:23.800 --> 1:03:26.520 + that you could possibly pose, + +1:03:26.520 --> 1:03:29.480 + that we actually care about given the laws of physics + +1:03:29.480 --> 1:03:32.480 + is also an infinite testimony, tiny little part. + +1:03:32.480 --> 1:03:35.440 + And amazingly, they're basically the same part. + +1:03:35.440 --> 1:03:37.560 + Yeah, it's almost like our world was created for, + +1:03:37.560 --> 1:03:39.000 + I mean, they kind of come together. + +1:03:39.000 --> 1:03:42.800 + Yeah, well, you could say maybe where the world was created + +1:03:42.800 --> 1:03:44.960 + for us, but I have a more modest interpretation, + +1:03:44.960 --> 1:03:46.680 + which is that the world was created for us, + +1:03:46.680 --> 1:03:48.040 + but I have a more modest interpretation, + +1:03:48.040 --> 1:03:50.360 + which is that instead evolution endowed us + +1:03:50.360 --> 1:03:53.120 + with neural networks precisely for that reason. + +1:03:53.120 --> 1:03:54.640 + Because this particular architecture, + +1:03:54.640 --> 1:03:56.040 + as opposed to the one in your laptop, + +1:03:56.040 --> 1:04:01.040 + is very, very well adapted to solving the kind of problems + +1:04:02.480 --> 1:04:05.560 + that nature kept presenting our ancestors with. + +1:04:05.560 --> 1:04:08.120 + So it makes sense that why do we have a brain + +1:04:08.120 --> 1:04:09.280 + in the first place? + +1:04:09.280 --> 1:04:11.880 + It's to be able to make predictions about the future + +1:04:11.880 --> 1:04:12.880 + and so on. + +1:04:12.880 --> 1:04:16.440 + So if we had a sucky system, which could never solve it, + +1:04:16.440 --> 1:04:18.280 + we wouldn't have a world. + +1:04:18.280 --> 1:04:23.280 + So this is, I think, a very beautiful fact. + +1:04:23.680 --> 1:04:24.520 + Yeah. + +1:04:24.520 --> 1:04:29.000 + We also realize that there's been earlier work + +1:04:29.000 --> 1:04:32.040 + on why deeper networks are good, + +1:04:32.040 --> 1:04:34.680 + but we were able to show an additional cool fact there, + +1:04:34.680 --> 1:04:38.360 + which is that even incredibly simple problems, + +1:04:38.360 --> 1:04:41.080 + like suppose I give you a thousand numbers + +1:04:41.080 --> 1:04:42.720 + and ask you to multiply them together, + +1:04:42.720 --> 1:04:46.680 + and you can write a few lines of code, boom, done, trivial. + +1:04:46.680 --> 1:04:49.520 + If you just try to do that with a neural network + +1:04:49.520 --> 1:04:52.440 + that has only one single hidden layer in it, + +1:04:52.440 --> 1:04:53.400 + you can do it, + +1:04:54.320 --> 1:04:57.360 + but you're going to need two to the power of a thousand + +1:04:57.360 --> 1:05:00.920 + neurons to multiply a thousand numbers, + +1:05:00.920 --> 1:05:02.520 + which is, again, more neurons than there are atoms + +1:05:02.520 --> 1:05:03.360 + in our universe. + +1:05:04.600 --> 1:05:05.480 + That's fascinating. + +1:05:05.480 --> 1:05:09.960 + But if you allow yourself to make it a deep network + +1:05:09.960 --> 1:05:13.240 + with many layers, you only need 4,000 neurons. + +1:05:13.240 --> 1:05:14.520 + It's perfectly feasible. + +1:05:16.400 --> 1:05:17.960 + That's really interesting. + +1:05:17.960 --> 1:05:18.800 + Yeah. + +1:05:18.800 --> 1:05:21.040 + So on another architecture type, + +1:05:21.040 --> 1:05:22.720 + I mean, you mentioned Schrodinger's equation, + +1:05:22.720 --> 1:05:26.360 + and what are your thoughts about quantum computing + +1:05:27.240 --> 1:05:32.240 + and the role of this kind of computational unit + +1:05:32.400 --> 1:05:34.880 + in creating an intelligence system? + +1:05:34.880 --> 1:05:39.520 + In some Hollywood movies that I will not mention by name + +1:05:39.520 --> 1:05:41.040 + because I don't want to spoil them. + +1:05:41.040 --> 1:05:44.240 + The way they get AGI is building a quantum computer. + +1:05:45.480 --> 1:05:47.600 + Because the word quantum sounds cool and so on. + +1:05:47.600 --> 1:05:48.440 + That's right. + +1:05:50.040 --> 1:05:52.880 + First of all, I think we don't need quantum computers + +1:05:52.880 --> 1:05:54.920 + to build AGI. + +1:05:54.920 --> 1:05:59.240 + I suspect your brain is not a quantum computer + +1:05:59.240 --> 1:06:00.640 + in any profound sense. + +1:06:01.600 --> 1:06:03.200 + So you don't even wrote a paper about that + +1:06:03.200 --> 1:06:04.560 + a lot many years ago. + +1:06:04.560 --> 1:06:08.120 + I calculated the so called decoherence time, + +1:06:08.120 --> 1:06:10.320 + how long it takes until the quantum computerness + +1:06:10.320 --> 1:06:13.400 + of what your neurons are doing gets erased + +1:06:15.320 --> 1:06:17.960 + by just random noise from the environment. + +1:06:17.960 --> 1:06:21.320 + And it's about 10 to the minus 21 seconds. + +1:06:21.320 --> 1:06:24.600 + So as cool as it would be to have a quantum computer + +1:06:24.600 --> 1:06:27.320 + in my head, I don't think that fast. + +1:06:27.320 --> 1:06:28.360 + On the other hand, + +1:06:28.360 --> 1:06:33.040 + there are very cool things you could do + +1:06:33.040 --> 1:06:34.200 + with quantum computers. + +1:06:35.240 --> 1:06:37.480 + Or I think we'll be able to do soon + +1:06:37.480 --> 1:06:39.360 + when we get bigger ones. + +1:06:39.360 --> 1:06:40.960 + That might actually help machine learning + +1:06:40.960 --> 1:06:43.160 + do even better than the brain. + +1:06:43.160 --> 1:06:45.640 + So for example, + +1:06:47.040 --> 1:06:50.760 + one, this is just a moonshot, + +1:06:50.760 --> 1:06:55.760 + but learning is very much same thing as search. + +1:07:01.800 --> 1:07:03.160 + If you're trying to train a neural network + +1:07:03.160 --> 1:07:06.240 + to get really learned to do something really well, + +1:07:06.240 --> 1:07:07.280 + you have some loss function, + +1:07:07.280 --> 1:07:10.360 + you have a bunch of knobs you can turn, + +1:07:10.360 --> 1:07:12.080 + represented by a bunch of numbers, + +1:07:12.080 --> 1:07:12.920 + and you're trying to tweak them + +1:07:12.920 --> 1:07:15.080 + so that it becomes as good as possible at this thing. + +1:07:15.080 --> 1:07:19.680 + So if you think of a landscape with some valley, + +1:07:20.720 --> 1:07:22.120 + where each dimension of the landscape + +1:07:22.120 --> 1:07:24.120 + corresponds to some number you can change, + +1:07:24.120 --> 1:07:25.640 + you're trying to find the minimum. + +1:07:25.640 --> 1:07:26.760 + And it's well known that + +1:07:26.760 --> 1:07:29.040 + if you have a very high dimensional landscape, + +1:07:29.040 --> 1:07:31.840 + complicated things, it's super hard to find the minimum. + +1:07:31.840 --> 1:07:35.840 + Quantum mechanics is amazingly good at this. + +1:07:35.840 --> 1:07:38.240 + Like if I want to know what's the lowest energy state + +1:07:38.240 --> 1:07:39.720 + this water can possibly have, + +1:07:41.720 --> 1:07:42.560 + incredibly hard to compute, + +1:07:42.560 --> 1:07:45.400 + but nature will happily figure this out for you + +1:07:45.400 --> 1:07:48.000 + if you just cool it down, make it very, very cold. + +1:07:49.800 --> 1:07:50.880 + If you put a ball somewhere, + +1:07:50.880 --> 1:07:52.240 + it'll roll down to its minimum. + +1:07:52.240 --> 1:07:54.280 + And this happens metaphorically + +1:07:54.280 --> 1:07:56.320 + at the energy landscape too. + +1:07:56.320 --> 1:07:59.280 + And quantum mechanics even uses some clever tricks, + +1:07:59.280 --> 1:08:02.520 + which today's machine learning systems don't. + +1:08:02.520 --> 1:08:04.160 + Like if you're trying to find the minimum + +1:08:04.160 --> 1:08:06.960 + and you get stuck in the little local minimum here, + +1:08:06.960 --> 1:08:08.760 + in quantum mechanics you can actually tunnel + +1:08:08.760 --> 1:08:11.840 + through the barrier and get unstuck again. + +1:08:13.480 --> 1:08:14.320 + That's really interesting. + +1:08:14.320 --> 1:08:16.120 + Yeah, so it may be, for example, + +1:08:16.120 --> 1:08:19.160 + that we'll one day use quantum computers + +1:08:19.160 --> 1:08:22.840 + that help train neural networks better. + +1:08:22.840 --> 1:08:23.680 + That's really interesting. + +1:08:23.680 --> 1:08:27.040 + Okay, so as a component of kind of the learning process, + +1:08:27.040 --> 1:08:27.880 + for example. + +1:08:27.880 --> 1:08:29.440 + Yeah. + +1:08:29.440 --> 1:08:33.080 + Let me ask sort of wrapping up here a little bit, + +1:08:33.080 --> 1:08:36.880 + let me return to the questions of our human nature + +1:08:36.880 --> 1:08:40.000 + and love, as I mentioned. + +1:08:40.000 --> 1:08:41.640 + So do you think, + +1:08:44.280 --> 1:08:46.000 + you mentioned sort of a helper robot, + +1:08:46.000 --> 1:08:48.640 + but you could think of also personal robots. + +1:08:48.640 --> 1:08:52.480 + Do you think the way we human beings fall in love + +1:08:52.480 --> 1:08:54.680 + and get connected to each other + +1:08:54.680 --> 1:08:58.040 + is possible to achieve in an AI system + +1:08:58.040 --> 1:09:00.360 + and human level AI intelligence system? + +1:09:00.360 --> 1:09:03.720 + Do you think we would ever see that kind of connection? + +1:09:03.720 --> 1:09:06.160 + Or, you know, in all this discussion + +1:09:06.160 --> 1:09:08.520 + about solving complex goals, + +1:09:08.520 --> 1:09:10.760 + is this kind of human social connection, + +1:09:10.760 --> 1:09:12.560 + do you think that's one of the goals + +1:09:12.560 --> 1:09:16.280 + on the peaks and valleys with the raising sea levels + +1:09:16.280 --> 1:09:17.360 + that we'll be able to achieve? + +1:09:17.360 --> 1:09:20.040 + Or do you think that's something that's ultimately, + +1:09:20.040 --> 1:09:21.760 + or at least in the short term, + +1:09:21.760 --> 1:09:23.640 + relative to the other goals is not achievable? + +1:09:23.640 --> 1:09:25.120 + I think it's all possible. + +1:09:25.120 --> 1:09:27.600 + And I mean, in recent, + +1:09:27.600 --> 1:09:30.840 + there's a very wide range of guesses, as you know, + +1:09:30.840 --> 1:09:33.720 + among AI researchers, when we're going to get AGI. + +1:09:35.120 --> 1:09:37.640 + Some people, you know, like our friend Rodney Brooks + +1:09:37.640 --> 1:09:41.040 + says it's going to be hundreds of years at least. + +1:09:41.040 --> 1:09:42.200 + And then there are many others + +1:09:42.200 --> 1:09:44.040 + who think it's going to happen much sooner. + +1:09:44.040 --> 1:09:45.520 + And recent polls, + +1:09:46.840 --> 1:09:48.640 + maybe half or so of AI researchers + +1:09:48.640 --> 1:09:50.920 + think we're going to get AGI within decades. + +1:09:50.920 --> 1:09:52.720 + So if that happens, of course, + +1:09:52.720 --> 1:09:55.040 + then I think these things are all possible. + +1:09:55.040 --> 1:09:56.840 + But in terms of whether it will happen, + +1:09:56.840 --> 1:10:00.600 + I think we shouldn't spend so much time asking + +1:10:00.600 --> 1:10:03.240 + what do we think will happen in the future? + +1:10:03.240 --> 1:10:05.160 + As if we are just some sort of pathetic, + +1:10:05.160 --> 1:10:07.040 + your passive bystanders, you know, + +1:10:07.040 --> 1:10:09.280 + waiting for the future to happen to us. + +1:10:09.280 --> 1:10:11.640 + Hey, we're the ones creating this future, right? + +1:10:11.640 --> 1:10:15.520 + So we should be proactive about it + +1:10:15.520 --> 1:10:16.920 + and ask ourselves what sort of future + +1:10:16.920 --> 1:10:18.240 + we would like to have happen. + +1:10:18.240 --> 1:10:19.920 + We're going to make it like that. + +1:10:19.920 --> 1:10:22.720 + Well, what I prefer is just some sort of incredibly boring, + +1:10:22.720 --> 1:10:24.320 + zombie like future where there's all these + +1:10:24.320 --> 1:10:26.040 + mechanical things happening and there's no passion, + +1:10:26.040 --> 1:10:28.040 + no emotion, no experience, maybe even. + +1:10:29.600 --> 1:10:32.040 + No, I would of course, much rather prefer it + +1:10:32.040 --> 1:10:35.240 + if all the things that we find that we value the most + +1:10:36.240 --> 1:10:40.680 + about humanity are our subjective experience, + +1:10:40.680 --> 1:10:43.000 + passion, inspiration, love, you know. + +1:10:43.000 --> 1:10:48.000 + If we can create a future where those things do happen, + +1:10:48.000 --> 1:10:50.840 + where those things do exist, you know, + +1:10:50.840 --> 1:10:54.560 + I think ultimately it's not our universe + +1:10:54.560 --> 1:10:57.960 + giving meaning to us, it's us giving meaning to our universe. + +1:10:57.960 --> 1:11:01.840 + And if we build more advanced intelligence, + +1:11:01.840 --> 1:11:03.680 + let's make sure we build it in such a way + +1:11:03.680 --> 1:11:08.680 + that meaning is part of it. + +1:11:09.120 --> 1:11:11.400 + A lot of people that seriously study this problem + +1:11:11.400 --> 1:11:13.600 + and think of it from different angles + +1:11:13.600 --> 1:11:16.880 + have trouble in the majority of cases, + +1:11:16.880 --> 1:11:19.160 + if they think through that happen, + +1:11:19.160 --> 1:11:22.520 + are the ones that are not beneficial to humanity. + +1:11:22.520 --> 1:11:25.560 + And so, yeah, so what are your thoughts? + +1:11:25.560 --> 1:11:29.400 + What's should people, you know, + +1:11:29.400 --> 1:11:32.040 + I really don't like people to be terrified. + +1:11:33.440 --> 1:11:35.040 + What's a way for people to think about it + +1:11:35.040 --> 1:11:39.600 + in a way we can solve it and we can make it better? + +1:11:39.600 --> 1:11:42.960 + No, I don't think panicking is going to help in any way. + +1:11:42.960 --> 1:11:44.840 + It's not going to increase chances + +1:11:44.840 --> 1:11:45.880 + of things going well either. + +1:11:45.880 --> 1:11:48.400 + Even if you are in a situation where there is a real threat, + +1:11:48.400 --> 1:11:51.080 + does it help if everybody just freaks out? + +1:11:51.080 --> 1:11:52.680 + No, of course, of course not. + +1:11:53.640 --> 1:11:56.600 + I think, yeah, there are of course ways + +1:11:56.600 --> 1:11:58.440 + in which things can go horribly wrong. + +1:11:59.560 --> 1:12:03.680 + First of all, it's important when we think about this thing, + +1:12:03.680 --> 1:12:05.280 + about the problems and risks, + +1:12:05.280 --> 1:12:07.160 + to also remember how huge the upsides can be + +1:12:07.160 --> 1:12:08.440 + if we get it right, right? + +1:12:08.440 --> 1:12:12.360 + Everything we love about society and civilization + +1:12:12.360 --> 1:12:13.400 + is a product of intelligence. + +1:12:13.400 --> 1:12:15.320 + So if we can amplify our intelligence + +1:12:15.320 --> 1:12:18.760 + with machine intelligence and not anymore lose our loved one + +1:12:18.760 --> 1:12:21.080 + to what we're told is an incurable disease + +1:12:21.080 --> 1:12:24.800 + and things like this, of course, we should aspire to that. + +1:12:24.800 --> 1:12:26.680 + So that can be a motivator, I think, + +1:12:26.680 --> 1:12:29.120 + reminding ourselves that the reason we try to solve problems + +1:12:29.120 --> 1:12:33.520 + is not just because we're trying to avoid gloom, + +1:12:33.520 --> 1:12:35.760 + but because we're trying to do something great. + +1:12:35.760 --> 1:12:37.680 + But then in terms of the risks, + +1:12:37.680 --> 1:12:42.680 + I think the really important question is to ask, + +1:12:42.680 --> 1:12:45.480 + what can we do today that will actually help + +1:12:45.480 --> 1:12:47.320 + make the outcome good, right? + +1:12:47.320 --> 1:12:49.880 + And dismissing the risk is not one of them. + +1:12:51.240 --> 1:12:54.800 + I find it quite funny often when I'm in discussion panels + +1:12:54.800 --> 1:12:55.960 + about these things, + +1:12:55.960 --> 1:13:00.960 + how the people who work for companies, + +1:13:01.200 --> 1:13:03.120 + always be like, oh, nothing to worry about, + +1:13:03.120 --> 1:13:04.760 + nothing to worry about, nothing to worry about. + +1:13:04.760 --> 1:13:09.600 + And it's only academics sometimes express concerns. + +1:13:09.600 --> 1:13:11.880 + That's not surprising at all if you think about it. + +1:13:11.880 --> 1:13:12.880 + Right. + +1:13:12.880 --> 1:13:15.200 + Upton Sinclair quipped, right, + +1:13:15.200 --> 1:13:18.040 + that it's hard to make a man believe in something + +1:13:18.040 --> 1:13:20.120 + when his income depends on not believing in it. + +1:13:20.120 --> 1:13:24.080 + And frankly, we know a lot of these people in companies + +1:13:24.080 --> 1:13:26.240 + that they're just as concerned as anyone else. + +1:13:26.240 --> 1:13:28.480 + But if you're the CEO of a company, + +1:13:28.480 --> 1:13:30.280 + that's not something you want to go on record saying + +1:13:30.280 --> 1:13:33.440 + when you have silly journalists who are gonna put a picture + +1:13:33.440 --> 1:13:35.720 + of a Terminator robot when they quote you. + +1:13:35.720 --> 1:13:39.040 + So the issues are real. + +1:13:39.040 --> 1:13:41.920 + And the way I think about what the issue is, + +1:13:41.920 --> 1:13:46.920 + is basically the real choice we have is, + +1:13:48.040 --> 1:13:50.840 + first of all, are we gonna just dismiss the risks + +1:13:50.840 --> 1:13:54.480 + and say, well, let's just go ahead and build machines + +1:13:54.480 --> 1:13:57.560 + that can do everything we can do better and cheaper. + +1:13:57.560 --> 1:14:00.200 + Let's just make ourselves obsolete as fast as possible. + +1:14:00.200 --> 1:14:01.720 + What could possibly go wrong? + +1:14:01.720 --> 1:14:03.440 + That's one attitude. + +1:14:03.440 --> 1:14:05.440 + The opposite attitude, I think, is to say, + +1:14:06.400 --> 1:14:08.800 + here's this incredible potential, + +1:14:08.800 --> 1:14:11.960 + let's think about what kind of future + +1:14:11.960 --> 1:14:14.640 + we're really, really excited about. + +1:14:14.640 --> 1:14:18.480 + What are the shared goals that we can really aspire towards? + +1:14:18.480 --> 1:14:19.960 + And then let's think really hard + +1:14:19.960 --> 1:14:22.000 + about how we can actually get there. + +1:14:22.000 --> 1:14:24.160 + So start with, don't start thinking about the risks, + +1:14:24.160 --> 1:14:26.720 + start thinking about the goals. + +1:14:26.720 --> 1:14:28.200 + And then when you do that, + +1:14:28.200 --> 1:14:30.480 + then you can think about the obstacles you want to avoid. + +1:14:30.480 --> 1:14:32.840 + I often get students coming in right here into my office + +1:14:32.840 --> 1:14:34.120 + for career advice. + +1:14:34.120 --> 1:14:35.560 + I always ask them this very question, + +1:14:35.560 --> 1:14:37.920 + where do you want to be in the future? + +1:14:37.920 --> 1:14:40.640 + If all she can say is, oh, maybe I'll have cancer, + +1:14:40.640 --> 1:14:42.480 + maybe I'll get run over by a truck. + +1:14:42.480 --> 1:14:44.280 + Yeah, focus on the obstacles instead of the goals. + +1:14:44.280 --> 1:14:46.880 + She's just going to end up a hypochondriac paranoid. + +1:14:47.920 --> 1:14:49.920 + Whereas if she comes in and fire in her eyes + +1:14:49.920 --> 1:14:51.840 + and is like, I want to be there. + +1:14:51.840 --> 1:14:53.960 + And then we can talk about the obstacles + +1:14:53.960 --> 1:14:55.760 + and see how we can circumvent them. + +1:14:55.760 --> 1:14:58.880 + That's, I think, a much, much healthier attitude. + +1:14:58.880 --> 1:15:03.880 + And I feel it's very challenging to come up with a vision + +1:15:03.880 --> 1:15:08.120 + for the future, which we are unequivocally excited about. + +1:15:08.120 --> 1:15:10.320 + I'm not just talking now in the vague terms, + +1:15:10.320 --> 1:15:12.360 + like, yeah, let's cure cancer, fine. + +1:15:12.360 --> 1:15:14.720 + I'm talking about what kind of society + +1:15:14.720 --> 1:15:15.840 + do we want to create? + +1:15:15.840 --> 1:15:20.360 + What do we want it to mean to be human in the age of AI, + +1:15:20.360 --> 1:15:21.720 + in the age of AGI? + +1:15:22.840 --> 1:15:25.360 + So if we can have this conversation, + +1:15:25.360 --> 1:15:28.200 + broad, inclusive conversation, + +1:15:28.200 --> 1:15:31.400 + and gradually start converging towards some, + +1:15:31.400 --> 1:15:34.240 + some future that with some direction, at least, + +1:15:34.240 --> 1:15:35.400 + that we want to steer towards, right, + +1:15:35.400 --> 1:15:38.160 + then we'll be much more motivated + +1:15:38.160 --> 1:15:39.960 + to constructively take on the obstacles. + +1:15:39.960 --> 1:15:43.560 + And I think if I had, if I had to, + +1:15:43.560 --> 1:15:46.640 + if I try to wrap this up in a more succinct way, + +1:15:46.640 --> 1:15:51.480 + I think we can all agree already now + +1:15:51.480 --> 1:15:56.160 + that we should aspire to build AGI + +1:15:56.160 --> 1:16:05.160 + that doesn't overpower us, but that empowers us. + +1:16:05.160 --> 1:16:08.560 + And think of the many various ways that can do that, + +1:16:08.560 --> 1:16:11.000 + whether that's from my side of the world + +1:16:11.000 --> 1:16:12.720 + of autonomous vehicles. + +1:16:12.720 --> 1:16:14.720 + I'm personally actually from the camp + +1:16:14.720 --> 1:16:16.800 + that believes this human level intelligence + +1:16:16.800 --> 1:16:20.480 + is required to achieve something like vehicles + +1:16:20.480 --> 1:16:23.880 + that would actually be something we would enjoy using + +1:16:23.880 --> 1:16:25.120 + and being part of. + +1:16:25.120 --> 1:16:27.040 + So that's one example, and certainly there's a lot + +1:16:27.040 --> 1:16:30.920 + of other types of robots and medicine and so on. + +1:16:30.920 --> 1:16:33.880 + So focusing on those and then coming up with the obstacles, + +1:16:33.880 --> 1:16:35.920 + coming up with the ways that that can go wrong + +1:16:35.920 --> 1:16:38.160 + and solving those one at a time. + +1:16:38.160 --> 1:16:41.520 + And just because you can build an autonomous vehicle, + +1:16:41.520 --> 1:16:42.800 + even if you could build one + +1:16:42.800 --> 1:16:45.080 + that would drive just fine without you, + +1:16:45.080 --> 1:16:46.720 + maybe there are some things in life + +1:16:46.720 --> 1:16:48.400 + that we would actually want to do ourselves. + +1:16:48.400 --> 1:16:49.240 + That's right. + +1:16:49.240 --> 1:16:51.400 + Right, like, for example, + +1:16:51.400 --> 1:16:53.040 + if you think of our society as a whole, + +1:16:53.040 --> 1:16:56.320 + there are some things that we find very meaningful to do. + +1:16:57.200 --> 1:16:59.640 + And that doesn't mean we have to stop doing them + +1:16:59.640 --> 1:17:02.000 + just because machines can do them better. + +1:17:02.000 --> 1:17:04.080 + I'm not gonna stop playing tennis + +1:17:04.080 --> 1:17:07.360 + just the day someone builds a tennis robot and beat me. + +1:17:07.360 --> 1:17:09.600 + People are still playing chess and even go. + +1:17:09.600 --> 1:17:14.600 + Yeah, and in the very near term even, + +1:17:14.600 --> 1:17:18.880 + some people are advocating basic income, replace jobs. + +1:17:18.880 --> 1:17:20.840 + But if the government is gonna be willing + +1:17:20.840 --> 1:17:24.040 + to just hand out cash to people for doing nothing, + +1:17:24.040 --> 1:17:25.840 + then one should also seriously consider + +1:17:25.840 --> 1:17:27.640 + whether the government should also hire + +1:17:27.640 --> 1:17:29.480 + a lot more teachers and nurses + +1:17:29.480 --> 1:17:32.160 + and the kind of jobs which people often + +1:17:32.160 --> 1:17:34.440 + find great fulfillment in doing, right? + +1:17:34.440 --> 1:17:36.320 + We get very tired of hearing politicians saying, + +1:17:36.320 --> 1:17:39.320 + oh, we can't afford hiring more teachers, + +1:17:39.320 --> 1:17:41.480 + but we're gonna maybe have basic income. + +1:17:41.480 --> 1:17:44.000 + If we can have more serious research and thought + +1:17:44.000 --> 1:17:46.200 + into what gives meaning to our lives, + +1:17:46.200 --> 1:17:48.960 + the jobs give so much more than income, right? + +1:17:48.960 --> 1:17:50.520 + Mm hmm. + +1:17:50.520 --> 1:17:53.320 + And then think about in the future, + +1:17:53.320 --> 1:17:58.320 + what are the roles that we wanna have people + +1:18:00.000 --> 1:18:03.040 + continually feeling empowered by machines? + +1:18:03.040 --> 1:18:06.120 + And I think sort of, I come from Russia, + +1:18:06.120 --> 1:18:07.240 + from the Soviet Union. + +1:18:07.240 --> 1:18:10.160 + And I think for a lot of people in the 20th century, + +1:18:10.160 --> 1:18:14.080 + going to the moon, going to space was an inspiring thing. + +1:18:14.080 --> 1:18:18.080 + I feel like the universe of the mind, + +1:18:18.080 --> 1:18:20.880 + so AI, understanding, creating intelligence + +1:18:20.880 --> 1:18:23.240 + is that for the 21st century. + +1:18:23.240 --> 1:18:24.400 + So it's really surprising. + +1:18:24.400 --> 1:18:25.640 + And I've heard you mention this. + +1:18:25.640 --> 1:18:27.400 + It's really surprising to me, + +1:18:27.400 --> 1:18:29.240 + both on the research funding side, + +1:18:29.240 --> 1:18:31.760 + that it's not funded as greatly as it could be, + +1:18:31.760 --> 1:18:34.760 + but most importantly, on the politician side, + +1:18:34.760 --> 1:18:36.520 + that it's not part of the public discourse + +1:18:36.520 --> 1:18:40.800 + except in the killer bots terminator kind of view, + +1:18:40.800 --> 1:18:44.880 + that people are not yet, I think, perhaps excited + +1:18:44.880 --> 1:18:46.680 + by the possible positive future + +1:18:46.680 --> 1:18:48.120 + that we can build together. + +1:18:48.120 --> 1:18:51.520 + So we should be, because politicians usually just focus + +1:18:51.520 --> 1:18:53.320 + on the next election cycle, right? + +1:18:54.480 --> 1:18:57.160 + The single most important thing I feel we humans have learned + +1:18:57.160 --> 1:18:59.320 + in the entire history of science + +1:18:59.320 --> 1:19:02.040 + is they were the masters of underestimation. + +1:19:02.040 --> 1:19:07.040 + We underestimated the size of our cosmos again and again, + +1:19:08.480 --> 1:19:10.200 + realizing that everything we thought existed + +1:19:10.200 --> 1:19:12.240 + was just a small part of something grander, right? + +1:19:12.240 --> 1:19:16.640 + Planet, solar system, the galaxy, clusters of galaxies. + +1:19:16.640 --> 1:19:17.560 + The universe. + +1:19:18.440 --> 1:19:23.120 + And we now know that the future has just + +1:19:23.120 --> 1:19:25.160 + so much more potential + +1:19:25.160 --> 1:19:27.640 + than our ancestors could ever have dreamt of. + +1:19:27.640 --> 1:19:32.360 + This cosmos, imagine if all of Earth + +1:19:33.600 --> 1:19:35.440 + was completely devoid of life, + +1:19:36.640 --> 1:19:38.520 + except for Cambridge, Massachusetts. + +1:19:39.560 --> 1:19:42.680 + Wouldn't it be kind of lame if all we ever aspired to + +1:19:42.680 --> 1:19:45.560 + was to stay in Cambridge, Massachusetts forever + +1:19:45.560 --> 1:19:47.160 + and then go extinct in one week, + +1:19:47.160 --> 1:19:49.760 + even though Earth was gonna continue on for longer? + +1:19:49.760 --> 1:19:52.800 + That sort of attitude I think we have now + +1:19:54.200 --> 1:19:57.800 + on the cosmic scale, life can flourish on Earth, + +1:19:57.800 --> 1:20:00.840 + not for four years, but for billions of years. + +1:20:00.840 --> 1:20:02.920 + I can even tell you about how to move it out of harm's way + +1:20:02.920 --> 1:20:04.840 + when the sun gets too hot. + +1:20:04.840 --> 1:20:09.520 + And then we have so much more resources out here, + +1:20:09.520 --> 1:20:12.480 + which today, maybe there are a lot of other planets + +1:20:12.480 --> 1:20:14.960 + with bacteria or cow like life on them, + +1:20:14.960 --> 1:20:19.880 + but most of this, all this opportunity seems, + +1:20:19.880 --> 1:20:22.440 + as far as we can tell, to be largely dead, + +1:20:22.440 --> 1:20:23.560 + like the Sahara Desert. + +1:20:23.560 --> 1:20:28.480 + And yet we have the opportunity to help life flourish + +1:20:28.480 --> 1:20:30.280 + around this for billions of years. + +1:20:30.280 --> 1:20:32.680 + So let's quit squabbling about + +1:20:34.080 --> 1:20:36.480 + whether some little border should be drawn + +1:20:36.480 --> 1:20:38.440 + one mile to the left or right, + +1:20:38.440 --> 1:20:41.080 + and look up into the skies and realize, + +1:20:41.080 --> 1:20:44.040 + hey, we can do such incredible things. + +1:20:44.040 --> 1:20:46.640 + Yeah, and that's, I think, why it's really exciting + +1:20:46.640 --> 1:20:49.440 + that you and others are connected + +1:20:49.440 --> 1:20:51.880 + with some of the work Elon Musk is doing, + +1:20:51.880 --> 1:20:54.480 + because he's literally going out into that space, + +1:20:54.480 --> 1:20:57.000 + really exploring our universe, and it's wonderful. + +1:20:57.000 --> 1:21:02.000 + That is exactly why Elon Musk is so misunderstood, right? + +1:21:02.000 --> 1:21:05.000 + Misconstrued him as some kind of pessimistic doomsayer. + +1:21:05.000 --> 1:21:07.640 + The reason he cares so much about AI safety + +1:21:07.640 --> 1:21:12.080 + is because he more than almost anyone else appreciates + +1:21:12.080 --> 1:21:14.280 + these amazing opportunities that we'll squander + +1:21:14.280 --> 1:21:16.640 + if we wipe out here on Earth. + +1:21:16.640 --> 1:21:19.680 + We're not just going to wipe out the next generation, + +1:21:19.680 --> 1:21:23.320 + all generations, and this incredible opportunity + +1:21:23.320 --> 1:21:25.400 + that's out there, and that would really be a waste. + +1:21:25.400 --> 1:21:30.080 + And AI, for people who think that it would be better + +1:21:30.080 --> 1:21:33.600 + to do without technology, let me just mention that + +1:21:34.680 --> 1:21:36.320 + if we don't improve our technology, + +1:21:36.320 --> 1:21:39.320 + the question isn't whether humanity is going to go extinct. + +1:21:39.320 --> 1:21:41.160 + The question is just whether we're going to get taken out + +1:21:41.160 --> 1:21:44.800 + by the next big asteroid or the next super volcano + +1:21:44.800 --> 1:21:48.280 + or something else dumb that we could easily prevent + +1:21:48.280 --> 1:21:49.840 + with more tech, right? + +1:21:49.840 --> 1:21:53.160 + And if we want life to flourish throughout the cosmos, + +1:21:53.160 --> 1:21:54.760 + AI is the key to it. + +1:21:56.120 --> 1:21:59.840 + As I mentioned in a lot of detail in my book right there, + +1:21:59.840 --> 1:22:04.840 + even many of the most inspired sci fi writers, + +1:22:04.880 --> 1:22:08.120 + I feel have totally underestimated the opportunities + +1:22:08.120 --> 1:22:11.240 + for space travel, especially at the other galaxies, + +1:22:11.240 --> 1:22:15.360 + because they weren't thinking about the possibility of AGI, + +1:22:15.360 --> 1:22:17.520 + which just makes it so much easier. + +1:22:17.520 --> 1:22:18.440 + Right, yeah. + +1:22:18.440 --> 1:22:23.440 + So that goes to your view of AGI that enables our progress, + +1:22:24.080 --> 1:22:25.760 + that enables a better life. + +1:22:25.760 --> 1:22:28.320 + So that's a beautiful way to put it + +1:22:28.320 --> 1:22:29.960 + and then something to strive for. + +1:22:29.960 --> 1:22:31.440 + So Max, thank you so much. + +1:22:31.440 --> 1:22:32.560 + Thank you for your time today. + +1:22:32.560 --> 1:22:33.560 + It's been awesome. + +1:22:33.560 --> 1:22:34.400 + Thank you so much. + +1:22:34.400 --> 1:22:35.240 + Thanks. + +1:22:35.240 --> 1:22:40.240 + Have a great day. +