diff --git "a/vtt/episode_017_small.vtt" "b/vtt/episode_017_small.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_017_small.vtt" @@ -0,0 +1,6362 @@ +WEBVTT + +00:00.000 --> 00:02.920 + The following is a conversation with Greg Brockman. + +00:02.920 --> 00:05.400 + He's the cofounder and CTO of OpenAI, + +00:05.400 --> 00:07.440 + a world class research organization + +00:07.440 --> 00:10.840 + developing ideas in AI with a goal of eventually + +00:10.840 --> 00:14.200 + creating a safe and friendly artificial general + +00:14.200 --> 00:18.840 + intelligence, one that benefits and empowers humanity. + +00:18.840 --> 00:22.520 + OpenAI is not only a source of publications, algorithms, + +00:22.520 --> 00:24.480 + tools, and data sets. + +00:24.480 --> 00:28.120 + Their mission is a catalyst for an important public discourse + +00:28.120 --> 00:32.720 + about our future with both narrow and general intelligence + +00:32.720 --> 00:34.000 + systems. + +00:34.000 --> 00:36.640 + This conversation is part of the Artificial Intelligence + +00:36.640 --> 00:39.480 + Podcast at MIT and beyond. + +00:39.480 --> 00:42.720 + If you enjoy it, subscribe on YouTube, iTunes, + +00:42.720 --> 00:44.520 + or simply connect with me on Twitter + +00:44.520 --> 00:48.040 + at Lex Friedman, spelled F R I D. + +00:48.040 --> 00:52.760 + And now here's my conversation with Greg Brockman. + +00:52.760 --> 00:54.640 + So in high school and right after you wrote + +00:54.640 --> 00:56.640 + a draft of a chemistry textbook, + +00:56.640 --> 00:58.120 + I saw that that covers everything + +00:58.120 --> 01:01.400 + from basic structure of the atom to quantum mechanics. + +01:01.400 --> 01:04.400 + So it's clear you have an intuition and a passion + +01:04.400 --> 01:09.880 + for both the physical world with chemistry and non robotics + +01:09.880 --> 01:13.680 + to the digital world with AI, deep learning, + +01:13.680 --> 01:15.360 + reinforcement learning, so on. + +01:15.360 --> 01:17.360 + Do you see the physical world and the digital world + +01:17.360 --> 01:18.600 + as different? + +01:18.600 --> 01:20.480 + And what do you think is the gap? + +01:20.480 --> 01:23.040 + A lot of it actually boils down to iteration speed, + +01:23.040 --> 01:25.240 + that I think that a lot of what really motivates me + +01:25.240 --> 01:27.720 + is building things, right? + +01:27.720 --> 01:29.320 + Think about mathematics, for example, + +01:29.320 --> 01:30.920 + where you think really hard about a problem. + +01:30.920 --> 01:31.680 + You understand it. + +01:31.680 --> 01:33.320 + You're right down in this very obscure form + +01:33.320 --> 01:34.560 + that we call a proof. + +01:34.560 --> 01:37.600 + But then this is in humanity's library, right? + +01:37.600 --> 01:38.400 + It's there forever. + +01:38.400 --> 01:40.400 + This is some truth that we've discovered. + +01:40.400 --> 01:43.040 + And maybe only five people in your field will ever read it. + +01:43.040 --> 01:45.440 + But somehow you've kind of moved humanity forward. + +01:45.440 --> 01:46.880 + And so I actually used to really think + +01:46.880 --> 01:48.680 + that I was going to be a mathematician. + +01:48.680 --> 01:51.400 + And then I actually started writing this chemistry textbook. + +01:51.400 --> 01:53.360 + One of my friends told me, you'll never publish it + +01:53.360 --> 01:54.880 + because you don't have a PhD. + +01:54.880 --> 01:58.000 + So instead, I decided to build a website + +01:58.000 --> 01:59.960 + and try to promote my ideas that way. + +01:59.960 --> 02:01.480 + And then I discovered programming. + +02:01.480 --> 02:05.320 + And in programming, you think hard about a problem. + +02:05.320 --> 02:06.080 + You understand it. + +02:06.080 --> 02:08.040 + You're right down in a very obscure form + +02:08.040 --> 02:09.720 + that we call a program. + +02:09.720 --> 02:12.240 + But then once again, it's in humanity's library, right? + +02:12.240 --> 02:14.120 + And anyone can get the benefit from it. + +02:14.120 --> 02:15.720 + And the scalability is massive. + +02:15.720 --> 02:17.720 + And so I think that the thing that really appeals to me + +02:17.720 --> 02:19.440 + about the digital world is that you + +02:19.440 --> 02:21.960 + can have this insane leverage, right? + +02:21.960 --> 02:24.960 + A single individual with an idea is able to affect + +02:24.960 --> 02:26.120 + the entire planet. + +02:26.120 --> 02:28.240 + And that's something I think is really hard to do + +02:28.240 --> 02:30.240 + if you're moving around physical atoms. + +02:30.240 --> 02:32.440 + But you said mathematics. + +02:32.440 --> 02:36.880 + So if you look at the wet thing over here, our mind, + +02:36.880 --> 02:39.800 + do you ultimately see it as just math, + +02:39.800 --> 02:41.800 + as just information processing? + +02:41.800 --> 02:44.880 + Or is there some other magic if you've + +02:44.880 --> 02:46.960 + seen through biology and chemistry and so on? + +02:46.960 --> 02:49.000 + I think it's really interesting to think about humans + +02:49.000 --> 02:51.000 + as just information processing systems. + +02:51.000 --> 02:54.080 + And it seems like it's actually a pretty good way + +02:54.080 --> 02:57.560 + of describing a lot of how the world works or a lot of what + +02:57.560 --> 03:01.000 + we're capable of to think that, again, if you just + +03:01.000 --> 03:03.640 + look at technological innovations over time, + +03:03.640 --> 03:05.920 + that in some ways, the most transformative innovation + +03:05.920 --> 03:07.760 + that we've had has been the computer, right? + +03:07.760 --> 03:10.560 + In some ways, the internet, what has the internet done? + +03:10.560 --> 03:12.760 + The internet is not about these physical cables. + +03:12.760 --> 03:14.560 + It's about the fact that I am suddenly + +03:14.560 --> 03:16.560 + able to instantly communicate with any other human + +03:16.560 --> 03:17.680 + on the planet. + +03:17.680 --> 03:19.680 + I'm able to retrieve any piece of knowledge + +03:19.680 --> 03:22.640 + that, in some ways, the human race has ever had, + +03:22.640 --> 03:26.080 + and that those are these insane transformations. + +03:26.080 --> 03:29.320 + Do you see our society as a whole the collective + +03:29.320 --> 03:31.240 + as another extension of the intelligence + +03:31.240 --> 03:32.280 + of the human being? + +03:32.280 --> 03:34.440 + So if you look at the human being as an information processing + +03:34.440 --> 03:36.920 + system, you mentioned the internet, the networking. + +03:36.920 --> 03:39.360 + Do you see us all together as a civilization + +03:39.360 --> 03:41.680 + as a kind of intelligent system? + +03:41.680 --> 03:43.560 + Yeah, I think this is actually a really interesting + +03:43.560 --> 03:45.840 + perspective to take and to think about + +03:45.840 --> 03:48.080 + that you sort of have this collective intelligence + +03:48.080 --> 03:49.520 + of all of society. + +03:49.520 --> 03:51.680 + The economy itself is this superhuman machine + +03:51.680 --> 03:54.440 + that is optimizing something, right? + +03:54.440 --> 03:56.400 + And it's almost, in some ways, a company + +03:56.400 --> 03:57.960 + has a will of its own, right? + +03:57.960 --> 03:59.400 + That you have all these individuals who are all + +03:59.400 --> 04:00.800 + pursuing their own individual goals + +04:00.800 --> 04:02.400 + and thinking really hard and thinking + +04:02.400 --> 04:04.640 + about the right things to do, but somehow the company does + +04:04.640 --> 04:07.880 + something that is this emergent thing + +04:07.880 --> 04:10.640 + and that it's a really useful abstraction. + +04:10.640 --> 04:12.440 + And so I think that in some ways, + +04:12.440 --> 04:14.880 + we think of ourselves as the most intelligent things + +04:14.880 --> 04:17.480 + on the planet and the most powerful things on the planet. + +04:17.480 --> 04:19.320 + But there are things that are bigger than us, + +04:19.320 --> 04:21.480 + that are these systems that we all contribute to. + +04:21.480 --> 04:25.000 + And so I think actually, it's interesting to think about, + +04:25.000 --> 04:27.440 + if you've read Asa Geismov's foundation, right, + +04:27.440 --> 04:30.160 + that there's this concept of psycho history in there, + +04:30.160 --> 04:31.920 + which is effectively this, that if you have trillions + +04:31.920 --> 04:35.200 + or quadrillions of beings, then maybe you could actually + +04:35.200 --> 04:39.080 + predict what that huge macro being will do + +04:39.080 --> 04:42.400 + and almost independent of what the individuals want. + +04:42.400 --> 04:44.240 + And I actually have a second angle on this + +04:44.240 --> 04:46.760 + that I think is interesting, which is thinking about + +04:46.760 --> 04:48.400 + technological determinism. + +04:48.400 --> 04:51.480 + One thing that I actually think a lot about with OpenAI + +04:51.480 --> 04:54.720 + is that we're kind of coming onto this insanely + +04:54.720 --> 04:57.400 + transformational technology of general intelligence + +04:57.400 --> 04:58.760 + that will happen at some point. + +04:58.760 --> 05:01.560 + And there's a question of how can you take actions + +05:01.560 --> 05:04.880 + that will actually steer it to go better rather than worse? + +05:04.880 --> 05:06.720 + And that I think one question you need to ask is, + +05:06.720 --> 05:09.320 + as a scientist, as an event or as a creator, + +05:09.320 --> 05:11.720 + what impact can you have in general? + +05:11.720 --> 05:12.880 + You look at things like the telephone + +05:12.880 --> 05:14.840 + invented by two people on the same day. + +05:14.840 --> 05:16.600 + Like what does that mean, like what does that mean + +05:16.600 --> 05:18.080 + about the shape of innovation? + +05:18.080 --> 05:20.160 + And I think that what's going on is everyone's building + +05:20.160 --> 05:21.720 + on the shoulders of the same giants. + +05:21.720 --> 05:23.840 + And so you can kind of, you can't really hope + +05:23.840 --> 05:25.720 + to create something no one else ever would. + +05:25.720 --> 05:27.040 + You know, if Einstein wasn't born, + +05:27.040 --> 05:29.200 + someone else would have come up with relativity. + +05:29.200 --> 05:31.000 + You know, he changed the timeline a bit, right? + +05:31.000 --> 05:33.000 + That maybe it would have taken another 20 years, + +05:33.000 --> 05:34.560 + but it wouldn't be that fundamentally humanity + +05:34.560 --> 05:37.360 + would never discover these fundamental truths. + +05:37.360 --> 05:40.440 + So there's some kind of invisible momentum + +05:40.440 --> 05:45.400 + that some people like Einstein or OpenAI is plugging into + +05:45.400 --> 05:47.800 + that anybody else can also plug into. + +05:47.800 --> 05:50.800 + And ultimately, that wave takes us into a certain direction. + +05:50.800 --> 05:51.840 + That's what you mean by digitalism? + +05:51.840 --> 05:52.840 + That's right, that's right. + +05:52.840 --> 05:54.240 + And you know, this kind of seems to play out + +05:54.240 --> 05:55.720 + in a bunch of different ways. + +05:55.720 --> 05:58.040 + That there's some exponential that is being written + +05:58.040 --> 05:59.960 + and that the exponential itself, which one it is, + +05:59.960 --> 06:01.520 + changes, think about Moore's Law, + +06:01.520 --> 06:04.800 + an entire industry set, it's clocked to it for 50 years. + +06:04.800 --> 06:06.200 + Like how can that be, right? + +06:06.200 --> 06:07.360 + How is that possible? + +06:07.360 --> 06:09.320 + And yet somehow it happened. + +06:09.320 --> 06:12.200 + And so I think you can't hope to ever invent something + +06:12.200 --> 06:13.360 + that no one else will. + +06:13.360 --> 06:15.360 + Maybe you can change the timeline a little bit. + +06:15.360 --> 06:17.400 + But if you really want to make a difference, + +06:17.400 --> 06:19.440 + I think that the thing that you really have to do, + +06:19.440 --> 06:21.320 + the only real degree of freedom you have + +06:21.320 --> 06:23.040 + is to set the initial conditions + +06:23.040 --> 06:24.960 + under which a technology is born. + +06:24.960 --> 06:26.680 + And so you think about the internet, right? + +06:26.680 --> 06:27.840 + That there are lots of other competitors + +06:27.840 --> 06:29.400 + trying to build similar things. + +06:29.400 --> 06:33.240 + And the internet one, and that the initial conditions + +06:33.240 --> 06:34.680 + where it was created by this group + +06:34.680 --> 06:37.760 + that really valued people being able to be, + +06:37.760 --> 06:39.120 + you know, anyone being able to plug in + +06:39.120 --> 06:42.480 + this very academic mindset of being open and connected. + +06:42.480 --> 06:44.400 + And I think that the internet for the next 40 years + +06:44.400 --> 06:46.360 + really played out that way. + +06:46.360 --> 06:47.680 + You know, maybe today, + +06:47.680 --> 06:49.840 + things are starting to shift in a different direction, + +06:49.840 --> 06:51.120 + but I think that those initial conditions + +06:51.120 --> 06:52.720 + were really important to determine + +06:52.720 --> 06:55.080 + the next 40 years worth of progress. + +06:55.080 --> 06:56.440 + That's really beautifully put. + +06:56.440 --> 06:58.800 + So another example of that I think about, + +06:58.800 --> 07:00.800 + you know, I recently looked at it. + +07:00.800 --> 07:03.800 + I looked at Wikipedia, the formation of Wikipedia. + +07:03.800 --> 07:05.520 + And I wonder what the internet would be like + +07:05.520 --> 07:07.760 + if Wikipedia had ads. + +07:07.760 --> 07:09.640 + You know, there's an interesting argument + +07:09.640 --> 07:14.280 + that why they chose not to put advertisement on Wikipedia. + +07:14.280 --> 07:17.800 + I think Wikipedia is one of the greatest resources + +07:17.800 --> 07:18.920 + we have on the internet. + +07:18.920 --> 07:21.280 + It's extremely surprising how well it works + +07:21.280 --> 07:22.960 + and how well it was able to aggregate + +07:22.960 --> 07:25.000 + all this kind of good information. + +07:25.000 --> 07:27.320 + And essentially the creator of Wikipedia, + +07:27.320 --> 07:29.360 + I don't know, there's probably some debates there, + +07:29.360 --> 07:31.200 + but set the initial conditions + +07:31.200 --> 07:33.240 + and now it carried itself forward. + +07:33.240 --> 07:34.080 + That's really interesting. + +07:34.080 --> 07:36.520 + So the way you're thinking about AGI + +07:36.520 --> 07:38.640 + or artificial intelligence is you're focused on + +07:38.640 --> 07:41.200 + setting the initial conditions for the progress. + +07:41.200 --> 07:42.320 + That's right. + +07:42.320 --> 07:43.160 + That's powerful. + +07:43.160 --> 07:45.560 + Okay, so look into the future. + +07:45.560 --> 07:48.160 + If you create an AGI system, + +07:48.160 --> 07:51.560 + like one that can ace the Turing test, natural language, + +07:51.560 --> 07:54.800 + what do you think would be the interactions + +07:54.800 --> 07:55.840 + you would have with it? + +07:55.840 --> 07:57.720 + What do you think are the questions you would ask? + +07:57.720 --> 08:00.560 + Like what would be the first question you would ask? + +08:00.560 --> 08:01.840 + It, her, him. + +08:01.840 --> 08:02.680 + That's right. + +08:02.680 --> 08:03.920 + I think that at that point, + +08:03.920 --> 08:05.960 + if you've really built a powerful system + +08:05.960 --> 08:08.480 + that is capable of shaping the future of humanity, + +08:08.480 --> 08:10.240 + the first question that you really should ask + +08:10.240 --> 08:12.280 + is how do we make sure that this plays out well? + +08:12.280 --> 08:13.960 + And so that's actually the first question + +08:13.960 --> 08:17.600 + that I would ask a powerful AGI system is. + +08:17.600 --> 08:19.160 + So you wouldn't ask your colleague, + +08:19.160 --> 08:22.280 + you wouldn't ask like Ilya, you would ask the AGI system. + +08:22.280 --> 08:24.640 + Oh, we've already had the conversation with Ilya, right? + +08:24.640 --> 08:25.720 + And everyone here. + +08:25.720 --> 08:27.480 + And so you want as many perspectives + +08:27.480 --> 08:29.720 + and a piece of wisdom as you can + +08:29.720 --> 08:31.200 + for answering this question. + +08:31.200 --> 08:33.120 + So I don't think you necessarily defer to + +08:33.120 --> 08:35.480 + whatever your powerful system tells you, + +08:35.480 --> 08:37.120 + but you use it as one input + +08:37.120 --> 08:39.280 + to try to figure out what to do. + +08:39.280 --> 08:40.920 + But, and I guess fundamentally, + +08:40.920 --> 08:42.160 + what it really comes down to is + +08:42.160 --> 08:43.960 + if you built something really powerful + +08:43.960 --> 08:45.280 + and you think about, think about, for example, + +08:45.280 --> 08:47.640 + the creation of, of shortly after + +08:47.640 --> 08:48.880 + the creation of nuclear weapons, right? + +08:48.880 --> 08:50.400 + The most important question in the world + +08:50.400 --> 08:52.800 + was what's the world we're going to be like? + +08:52.800 --> 08:54.880 + How do we set ourselves up in a place + +08:54.880 --> 08:58.320 + where we're going to be able to survive as a species? + +08:58.320 --> 09:00.640 + With AGI, I think the question is slightly different, right? + +09:00.640 --> 09:02.720 + That there is a question of how do we make sure + +09:02.720 --> 09:04.440 + that we don't get the negative effects? + +09:04.440 --> 09:06.240 + But there's also the positive side, right? + +09:06.240 --> 09:08.040 + You imagine that, you know, like, + +09:08.040 --> 09:09.720 + like what will AGI be like? + +09:09.720 --> 09:11.280 + Like what will it be capable of? + +09:11.280 --> 09:13.520 + And I think that one of the core reasons + +09:13.520 --> 09:15.760 + that an AGI can be powerful and transformative + +09:15.760 --> 09:18.920 + is actually due to technological development, right? + +09:18.920 --> 09:20.560 + If you have something that's capable, + +09:20.560 --> 09:23.880 + that's capable as a human and that it's much more scalable, + +09:23.880 --> 09:25.880 + that you absolutely want that thing + +09:25.880 --> 09:27.640 + to go read the whole scientific literature + +09:27.640 --> 09:30.000 + and think about how to create cures for all the diseases, right? + +09:30.000 --> 09:31.480 + You want it to think about how to go + +09:31.480 --> 09:33.360 + and build technologies to help us + +09:33.360 --> 09:37.320 + create material abundance and to figure out societal problems + +09:37.320 --> 09:38.160 + that we have trouble with, + +09:38.160 --> 09:40.000 + like how are we supposed to clean up the environment? + +09:40.000 --> 09:42.200 + And, you know, maybe you want this + +09:42.200 --> 09:44.120 + to go and invent a bunch of little robots that will go out + +09:44.120 --> 09:47.280 + and be biodegradable and turn ocean debris + +09:47.280 --> 09:49.640 + into harmless molecules. + +09:49.640 --> 09:54.040 + And I think that that positive side + +09:54.040 --> 09:55.720 + is something that I think people miss + +09:55.720 --> 09:58.160 + sometimes when thinking about what an AGI will be like. + +09:58.160 --> 10:00.280 + And so I think that if you have a system + +10:00.280 --> 10:01.640 + that's capable of all of that, + +10:01.640 --> 10:03.960 + you absolutely want its advice about how do I make sure + +10:03.960 --> 10:07.600 + that we're using your capabilities + +10:07.600 --> 10:09.200 + in a positive way for humanity. + +10:09.200 --> 10:11.400 + So what do you think about that psychology + +10:11.400 --> 10:14.800 + that looks at all the different possible trajectories + +10:14.800 --> 10:17.520 + of an AGI system, many of which, + +10:17.520 --> 10:19.960 + perhaps the majority of which are positive + +10:19.960 --> 10:23.320 + and nevertheless focuses on the negative trajectories? + +10:23.320 --> 10:24.720 + I mean, you get to interact with folks, + +10:24.720 --> 10:28.840 + you get to think about this maybe within yourself as well. + +10:28.840 --> 10:30.560 + You look at Sam Harris and so on. + +10:30.560 --> 10:32.720 + It seems to be, sorry to put it this way, + +10:32.720 --> 10:36.720 + but almost more fun to think about the negative possibilities. + +10:37.800 --> 10:39.560 + Whatever that's deep in our psychology, + +10:39.560 --> 10:40.760 + what do you think about that? + +10:40.760 --> 10:41.920 + And how do we deal with it? + +10:41.920 --> 10:44.400 + Because we want AI to help us. + +10:44.400 --> 10:47.880 + So I think there's kind of two problems + +10:47.880 --> 10:49.960 + entailed in that question. + +10:49.960 --> 10:52.360 + The first is more of the question of, + +10:52.360 --> 10:54.600 + how can you even picture what a world + +10:54.600 --> 10:56.600 + with a new technology will be like? + +10:56.600 --> 10:57.880 + Now imagine we're in 1950 + +10:57.880 --> 11:01.040 + and I'm trying to describe Uber to someone. + +11:01.040 --> 11:05.360 + Aps and the internet. + +11:05.360 --> 11:08.920 + Yeah, I mean, that's going to be extremely complicated, + +11:08.920 --> 11:10.160 + but it's imaginable. + +11:10.160 --> 11:11.400 + It's imaginable, right? + +11:11.400 --> 11:14.000 + But, and now imagine being in 1950 + +11:14.000 --> 11:15.280 + and predicting Uber, right? + +11:15.280 --> 11:17.680 + And you need to describe the internet, + +11:17.680 --> 11:18.720 + you need to describe GPS, + +11:18.720 --> 11:20.280 + you need to describe the fact + +11:20.280 --> 11:23.920 + that everyone's going to have this phone in their pocket. + +11:23.920 --> 11:26.160 + And so I think that just the first truth + +11:26.160 --> 11:28.040 + is that it is hard to picture + +11:28.040 --> 11:31.160 + how a transformative technology will play out in the world. + +11:31.160 --> 11:32.760 + We've seen that before with technologies + +11:32.760 --> 11:35.560 + that are far less transformative than AGI will be. + +11:35.560 --> 11:37.480 + And so I think that one piece + +11:37.480 --> 11:39.560 + is that it's just even hard to imagine + +11:39.560 --> 11:41.640 + and to really put yourself in a world + +11:41.640 --> 11:44.600 + where you can predict what that positive vision + +11:44.600 --> 11:45.760 + would be like. + +11:46.920 --> 11:49.520 + And I think the second thing is that it is, + +11:49.520 --> 11:53.280 + I think it is always easier to support + +11:53.280 --> 11:55.080 + the negative side than the positive side. + +11:55.080 --> 11:57.120 + It's always easier to destroy than create. + +11:58.200 --> 12:00.800 + And, you know, less in a physical sense + +12:00.800 --> 12:03.080 + and more just in an intellectual sense, right? + +12:03.080 --> 12:05.680 + Because, you know, I think that with creating something, + +12:05.680 --> 12:07.440 + you need to just get a bunch of things right + +12:07.440 --> 12:10.280 + and to destroy, you just need to get one thing wrong. + +12:10.280 --> 12:12.080 + And so I think that what that means + +12:12.080 --> 12:14.240 + is that I think a lot of people's thinking dead ends + +12:14.240 --> 12:16.880 + as soon as they see the negative story. + +12:16.880 --> 12:20.360 + But that being said, I actually have some hope, right? + +12:20.360 --> 12:23.160 + I think that the positive vision + +12:23.160 --> 12:26.000 + is something that I think can be, + +12:26.000 --> 12:27.600 + is something that we can talk about. + +12:27.600 --> 12:30.200 + I think that just simply saying this fact of, + +12:30.200 --> 12:32.000 + yeah, like there's positive, there's negatives, + +12:32.000 --> 12:33.600 + everyone likes to dwell on the negative, + +12:33.600 --> 12:35.360 + people actually respond well to that message and say, + +12:35.360 --> 12:37.040 + huh, you're right, there's a part of this + +12:37.040 --> 12:39.640 + that we're not talking about, not thinking about. + +12:39.640 --> 12:41.240 + And that's actually something that's, + +12:41.240 --> 12:43.800 + I think really been a key part + +12:43.800 --> 12:46.640 + of how we think about AGI at OpenAI, right? + +12:46.640 --> 12:48.160 + You can kind of look at it as like, okay, + +12:48.160 --> 12:51.000 + like OpenAI talks about the fact that there are risks + +12:51.000 --> 12:53.160 + and yet they're trying to build this system. + +12:53.160 --> 12:56.080 + Like how do you square those two facts? + +12:56.080 --> 12:59.120 + So do you share the intuition that some people have, + +12:59.120 --> 13:02.680 + I mean, from Sam Harris to even Elon Musk himself, + +13:02.680 --> 13:06.600 + that it's tricky as you develop AGI + +13:06.600 --> 13:10.400 + to keep it from slipping into the existential threats, + +13:10.400 --> 13:11.760 + into the negative. + +13:11.760 --> 13:13.640 + What's your intuition about, + +13:13.640 --> 13:17.720 + how hard is it to keep AI development + +13:17.720 --> 13:19.640 + on the positive track? + +13:19.640 --> 13:20.680 + What's your intuition there? + +13:20.680 --> 13:21.560 + To answer that question, + +13:21.560 --> 13:23.960 + you can really look at how we structure OpenAI. + +13:23.960 --> 13:25.840 + So we really have three main arms. + +13:25.840 --> 13:26.960 + So we have capabilities, + +13:26.960 --> 13:29.040 + which is actually doing the technical work + +13:29.040 --> 13:31.160 + and pushing forward what these systems can do. + +13:31.160 --> 13:35.120 + There's safety, which is working on technical mechanisms + +13:35.120 --> 13:36.920 + to ensure that the systems we build + +13:36.920 --> 13:38.480 + are aligned with human values. + +13:38.480 --> 13:39.640 + And then there's policy, + +13:39.640 --> 13:42.040 + which is making sure that we have governance mechanisms, + +13:42.040 --> 13:45.280 + answering that question of, well, whose values? + +13:45.280 --> 13:47.360 + And so I think that the technical safety one + +13:47.360 --> 13:50.480 + is the one that people kind of talk about the most, right? + +13:50.480 --> 13:52.080 + You talk about, like think about, + +13:52.080 --> 13:54.200 + you know, all of the dystopic AI movies, + +13:54.200 --> 13:55.960 + a lot of that is about not having good + +13:55.960 --> 13:57.520 + technical safety in place. + +13:57.520 --> 13:59.960 + And what we've been finding is that, you know, + +13:59.960 --> 14:01.360 + I think that actually a lot of people + +14:01.360 --> 14:02.680 + look at the technical safety problem + +14:02.680 --> 14:05.400 + and think it's just intractable, right? + +14:05.400 --> 14:07.840 + This question of what do humans want? + +14:07.840 --> 14:09.160 + How am I supposed to write that down? + +14:09.160 --> 14:11.240 + Can I even write down what I want? + +14:11.240 --> 14:12.080 + No way. + +14:13.040 --> 14:14.800 + And then they stop there. + +14:14.800 --> 14:16.880 + But the thing is we've already built systems + +14:16.880 --> 14:20.920 + that are able to learn things that humans can't specify. + +14:20.920 --> 14:22.920 + You know, even the rules for how to recognize + +14:22.920 --> 14:25.000 + if there's a cat or a dog in an image. + +14:25.000 --> 14:26.520 + Turns out it's intractable to write that down + +14:26.520 --> 14:28.400 + and yet we're able to learn it. + +14:28.400 --> 14:31.040 + And that what we're seeing with systems we build at OpenAI + +14:31.040 --> 14:33.800 + and they're still in early proof of concept stage + +14:33.800 --> 14:36.320 + is that you are able to learn human preferences. + +14:36.320 --> 14:38.920 + You're able to learn what humans want from data. + +14:38.920 --> 14:40.400 + And so that's kind of the core focus + +14:40.400 --> 14:41.760 + for our technical safety team. + +14:41.760 --> 14:43.800 + And I think that they're actually, + +14:43.800 --> 14:45.640 + we've had some pretty encouraging updates + +14:45.640 --> 14:48.040 + in terms of what we've been able to make work. + +14:48.040 --> 14:51.680 + So you have an intuition and a hope that from data, + +14:51.680 --> 14:53.640 + you know, looking at the value alignment problem, + +14:53.640 --> 14:57.040 + from data we can build systems that align + +14:57.040 --> 15:00.600 + with the collective better angels of our nature. + +15:00.600 --> 15:04.600 + So align with the ethics and the morals of human beings. + +15:04.600 --> 15:05.880 + To even say this in a different way, + +15:05.880 --> 15:08.560 + I mean, think about how do we align humans, right? + +15:08.560 --> 15:10.400 + Think about like a human baby can grow up + +15:10.400 --> 15:12.880 + to be an evil person or a great person. + +15:12.880 --> 15:15.200 + And a lot of that is from learning from data, right? + +15:15.200 --> 15:17.720 + That you have some feedback as a child is growing up. + +15:17.720 --> 15:19.160 + They get to see positive examples. + +15:19.160 --> 15:23.120 + And so I think that just like the only example + +15:23.120 --> 15:25.400 + we have of a general intelligence + +15:25.400 --> 15:28.040 + that is able to learn from data + +15:28.040 --> 15:31.440 + to align with human values and to learn values, + +15:31.440 --> 15:32.880 + I think we shouldn't be surprised + +15:32.880 --> 15:36.040 + that we can do the same sorts of techniques + +15:36.040 --> 15:37.440 + or whether the same sort of techniques + +15:37.440 --> 15:41.080 + end up being how we solve value alignment for AGI's. + +15:41.080 --> 15:42.680 + So let's go even higher. + +15:42.680 --> 15:44.800 + I don't know if you've read the book, Sapiens. + +15:44.800 --> 15:48.320 + But there's an idea that, you know, + +15:48.320 --> 15:50.000 + that as a collective, as us human beings, + +15:50.000 --> 15:54.720 + we kind of develop together ideas that we hold. + +15:54.720 --> 15:57.920 + There's no, in that context, objective truth. + +15:57.920 --> 16:00.000 + We just kind of all agree to certain ideas + +16:00.000 --> 16:01.440 + and hold them as a collective. + +16:01.440 --> 16:03.480 + Did you have a sense that there is + +16:03.480 --> 16:05.360 + in the world of good and evil, + +16:05.360 --> 16:07.560 + do you have a sense that to the first approximation, + +16:07.560 --> 16:10.280 + there are some things that are good + +16:10.280 --> 16:14.520 + and that you could teach systems to behave to be good? + +16:14.520 --> 16:18.440 + So I think that this actually blends into our third team, + +16:18.440 --> 16:19.880 + which is the policy team. + +16:19.880 --> 16:22.320 + And this is the one, the aspect that I think people + +16:22.320 --> 16:25.280 + really talk about way less than they should. + +16:25.280 --> 16:27.640 + Because imagine that we build super powerful systems + +16:27.640 --> 16:29.720 + that we've managed to figure out all the mechanisms + +16:29.720 --> 16:32.800 + for these things to do whatever the operator wants. + +16:32.800 --> 16:34.480 + The most important question becomes, + +16:34.480 --> 16:36.720 + who's the operator, what do they want, + +16:36.720 --> 16:39.400 + and how is that going to affect everyone else? + +16:39.400 --> 16:43.080 + And I think that this question of what is good, + +16:43.080 --> 16:44.720 + what are those values, I mean, + +16:44.720 --> 16:45.960 + I think you don't even have to go + +16:45.960 --> 16:48.400 + to those very grand existential places + +16:48.400 --> 16:50.920 + to realize how hard this problem is. + +16:50.920 --> 16:52.880 + You just look at different countries + +16:52.880 --> 16:54.520 + and cultures across the world. + +16:54.520 --> 16:57.120 + And that there's a very different conception + +16:57.120 --> 17:01.920 + of how the world works and what kinds of ways + +17:01.920 --> 17:03.400 + that society wants to operate. + +17:03.400 --> 17:07.000 + And so I think that the really core question + +17:07.000 --> 17:09.560 + is actually very concrete. + +17:09.560 --> 17:10.960 + And I think it's not a question + +17:10.960 --> 17:12.880 + that we have ready answers to, + +17:12.880 --> 17:16.560 + how do you have a world where all the different countries + +17:16.560 --> 17:19.720 + that we have, United States, China, Russia, + +17:19.720 --> 17:22.720 + and the hundreds of other countries out there + +17:22.720 --> 17:26.600 + are able to continue to not just operate + +17:26.600 --> 17:28.440 + in the way that they see fit, + +17:28.440 --> 17:32.520 + but in the world that emerges in these, + +17:32.520 --> 17:34.680 + where you have these very powerful systems, + +17:36.040 --> 17:37.800 + operating alongside humans, + +17:37.800 --> 17:39.800 + ends up being something that empowers humans more, + +17:39.800 --> 17:44.120 + that makes human existence be a more meaningful thing + +17:44.120 --> 17:46.400 + and that people are happier and wealthier + +17:46.400 --> 17:48.960 + and able to live more fulfilling lives. + +17:48.960 --> 17:51.560 + It's not an obvious thing for how to design that world + +17:51.560 --> 17:53.600 + once you have that very powerful system. + +17:53.600 --> 17:55.800 + So if we take a little step back, + +17:55.800 --> 17:58.200 + and we're having like a fascinating conversation + +17:58.200 --> 18:01.880 + and open as in many ways a tech leader in the world, + +18:01.880 --> 18:05.440 + and yet we're thinking about these big existential questions + +18:05.440 --> 18:07.000 + which is fascinating, really important. + +18:07.000 --> 18:09.160 + I think you're a leader in that space + +18:09.160 --> 18:10.840 + and that's a really important space + +18:10.840 --> 18:13.080 + of just thinking how AI affects society + +18:13.080 --> 18:14.360 + in a big picture view. + +18:14.360 --> 18:17.320 + So Oscar Wilde said, we're all in the gutter, + +18:17.320 --> 18:19.000 + but some of us are looking at the stars + +18:19.000 --> 18:22.320 + and I think OpenAI has a charter + +18:22.320 --> 18:24.600 + that looks to the stars, I would say, + +18:24.600 --> 18:26.880 + to create intelligence, to create general intelligence, + +18:26.880 --> 18:29.440 + make it beneficial, safe, and collaborative. + +18:29.440 --> 18:33.680 + So can you tell me how that came about? + +18:33.680 --> 18:36.320 + How a mission like that and the path + +18:36.320 --> 18:39.120 + to creating a mission like that at OpenAI was founded? + +18:39.120 --> 18:41.640 + Yeah, so I think that in some ways + +18:41.640 --> 18:45.040 + it really boils down to taking a look at the landscape. + +18:45.040 --> 18:47.040 + So if you think about the history of AI + +18:47.040 --> 18:49.920 + that basically for the past 60 or 70 years, + +18:49.920 --> 18:51.640 + people have thought about this goal + +18:51.640 --> 18:53.960 + of what could happen if you could automate + +18:53.960 --> 18:55.640 + human intellectual labor. + +18:56.680 --> 18:58.280 + Imagine you can build a computer system + +18:58.280 --> 19:00.560 + that could do that, what becomes possible? + +19:00.560 --> 19:02.400 + We have a lot of sci fi that tells stories + +19:02.400 --> 19:04.920 + of various dystopias and increasingly you have movies + +19:04.920 --> 19:06.480 + like Her that tell you a little bit about + +19:06.480 --> 19:09.440 + maybe more of a little bit utopic vision. + +19:09.440 --> 19:12.560 + You think about the impacts that we've seen + +19:12.560 --> 19:16.280 + from being able to have bicycles for our minds + +19:16.280 --> 19:20.360 + and computers and that I think that the impact + +19:20.360 --> 19:23.480 + of computers and the internet has just far outstripped + +19:23.480 --> 19:26.200 + what anyone really could have predicted. + +19:26.200 --> 19:27.400 + And so I think that it's very clear + +19:27.400 --> 19:29.360 + that if you can build an AGI, + +19:29.360 --> 19:31.600 + it will be the most transformative technology + +19:31.600 --> 19:33.040 + that humans will ever create. + +19:33.040 --> 19:36.840 + And so what it boils down to then is a question of, + +19:36.840 --> 19:38.680 + well, is there a path? + +19:38.680 --> 19:39.520 + Is there hope? + +19:39.520 --> 19:41.680 + Is there a way to build such a system? + +19:41.680 --> 19:43.640 + And I think that for 60 or 70 years + +19:43.640 --> 19:48.040 + that people got excited and that ended up not being able + +19:48.040 --> 19:51.480 + to deliver on the hopes that people had pinned on them. + +19:51.480 --> 19:54.880 + And I think that then, that after two winters + +19:54.880 --> 19:57.600 + of AI development, that people, + +19:57.600 --> 20:00.560 + I think kind of almost stopped daring to dream, right? + +20:00.560 --> 20:03.280 + That really talking about AGI or thinking about AGI + +20:03.280 --> 20:05.640 + became almost this taboo in the community. + +20:06.640 --> 20:08.720 + But I actually think that people took the wrong lesson + +20:08.720 --> 20:10.080 + from AI history. + +20:10.080 --> 20:12.400 + And if you look back, starting in 1959 + +20:12.400 --> 20:14.240 + is when the Perceptron was released. + +20:14.240 --> 20:17.720 + And this is basically one of the earliest neural networks. + +20:17.720 --> 20:19.280 + It was released to what was perceived + +20:19.280 --> 20:20.840 + as this massive overhype. + +20:20.840 --> 20:22.360 + So in the New York Times in 1959, + +20:22.360 --> 20:26.400 + you have this article saying that the Perceptron + +20:26.400 --> 20:29.160 + will one day recognize people, call out their names, + +20:29.160 --> 20:31.480 + instantly translate speech between languages. + +20:31.480 --> 20:33.800 + And people at the time looked at this and said, + +20:33.800 --> 20:36.120 + this is, your system can't do any of that. + +20:36.120 --> 20:38.080 + And basically spent 10 years trying to discredit + +20:38.080 --> 20:40.640 + the whole Perceptron direction and succeeded. + +20:40.640 --> 20:41.840 + And all the funding dried up. + +20:41.840 --> 20:44.960 + And people kind of went in other directions. + +20:44.960 --> 20:46.920 + And in the 80s, there was this resurgence. + +20:46.920 --> 20:49.320 + And I'd always heard that the resurgence in the 80s + +20:49.320 --> 20:51.520 + was due to the invention of back propagation + +20:51.520 --> 20:53.720 + and these algorithms that got people excited. + +20:53.720 --> 20:55.760 + But actually the causality was due to people + +20:55.760 --> 20:57.200 + building larger computers. + +20:57.200 --> 20:59.280 + That you can find these articles from the 80s saying + +20:59.280 --> 21:01.760 + that the democratization of computing power + +21:01.760 --> 21:04.040 + suddenly meant that you could run these larger neural networks. + +21:04.040 --> 21:06.280 + And then people started to do all these amazing things, + +21:06.280 --> 21:08.000 + back propagation algorithm was invented. + +21:08.000 --> 21:10.120 + And the neural nets people were running + +21:10.120 --> 21:13.000 + were these tiny little like 20 neuron neural nets. + +21:13.000 --> 21:15.160 + What are you supposed to learn with 20 neurons? + +21:15.160 --> 21:18.640 + And so of course they weren't able to get great results. + +21:18.640 --> 21:21.960 + And it really wasn't until 2012 that this approach, + +21:21.960 --> 21:24.680 + that's almost the most simple, natural approach + +21:24.680 --> 21:27.720 + that people had come up with in the 50s, right? + +21:27.720 --> 21:30.360 + In some ways, even in the 40s before there were computers + +21:30.360 --> 21:32.000 + with the Pits McCullin neuron, + +21:33.040 --> 21:37.480 + suddenly this became the best way of solving problems, right? + +21:37.480 --> 21:39.280 + And I think there are three core properties + +21:39.280 --> 21:42.120 + that deep learning has that I think + +21:42.120 --> 21:44.120 + are very worth paying attention to. + +21:44.120 --> 21:45.920 + The first is generality. + +21:45.920 --> 21:48.760 + We have a very small number of deep learning tools, + +21:48.760 --> 21:52.360 + SGD, deep neural net, maybe some, you know, RL. + +21:52.360 --> 21:55.600 + And it solves this huge variety of problems, + +21:55.600 --> 21:57.240 + speech recognition, machine translation, + +21:57.240 --> 22:00.200 + game playing, all of these problems, + +22:00.200 --> 22:01.040 + small set of tools. + +22:01.040 --> 22:02.760 + So there's the generality. + +22:02.760 --> 22:05.000 + There's a second piece, which is the competence. + +22:05.000 --> 22:07.040 + You wanna solve any of those problems? + +22:07.040 --> 22:10.640 + Throughout 40 years worth of normal computer vision research + +22:10.640 --> 22:13.640 + replaced with a deep neural net, it's gonna work better. + +22:13.640 --> 22:16.320 + And there's a third piece, which is the scalability, right? + +22:16.320 --> 22:18.720 + That one thing that has been shown time and time again + +22:18.720 --> 22:21.760 + is that you, if you have a larger neural network, + +22:21.760 --> 22:25.120 + throw more compute, more data at it, it will work better. + +22:25.120 --> 22:28.880 + Those three properties together feel like essential parts + +22:28.880 --> 22:30.800 + of building a general intelligence. + +22:30.800 --> 22:33.000 + Now, it doesn't just mean that if we scale up + +22:33.000 --> 22:35.200 + what we have, that we will have an AGI, right? + +22:35.200 --> 22:36.800 + There are clearly missing pieces. + +22:36.800 --> 22:38.000 + There are missing ideas. + +22:38.000 --> 22:40.000 + We need to have answers for reasoning. + +22:40.000 --> 22:44.800 + But I think that the core here is that for the first time, + +22:44.800 --> 22:46.880 + it feels that we have a paradigm + +22:46.880 --> 22:48.960 + that gives us hope that general intelligence + +22:48.960 --> 22:50.560 + can be achievable. + +22:50.560 --> 22:52.160 + And so as soon as you believe that, + +22:52.160 --> 22:54.480 + everything else becomes into focus, right? + +22:54.480 --> 22:56.560 + If you imagine that you may be able to, + +22:56.560 --> 22:59.920 + and that the timeline I think remains uncertain, + +22:59.920 --> 23:02.200 + but I think that certainly within our lifetimes + +23:02.200 --> 23:04.640 + and possibly within a much shorter period of time + +23:04.640 --> 23:06.560 + than people would expect, + +23:06.560 --> 23:09.360 + if you can really build the most transformative technology + +23:09.360 --> 23:11.720 + that will ever exist, you stop thinking about yourself + +23:11.720 --> 23:12.560 + so much, right? + +23:12.560 --> 23:14.240 + And you start thinking about just like, + +23:14.240 --> 23:16.440 + how do you have a world where this goes well? + +23:16.440 --> 23:18.160 + And that you need to think about the practicalities + +23:18.160 --> 23:19.560 + of how do you build an organization + +23:19.560 --> 23:22.000 + and get together a bunch of people and resources + +23:22.000 --> 23:25.160 + and to make sure that people feel motivated + +23:25.160 --> 23:26.800 + and ready to do it. + +23:28.080 --> 23:30.720 + But I think that then you start thinking about, + +23:30.720 --> 23:32.080 + well, what if we succeed? + +23:32.080 --> 23:34.280 + And how do we make sure that when we succeed, + +23:34.280 --> 23:35.600 + that the world is actually the place + +23:35.600 --> 23:38.200 + that we want ourselves to exist in? + +23:38.200 --> 23:41.080 + And almost in the Rawlsian Vale sense of the word. + +23:41.080 --> 23:43.880 + And so that's kind of the broader landscape. + +23:43.880 --> 23:46.680 + And Open AI was really formed in 2015 + +23:46.680 --> 23:51.480 + with that high level picture of AGI might be possible + +23:51.480 --> 23:52.880 + sooner than people think + +23:52.880 --> 23:55.840 + and that we need to try to do our best + +23:55.840 --> 23:57.480 + to make sure it's going to go well. + +23:57.480 --> 23:59.360 + And then we spent the next couple of years + +23:59.360 --> 24:00.840 + really trying to figure out what does that mean? + +24:00.840 --> 24:01.960 + How do we do it? + +24:01.960 --> 24:04.800 + And I think that typically with a company, + +24:04.800 --> 24:07.320 + you start out very small. + +24:07.320 --> 24:09.000 + So you want a cofounder and you build a product, + +24:09.000 --> 24:11.360 + you get some users, you get a product market fit, + +24:11.360 --> 24:13.320 + then at some point you raise some money, + +24:13.320 --> 24:14.840 + you hire people, you scale, + +24:14.840 --> 24:17.440 + and then down the road, then the big companies + +24:17.440 --> 24:19.080 + realize you exist and try to kill you. + +24:19.080 --> 24:21.520 + And for Open AI, it was basically everything + +24:21.520 --> 24:22.960 + in exactly the opposite order. + +24:25.480 --> 24:26.760 + Let me just pause for a second. + +24:26.760 --> 24:27.520 + He said a lot of things. + +24:27.520 --> 24:31.240 + And let me just admire the jarring aspect + +24:31.240 --> 24:35.160 + of what Open AI stands for, which is daring to dream. + +24:35.160 --> 24:37.120 + I mean, you said it's pretty powerful. + +24:37.120 --> 24:40.080 + You caught me off guard because I think that's very true. + +24:40.080 --> 24:44.040 + The step of just daring to dream + +24:44.040 --> 24:46.720 + about the possibilities of creating intelligence + +24:46.720 --> 24:48.760 + in a positive and a safe way, + +24:48.760 --> 24:50.640 + but just even creating intelligence + +24:50.640 --> 24:55.640 + is a much needed, refreshing catalyst + +24:56.280 --> 24:57.360 + for the AI community. + +24:57.360 --> 24:58.800 + So that's the starting point. + +24:58.800 --> 25:02.840 + Okay, so then formation of Open AI, what's your point? + +25:02.840 --> 25:05.640 + I would just say that when we were starting Open AI, + +25:05.640 --> 25:07.760 + that kind of the first question that we had is, + +25:07.760 --> 25:12.000 + is it too late to start a lab with a bunch of the best people? + +25:12.000 --> 25:13.160 + Right, is that even possible? + +25:13.160 --> 25:14.320 + That was an actual question. + +25:14.320 --> 25:17.280 + That was the core question of, + +25:17.280 --> 25:19.320 + we had this dinner in July of 2015, + +25:19.320 --> 25:21.240 + and that was really what we spent the whole time + +25:21.240 --> 25:22.320 + talking about. + +25:22.320 --> 25:26.800 + And because you think about kind of where AI was, + +25:26.800 --> 25:30.200 + is that it transitioned from being an academic pursuit + +25:30.200 --> 25:32.240 + to an industrial pursuit. + +25:32.240 --> 25:34.240 + And so a lot of the best people were in these big + +25:34.240 --> 25:37.000 + research labs and that we wanted to start our own one + +25:37.000 --> 25:40.560 + that no matter how much resources we could accumulate + +25:40.560 --> 25:43.520 + would be pale in comparison to the big tech companies. + +25:43.520 --> 25:44.720 + And we knew that. + +25:44.720 --> 25:45.800 + And there's a question of, + +25:45.800 --> 25:47.720 + are we going to be actually able to get this thing + +25:47.720 --> 25:48.720 + off the ground? + +25:48.720 --> 25:49.760 + You need critical mass. + +25:49.760 --> 25:52.120 + You can't just do you and a cofounder build a product, right? + +25:52.120 --> 25:55.600 + You really need to have a group of five to 10 people. + +25:55.600 --> 25:59.480 + And we kind of concluded it wasn't obviously impossible. + +25:59.480 --> 26:00.840 + So it seemed worth trying. + +26:02.240 --> 26:04.800 + Well, you're also a dreamer, so who knows, right? + +26:04.800 --> 26:05.640 + That's right. + +26:05.640 --> 26:07.720 + Okay, so speaking of that, + +26:07.720 --> 26:10.520 + competing with the big players, + +26:11.520 --> 26:14.080 + let's talk about some of the tricky things + +26:14.080 --> 26:17.480 + as you think through this process of growing, + +26:17.480 --> 26:20.080 + of seeing how you can develop these systems + +26:20.080 --> 26:22.640 + at a scale that competes. + +26:22.640 --> 26:25.720 + So you recently formed OpenAI LP, + +26:26.560 --> 26:30.800 + a new cap profit company that now carries the name OpenAI. + +26:30.800 --> 26:33.280 + So OpenAI is now this official company. + +26:33.280 --> 26:36.520 + The original nonprofit company still exists + +26:36.520 --> 26:39.800 + and carries the OpenAI nonprofit name. + +26:39.800 --> 26:42.000 + So can you explain what this company is, + +26:42.000 --> 26:44.280 + what the purpose of its creation is, + +26:44.280 --> 26:48.800 + and how did you arrive at the decision to create it? + +26:48.800 --> 26:53.280 + OpenAI, the whole entity and OpenAI LP as a vehicle + +26:53.280 --> 26:55.560 + is trying to accomplish the mission + +26:55.560 --> 26:57.520 + of ensuring that artificial general intelligence + +26:57.520 --> 26:58.800 + benefits everyone. + +26:58.800 --> 27:00.240 + And the main way that we're trying to do that + +27:00.240 --> 27:01.840 + is by actually trying to build + +27:01.840 --> 27:03.240 + general intelligence to ourselves + +27:03.240 --> 27:05.920 + and make sure the benefits are distributed to the world. + +27:05.920 --> 27:07.200 + That's the primary way. + +27:07.200 --> 27:09.600 + We're also fine if someone else does this, right? + +27:09.600 --> 27:10.640 + It doesn't have to be us. + +27:10.640 --> 27:12.640 + If someone else is going to build an AGI + +27:12.640 --> 27:14.840 + and make sure that the benefits don't get locked up + +27:14.840 --> 27:18.160 + in one company or with one set of people, + +27:19.280 --> 27:21.160 + like we're actually fine with that. + +27:21.160 --> 27:25.400 + And so those ideas are baked into our charter, + +27:25.400 --> 27:28.400 + which is kind of the foundational document + +27:28.400 --> 27:31.920 + that describes kind of our values and how we operate. + +27:31.920 --> 27:36.360 + And it's also really baked into the structure of OpenAI LP. + +27:36.360 --> 27:37.960 + And so the way that we've set up OpenAI LP + +27:37.960 --> 27:42.160 + is that in the case where we succeed, right? + +27:42.160 --> 27:45.320 + If we actually build what we're trying to build, + +27:45.320 --> 27:47.800 + then investors are able to get a return, + +27:47.800 --> 27:50.400 + and but that return is something that is capped. + +27:50.400 --> 27:53.000 + And so if you think of AGI in terms of the value + +27:53.000 --> 27:54.160 + that you could really create, + +27:54.160 --> 27:56.320 + you're talking about the most transformative technology + +27:56.320 --> 27:58.000 + ever created, it's gonna create, + +27:58.000 --> 28:01.880 + or does the magnitude more value than any existing company? + +28:01.880 --> 28:05.960 + And that all of that value will be owned by the world, + +28:05.960 --> 28:07.880 + like legally titled to the nonprofit + +28:07.880 --> 28:09.560 + to fulfill that mission. + +28:09.560 --> 28:12.800 + And so that's the structure. + +28:12.800 --> 28:15.200 + So the mission is a powerful one, + +28:15.200 --> 28:18.920 + and it's one that I think most people would agree with. + +28:18.920 --> 28:22.960 + It's how we would hope AI progresses. + +28:22.960 --> 28:25.440 + And so how do you tie yourself to that mission? + +28:25.440 --> 28:29.240 + How do you make sure you do not deviate from that mission + +28:29.240 --> 28:34.240 + that other incentives that are profit driven + +28:34.560 --> 28:36.800 + wouldn't don't interfere with the mission? + +28:36.800 --> 28:39.560 + So this was actually a really core question for us + +28:39.560 --> 28:40.920 + for the past couple of years, + +28:40.920 --> 28:43.560 + because I'd say that the way that our history went + +28:43.560 --> 28:44.960 + was that for the first year, + +28:44.960 --> 28:46.240 + we were getting off the ground, right? + +28:46.240 --> 28:47.960 + We had this high level picture, + +28:47.960 --> 28:51.880 + but we didn't know exactly how we wanted to accomplish it. + +28:51.880 --> 28:53.440 + And really two years ago, + +28:53.440 --> 28:55.040 + it's when we first started realizing + +28:55.040 --> 28:56.160 + in order to build AGI, + +28:56.160 --> 28:58.720 + we're just gonna need to raise way more money + +28:58.720 --> 29:00.680 + than we can as a nonprofit. + +29:00.680 --> 29:02.800 + We're talking many billions of dollars. + +29:02.800 --> 29:05.440 + And so the first question is, + +29:05.440 --> 29:06.840 + how are you supposed to do that + +29:06.840 --> 29:08.680 + and stay true to this mission? + +29:08.680 --> 29:10.560 + And we looked at every legal structure out there + +29:10.560 --> 29:11.960 + and included none of them were quite right + +29:11.960 --> 29:13.400 + for what we wanted to do. + +29:13.400 --> 29:14.600 + And I guess it shouldn't be too surprising + +29:14.600 --> 29:16.920 + if you're gonna do some crazy unprecedented technology + +29:16.920 --> 29:17.920 + that you're gonna have to come + +29:17.920 --> 29:20.320 + with some crazy unprecedented structure to do it in. + +29:20.320 --> 29:25.320 + And a lot of our conversation was with people at OpenAI, + +29:26.080 --> 29:27.240 + the people who really joined + +29:27.240 --> 29:29.160 + because they believe so much in this mission + +29:29.160 --> 29:32.120 + and thinking about how do we actually raise the resources + +29:32.120 --> 29:35.920 + to do it and also stay true to what we stand for. + +29:35.920 --> 29:38.000 + And the place you gotta start is to really align + +29:38.000 --> 29:39.560 + on what is it that we stand for, right? + +29:39.560 --> 29:40.560 + What are those values? + +29:40.560 --> 29:41.840 + What's really important to us? + +29:41.840 --> 29:43.760 + And so I'd say that we spent about a year + +29:43.760 --> 29:46.240 + really compiling the OpenAI charter. + +29:46.240 --> 29:47.560 + And that determines, + +29:47.560 --> 29:50.240 + and if you even look at the first line item in there, + +29:50.240 --> 29:52.360 + it says that, look, we expect we're gonna have to marshal + +29:52.360 --> 29:53.760 + huge amounts of resources, + +29:53.760 --> 29:55.160 + but we're going to make sure + +29:55.160 --> 29:57.920 + that we minimize conflict of interest with the mission. + +29:57.920 --> 30:00.720 + And that kind of aligning on all of those pieces + +30:00.720 --> 30:04.240 + was the most important step towards figuring out + +30:04.240 --> 30:06.040 + how do we structure a company + +30:06.040 --> 30:08.240 + that can actually raise the resources + +30:08.240 --> 30:10.360 + to do what we need to do. + +30:10.360 --> 30:14.760 + I imagine OpenAI, the decision to create OpenAI LP + +30:14.760 --> 30:16.360 + was a really difficult one. + +30:16.360 --> 30:17.920 + And there was a lot of discussions + +30:17.920 --> 30:19.640 + as you mentioned for a year. + +30:19.640 --> 30:22.760 + And there was different ideas, + +30:22.760 --> 30:25.120 + perhaps detractors within OpenAI, + +30:26.120 --> 30:28.920 + sort of different paths that you could have taken. + +30:28.920 --> 30:30.240 + What were those concerns? + +30:30.240 --> 30:32.040 + What were the different paths considered? + +30:32.040 --> 30:34.080 + What was that process of making that decision like? + +30:34.080 --> 30:35.000 + Yep. + +30:35.000 --> 30:37.200 + But so if you look actually at the OpenAI charter, + +30:37.200 --> 30:40.880 + that there's almost two paths embedded within it. + +30:40.880 --> 30:44.880 + There is, we are primarily trying to build AGI ourselves, + +30:44.880 --> 30:47.360 + but we're also okay if someone else does it. + +30:47.360 --> 30:49.040 + And this is a weird thing for a company. + +30:49.040 --> 30:50.480 + It's really interesting, actually. + +30:50.480 --> 30:51.320 + Yeah. + +30:51.320 --> 30:53.280 + But there is an element of competition + +30:53.280 --> 30:56.680 + that you do want to be the one that does it, + +30:56.680 --> 30:59.040 + but at the same time, you're okay if somebody else doesn't. + +30:59.040 --> 31:01.000 + We'll talk about that a little bit, that trade off, + +31:01.000 --> 31:02.960 + that dance that's really interesting. + +31:02.960 --> 31:04.600 + And I think this was the core tension + +31:04.600 --> 31:06.360 + as we were designing OpenAI LP + +31:06.360 --> 31:08.240 + and really the OpenAI strategy, + +31:08.240 --> 31:11.080 + is how do you make sure that both you have a shot + +31:11.080 --> 31:12.640 + at being a primary actor, + +31:12.640 --> 31:15.840 + which really requires building an organization, + +31:15.840 --> 31:17.720 + raising massive resources, + +31:17.720 --> 31:19.440 + and really having the will to go + +31:19.440 --> 31:22.000 + and execute on some really, really hard vision, right? + +31:22.000 --> 31:23.760 + You need to really sign up for a long period + +31:23.760 --> 31:27.120 + to go and take on a lot of pain and a lot of risk. + +31:27.120 --> 31:29.000 + And to do that, + +31:29.000 --> 31:31.720 + normally you just import the startup mindset, right? + +31:31.720 --> 31:32.760 + And that you think about, okay, + +31:32.760 --> 31:34.240 + like how do we out execute everyone? + +31:34.240 --> 31:36.160 + You have this very competitive angle. + +31:36.160 --> 31:38.120 + But you also have the second angle of saying that, + +31:38.120 --> 31:41.600 + well, the true mission isn't for OpenAI to build AGI. + +31:41.600 --> 31:45.080 + The true mission is for AGI to go well for humanity. + +31:45.080 --> 31:48.080 + And so how do you take all of those first actions + +31:48.080 --> 31:51.320 + and make sure you don't close the door on outcomes + +31:51.320 --> 31:54.480 + that would actually be positive and fulfill the mission? + +31:54.480 --> 31:56.680 + And so I think it's a very delicate balance, right? + +31:56.680 --> 31:59.560 + And I think that going 100% one direction or the other + +31:59.560 --> 32:01.320 + is clearly not the correct answer. + +32:01.320 --> 32:03.920 + And so I think that even in terms of just how we talk about + +32:03.920 --> 32:05.400 + OpenAI and think about it, + +32:05.400 --> 32:07.600 + there's just like one thing that's always + +32:07.600 --> 32:09.680 + in the back of my mind is to make sure + +32:09.680 --> 32:12.120 + that we're not just saying OpenAI's goal + +32:12.120 --> 32:14.000 + is to build AGI, right? + +32:14.000 --> 32:15.560 + That it's actually much broader than that, right? + +32:15.560 --> 32:19.360 + That first of all, it's not just AGI, it's safe AGI + +32:19.360 --> 32:20.320 + that's very important. + +32:20.320 --> 32:23.120 + But secondly, our goal isn't to be the ones to build it, + +32:23.120 --> 32:24.720 + our goal is to make sure it goes well for the world. + +32:24.720 --> 32:26.120 + And so I think that figuring out, + +32:26.120 --> 32:27.960 + how do you balance all of those + +32:27.960 --> 32:30.280 + and to get people to really come to the table + +32:30.280 --> 32:35.280 + and compile a single document that encompasses all of that + +32:36.360 --> 32:37.560 + wasn't trivial. + +32:37.560 --> 32:41.680 + So part of the challenge here is your mission is, + +32:41.680 --> 32:44.240 + I would say, beautiful, empowering, + +32:44.240 --> 32:47.520 + and a beacon of hope for people in the research community + +32:47.520 --> 32:49.200 + and just people thinking about AI. + +32:49.200 --> 32:51.880 + So your decisions are scrutinized + +32:51.880 --> 32:55.920 + more than, I think, a regular profit driven company. + +32:55.920 --> 32:57.400 + Do you feel the burden of this + +32:57.400 --> 32:58.560 + in the creation of the charter + +32:58.560 --> 33:00.200 + and just in the way you operate? + +33:00.200 --> 33:01.040 + Yes. + +33:03.040 --> 33:05.920 + So why do you lean into the burden + +33:07.040 --> 33:08.640 + by creating such a charter? + +33:08.640 --> 33:10.440 + Why not keep it quiet? + +33:10.440 --> 33:12.920 + I mean, it just boils down to the mission, right? + +33:12.920 --> 33:15.200 + Like, I'm here and everyone else is here + +33:15.200 --> 33:17.880 + because we think this is the most important mission, right? + +33:17.880 --> 33:19.000 + Dare to dream. + +33:19.000 --> 33:23.360 + All right, so do you think you can be good for the world + +33:23.360 --> 33:26.000 + or create an AGI system that's good + +33:26.000 --> 33:28.320 + when you're a for profit company? + +33:28.320 --> 33:32.920 + From my perspective, I don't understand why profit + +33:32.920 --> 33:37.640 + interferes with positive impact on society. + +33:37.640 --> 33:40.760 + I don't understand why Google + +33:40.760 --> 33:42.920 + that makes most of its money from ads + +33:42.920 --> 33:45.040 + can't also do good for the world + +33:45.040 --> 33:47.520 + or other companies, Facebook, anything. + +33:47.520 --> 33:50.240 + I don't understand why those have to interfere. + +33:50.240 --> 33:55.120 + You know, you can, profit isn't the thing in my view + +33:55.120 --> 33:57.240 + that affects the impact of a company. + +33:57.240 --> 34:00.360 + What affects the impact of the company is the charter, + +34:00.360 --> 34:04.160 + is the culture, is the people inside + +34:04.160 --> 34:07.360 + and profit is the thing that just fuels those people. + +34:07.360 --> 34:08.760 + What are your views there? + +34:08.760 --> 34:10.920 + Yeah, so I think that's a really good question + +34:10.920 --> 34:14.200 + and there's some real like longstanding debates + +34:14.200 --> 34:16.520 + in human society that are wrapped up in it. + +34:16.520 --> 34:18.680 + The way that I think about it is just think about + +34:18.680 --> 34:21.520 + what are the most impactful nonprofits in the world? + +34:24.000 --> 34:26.760 + What are the most impactful for profits in the world? + +34:26.760 --> 34:29.280 + Right, it's much easier to list the for profits. + +34:29.280 --> 34:30.120 + That's right. + +34:30.120 --> 34:32.400 + And I think that there's some real truth here + +34:32.400 --> 34:34.600 + that the system that we set up, + +34:34.600 --> 34:38.320 + the system for kind of how today's world is organized + +34:38.320 --> 34:41.760 + is one that really allows for huge impact + +34:41.760 --> 34:45.400 + and that kind of part of that is that you need to be, + +34:45.400 --> 34:48.080 + that for profits are self sustaining + +34:48.080 --> 34:51.200 + and able to kind of build on their own momentum. + +34:51.200 --> 34:53.080 + And I think that's a really powerful thing. + +34:53.080 --> 34:55.880 + It's something that when it turns out + +34:55.880 --> 34:57.920 + that we haven't set the guardrails correctly, + +34:57.920 --> 34:58.840 + causes problems, right? + +34:58.840 --> 35:02.720 + Think about logging companies that go into the rainforest, + +35:02.720 --> 35:04.680 + that's really bad, we don't want that. + +35:04.680 --> 35:06.520 + And it's actually really interesting to me + +35:06.520 --> 35:08.480 + that kind of this question of + +35:08.480 --> 35:11.400 + how do you get positive benefits out of a for profit company? + +35:11.400 --> 35:12.600 + It's actually very similar to + +35:12.600 --> 35:15.800 + how do you get positive benefits out of an AGI, right? + +35:15.800 --> 35:18.000 + That you have this like very powerful system, + +35:18.000 --> 35:19.680 + it's more powerful than any human + +35:19.680 --> 35:21.760 + and it's kind of autonomous in some ways. + +35:21.760 --> 35:23.800 + You know, it's super human in a lot of axes + +35:23.800 --> 35:25.400 + and somehow you have to set the guardrails + +35:25.400 --> 35:26.800 + to get good things to happen. + +35:26.800 --> 35:29.360 + But when you do, the benefits are massive. + +35:29.360 --> 35:32.920 + And so I think that when I think about nonprofit + +35:32.920 --> 35:36.120 + versus for profit, I think just not enough happens + +35:36.120 --> 35:37.800 + in nonprofits, they're very pure, + +35:37.800 --> 35:39.200 + but it's just kind of, you know, + +35:39.200 --> 35:40.840 + it's just hard to do things there. + +35:40.840 --> 35:44.000 + And for profits in some ways, like too much happens, + +35:44.000 --> 35:46.440 + but if kind of shaped in the right way, + +35:46.440 --> 35:47.840 + it can actually be very positive. + +35:47.840 --> 35:52.160 + And so with OpenILP, we're picking a road in between. + +35:52.160 --> 35:54.880 + Now, the thing that I think is really important to recognize + +35:54.880 --> 35:57.160 + is that the way that we think about OpenILP + +35:57.160 --> 36:00.440 + is that in the world where AGI actually happens, right? + +36:00.440 --> 36:01.720 + In a world where we are successful, + +36:01.720 --> 36:03.800 + we build the most transformative technology ever, + +36:03.800 --> 36:06.600 + the amount of value we're going to create will be astronomical. + +36:07.600 --> 36:12.600 + And so then in that case, that the cap that we have + +36:12.760 --> 36:15.520 + will be a small fraction of the value we create. + +36:15.520 --> 36:17.800 + And the amount of value that goes back to investors + +36:17.800 --> 36:20.000 + and employees looks pretty similar to what would happen + +36:20.000 --> 36:21.680 + in a pretty successful startup. + +36:23.760 --> 36:26.520 + And that's really the case that we're optimizing for, right? + +36:26.520 --> 36:28.560 + That we're thinking about in the success case, + +36:28.560 --> 36:32.120 + making sure that the value we create doesn't get locked up. + +36:32.120 --> 36:34.920 + And I expect that in other for profit companies + +36:34.920 --> 36:37.800 + that it's possible to do something like that. + +36:37.800 --> 36:39.720 + I think it's not obvious how to do it, right? + +36:39.720 --> 36:41.440 + And I think that as a for profit company, + +36:41.440 --> 36:44.240 + you have a lot of fiduciary duty to your shareholders + +36:44.240 --> 36:45.640 + and that there are certain decisions + +36:45.640 --> 36:47.520 + that you just cannot make. + +36:47.520 --> 36:49.080 + In our structure, we've set it up + +36:49.080 --> 36:52.440 + so that we have a fiduciary duty to the charter, + +36:52.440 --> 36:54.400 + that we always get to make the decision + +36:54.400 --> 36:56.720 + that is right for the charter, + +36:56.720 --> 36:58.800 + rather than even if it comes at the expense + +36:58.800 --> 37:00.680 + of our own stakeholders. + +37:00.680 --> 37:03.400 + And so I think that when I think about + +37:03.400 --> 37:04.360 + what's really important, + +37:04.360 --> 37:06.280 + it's not really about nonprofit versus for profit. + +37:06.280 --> 37:09.600 + It's really a question of if you build a GI + +37:09.600 --> 37:10.600 + and you kind of, you know, + +37:10.600 --> 37:13.080 + humanity is now at this new age, + +37:13.080 --> 37:15.760 + who benefits, whose lives are better? + +37:15.760 --> 37:17.120 + And I think that what's really important + +37:17.120 --> 37:20.320 + is to have an answer that is everyone. + +37:20.320 --> 37:23.400 + Yeah, which is one of the core aspects of the charter. + +37:23.400 --> 37:26.520 + So one concern people have, not just with OpenAI, + +37:26.520 --> 37:28.400 + but with Google, Facebook, Amazon, + +37:28.400 --> 37:33.400 + anybody really that's creating impact at scale + +37:35.000 --> 37:37.680 + is how do we avoid, as your charter says, + +37:37.680 --> 37:40.080 + avoid enabling the use of AI or AGI + +37:40.080 --> 37:43.640 + to unduly concentrate power? + +37:43.640 --> 37:45.920 + Why would not a company like OpenAI + +37:45.920 --> 37:48.640 + keep all the power of an AGI system to itself? + +37:48.640 --> 37:49.520 + The charter. + +37:49.520 --> 37:50.360 + The charter. + +37:50.360 --> 37:51.960 + So, you know, how does the charter + +37:53.120 --> 37:57.240 + actualize itself in day to day? + +37:57.240 --> 38:00.480 + So I think that first to zoom out, right, + +38:00.480 --> 38:01.880 + that the way that we structure the company + +38:01.880 --> 38:04.560 + is so that the power for sort of, you know, + +38:04.560 --> 38:06.720 + dictating the actions that OpenAI takes + +38:06.720 --> 38:08.600 + ultimately rests with the board, right? + +38:08.600 --> 38:11.720 + The board of the nonprofit and the board is set up + +38:11.720 --> 38:13.480 + in certain ways, with certain restrictions + +38:13.480 --> 38:16.280 + that you can read about in the OpenAI LP blog post. + +38:16.280 --> 38:19.200 + But effectively the board is the governing body + +38:19.200 --> 38:21.200 + for OpenAI LP. + +38:21.200 --> 38:24.400 + And the board has a duty to fulfill the mission + +38:24.400 --> 38:26.360 + of the nonprofit. + +38:26.360 --> 38:28.800 + And so that's kind of how we tie, + +38:28.800 --> 38:30.960 + how we thread all these things together. + +38:30.960 --> 38:32.880 + Now there's a question of so day to day, + +38:32.880 --> 38:34.800 + how do people, the individuals, + +38:34.800 --> 38:36.960 + who in some ways are the most empowered ones, right? + +38:36.960 --> 38:38.800 + You know, the board sort of gets to call the shots + +38:38.800 --> 38:41.920 + at the high level, but the people who are actually executing + +38:41.920 --> 38:43.120 + are the employees, right? + +38:43.120 --> 38:45.480 + The people here on a day to day basis who have the, + +38:45.480 --> 38:47.720 + you know, the keys to the technical kingdom. + +38:48.960 --> 38:51.720 + And there I think that the answer looks a lot like, + +38:51.720 --> 38:55.120 + well, how does any company's values get actualized, right? + +38:55.120 --> 38:56.720 + And I think that a lot of that comes down to + +38:56.720 --> 38:58.160 + that you need people who are here + +38:58.160 --> 39:01.320 + because they really believe in that mission + +39:01.320 --> 39:02.800 + and they believe in the charter + +39:02.800 --> 39:05.440 + and that they are willing to take actions + +39:05.440 --> 39:08.600 + that maybe are worse for them, but are better for the charter. + +39:08.600 --> 39:11.440 + And that's something that's really baked into the culture. + +39:11.440 --> 39:13.200 + And honestly, I think it's, you know, + +39:13.200 --> 39:14.560 + I think that that's one of the things + +39:14.560 --> 39:18.200 + that we really have to work to preserve as time goes on. + +39:18.200 --> 39:20.760 + And that's a really important part of how we think + +39:20.760 --> 39:23.040 + about hiring people and bringing people into OpenAI. + +39:23.040 --> 39:25.320 + So there's people here, there's people here + +39:25.320 --> 39:30.320 + who could speak up and say, like, hold on a second, + +39:30.840 --> 39:34.600 + this is totally against what we stand for, culture wise. + +39:34.600 --> 39:35.440 + Yeah, yeah, for sure. + +39:35.440 --> 39:37.120 + I mean, I think that we actually have, + +39:37.120 --> 39:38.760 + I think that's like a pretty important part + +39:38.760 --> 39:41.920 + of how we operate and how we have, + +39:41.920 --> 39:44.160 + even again with designing the charter + +39:44.160 --> 39:46.680 + and designing OpenAI in the first place, + +39:46.680 --> 39:48.760 + that there has been a lot of conversation + +39:48.760 --> 39:50.480 + with employees here and a lot of times + +39:50.480 --> 39:52.400 + where employees said, wait a second, + +39:52.400 --> 39:53.920 + this seems like it's going in the wrong direction + +39:53.920 --> 39:55.120 + and let's talk about it. + +39:55.120 --> 39:57.360 + And so I think one thing that's, I think are really, + +39:57.360 --> 39:58.880 + and you know, here's actually one thing + +39:58.880 --> 40:02.080 + that I think is very unique about us as a small company, + +40:02.080 --> 40:04.360 + is that if you're at a massive tech giant, + +40:04.360 --> 40:05.680 + that's a little bit hard for someone + +40:05.680 --> 40:08.120 + who's a line employee to go and talk to the CEO + +40:08.120 --> 40:10.520 + and say, I think that we're doing this wrong. + +40:10.520 --> 40:13.040 + And you know, you'll get companies like Google + +40:13.040 --> 40:15.720 + that have had some collective action from employees + +40:15.720 --> 40:19.400 + to make ethical change around things like Maven. + +40:19.400 --> 40:20.680 + And so maybe there are mechanisms + +40:20.680 --> 40:22.240 + that other companies that work, + +40:22.240 --> 40:24.480 + but here, super easy for anyone to pull me aside, + +40:24.480 --> 40:26.320 + to pull Sam aside, to pull Eli aside, + +40:26.320 --> 40:27.800 + and people do it all the time. + +40:27.800 --> 40:29.800 + One of the interesting things in the charter + +40:29.800 --> 40:31.640 + is this idea that it'd be great + +40:31.640 --> 40:34.240 + if you could try to describe or untangle + +40:34.240 --> 40:36.440 + switching from competition to collaboration + +40:36.440 --> 40:38.920 + and late stage AGI development. + +40:38.920 --> 40:39.760 + It's really interesting, + +40:39.760 --> 40:42.160 + this dance between competition and collaboration, + +40:42.160 --> 40:43.400 + how do you think about that? + +40:43.400 --> 40:45.000 + Yeah, assuming that you can actually do + +40:45.000 --> 40:47.040 + the technical side of AGI development, + +40:47.040 --> 40:48.960 + I think there's going to be two key problems + +40:48.960 --> 40:50.400 + with figuring out how do you actually deploy it + +40:50.400 --> 40:51.520 + and make it go well. + +40:51.520 --> 40:53.160 + The first one of these is the run up + +40:53.160 --> 40:56.360 + to building the first AGI. + +40:56.360 --> 40:58.920 + You look at how self driving cars are being developed, + +40:58.920 --> 41:00.680 + and it's a competitive race. + +41:00.680 --> 41:02.560 + And the thing that always happens in competitive race + +41:02.560 --> 41:04.160 + is that you have huge amounts of pressure + +41:04.160 --> 41:05.600 + to get rid of safety. + +41:06.800 --> 41:08.920 + And so that's one thing we're very concerned about, right? + +41:08.920 --> 41:12.000 + Is that people, multiple teams figuring out, + +41:12.000 --> 41:13.600 + we can actually get there, + +41:13.600 --> 41:16.680 + but you know, if we took the slower path + +41:16.680 --> 41:20.240 + that is more guaranteed to be safe, we will lose. + +41:20.240 --> 41:22.360 + And so we're going to take the fast path. + +41:22.360 --> 41:25.480 + And so the more that we can, both ourselves, + +41:25.480 --> 41:27.280 + be in a position where we don't generate + +41:27.280 --> 41:29.000 + that competitive race, where we say, + +41:29.000 --> 41:31.520 + if the race is being run and that someone else + +41:31.520 --> 41:33.280 + is further ahead than we are, + +41:33.280 --> 41:35.600 + we're not going to try to leapfrog. + +41:35.600 --> 41:37.200 + We're going to actually work with them, right? + +41:37.200 --> 41:38.800 + We will help them succeed. + +41:38.800 --> 41:40.440 + As long as what they're trying to do + +41:40.440 --> 41:42.920 + is to fulfill our mission, then we're good. + +41:42.920 --> 41:44.800 + We don't have to build AGI ourselves. + +41:44.800 --> 41:47.080 + And I think that's a really important commitment from us, + +41:47.080 --> 41:49.080 + but it can't just be unilateral, right? + +41:49.080 --> 41:50.400 + I think that it's really important + +41:50.400 --> 41:53.120 + that other players who are serious about building AGI + +41:53.120 --> 41:54.680 + make similar commitments, right? + +41:54.680 --> 41:56.640 + And I think that, you know, again, + +41:56.640 --> 41:57.840 + to the extent that everyone believes + +41:57.840 --> 42:00.080 + that AGI should be something to benefit everyone, + +42:00.080 --> 42:01.240 + then it actually really shouldn't matter + +42:01.240 --> 42:02.440 + which company builds it. + +42:02.440 --> 42:04.160 + And we should all be concerned about the case + +42:04.160 --> 42:06.080 + where we just race so hard to get there + +42:06.080 --> 42:07.640 + that something goes wrong. + +42:07.640 --> 42:09.600 + So what role do you think government, + +42:10.560 --> 42:13.840 + our favorite entity has in setting policy and rules + +42:13.840 --> 42:18.320 + about this domain, from research to the development + +42:18.320 --> 42:22.880 + to early stage, to late stage AI and AGI development? + +42:22.880 --> 42:25.640 + So I think that, first of all, + +42:25.640 --> 42:28.080 + it's really important that government's in there, right? + +42:28.080 --> 42:29.800 + In some way, shape, or form, you know, + +42:29.800 --> 42:30.920 + at the end of the day, we're talking about + +42:30.920 --> 42:35.080 + building technology that will shape how the world operates + +42:35.080 --> 42:39.040 + and that there needs to be government as part of that answer. + +42:39.040 --> 42:42.160 + And so that's why we've done a number + +42:42.160 --> 42:43.600 + of different congressional testimonies. + +42:43.600 --> 42:46.440 + We interact with a number of different lawmakers + +42:46.440 --> 42:50.040 + and that right now, a lot of our message to them + +42:50.040 --> 42:54.360 + is that it's not the time for regulation, + +42:54.360 --> 42:56.400 + it is the time for measurement, right? + +42:56.400 --> 42:59.080 + That our main policy recommendation is that people, + +42:59.080 --> 43:00.680 + and you know, the government does this all the time + +43:00.680 --> 43:04.880 + with bodies like NIST, spend time trying to figure out + +43:04.880 --> 43:07.920 + just where the technology is, how fast it's moving, + +43:07.920 --> 43:11.200 + and can really become literate and up to speed + +43:11.200 --> 43:13.520 + with respect to what to expect. + +43:13.520 --> 43:15.240 + So I think that today, the answer really + +43:15.240 --> 43:17.320 + is about measurement. + +43:17.320 --> 43:20.160 + And I think that there will be a time and place + +43:20.160 --> 43:21.720 + where that will change. + +43:21.720 --> 43:24.840 + And I think it's a little bit hard to predict exactly + +43:24.840 --> 43:27.120 + what exactly that trajectory should look like. + +43:27.120 --> 43:31.080 + So there will be a point at which regulation, + +43:31.080 --> 43:34.200 + federal in the United States, the government steps in + +43:34.200 --> 43:39.200 + and helps be the, I don't wanna say the adult in the room, + +43:39.520 --> 43:42.400 + to make sure that there is strict rules, + +43:42.400 --> 43:45.200 + maybe conservative rules that nobody can cross. + +43:45.200 --> 43:47.400 + Well, I think there's kind of maybe two angles to it. + +43:47.400 --> 43:49.800 + So today with narrow AI applications, + +43:49.800 --> 43:51.960 + that I think there are already existing bodies + +43:51.960 --> 43:54.880 + that are responsible and should be responsible for regulation. + +43:54.880 --> 43:57.040 + You think about, for example, with self driving cars, + +43:57.040 --> 43:59.440 + that you want the national highway. + +44:00.720 --> 44:02.920 + Yeah, exactly to be regulated in that. + +44:02.920 --> 44:04.040 + That makes sense, right? + +44:04.040 --> 44:04.960 + That basically what we're saying + +44:04.960 --> 44:08.120 + is that we're going to have these technological systems + +44:08.120 --> 44:10.600 + that are going to be performing applications + +44:10.600 --> 44:12.280 + that humans already do. + +44:12.280 --> 44:14.800 + Great, we already have ways of thinking about standards + +44:14.800 --> 44:16.160 + and safety for those. + +44:16.160 --> 44:18.880 + So I think actually empowering those regulators today + +44:18.880 --> 44:20.040 + is also pretty important. + +44:20.040 --> 44:24.760 + And then I think for AGI, that there's going to be a point + +44:24.760 --> 44:26.040 + where we'll have better answers. + +44:26.040 --> 44:27.640 + And I think that maybe a similar approach + +44:27.640 --> 44:30.520 + of first measurement and start thinking about + +44:30.520 --> 44:31.640 + what the rules should be. + +44:31.640 --> 44:32.640 + I think it's really important + +44:32.640 --> 44:36.280 + that we don't prematurely squash progress. + +44:36.280 --> 44:40.160 + I think it's very easy to kind of smother a budding field. + +44:40.160 --> 44:42.160 + And I think that's something to really avoid. + +44:42.160 --> 44:43.760 + But I don't think that the right way of doing it + +44:43.760 --> 44:46.920 + is to say, let's just try to blaze ahead + +44:46.920 --> 44:50.280 + and not involve all these other stakeholders. + +44:51.480 --> 44:56.240 + So you've recently released a paper on GPT2 + +44:56.240 --> 45:01.240 + language modeling, but did not release the full model + +45:02.040 --> 45:05.280 + because you had concerns about the possible negative effects + +45:05.280 --> 45:07.480 + of the availability of such model. + +45:07.480 --> 45:10.680 + It's outside of just that decision, + +45:10.680 --> 45:14.360 + and it's super interesting because of the discussion + +45:14.360 --> 45:17.000 + at a societal level, the discourse it creates. + +45:17.000 --> 45:19.320 + So it's fascinating in that aspect. + +45:19.320 --> 45:22.880 + But if you think that's the specifics here at first, + +45:22.880 --> 45:25.920 + what are some negative effects that you envisioned? + +45:25.920 --> 45:28.600 + And of course, what are some of the positive effects? + +45:28.600 --> 45:30.640 + Yeah, so again, I think to zoom out, + +45:30.640 --> 45:34.040 + like the way that we thought about GPT2 + +45:34.040 --> 45:35.800 + is that with language modeling, + +45:35.800 --> 45:38.560 + we are clearly on a trajectory right now + +45:38.560 --> 45:40.880 + where we scale up our models + +45:40.880 --> 45:44.480 + and we get qualitatively better performance, right? + +45:44.480 --> 45:47.360 + GPT2 itself was actually just a scale up + +45:47.360 --> 45:50.680 + of a model that we've released in the previous June, right? + +45:50.680 --> 45:52.880 + And we just ran it at much larger scale + +45:52.880 --> 45:53.880 + and we got these results + +45:53.880 --> 45:57.240 + where suddenly starting to write coherent pros, + +45:57.240 --> 46:00.040 + which was not something we'd seen previously. + +46:00.040 --> 46:01.320 + And what are we doing now? + +46:01.320 --> 46:05.760 + Well, we're gonna scale up GPT2 by 10x by 100x by 1000x + +46:05.760 --> 46:07.840 + and we don't know what we're gonna get. + +46:07.840 --> 46:10.120 + And so it's very clear that the model + +46:10.120 --> 46:12.840 + that we released last June, + +46:12.840 --> 46:16.440 + I think it's kind of like, it's a good academic toy. + +46:16.440 --> 46:18.920 + It's not something that we think is something + +46:18.920 --> 46:20.440 + that can really have negative applications + +46:20.440 --> 46:21.680 + or to the extent that it can, + +46:21.680 --> 46:24.360 + that the positive of people being able to play with it + +46:24.360 --> 46:28.280 + is far outweighs the possible harms. + +46:28.280 --> 46:32.600 + You fast forward to not GPT2, but GPT20, + +46:32.600 --> 46:34.720 + and you think about what that's gonna be like. + +46:34.720 --> 46:38.200 + And I think that the capabilities are going to be substantive. + +46:38.200 --> 46:41.120 + And so there needs to be a point in between the two + +46:41.120 --> 46:43.480 + where you say, this is something + +46:43.480 --> 46:45.200 + where we are drawing the line + +46:45.200 --> 46:48.000 + and that we need to start thinking about the safety aspects. + +46:48.000 --> 46:50.160 + And I think for GPT2, we could have gone either way. + +46:50.160 --> 46:52.720 + And in fact, when we had conversations internally + +46:52.720 --> 46:54.760 + that we had a bunch of pros and cons + +46:54.760 --> 46:58.160 + and it wasn't clear which one outweighed the other. + +46:58.160 --> 46:59.840 + And I think that when we announced + +46:59.840 --> 47:02.160 + that, hey, we decide not to release this model, + +47:02.160 --> 47:03.600 + then there was a bunch of conversation + +47:03.600 --> 47:05.200 + where various people said it's so obvious + +47:05.200 --> 47:06.360 + that you should have just released it. + +47:06.360 --> 47:07.520 + There are other people that said it's so obvious + +47:07.520 --> 47:08.840 + you should not have released it. + +47:08.840 --> 47:10.960 + And I think that that almost definitionally means + +47:10.960 --> 47:13.800 + that holding it back was the correct decision. + +47:13.800 --> 47:17.000 + If it's not obvious whether something is beneficial + +47:17.000 --> 47:19.720 + or not, you should probably default to caution. + +47:19.720 --> 47:22.440 + And so I think that the overall landscape + +47:22.440 --> 47:23.760 + for how we think about it + +47:23.760 --> 47:25.920 + is that this decision could have gone either way. + +47:25.920 --> 47:27.960 + There are great arguments in both directions. + +47:27.960 --> 47:30.080 + But for future models down the road, + +47:30.080 --> 47:32.320 + and possibly sooner than you'd expect, + +47:32.320 --> 47:33.880 + because scaling these things up doesn't actually + +47:33.880 --> 47:36.800 + take that long, those ones, + +47:36.800 --> 47:39.600 + you're definitely not going to want to release into the wild. + +47:39.600 --> 47:42.640 + And so I think that we almost view this as a test case + +47:42.640 --> 47:45.360 + and to see, can we even design, + +47:45.360 --> 47:47.960 + how do you have a society or how do you have a system + +47:47.960 --> 47:50.520 + that goes from having no concept of responsible disclosure + +47:50.520 --> 47:53.440 + where the mere idea of not releasing something + +47:53.440 --> 47:55.960 + for safety reasons is unfamiliar + +47:55.960 --> 47:57.440 + to a world where you say, okay, + +47:57.440 --> 47:58.720 + we have a powerful model. + +47:58.720 --> 47:59.720 + Let's at least think about it. + +47:59.720 --> 48:01.280 + Let's go through some process. + +48:01.280 --> 48:02.680 + And you think about the security community. + +48:02.680 --> 48:03.880 + It took them a long time + +48:03.880 --> 48:05.960 + to design responsible disclosure. + +48:05.960 --> 48:07.200 + You think about this question of, + +48:07.200 --> 48:08.800 + well, I have a security exploit. + +48:08.800 --> 48:09.760 + I send it to the company. + +48:09.760 --> 48:12.000 + The company is like, tries to prosecute me + +48:12.000 --> 48:14.760 + or just ignores it. + +48:14.760 --> 48:16.080 + What do I do? + +48:16.080 --> 48:17.320 + And so the alternatives of, + +48:17.320 --> 48:19.120 + oh, I just always publish your exploits. + +48:19.120 --> 48:20.200 + That doesn't seem good either. + +48:20.200 --> 48:21.600 + And so it really took a long time + +48:21.600 --> 48:25.320 + and it was bigger than any individual. + +48:25.320 --> 48:27.080 + It's really about building a whole community + +48:27.080 --> 48:28.760 + that believe that, okay, we'll have this process + +48:28.760 --> 48:30.160 + where you send it to the company + +48:30.160 --> 48:31.680 + if they don't act at a certain time, + +48:31.680 --> 48:33.120 + then you can go public + +48:33.120 --> 48:34.440 + and you're not a bad person. + +48:34.440 --> 48:36.240 + You've done the right thing. + +48:36.240 --> 48:38.680 + And I think that in AI, + +48:38.680 --> 48:41.400 + part of the response to GPT2 just proves + +48:41.400 --> 48:44.200 + that we don't have any concept of this. + +48:44.200 --> 48:47.080 + So that's the high level picture. + +48:47.080 --> 48:48.720 + And so I think that, + +48:48.720 --> 48:51.240 + I think this was a really important move to make. + +48:51.240 --> 48:54.000 + And we could have maybe delayed it for GPT3, + +48:54.000 --> 48:56.080 + but I'm really glad we did it for GPT2. + +48:56.080 --> 48:57.760 + And so now you look at GPT2 itself + +48:57.760 --> 48:59.440 + and you think about the substance of, okay, + +48:59.440 --> 49:01.320 + what are potential negative applications? + +49:01.320 --> 49:04.120 + So you have this model that's been trained on the internet, + +49:04.120 --> 49:06.520 + which is also going to be a bunch of very biased data, + +49:06.520 --> 49:09.600 + a bunch of very offensive content in there. + +49:09.600 --> 49:13.240 + And you can ask it to generate content for you + +49:13.240 --> 49:14.600 + on basically any topic, right? + +49:14.600 --> 49:15.440 + You just give it a prompt + +49:15.440 --> 49:16.800 + and it'll just start writing + +49:16.800 --> 49:19.120 + and it writes content like you see on the internet, + +49:19.120 --> 49:21.960 + you know, even down to like saying advertisement + +49:21.960 --> 49:24.200 + in the middle of some of its generations. + +49:24.200 --> 49:26.200 + And you think about the possibilities + +49:26.200 --> 49:29.280 + for generating fake news or abusive content. + +49:29.280 --> 49:30.120 + And, you know, it's interesting + +49:30.120 --> 49:31.880 + seeing what people have done with, you know, + +49:31.880 --> 49:34.400 + we released a smaller version of GPT2 + +49:34.400 --> 49:37.480 + and the people have done things like try to generate, + +49:37.480 --> 49:40.760 + you know, take my own Facebook message history + +49:40.760 --> 49:43.360 + and generate more Facebook messages like me + +49:43.360 --> 49:47.360 + and people generating fake politician content + +49:47.360 --> 49:49.520 + or, you know, there's a bunch of things there + +49:49.520 --> 49:51.920 + where you at least have to think, + +49:51.920 --> 49:54.720 + is this going to be good for the world? + +49:54.720 --> 49:56.320 + There's the flip side, which is I think + +49:56.320 --> 49:57.840 + that there's a lot of awesome applications + +49:57.840 --> 50:01.640 + that we really want to see like creative applications + +50:01.640 --> 50:04.000 + in terms of if you have sci fi authors + +50:04.000 --> 50:06.760 + that can work with this tool and come with cool ideas, + +50:06.760 --> 50:09.720 + like that seems awesome if we can write better sci fi + +50:09.720 --> 50:11.360 + through the use of these tools. + +50:11.360 --> 50:13.080 + And we've actually had a bunch of people right into us + +50:13.080 --> 50:16.160 + asking, hey, can we use it for, you know, + +50:16.160 --> 50:18.360 + a variety of different creative applications? + +50:18.360 --> 50:21.880 + So the positive are actually pretty easy to imagine. + +50:21.880 --> 50:26.880 + There are, you know, the usual NLP applications + +50:26.880 --> 50:30.960 + that are really interesting, but let's go there. + +50:30.960 --> 50:32.960 + It's kind of interesting to think about a world + +50:32.960 --> 50:37.960 + where, look at Twitter, where not just fake news + +50:37.960 --> 50:42.960 + but smarter and smarter bots being able to spread + +50:43.040 --> 50:47.400 + in an interesting complex networking way in information + +50:47.400 --> 50:50.800 + that just floods out us regular human beings + +50:50.800 --> 50:52.880 + with our original thoughts. + +50:52.880 --> 50:57.880 + So what are your views of this world with GPT 20? + +50:58.760 --> 51:01.600 + Right, how do we think about, again, + +51:01.600 --> 51:03.560 + it's like one of those things about in the 50s + +51:03.560 --> 51:08.560 + trying to describe the internet or the smartphone. + +51:08.720 --> 51:09.960 + What do you think about that world, + +51:09.960 --> 51:11.400 + the nature of information? + +51:12.920 --> 51:16.760 + One possibility is that we'll always try to design systems + +51:16.760 --> 51:19.680 + that identify a robot versus human + +51:19.680 --> 51:21.280 + and we'll do so successfully. + +51:21.280 --> 51:24.600 + And so we'll authenticate that we're still human. + +51:24.600 --> 51:27.520 + And the other world is that we just accept the fact + +51:27.520 --> 51:30.360 + that we're swimming in a sea of fake news + +51:30.360 --> 51:32.120 + and just learn to swim there. + +51:32.120 --> 51:34.800 + Well, have you ever seen the, there's a, you know, + +51:34.800 --> 51:39.800 + popular meme of a robot with a physical arm and pen + +51:41.520 --> 51:43.440 + clicking the I'm not a robot button? + +51:43.440 --> 51:44.280 + Yeah. + +51:44.280 --> 51:48.560 + I think the truth is that really trying to distinguish + +51:48.560 --> 51:52.160 + between robot and human is a losing battle. + +51:52.160 --> 51:53.800 + Ultimately, you think it's a losing battle? + +51:53.800 --> 51:55.520 + I think it's a losing battle ultimately, right? + +51:55.520 --> 51:57.800 + I think that that is that in terms of the content, + +51:57.800 --> 51:59.360 + in terms of the actions that you can take. + +51:59.360 --> 52:01.200 + I mean, think about how captures have gone, right? + +52:01.200 --> 52:02.920 + The captures used to be a very nice, simple. + +52:02.920 --> 52:06.320 + You just have this image, all of our OCR is terrible. + +52:06.320 --> 52:08.880 + You put a couple of artifacts in it, you know, + +52:08.880 --> 52:11.040 + humans are gonna be able to tell what it is + +52:11.040 --> 52:13.840 + an AI system wouldn't be able to today. + +52:13.840 --> 52:15.720 + Like I could barely do captures. + +52:15.720 --> 52:18.360 + And I think that this is just kind of where we're going. + +52:18.360 --> 52:20.400 + I think captures where we're a moment in time thing. + +52:20.400 --> 52:22.520 + And as AI systems become more powerful, + +52:22.520 --> 52:25.520 + that there being human capabilities that can be measured + +52:25.520 --> 52:29.360 + in a very easy automated way that the AIs will not be + +52:29.360 --> 52:31.120 + capable of, I think that's just like, + +52:31.120 --> 52:34.160 + it's just an increasingly hard technical battle. + +52:34.160 --> 52:36.240 + But it's not that all hope is lost, right? + +52:36.240 --> 52:39.760 + And you think about how do we already authenticate + +52:39.760 --> 52:40.600 + ourselves, right? + +52:40.600 --> 52:41.760 + That, you know, we have systems. + +52:41.760 --> 52:43.440 + We have social security numbers. + +52:43.440 --> 52:46.560 + If you're in the U S or, you know, you have, you have, + +52:46.560 --> 52:48.920 + you know, ways of identifying individual people + +52:48.920 --> 52:51.880 + and having real world identity tied to digital identity + +52:51.880 --> 52:54.880 + seems like a step towards, you know, + +52:54.880 --> 52:56.200 + authenticating the source of content + +52:56.200 --> 52:58.240 + rather than the content itself. + +52:58.240 --> 53:00.000 + Now, there are problems with that. + +53:00.000 --> 53:03.000 + How can you have privacy and anonymity in a world + +53:03.000 --> 53:05.440 + where the only content you can really trust is, + +53:05.440 --> 53:06.560 + or the only way you can trust content + +53:06.560 --> 53:08.560 + is by looking at where it comes from. + +53:08.560 --> 53:11.400 + And so I think that building out good reputation networks + +53:11.400 --> 53:14.080 + maybe one possible solution. + +53:14.080 --> 53:16.280 + But yeah, I think that this question is not + +53:16.280 --> 53:17.720 + an obvious one. + +53:17.720 --> 53:19.320 + And I think that we, you know, + +53:19.320 --> 53:20.880 + maybe sooner than we think we'll be in a world + +53:20.880 --> 53:23.800 + where, you know, today I often will read a tweet + +53:23.800 --> 53:25.960 + and be like, do I feel like a real human wrote this? + +53:25.960 --> 53:27.560 + Or, you know, do I feel like this was like genuine? + +53:27.560 --> 53:30.160 + I feel like I can kind of judge the content a little bit. + +53:30.160 --> 53:32.640 + And I think in the future, it just won't be the case. + +53:32.640 --> 53:36.880 + You look at, for example, the FCC comments on net neutrality. + +53:36.880 --> 53:39.880 + It came out later that millions of those were auto generated + +53:39.880 --> 53:41.960 + and that the researchers were able to do various + +53:41.960 --> 53:44.040 + statistical techniques to do that. + +53:44.040 --> 53:47.160 + What do you do in a world where those statistical techniques + +53:47.160 --> 53:48.000 + don't exist? + +53:48.000 --> 53:49.120 + It's just impossible to tell the difference + +53:49.120 --> 53:50.640 + between humans and AI's. + +53:50.640 --> 53:53.960 + And in fact, the most persuasive arguments + +53:53.960 --> 53:57.200 + are written by AI, all that stuff. + +53:57.200 --> 53:58.600 + It's not sci fi anymore. + +53:58.600 --> 54:01.320 + You look at GPT2 making a great argument for why recycling + +54:01.320 --> 54:02.560 + is bad for the world. + +54:02.560 --> 54:04.440 + You got to read that and be like, huh, you're right. + +54:04.440 --> 54:06.520 + We are addressing just the symptoms. + +54:06.520 --> 54:08.120 + Yeah, that's quite interesting. + +54:08.120 --> 54:11.320 + I mean, ultimately it boils down to the physical world + +54:11.320 --> 54:13.680 + being the last frontier of proving. + +54:13.680 --> 54:16.080 + So you said like basically networks of people, + +54:16.080 --> 54:19.400 + humans vouching for humans in the physical world. + +54:19.400 --> 54:22.960 + And somehow the authentication ends there. + +54:22.960 --> 54:24.560 + I mean, if I had to ask you, + +54:25.520 --> 54:28.160 + I mean, you're way too eloquent for a human. + +54:28.160 --> 54:31.240 + So if I had to ask you to authenticate, + +54:31.240 --> 54:33.120 + like prove how do I know you're not a robot + +54:33.120 --> 54:34.920 + and how do you know I'm not a robot? + +54:34.920 --> 54:35.760 + Yeah. + +54:35.760 --> 54:40.520 + I think that's so far were in this space, + +54:40.520 --> 54:42.120 + this conversation we just had, + +54:42.120 --> 54:44.000 + the physical movements we did + +54:44.000 --> 54:47.040 + is the biggest gap between us and AI systems + +54:47.040 --> 54:49.360 + is the physical manipulation. + +54:49.360 --> 54:51.280 + So maybe that's the last frontier. + +54:51.280 --> 54:53.040 + Well, here's another question is, + +54:53.040 --> 54:57.320 + why is solving this problem important, right? + +54:57.320 --> 54:59.080 + Like what aspects are really important to us? + +54:59.080 --> 55:01.200 + And I think that probably where we'll end up + +55:01.200 --> 55:03.600 + is we'll hone in on what do we really want + +55:03.600 --> 55:06.400 + out of knowing if we're talking to a human. + +55:06.400 --> 55:09.480 + And I think that again, this comes down to identity. + +55:09.480 --> 55:11.760 + And so I think that the internet of the future, + +55:11.760 --> 55:14.840 + I expect to be one that will have lots of agents out there + +55:14.840 --> 55:16.320 + that will interact with you. + +55:16.320 --> 55:17.880 + But I think that the question of, + +55:17.880 --> 55:21.520 + is this real flesh and blood human + +55:21.520 --> 55:23.800 + or is this an automated system? + +55:23.800 --> 55:25.800 + May actually just be less important. + +55:25.800 --> 55:27.360 + Let's actually go there. + +55:27.360 --> 55:32.360 + It's GPT2 is impressive and let's look at GPT20. + +55:32.440 --> 55:37.440 + Why is it so bad that all my friends are GPT20? + +55:37.440 --> 55:42.440 + Why is it so important on the internet? + +55:43.320 --> 55:47.360 + Do you think to interact with only human beings? + +55:47.360 --> 55:50.640 + Why can't we live in a world where ideas can come + +55:50.640 --> 55:52.960 + from models trained on human data? + +55:52.960 --> 55:55.720 + Yeah, I think this is actually a really interesting question. + +55:55.720 --> 55:56.560 + This comes back to the, + +55:56.560 --> 55:59.560 + how do you even picture a world with some new technology? + +55:59.560 --> 56:02.080 + And I think that one thing that I think is important + +56:02.080 --> 56:04.760 + is, you know, let's say honesty. + +56:04.760 --> 56:07.520 + And I think that if you have, you know, almost in the + +56:07.520 --> 56:11.120 + Turing test style sense of technology, + +56:11.120 --> 56:13.200 + you have AIs that are pretending to be humans + +56:13.200 --> 56:15.800 + and deceiving you, I think that is, you know, + +56:15.800 --> 56:17.560 + that feels like a bad thing, right? + +56:17.560 --> 56:19.720 + I think that it's really important that we feel like + +56:19.720 --> 56:21.280 + we're in control of our environment, right? + +56:21.280 --> 56:23.400 + That we understand who we're interacting with. + +56:23.400 --> 56:25.880 + And if it's an AI or a human, + +56:25.880 --> 56:28.680 + that that's not something that we're being deceived about. + +56:28.680 --> 56:30.240 + But I think that the flip side of, + +56:30.240 --> 56:32.680 + can I have as meaningful of an interaction with an AI + +56:32.680 --> 56:34.240 + as I can with a human? + +56:34.240 --> 56:36.880 + Well, I actually think here you can turn to sci fi. + +56:36.880 --> 56:40.040 + And her, I think is a great example of asking this very + +56:40.040 --> 56:40.880 + question, right? + +56:40.880 --> 56:42.800 + And one thing I really love about her is it really starts + +56:42.800 --> 56:45.800 + out almost by asking how meaningful are human + +56:45.800 --> 56:47.280 + virtual relationships, right? + +56:47.280 --> 56:51.200 + And then you have a human who has a relationship with an AI + +56:51.200 --> 56:54.320 + and that you really start to be drawn into that, right? + +56:54.320 --> 56:56.960 + And that all of your emotional buttons get triggered + +56:56.960 --> 56:59.000 + in the same way as if there was a real human that was on + +56:59.000 --> 57:00.400 + the other side of that phone. + +57:00.400 --> 57:03.800 + And so I think that this is one way of thinking about it, + +57:03.800 --> 57:07.160 + is that I think that we can have meaningful interactions + +57:07.160 --> 57:09.720 + and that if there's a funny joke, + +57:09.720 --> 57:11.320 + some sense it doesn't really matter if it was written + +57:11.320 --> 57:14.600 + by a human or an AI, but what you don't want in a way + +57:14.600 --> 57:17.360 + where I think we should really draw hard lines is deception. + +57:17.360 --> 57:20.200 + And I think that as long as we're in a world where, + +57:20.200 --> 57:22.640 + you know, why do we build AI systems at all, right? + +57:22.640 --> 57:25.000 + The reason we want to build them is to enhance human lives, + +57:25.000 --> 57:26.680 + to make humans be able to do more things, + +57:26.680 --> 57:29.040 + to have humans feel more fulfilled. + +57:29.040 --> 57:32.040 + And if we can build AI systems that do that, + +57:32.040 --> 57:33.200 + you know, sign me up. + +57:33.200 --> 57:35.160 + So the process of language modeling, + +57:37.120 --> 57:38.760 + how far do you think it take us? + +57:38.760 --> 57:40.680 + Let's look at movie HER. + +57:40.680 --> 57:45.040 + Do you think a dialogue, natural language conversation + +57:45.040 --> 57:47.840 + is formulated by the Turing test, for example, + +57:47.840 --> 57:50.760 + do you think that process could be achieved through + +57:50.760 --> 57:53.160 + this kind of unsupervised language modeling? + +57:53.160 --> 57:56.960 + So I think the Turing test in its real form + +57:56.960 --> 57:58.680 + isn't just about language, right? + +57:58.680 --> 58:00.560 + It's really about reasoning too, right? + +58:00.560 --> 58:01.920 + That to really pass the Turing test, + +58:01.920 --> 58:03.880 + I should be able to teach calculus + +58:03.880 --> 58:05.520 + to whoever's on the other side + +58:05.520 --> 58:07.480 + and have it really understand calculus + +58:07.480 --> 58:09.320 + and be able to, you know, go and solve + +58:09.320 --> 58:11.280 + new calculus problems. + +58:11.280 --> 58:13.960 + And so I think that to really solve the Turing test, + +58:13.960 --> 58:16.440 + we need more than what we're seeing with language models. + +58:16.440 --> 58:18.720 + We need some way of plugging in reasoning. + +58:18.720 --> 58:22.400 + Now, how different will that be from what we already do? + +58:22.400 --> 58:23.880 + That's an open question, right? + +58:23.880 --> 58:25.480 + It might be that we need some sequence + +58:25.480 --> 58:27.200 + of totally radical new ideas, + +58:27.200 --> 58:29.560 + or it might be that we just need to kind of shape + +58:29.560 --> 58:31.920 + our existing systems in a slightly different way. + +58:33.040 --> 58:34.640 + But I think that in terms of how far + +58:34.640 --> 58:35.920 + language modeling will go, + +58:35.920 --> 58:37.520 + it's already gone way further + +58:37.520 --> 58:39.760 + than many people would have expected, right? + +58:39.760 --> 58:40.960 + I think that things like, + +58:40.960 --> 58:42.720 + and I think there's a lot of really interesting angles + +58:42.720 --> 58:45.920 + to poke in terms of how much does GPT2 + +58:45.920 --> 58:47.880 + understand physical world? + +58:47.880 --> 58:49.360 + Like, you know, you read a little bit + +58:49.360 --> 58:52.360 + about fire underwater in GPT2. + +58:52.360 --> 58:54.200 + So it's like, okay, maybe it doesn't quite understand + +58:54.200 --> 58:55.680 + what these things are. + +58:55.680 --> 58:58.560 + But at the same time, I think that you also see + +58:58.560 --> 59:00.640 + various things like smoke coming from flame, + +59:00.640 --> 59:02.680 + and you know, a bunch of these things that GPT2, + +59:02.680 --> 59:04.880 + it has no body, it has no physical experience, + +59:04.880 --> 59:07.280 + it's just statically read data. + +59:07.280 --> 59:11.680 + And I think that if the answer is like, + +59:11.680 --> 59:14.600 + we don't know yet, and these questions though, + +59:14.600 --> 59:16.240 + we're starting to be able to actually ask them + +59:16.240 --> 59:18.720 + to physical systems, to real systems that exist, + +59:18.720 --> 59:19.880 + and that's very exciting. + +59:19.880 --> 59:21.160 + Do you think, what's your intuition? + +59:21.160 --> 59:24.040 + Do you think if you just scale language modeling, + +59:24.040 --> 59:29.040 + like significantly scale, that reasoning can emerge + +59:29.320 --> 59:31.320 + from the same exact mechanisms? + +59:31.320 --> 59:34.960 + I think it's unlikely that if we just scale GPT2, + +59:34.960 --> 59:38.600 + that we'll have reasoning in the full fledged way. + +59:38.600 --> 59:39.760 + And I think that there's like, + +59:39.760 --> 59:41.520 + the type signature is a little bit wrong, right? + +59:41.520 --> 59:44.560 + That like, there's something we do with, + +59:44.560 --> 59:45.800 + that we call thinking, right? + +59:45.800 --> 59:47.640 + Where we spend a lot of compute, + +59:47.640 --> 59:49.160 + like a variable amount of compute + +59:49.160 --> 59:50.680 + to get to better answers, right? + +59:50.680 --> 59:53.040 + I think a little bit harder, I get a better answer. + +59:53.040 --> 59:55.160 + And that that kind of type signature + +59:55.160 --> 59:58.880 + isn't quite encoded in a GPT, right? + +59:58.880 --> 1:00:01.880 + GPT will kind of like, it's spent a long time + +1:00:01.880 --> 1:00:03.640 + in it's like evolutionary history, + +1:00:03.640 --> 1:00:04.680 + baking and all this information, + +1:00:04.680 --> 1:00:07.000 + getting very, very good at this predictive process. + +1:00:07.000 --> 1:00:10.320 + And then at runtime, I just kind of do one forward pass + +1:00:10.320 --> 1:00:13.240 + and am able to generate stuff. + +1:00:13.240 --> 1:00:15.560 + And so, there might be small tweaks + +1:00:15.560 --> 1:00:18.040 + to what we do in order to get the type signature, right? + +1:00:18.040 --> 1:00:21.040 + For example, well, it's not really one forward pass, right? + +1:00:21.040 --> 1:00:22.640 + You generate symbol by symbol. + +1:00:22.640 --> 1:00:25.560 + And so, maybe you generate like a whole sequence of thoughts + +1:00:25.560 --> 1:00:28.200 + and you only keep like the last bit or something. + +1:00:28.200 --> 1:00:29.840 + But I think that at the very least, + +1:00:29.840 --> 1:00:32.160 + I would expect you have to make changes like that. + +1:00:32.160 --> 1:00:35.520 + Yeah, just exactly how we, you said think + +1:00:35.520 --> 1:00:38.400 + is the process of generating thought by thought + +1:00:38.400 --> 1:00:40.360 + in the same kind of way, like you said, + +1:00:40.360 --> 1:00:43.640 + keep the last bit, the thing that we converge towards. + +1:00:45.000 --> 1:00:47.280 + And I think there's another piece which is interesting, + +1:00:47.280 --> 1:00:50.240 + which is this out of distribution generalization, right? + +1:00:50.240 --> 1:00:52.600 + That like thinking somehow lets us do that, right? + +1:00:52.600 --> 1:00:54.400 + That we have an experience of thing + +1:00:54.400 --> 1:00:56.080 + and yet somehow we just kind of keep refining + +1:00:56.080 --> 1:00:58.040 + our mental model of it. + +1:00:58.040 --> 1:01:01.160 + This is again, something that feels tied to + +1:01:01.160 --> 1:01:03.360 + whatever reasoning is. + +1:01:03.360 --> 1:01:05.720 + And maybe it's a small tweak to what we do. + +1:01:05.720 --> 1:01:08.080 + Maybe it's many ideas and we'll take as many decades. + +1:01:08.080 --> 1:01:11.920 + Yeah, so the assumption there, generalization + +1:01:11.920 --> 1:01:14.160 + out of distribution is that it's possible + +1:01:14.160 --> 1:01:16.880 + to create new ideas. + +1:01:18.160 --> 1:01:20.840 + It's possible that nobody's ever created any new ideas. + +1:01:20.840 --> 1:01:25.360 + And then with scaling GPT2 to GPT20, + +1:01:25.360 --> 1:01:30.360 + you would essentially generalize to all possible thoughts + +1:01:30.520 --> 1:01:34.200 + as humans can have, just to play devil's advocate. + +1:01:34.200 --> 1:01:37.280 + Right, I mean, how many new story ideas + +1:01:37.280 --> 1:01:39.120 + have we come up with since Shakespeare, right? + +1:01:39.120 --> 1:01:40.160 + Yeah, exactly. + +1:01:41.600 --> 1:01:44.680 + It's just all different forms of love and drama and so on. + +1:01:44.680 --> 1:01:45.800 + Okay. + +1:01:45.800 --> 1:01:47.520 + Not sure if you read Biddle Lesson, + +1:01:47.520 --> 1:01:49.400 + a recent blog post by Rich Sutton. + +1:01:49.400 --> 1:01:50.880 + Yep, I have. + +1:01:50.880 --> 1:01:53.720 + He basically says something that echoes + +1:01:53.720 --> 1:01:55.480 + some of the ideas that you've been talking about, + +1:01:55.480 --> 1:01:58.320 + which is, he says the biggest lesson + +1:01:58.320 --> 1:02:00.680 + that can be read from 70 years of AI research + +1:02:00.680 --> 1:02:03.880 + is that general methods that leverage computation + +1:02:03.880 --> 1:02:07.920 + are ultimately going to ultimately win out. + +1:02:07.920 --> 1:02:08.960 + Do you agree with this? + +1:02:08.960 --> 1:02:13.520 + So basically open AI in general about the ideas + +1:02:13.520 --> 1:02:15.880 + you're exploring about coming up with methods, + +1:02:15.880 --> 1:02:20.120 + whether it's GPT2 modeling or whether it's open AI5, + +1:02:20.120 --> 1:02:23.160 + playing Dota, where a general method + +1:02:23.160 --> 1:02:27.160 + is better than a more fine tuned, expert tuned method. + +1:02:29.760 --> 1:02:32.200 + Yeah, so I think that, well, one thing that I think + +1:02:32.200 --> 1:02:33.800 + was really interesting about the reaction + +1:02:33.800 --> 1:02:36.480 + to that blog post was that a lot of people have read this + +1:02:36.480 --> 1:02:39.440 + as saying that compute is all that matters. + +1:02:39.440 --> 1:02:41.360 + And that's a very threatening idea, right? + +1:02:41.360 --> 1:02:43.720 + And I don't think it's a true idea either, right? + +1:02:43.720 --> 1:02:45.800 + It's very clear that we have algorithmic ideas + +1:02:45.800 --> 1:02:47.920 + that have been very important for making progress. + +1:02:47.920 --> 1:02:50.720 + And to really build AI, you wanna push as far as you can + +1:02:50.720 --> 1:02:52.760 + on the computational scale, and you wanna push + +1:02:52.760 --> 1:02:55.520 + as far as you can on human ingenuity. + +1:02:55.520 --> 1:02:57.040 + And so I think you need both. + +1:02:57.040 --> 1:02:58.320 + But I think the way that you phrase the question + +1:02:58.320 --> 1:02:59.640 + is actually very good, right? + +1:02:59.640 --> 1:03:02.200 + That it's really about what kind of ideas + +1:03:02.200 --> 1:03:04.040 + should we be striving for? + +1:03:04.040 --> 1:03:07.600 + And absolutely, if you can find a scalable idea, + +1:03:07.600 --> 1:03:08.640 + you pour more compute into it, + +1:03:08.640 --> 1:03:11.400 + you pour more data into it, it gets better. + +1:03:11.400 --> 1:03:13.800 + Like that's the real Holy Grail. + +1:03:13.800 --> 1:03:16.600 + And so I think that the answer to the question, + +1:03:16.600 --> 1:03:19.920 + I think is yes, that's really how we think about it. + +1:03:19.920 --> 1:03:22.760 + And that part of why we're excited about the power + +1:03:22.760 --> 1:03:25.320 + of deep learning and the potential for building AGI + +1:03:25.320 --> 1:03:27.600 + is because we look at the systems that exist + +1:03:27.600 --> 1:03:29.720 + in the most successful AI systems, + +1:03:29.720 --> 1:03:32.680 + and we realize that you scale those up, + +1:03:32.680 --> 1:03:34.000 + they're gonna work better. + +1:03:34.000 --> 1:03:36.320 + And I think that that scalability is something + +1:03:36.320 --> 1:03:37.160 + that really gives us hope + +1:03:37.160 --> 1:03:39.600 + for being able to build transformative systems. + +1:03:39.600 --> 1:03:43.240 + So I'll tell you, this is partially an emotional, + +1:03:43.240 --> 1:03:45.760 + you know, a thing that a response that people often have, + +1:03:45.760 --> 1:03:49.280 + if compute is so important for state of the art performance, + +1:03:49.280 --> 1:03:50.760 + you know, individual developers, + +1:03:50.760 --> 1:03:52.960 + maybe a 13 year old sitting somewhere in Kansas + +1:03:52.960 --> 1:03:55.040 + or something like that, you know, they're sitting, + +1:03:55.040 --> 1:03:56.760 + they might not even have a GPU + +1:03:56.760 --> 1:04:00.080 + and or may have a single GPU, a 1080 or something like that. + +1:04:00.080 --> 1:04:02.640 + And there's this feeling like, well, + +1:04:02.640 --> 1:04:07.280 + how can I possibly compete or contribute to this world of AI + +1:04:07.280 --> 1:04:09.840 + if scale is so important? + +1:04:09.840 --> 1:04:11.920 + So if you can comment on that, + +1:04:11.920 --> 1:04:14.320 + and in general, do you think we need to also + +1:04:14.320 --> 1:04:18.800 + in the future focus on democratizing compute resources + +1:04:18.800 --> 1:04:22.680 + more or as much as we democratize the algorithms? + +1:04:22.680 --> 1:04:23.960 + Well, so the way that I think about it + +1:04:23.960 --> 1:04:28.880 + is that there's this space of possible progress, right? + +1:04:28.880 --> 1:04:30.920 + There's a space of ideas and sort of systems + +1:04:30.920 --> 1:04:32.960 + that will work, that will move us forward. + +1:04:32.960 --> 1:04:34.840 + And there's a portion of that space, + +1:04:34.840 --> 1:04:35.760 + and to some extent, + +1:04:35.760 --> 1:04:37.960 + an increasingly significant portion of that space + +1:04:37.960 --> 1:04:41.080 + that does just require massive compute resources. + +1:04:41.080 --> 1:04:44.760 + And for that, I think that the answer is kind of clear + +1:04:44.760 --> 1:04:47.960 + and that part of why we have the structure that we do + +1:04:47.960 --> 1:04:49.640 + is because we think it's really important + +1:04:49.640 --> 1:04:50.600 + to be pushing the scale + +1:04:50.600 --> 1:04:53.840 + and to be building these large clusters and systems. + +1:04:53.840 --> 1:04:55.920 + But there's another portion of the space + +1:04:55.920 --> 1:04:57.880 + that isn't about the large scale compute, + +1:04:57.880 --> 1:04:59.960 + that are these ideas that, and again, + +1:04:59.960 --> 1:05:02.200 + I think that for the ideas to really be impactful + +1:05:02.200 --> 1:05:04.200 + and really shine, that they should be ideas + +1:05:04.200 --> 1:05:05.840 + that if you scale them up, + +1:05:05.840 --> 1:05:08.840 + would work way better than they do at small scale. + +1:05:08.840 --> 1:05:11.160 + But you can discover them without massive + +1:05:11.160 --> 1:05:12.760 + computational resources. + +1:05:12.760 --> 1:05:15.200 + And if you look at the history of recent developments, + +1:05:15.200 --> 1:05:17.680 + you think about things like the GAN or the VAE, + +1:05:17.680 --> 1:05:20.920 + that these are ones that I think you could come up with them + +1:05:20.920 --> 1:05:22.720 + without having, and in practice, + +1:05:22.720 --> 1:05:24.520 + people did come up with them without having + +1:05:24.520 --> 1:05:26.560 + massive, massive computational resources. + +1:05:26.560 --> 1:05:28.000 + Right, I just talked to Ian Goodfellow, + +1:05:28.000 --> 1:05:31.600 + but the thing is the initial GAN + +1:05:31.600 --> 1:05:34.200 + produced pretty terrible results, right? + +1:05:34.200 --> 1:05:36.880 + So only because it was in a very specific, + +1:05:36.880 --> 1:05:38.640 + only because they're smart enough to know + +1:05:38.640 --> 1:05:41.520 + that this is quite surprising to generate anything + +1:05:41.520 --> 1:05:43.160 + that they know. + +1:05:43.160 --> 1:05:46.040 + Do you see a world, or is that too optimistic and dreamer, + +1:05:46.040 --> 1:05:49.760 + like, to imagine that the compute resources + +1:05:49.760 --> 1:05:52.200 + are something that's owned by governments + +1:05:52.200 --> 1:05:55.040 + and provided as a utility? + +1:05:55.040 --> 1:05:57.120 + Actually, to some extent, this question reminds me + +1:05:57.120 --> 1:06:00.280 + of a blog post from one of my former professors + +1:06:00.280 --> 1:06:02.440 + at Harvard, this guy, Matt Welch, + +1:06:02.440 --> 1:06:03.760 + who was a systems professor. + +1:06:03.760 --> 1:06:05.280 + I remember sitting in his tenure talk, right, + +1:06:05.280 --> 1:06:08.800 + and that he had literally just gotten tenure. + +1:06:08.800 --> 1:06:10.960 + He went to Google for the summer, + +1:06:10.960 --> 1:06:15.680 + and then decided he wasn't going back to academia, right? + +1:06:15.680 --> 1:06:17.760 + And kind of in his blog post, he makes this point + +1:06:17.760 --> 1:06:20.800 + that, look, as a systems researcher, + +1:06:20.800 --> 1:06:23.040 + that I come up with these cool system ideas, + +1:06:23.040 --> 1:06:25.080 + right, and kind of build a little proof of concept, + +1:06:25.080 --> 1:06:27.080 + and the best thing I could hope for + +1:06:27.080 --> 1:06:30.120 + is that the people at Google or Yahoo, + +1:06:30.120 --> 1:06:32.600 + which was around at the time, + +1:06:32.600 --> 1:06:35.400 + will implement it and actually make it work at scale, right? + +1:06:35.400 --> 1:06:36.640 + That's like the dream for me, right? + +1:06:36.640 --> 1:06:38.000 + I build the little thing, and they turn it into + +1:06:38.000 --> 1:06:40.000 + the big thing that's actually working. + +1:06:40.000 --> 1:06:43.360 + And for him, he said, I'm done with that. + +1:06:43.360 --> 1:06:45.320 + I want to be the person who's actually doing + +1:06:45.320 --> 1:06:47.200 + building and deploying. + +1:06:47.200 --> 1:06:49.560 + And I think that there's a similar dichotomy here, right? + +1:06:49.560 --> 1:06:52.400 + I think that there are people who really actually + +1:06:52.400 --> 1:06:55.240 + find value, and I think it is a valuable thing to do, + +1:06:55.240 --> 1:06:57.440 + to be the person who produces those ideas, right, + +1:06:57.440 --> 1:06:58.840 + who builds the proof of concept. + +1:06:58.840 --> 1:07:00.600 + And yeah, you don't get to generate + +1:07:00.600 --> 1:07:02.760 + the coolest possible GAN images, + +1:07:02.760 --> 1:07:04.480 + but you invented the GAN, right? + +1:07:04.480 --> 1:07:07.560 + And so there's a real trade off there. + +1:07:07.560 --> 1:07:09.040 + And I think that that's a very personal choice, + +1:07:09.040 --> 1:07:10.840 + but I think there's value in both sides. + +1:07:10.840 --> 1:07:14.600 + So do you think creating AGI, something, + +1:07:14.600 --> 1:07:19.600 + or some new models, we would see echoes of the brilliance + +1:07:20.440 --> 1:07:22.240 + even at the prototype level. + +1:07:22.240 --> 1:07:24.080 + So you would be able to develop those ideas + +1:07:24.080 --> 1:07:27.240 + without scale, the initial seeds. + +1:07:27.240 --> 1:07:30.680 + So take a look at, I always like to look at examples + +1:07:30.680 --> 1:07:32.680 + that exist, right, look at real precedent. + +1:07:32.680 --> 1:07:36.240 + And so take a look at the June 2018 model + +1:07:36.240 --> 1:07:39.200 + that we released that we scaled up to turn to GPT2. + +1:07:39.200 --> 1:07:41.280 + And you can see that at small scale, + +1:07:41.280 --> 1:07:42.800 + it set some records, right? + +1:07:42.800 --> 1:07:44.800 + This was the original GPT. + +1:07:44.800 --> 1:07:46.840 + We actually had some cool generations. + +1:07:46.840 --> 1:07:49.840 + They weren't nearly as amazing and really stunning + +1:07:49.840 --> 1:07:52.000 + as the GPT2 ones, but it was promising. + +1:07:52.000 --> 1:07:53.040 + It was interesting. + +1:07:53.040 --> 1:07:55.280 + And so I think it is the case that with a lot + +1:07:55.280 --> 1:07:58.280 + of these ideas that you see promise at small scale, + +1:07:58.280 --> 1:08:00.800 + but there isn't an asterisk here, a very big asterisk, + +1:08:00.800 --> 1:08:05.240 + which is sometimes we see behaviors that emerge + +1:08:05.240 --> 1:08:07.280 + that are qualitatively different + +1:08:07.280 --> 1:08:09.080 + from anything we saw at small scale. + +1:08:09.080 --> 1:08:12.600 + And that the original inventor of whatever algorithm + +1:08:12.600 --> 1:08:15.520 + looks at and says, I didn't think it could do that. + +1:08:15.520 --> 1:08:17.400 + This is what we saw in Dota, right? + +1:08:17.400 --> 1:08:19.320 + So PPO was created by John Shulman, + +1:08:19.320 --> 1:08:20.560 + who's a researcher here. + +1:08:20.560 --> 1:08:24.680 + And with Dota, we basically just ran PPO + +1:08:24.680 --> 1:08:26.520 + at massive, massive scale. + +1:08:26.520 --> 1:08:29.120 + And there's some tweaks in order to make it work, + +1:08:29.120 --> 1:08:31.520 + but fundamentally it's PPO at the core. + +1:08:31.520 --> 1:08:35.280 + And we were able to get this longterm planning, + +1:08:35.280 --> 1:08:38.680 + these behaviors to really play out on a time scale + +1:08:38.680 --> 1:08:40.760 + that we just thought was not possible. + +1:08:40.760 --> 1:08:42.680 + And John looked at that and was like, + +1:08:42.680 --> 1:08:44.240 + I didn't think it could do that. + +1:08:44.240 --> 1:08:45.480 + That's what happens when you're at three orders + +1:08:45.480 --> 1:08:48.400 + of magnitude more scale than you tested at. + +1:08:48.400 --> 1:08:50.600 + Yeah, but it still has the same flavors of, + +1:08:50.600 --> 1:08:55.600 + you know, at least echoes of the expected billions. + +1:08:56.000 --> 1:08:57.880 + Although I suspect with GPT, + +1:08:57.880 --> 1:09:01.800 + it's scaled more and more, you might get surprising things. + +1:09:01.800 --> 1:09:03.200 + So yeah, you're right. + +1:09:03.200 --> 1:09:06.360 + It's interesting that it's difficult to see + +1:09:06.360 --> 1:09:09.320 + how far an idea will go when it's scaled. + +1:09:09.320 --> 1:09:11.080 + It's an open question. + +1:09:11.080 --> 1:09:13.080 + Well, so to that point with Dota and PPO, + +1:09:13.080 --> 1:09:15.040 + like I mean, here's a very concrete one, right? + +1:09:15.040 --> 1:09:16.680 + It's like, it's actually one thing + +1:09:16.680 --> 1:09:17.720 + that's very surprising about Dota + +1:09:17.720 --> 1:09:20.400 + that I think people don't really pay that much attention to. + +1:09:20.400 --> 1:09:22.360 + Is the decree of generalization + +1:09:22.360 --> 1:09:24.560 + out of distribution that happens, right? + +1:09:24.560 --> 1:09:26.320 + That you have this AI that's trained + +1:09:26.320 --> 1:09:28.880 + against other bots for its entirety, + +1:09:28.880 --> 1:09:30.360 + the entirety of its existence. + +1:09:30.360 --> 1:09:31.440 + Sorry to take a step back. + +1:09:31.440 --> 1:09:36.440 + Can you talk through, you know, a story of Dota, + +1:09:37.240 --> 1:09:42.040 + a story of leading up to opening I5 and that past, + +1:09:42.040 --> 1:09:43.920 + and what was the process of self playing + +1:09:43.920 --> 1:09:45.440 + and so on of training on this? + +1:09:45.440 --> 1:09:46.280 + Yeah, yeah, yeah. + +1:09:46.280 --> 1:09:47.120 + So with Dota. + +1:09:47.120 --> 1:09:47.960 + What is Dota? + +1:09:47.960 --> 1:09:50.000 + Dota is a complex video game + +1:09:50.000 --> 1:09:51.320 + and we started training, + +1:09:51.320 --> 1:09:52.720 + we started trying to solve Dota + +1:09:52.720 --> 1:09:55.680 + because we felt like this was a step towards the real world + +1:09:55.680 --> 1:09:58.040 + relative to other games like Chess or Go, right? + +1:09:58.040 --> 1:09:59.160 + Those very cerebral games + +1:09:59.160 --> 1:10:00.480 + where you just kind of have this board + +1:10:00.480 --> 1:10:01.880 + of very discreet moves. + +1:10:01.880 --> 1:10:04.040 + Dota starts to be much more continuous time. + +1:10:04.040 --> 1:10:06.200 + So you have this huge variety of different actions + +1:10:06.200 --> 1:10:07.680 + that you have a 45 minute game + +1:10:07.680 --> 1:10:09.360 + with all these different units + +1:10:09.360 --> 1:10:11.840 + and it's got a lot of messiness to it + +1:10:11.840 --> 1:10:14.480 + that really hasn't been captured by previous games. + +1:10:14.480 --> 1:10:17.320 + And famously all of the hard coded bots for Dota + +1:10:17.320 --> 1:10:18.400 + were terrible, right? + +1:10:18.400 --> 1:10:19.920 + It's just impossible to write anything good for it + +1:10:19.920 --> 1:10:21.240 + because it's so complex. + +1:10:21.240 --> 1:10:23.280 + And so this seemed like a really good place + +1:10:23.280 --> 1:10:25.240 + to push what's the state of the art + +1:10:25.240 --> 1:10:26.800 + in reinforcement learning. + +1:10:26.800 --> 1:10:29.000 + And so we started by focusing on the one versus one + +1:10:29.000 --> 1:10:32.360 + version of the game and we're able to solve that. + +1:10:32.360 --> 1:10:33.880 + We're able to beat the world champions + +1:10:33.880 --> 1:10:37.240 + and the learning, the skill curve + +1:10:37.240 --> 1:10:38.960 + was this crazy exponential, right? + +1:10:38.960 --> 1:10:41.000 + It was like constantly we were just scaling up, + +1:10:41.000 --> 1:10:43.240 + that we were fixing bugs and that you look + +1:10:43.240 --> 1:10:46.600 + at the skill curve and it was really a very, very smooth one. + +1:10:46.600 --> 1:10:47.440 + So it's actually really interesting + +1:10:47.440 --> 1:10:50.000 + to see how that like human iteration loop + +1:10:50.000 --> 1:10:52.680 + yielded very steady exponential progress. + +1:10:52.680 --> 1:10:55.160 + And to one side note, first of all, + +1:10:55.160 --> 1:10:57.080 + it's an exceptionally popular video game. + +1:10:57.080 --> 1:10:59.400 + The side effect is that there's a lot + +1:10:59.400 --> 1:11:01.920 + of incredible human experts at that video game. + +1:11:01.920 --> 1:11:05.200 + So the benchmark that you're trying to reach is very high. + +1:11:05.200 --> 1:11:07.840 + And the other, can you talk about the approach + +1:11:07.840 --> 1:11:10.600 + that was used initially and throughout training + +1:11:10.600 --> 1:11:12.040 + these agents to play this game? + +1:11:12.040 --> 1:11:12.880 + Yep. + +1:11:12.880 --> 1:11:14.400 + And so the approach that we used is self play. + +1:11:14.400 --> 1:11:17.320 + And so you have two agents that don't know anything. + +1:11:17.320 --> 1:11:18.640 + They battle each other, + +1:11:18.640 --> 1:11:20.760 + they discover something a little bit good + +1:11:20.760 --> 1:11:22.000 + and now they both know it. + +1:11:22.000 --> 1:11:24.520 + And they just get better and better and better without bound. + +1:11:24.520 --> 1:11:27.040 + And that's a really powerful idea, right? + +1:11:27.040 --> 1:11:30.160 + That we then went from the one versus one version + +1:11:30.160 --> 1:11:32.400 + of the game and scaled up to five versus five, right? + +1:11:32.400 --> 1:11:34.280 + So you think about kind of like with basketball + +1:11:34.280 --> 1:11:35.440 + where you have this like team sport + +1:11:35.440 --> 1:11:37.640 + and you need to do all this coordination + +1:11:37.640 --> 1:11:40.920 + and we were able to push the same idea, + +1:11:40.920 --> 1:11:45.920 + the same self play to really get to the professional level + +1:11:45.920 --> 1:11:48.880 + at the full five versus five version of the game. + +1:11:48.880 --> 1:11:52.400 + And the things that I think are really interesting here + +1:11:52.400 --> 1:11:54.760 + is that these agents in some ways + +1:11:54.760 --> 1:11:56.760 + they're almost like an insect like intelligence, right? + +1:11:56.760 --> 1:11:59.920 + Where they have a lot in common with how an insect is trained, + +1:11:59.920 --> 1:12:00.760 + right? + +1:12:00.760 --> 1:12:02.640 + An insect kind of lives in this environment for a very long time + +1:12:02.640 --> 1:12:05.280 + or the ancestors of this insect have been around + +1:12:05.280 --> 1:12:07.000 + for a long time and had a lot of experience. + +1:12:07.000 --> 1:12:09.680 + I think it's baked into this agent. + +1:12:09.680 --> 1:12:12.720 + And it's not really smart in the sense of a human, right? + +1:12:12.720 --> 1:12:14.560 + It's not able to go and learn calculus, + +1:12:14.560 --> 1:12:17.000 + but it's able to navigate its environment extremely well. + +1:12:17.000 --> 1:12:18.480 + And it's able to handle unexpected things + +1:12:18.480 --> 1:12:22.080 + in the environment that's never seen before, pretty well. + +1:12:22.080 --> 1:12:24.800 + And we see the same sort of thing with our Dota bots, right? + +1:12:24.800 --> 1:12:26.720 + That they're able to, within this game, + +1:12:26.720 --> 1:12:28.440 + they're able to play against humans, + +1:12:28.440 --> 1:12:30.000 + which is something that never existed + +1:12:30.000 --> 1:12:31.360 + in its evolutionary environment. + +1:12:31.360 --> 1:12:34.400 + Totally different play styles from humans versus the bots. + +1:12:34.400 --> 1:12:37.200 + And yet it's able to handle it extremely well. + +1:12:37.200 --> 1:12:40.400 + And that's something that I think was very surprising to us + +1:12:40.400 --> 1:12:43.440 + was something that doesn't really emerge + +1:12:43.440 --> 1:12:47.200 + from what we've seen with PPO at smaller scale, right? + +1:12:47.200 --> 1:12:48.560 + And the kind of scale we're running this stuff at + +1:12:48.560 --> 1:12:51.920 + was I could take 100,000 CPU cores, + +1:12:51.920 --> 1:12:54.040 + running with like hundreds of GPUs. + +1:12:54.040 --> 1:12:59.040 + It was probably about something like hundreds of years + +1:12:59.040 --> 1:13:03.800 + of experience going into this bot every single real day. + +1:13:03.800 --> 1:13:06.200 + And so that scale is massive. + +1:13:06.200 --> 1:13:08.400 + And we start to see very different kinds of behaviors + +1:13:08.400 --> 1:13:10.760 + out of the algorithms that we all know and love. + +1:13:10.760 --> 1:13:15.160 + Dota, you mentioned, beat the world expert 1v1. + +1:13:15.160 --> 1:13:21.160 + And then you weren't able to win 5v5 this year + +1:13:21.160 --> 1:13:24.080 + at the best players in the world. + +1:13:24.080 --> 1:13:26.640 + So what's the comeback story? + +1:13:26.640 --> 1:13:27.680 + First of all, talk through that. + +1:13:27.680 --> 1:13:29.480 + That was an exceptionally exciting event. + +1:13:29.480 --> 1:13:33.160 + And what's the following months in this year look like? + +1:13:33.160 --> 1:13:33.760 + Yeah, yeah. + +1:13:33.760 --> 1:13:38.640 + So one thing that's interesting is that we lose all the time. + +1:13:38.640 --> 1:13:40.040 + Because we play here. + +1:13:40.040 --> 1:13:42.840 + So the Dota team at OpenAI, we play the bot + +1:13:42.840 --> 1:13:45.800 + against better players than our system all the time. + +1:13:45.800 --> 1:13:47.400 + Or at least we used to, right? + +1:13:47.400 --> 1:13:50.680 + Like the first time we lost publicly was we went up + +1:13:50.680 --> 1:13:53.480 + on stage at the international and we played against some + +1:13:53.480 --> 1:13:54.800 + of the best teams in the world. + +1:13:54.800 --> 1:13:56.320 + And we ended up losing both games. + +1:13:56.320 --> 1:13:58.520 + But we give them a run for their money, right? + +1:13:58.520 --> 1:14:01.440 + That both games were kind of 30 minutes, 25 minutes. + +1:14:01.440 --> 1:14:04.200 + And they went back and forth, back and forth, back and forth. + +1:14:04.200 --> 1:14:06.360 + And so I think that really shows that we're + +1:14:06.360 --> 1:14:08.280 + at the professional level. + +1:14:08.280 --> 1:14:09.640 + And that kind of looking at those games, + +1:14:09.640 --> 1:14:12.280 + we think that the coin could have gone a different direction + +1:14:12.280 --> 1:14:13.560 + and we could have had some wins. + +1:14:13.560 --> 1:14:16.200 + And so that was actually very encouraging for us. + +1:14:16.200 --> 1:14:18.360 + And you know, it's interesting because the international was + +1:14:18.360 --> 1:14:19.720 + at a fixed time, right? + +1:14:19.720 --> 1:14:22.680 + So we knew exactly what day we were going to be playing. + +1:14:22.680 --> 1:14:25.480 + And we pushed as far as we could, as fast as we could. + +1:14:25.480 --> 1:14:28.040 + Two weeks later, we had a bot that had an 80% win rate + +1:14:28.040 --> 1:14:30.120 + versus the one that played at TI. + +1:14:30.120 --> 1:14:31.720 + So the March of Progress, you know, + +1:14:31.720 --> 1:14:33.480 + that you should think of as a snapshot rather + +1:14:33.480 --> 1:14:34.920 + than as an end state. + +1:14:34.920 --> 1:14:39.000 + And so in fact, we'll be announcing our finals pretty soon. + +1:14:39.000 --> 1:14:42.760 + I actually think that we'll announce our final match + +1:14:42.760 --> 1:14:45.240 + prior to this podcast being released. + +1:14:45.240 --> 1:14:49.240 + So there should be, we'll be playing against the world + +1:14:49.240 --> 1:14:49.720 + champions. + +1:14:49.720 --> 1:14:52.520 + And you know, for us, it's really less about, + +1:14:52.520 --> 1:14:55.400 + like the way that we think about what's upcoming + +1:14:55.400 --> 1:14:59.000 + is the final milestone, the final competitive milestone + +1:14:59.000 --> 1:15:00.280 + for the project, right? + +1:15:00.280 --> 1:15:02.760 + That our goal in all of this isn't really + +1:15:02.760 --> 1:15:05.160 + about beating humans at Dota. + +1:15:05.160 --> 1:15:06.760 + Our goal is to push the state of the art + +1:15:06.760 --> 1:15:07.800 + in reinforcement learning. + +1:15:07.800 --> 1:15:08.920 + And we've done that, right? + +1:15:08.920 --> 1:15:10.680 + And we've actually learned a lot from our system + +1:15:10.680 --> 1:15:13.320 + and that we have, you know, I think a lot of exciting + +1:15:13.320 --> 1:15:14.680 + next steps that we want to take. + +1:15:14.680 --> 1:15:16.440 + And so, you know, kind of the final showcase + +1:15:16.440 --> 1:15:18.760 + of what we built, we're going to do this match. + +1:15:18.760 --> 1:15:21.240 + But for us, it's not really the success or failure + +1:15:21.240 --> 1:15:23.800 + to see, you know, do we have the coin flip go + +1:15:23.800 --> 1:15:24.840 + in our direction or against. + +1:15:25.880 --> 1:15:28.680 + Where do you see the field of deep learning + +1:15:28.680 --> 1:15:30.680 + heading in the next few years? + +1:15:31.720 --> 1:15:35.480 + Where do you see the work in reinforcement learning + +1:15:35.480 --> 1:15:40.360 + perhaps heading and more specifically with OpenAI, + +1:15:41.160 --> 1:15:43.480 + all the exciting projects that you're working on, + +1:15:44.280 --> 1:15:46.360 + what does 2019 hold for you? + +1:15:46.360 --> 1:15:47.400 + Massive scale. + +1:15:47.400 --> 1:15:47.880 + Scale. + +1:15:47.880 --> 1:15:49.480 + I will put an atrocious on that and just say, + +1:15:49.480 --> 1:15:52.200 + you know, I think that it's about ideas plus scale. + +1:15:52.200 --> 1:15:52.840 + You need both. + +1:15:52.840 --> 1:15:54.920 + So that's a really good point. + +1:15:54.920 --> 1:15:57.720 + So the question, in terms of ideas, + +1:15:58.520 --> 1:16:02.200 + you have a lot of projects that are exploring + +1:16:02.200 --> 1:16:04.280 + different areas of intelligence. + +1:16:04.280 --> 1:16:07.480 + And the question is, when you think of scale, + +1:16:07.480 --> 1:16:09.560 + do you think about growing the scale + +1:16:09.560 --> 1:16:10.680 + of those individual projects, + +1:16:10.680 --> 1:16:12.600 + or do you think about adding new projects? + +1:16:13.160 --> 1:16:17.320 + And sorry, if you were thinking about adding new projects, + +1:16:17.320 --> 1:16:19.800 + or if you look at the past, what's the process + +1:16:19.800 --> 1:16:21.960 + of coming up with new projects and new ideas? + +1:16:21.960 --> 1:16:22.680 + Yep. + +1:16:22.680 --> 1:16:24.600 + So we really have a life cycle of project here. + +1:16:25.240 --> 1:16:27.320 + So we start with a few people just working + +1:16:27.320 --> 1:16:28.440 + on a small scale idea. + +1:16:28.440 --> 1:16:30.520 + And language is actually a very good example of this, + +1:16:30.520 --> 1:16:32.440 + that it was really, you know, one person here + +1:16:32.440 --> 1:16:34.840 + who was pushing on language for a long time. + +1:16:34.840 --> 1:16:36.680 + I mean, then you get signs of life, right? + +1:16:36.680 --> 1:16:38.440 + And so this is like, let's say, you know, + +1:16:38.440 --> 1:16:42.600 + with the original GPT, we had something that was interesting. + +1:16:42.600 --> 1:16:44.760 + And we said, okay, it's time to scale this, right? + +1:16:44.760 --> 1:16:45.960 + It's time to put more people on it, + +1:16:45.960 --> 1:16:48.120 + put more computational resources behind it, + +1:16:48.120 --> 1:16:51.560 + and then we just kind of keep pushing and keep pushing. + +1:16:51.560 --> 1:16:52.920 + And the end state is something that looks like + +1:16:52.920 --> 1:16:55.400 + Dota or Robotics, where you have a large team of, + +1:16:55.400 --> 1:16:57.800 + you know, 10 or 15 people that are running things + +1:16:57.800 --> 1:17:00.680 + at very large scale, and that you're able to really have + +1:17:00.680 --> 1:17:04.280 + material engineering and, you know, + +1:17:04.280 --> 1:17:06.520 + sort of machine learning science coming together + +1:17:06.520 --> 1:17:10.200 + to make systems that work and get material results + +1:17:10.200 --> 1:17:11.560 + that just would have been impossible otherwise. + +1:17:12.200 --> 1:17:13.560 + So we do that whole life cycle. + +1:17:13.560 --> 1:17:16.600 + We've done it a number of times, you know, typically end to end. + +1:17:16.600 --> 1:17:19.960 + It's probably two years or so to do it. + +1:17:19.960 --> 1:17:21.720 + You know, the organization's been around for three years, + +1:17:21.720 --> 1:17:23.000 + so maybe we'll find that we also have + +1:17:23.000 --> 1:17:24.760 + longer life cycle projects. + +1:17:24.760 --> 1:17:27.480 + But, you know, we work up to those. + +1:17:27.480 --> 1:17:30.280 + So one team that we're actually just starting, + +1:17:30.280 --> 1:17:32.200 + Illy and I, are kicking off a new team + +1:17:32.200 --> 1:17:35.080 + called the Reasoning Team, and this is to really try to tackle + +1:17:35.080 --> 1:17:37.400 + how do you get neural networks to reason? + +1:17:37.400 --> 1:17:41.400 + And we think that this will be a long term project. + +1:17:41.400 --> 1:17:42.840 + It's one that we're very excited about. + +1:17:42.840 --> 1:17:46.200 + In terms of reasoning, super exciting topic, + +1:17:47.400 --> 1:17:52.200 + what kind of benchmarks, what kind of tests of reasoning + +1:17:52.200 --> 1:17:53.800 + do you envision? + +1:17:53.800 --> 1:17:55.880 + What would, if you set back, + +1:17:55.880 --> 1:17:59.240 + whatever drink, and you would be impressed + +1:17:59.240 --> 1:18:01.640 + that this system is able to do something, + +1:18:01.640 --> 1:18:02.760 + what would that look like? + +1:18:02.760 --> 1:18:03.800 + Theorem proving. + +1:18:03.800 --> 1:18:04.840 + Theorem proving. + +1:18:04.840 --> 1:18:09.480 + So some kind of logic, and especially mathematical logic. + +1:18:09.480 --> 1:18:10.440 + I think so, right? + +1:18:10.440 --> 1:18:12.440 + And I think that there's kind of other problems + +1:18:12.440 --> 1:18:14.520 + that are dual to theorem proving in particular. + +1:18:14.520 --> 1:18:16.840 + You know, you think about programming, + +1:18:16.840 --> 1:18:19.960 + you think about even like security analysis of code, + +1:18:19.960 --> 1:18:24.200 + that these all kind of capture the same sorts of core reasoning + +1:18:24.200 --> 1:18:27.480 + and being able to do some out of distribution generalization. + +1:18:28.440 --> 1:18:31.880 + It would be quite exciting if OpenAI Reasoning Team + +1:18:31.880 --> 1:18:33.880 + was able to prove that P equals NP. + +1:18:33.880 --> 1:18:35.080 + That would be very nice. + +1:18:35.080 --> 1:18:37.720 + It would be very, very exciting especially. + +1:18:37.720 --> 1:18:39.080 + If it turns out that P equals NP, + +1:18:39.080 --> 1:18:40.120 + that'll be interesting too. + +1:18:40.120 --> 1:18:45.160 + It would be ironic and humorous. + +1:18:45.160 --> 1:18:51.800 + So what problem stands out to you as the most exciting + +1:18:51.800 --> 1:18:55.720 + and challenging impactful to the work for us as a community + +1:18:55.720 --> 1:18:58.440 + in general and for OpenAI this year? + +1:18:58.440 --> 1:18:59.480 + You mentioned reasoning. + +1:18:59.480 --> 1:19:01.320 + I think that's a heck of a problem. + +1:19:01.320 --> 1:19:01.480 + Yeah. + +1:19:01.480 --> 1:19:02.760 + So I think reasoning is an important one. + +1:19:02.760 --> 1:19:04.840 + I think it's going to be hard to get good results in 2019. + +1:19:05.480 --> 1:19:07.480 + You know, again, just like we think about the lifecycle, + +1:19:07.480 --> 1:19:07.960 + takes time. + +1:19:08.600 --> 1:19:11.320 + I think for 2019, language modeling seems to be kind of + +1:19:11.320 --> 1:19:12.520 + on that ramp, right? + +1:19:12.520 --> 1:19:14.760 + It's at the point that we have a technique that works. + +1:19:14.760 --> 1:19:17.000 + We want to scale 100x, 1000x, see what happens. + +1:19:18.040 --> 1:19:18.360 + Awesome. + +1:19:18.360 --> 1:19:21.800 + Do you think we're living in a simulation? + +1:19:21.800 --> 1:19:24.520 + I think it's hard to have a real opinion about it. + +1:19:25.560 --> 1:19:26.200 + It's actually interesting. + +1:19:26.200 --> 1:19:29.960 + I separate out things that I think can have yield + +1:19:29.960 --> 1:19:31.880 + materially different predictions about the world + +1:19:32.520 --> 1:19:35.640 + from ones that are just kind of fun to speculate about. + +1:19:35.640 --> 1:19:37.800 + And I kind of view simulation as more like, + +1:19:37.800 --> 1:19:40.200 + is there a flying teapot between Mars and Jupiter? + +1:19:40.200 --> 1:19:43.800 + Like, maybe, but it's a little bit hard to know + +1:19:43.800 --> 1:19:45.000 + what that would mean for my life. + +1:19:45.000 --> 1:19:46.360 + So there is something actionable. + +1:19:46.360 --> 1:19:50.680 + So some of the best work opening as done + +1:19:50.680 --> 1:19:52.200 + is in the field of reinforcement learning. + +1:19:52.760 --> 1:19:56.520 + And some of the success of reinforcement learning + +1:19:56.520 --> 1:19:59.080 + come from being able to simulate the problem you're trying + +1:19:59.080 --> 1:20:00.040 + to solve. + +1:20:00.040 --> 1:20:03.560 + So do you have a hope for reinforcement, + +1:20:03.560 --> 1:20:05.160 + for the future of reinforcement learning + +1:20:05.160 --> 1:20:06.920 + and for the future of simulation? + +1:20:06.920 --> 1:20:09.000 + Like, whether we're talking about autonomous vehicles + +1:20:09.000 --> 1:20:12.760 + or any kind of system, do you see that scaling? + +1:20:12.760 --> 1:20:16.280 + So we'll be able to simulate systems and, hence, + +1:20:16.280 --> 1:20:19.400 + be able to create a simulator that echoes our real world + +1:20:19.400 --> 1:20:22.520 + and proving once and for all, even though you're denying it + +1:20:22.520 --> 1:20:23.800 + that we're living in a simulation. + +1:20:24.840 --> 1:20:26.360 + I feel like I've used that for questions, right? + +1:20:26.360 --> 1:20:28.200 + So, you know, kind of at the core there of, like, + +1:20:28.200 --> 1:20:30.280 + can we use simulation for self driving cars? + +1:20:31.080 --> 1:20:33.720 + Take a look at our robotic system, DACTL, right? + +1:20:33.720 --> 1:20:37.720 + That was trained in simulation using the Dota system, in fact. + +1:20:37.720 --> 1:20:39.560 + And it transfers to a physical robot. + +1:20:40.280 --> 1:20:42.120 + And I think everyone looks at our Dota system, + +1:20:42.120 --> 1:20:43.400 + they're like, okay, it's just a game. + +1:20:43.400 --> 1:20:45.080 + How are you ever going to escape to the real world? + +1:20:45.080 --> 1:20:47.320 + And the answer is, well, we did it with the physical robot, + +1:20:47.320 --> 1:20:48.600 + the no one could program. + +1:20:48.600 --> 1:20:50.840 + And so I think the answer is simulation goes a lot further + +1:20:50.840 --> 1:20:53.400 + than you think if you apply the right techniques to it. + +1:20:54.040 --> 1:20:55.400 + Now, there's a question of, you know, + +1:20:55.400 --> 1:20:57.400 + are the beings in that simulation going to wake up + +1:20:57.400 --> 1:20:58.520 + and have consciousness? + +1:20:59.480 --> 1:21:02.840 + I think that one seems a lot harder to, again, reason about. + +1:21:02.840 --> 1:21:05.240 + I think that, you know, you really should think about, like, + +1:21:05.240 --> 1:21:07.800 + where exactly does human consciousness come from + +1:21:07.800 --> 1:21:09.000 + in our own self awareness? + +1:21:09.000 --> 1:21:10.600 + And, you know, is it just that, like, + +1:21:10.600 --> 1:21:12.280 + once you have, like, a complicated enough neural net, + +1:21:12.280 --> 1:21:14.440 + do you have to worry about the agent's feeling pain? + +1:21:15.720 --> 1:21:17.560 + And, you know, I think there's, like, + +1:21:17.560 --> 1:21:19.320 + interesting speculation to do there. + +1:21:19.320 --> 1:21:22.920 + But, you know, again, I think it's a little bit hard to know for sure. + +1:21:22.920 --> 1:21:24.840 + Well, let me just keep with the speculation. + +1:21:24.840 --> 1:21:28.040 + Do you think to create intelligence, general intelligence, + +1:21:28.600 --> 1:21:33.000 + you need one consciousness and two a body? + +1:21:33.000 --> 1:21:34.920 + Do you think any of those elements are needed, + +1:21:34.920 --> 1:21:38.360 + or is intelligence something that's orthogonal to those? + +1:21:38.360 --> 1:21:41.560 + I'll stick to the kind of, like, the non grand answer first, + +1:21:41.560 --> 1:21:41.720 + right? + +1:21:41.720 --> 1:21:43.960 + So the non grand answer is just to look at, + +1:21:43.960 --> 1:21:45.560 + you know, what are we already making work? + +1:21:45.560 --> 1:21:47.640 + You look at GPT2, a lot of people would have said + +1:21:47.640 --> 1:21:49.320 + that to even get these kinds of results, + +1:21:49.320 --> 1:21:50.920 + you need real world experience. + +1:21:50.920 --> 1:21:52.440 + You need a body, you need grounding. + +1:21:52.440 --> 1:21:54.920 + How are you supposed to reason about any of these things? + +1:21:54.920 --> 1:21:56.360 + How are you supposed to, like, even kind of know + +1:21:56.360 --> 1:21:57.960 + about smoke and fire and those things + +1:21:57.960 --> 1:21:59.560 + if you've never experienced them? + +1:21:59.560 --> 1:22:03.000 + And GPT2 shows that you can actually go way further + +1:22:03.000 --> 1:22:05.640 + than that kind of reasoning would predict. + +1:22:05.640 --> 1:22:09.240 + So I think that in terms of, do we need consciousness? + +1:22:09.240 --> 1:22:10.360 + Do we need a body? + +1:22:10.360 --> 1:22:11.880 + It seems the answer is probably not, right? + +1:22:11.880 --> 1:22:13.640 + That we could probably just continue to push + +1:22:13.640 --> 1:22:14.680 + kind of the systems we have. + +1:22:14.680 --> 1:22:16.520 + They already feel general. + +1:22:16.520 --> 1:22:19.080 + They're not as competent or as general + +1:22:19.080 --> 1:22:21.640 + or able to learn as quickly as an AGI would, + +1:22:21.640 --> 1:22:24.680 + but, you know, they're at least like kind of proto AGI + +1:22:24.680 --> 1:22:28.040 + in some way, and they don't need any of those things. + +1:22:28.040 --> 1:22:31.640 + Now, let's move to the grand answer, which is, you know, + +1:22:31.640 --> 1:22:34.840 + if our neural nets consciousness, + +1:22:34.840 --> 1:22:37.240 + nets conscious already, would we ever know? + +1:22:37.240 --> 1:22:38.680 + How can we tell, right? + +1:22:38.680 --> 1:22:40.920 + And, you know, here's where the speculation starts + +1:22:40.920 --> 1:22:44.760 + to become, you know, at least interesting or fun + +1:22:44.760 --> 1:22:46.200 + and maybe a little bit disturbing, + +1:22:46.200 --> 1:22:47.880 + depending on where you take it. + +1:22:47.880 --> 1:22:51.080 + But it certainly seems that when we think about animals, + +1:22:51.080 --> 1:22:53.080 + that there's some continuum of consciousness. + +1:22:53.080 --> 1:22:56.040 + You know, my cat, I think, is conscious in some way, right? + +1:22:56.040 --> 1:22:58.040 + You know, not as conscious as a human. + +1:22:58.040 --> 1:22:59.880 + And you could imagine that you could build + +1:22:59.880 --> 1:23:01.000 + a little consciousness meter, right? + +1:23:01.000 --> 1:23:02.840 + You point at a cat, it gives you a little reading, + +1:23:02.840 --> 1:23:06.200 + you point at a human, it gives you much bigger reading. + +1:23:06.200 --> 1:23:07.960 + What would happen if you pointed one of those + +1:23:07.960 --> 1:23:09.800 + at a Dota neural net? + +1:23:09.800 --> 1:23:11.960 + And if you're training this massive simulation, + +1:23:11.960 --> 1:23:14.600 + do the neural nets feel pain? + +1:23:14.600 --> 1:23:16.760 + You know, it becomes pretty hard to know + +1:23:16.760 --> 1:23:20.040 + that the answer is no, and it becomes pretty hard + +1:23:20.040 --> 1:23:22.360 + to really think about what that would mean + +1:23:22.360 --> 1:23:25.160 + if the answer were yes. + +1:23:25.160 --> 1:23:27.400 + And it's very possible, you know, for example, + +1:23:27.400 --> 1:23:29.400 + you could imagine that maybe the reason + +1:23:29.400 --> 1:23:31.400 + that humans have consciousness + +1:23:31.400 --> 1:23:35.000 + is because it's a convenient computational shortcut, right? + +1:23:35.000 --> 1:23:36.920 + If you think about it, if you have a being + +1:23:36.920 --> 1:23:39.320 + that wants to avoid pain, which seems pretty important + +1:23:39.320 --> 1:23:41.000 + to survive in this environment + +1:23:41.000 --> 1:23:43.640 + and wants to, like, you know, eat food, + +1:23:43.640 --> 1:23:45.400 + then maybe the best way of doing it + +1:23:45.400 --> 1:23:47.080 + is to have a being that's conscious, right? + +1:23:47.080 --> 1:23:49.480 + That, you know, in order to succeed in the environment, + +1:23:49.480 --> 1:23:51.080 + you need to have those properties + +1:23:51.080 --> 1:23:52.600 + and how are you supposed to implement them? + +1:23:52.600 --> 1:23:55.240 + And maybe this consciousness is a way of doing that. + +1:23:55.240 --> 1:23:57.720 + If that's true, then actually maybe we should expect + +1:23:57.720 --> 1:23:59.880 + that really competent reinforcement learning agents + +1:23:59.880 --> 1:24:01.960 + will also have consciousness. + +1:24:01.960 --> 1:24:03.240 + But, you know, that's a big if. + +1:24:03.240 --> 1:24:04.760 + And I think there are a lot of other arguments + +1:24:04.760 --> 1:24:05.880 + that you can make in other directions. + +1:24:06.680 --> 1:24:08.360 + I think that's a really interesting idea + +1:24:08.360 --> 1:24:11.400 + that even GPT2 has some degree of consciousness. + +1:24:11.400 --> 1:24:14.200 + That's something that's actually not as crazy + +1:24:14.200 --> 1:24:14.760 + to think about. + +1:24:14.760 --> 1:24:17.720 + It's useful to think about as we think about + +1:24:17.720 --> 1:24:19.800 + what it means to create intelligence of a dog, + +1:24:19.800 --> 1:24:24.360 + intelligence of a cat, and the intelligence of a human. + +1:24:24.360 --> 1:24:30.760 + So, last question, do you think we will ever fall in love, + +1:24:30.760 --> 1:24:33.560 + like in the movie, Her, with an artificial intelligence system + +1:24:34.360 --> 1:24:36.200 + or an artificial intelligence system + +1:24:36.200 --> 1:24:38.440 + falling in love with a human? + +1:24:38.440 --> 1:24:38.920 + I hope so. + +1:24:40.120 --> 1:24:43.640 + If there's any better way to end it is on love. + +1:24:43.640 --> 1:24:45.560 + So, Greg, thanks so much for talking today. + +1:24:45.560 --> 1:24:55.560 + Thank you for having me. +