Datasets:
Languages:
English
Multilinguality:
monolingual
Size Categories:
n<1K
Language Creators:
found
Source Datasets:
original
Tags:
karpathy,whisper,openai
WEBVTT | |
00:00.000 --> 00:02.520 | |
The following is a conversation with George Hotz. | |
00:02.520 --> 00:04.480 | |
He's the founder of Comma AI, | |
00:04.480 --> 00:07.400 | |
a machine learning based vehicle automation company. | |
00:07.400 --> 00:10.200 | |
He is most certainly an outspoken personality | |
00:10.200 --> 00:13.160 | |
in the field of AI and technology in general. | |
00:13.160 --> 00:16.240 | |
He first gained recognition for being the first person | |
00:16.240 --> 00:18.400 | |
to carry on lock and iPhone. | |
00:18.400 --> 00:21.280 | |
And since then, he's done quite a few interesting things | |
00:21.280 --> 00:24.400 | |
at the intersection of hardware and software. | |
00:24.400 --> 00:27.440 | |
This is the artificial intelligence podcast. | |
00:27.440 --> 00:29.560 | |
If you enjoy it, subscribe on YouTube, | |
00:29.560 --> 00:32.920 | |
give it five stars on iTunes, support it on Patreon, | |
00:32.920 --> 00:34.920 | |
or simply connect with me on Twitter. | |
00:34.920 --> 00:39.120 | |
Alex Friedman, spelled F R I D M A N. | |
00:39.120 --> 00:42.000 | |
And I'd like to give a special thank you to Jennifer | |
00:42.000 --> 00:45.880 | |
from Canada for her support of the podcast on Patreon. | |
00:45.880 --> 00:47.720 | |
Merci beaucoup, Jennifer. | |
00:47.720 --> 00:50.600 | |
She's been a friend and an engineering colleague | |
00:50.600 --> 00:52.800 | |
for many years since I was in grad school. | |
00:52.800 --> 00:55.520 | |
Your support means a lot and inspires me | |
00:55.520 --> 00:57.920 | |
to keep this series going. | |
00:57.920 --> 01:01.600 | |
And now here's my conversation with George Hotz. | |
01:02.720 --> 01:04.720 | |
Do you think we're living in a simulation? | |
01:06.480 --> 01:10.080 | |
Yes, but it may be unfalsifiable. | |
01:10.080 --> 01:12.440 | |
What do you mean by unfalsifiable? | |
01:12.440 --> 01:16.840 | |
So if the simulation is designed in such a way | |
01:16.840 --> 01:19.640 | |
that they did like a formal proof | |
01:19.640 --> 01:22.320 | |
to show that no information can get in and out. | |
01:22.320 --> 01:25.200 | |
And if their hardware is designed for the anything | |
01:25.200 --> 01:27.880 | |
in the simulation to always keep the hardware in spec, | |
01:27.880 --> 01:29.480 | |
it may be impossible to prove | |
01:29.480 --> 01:31.280 | |
whether we're in a simulation or not. | |
01:32.600 --> 01:35.680 | |
So they've designed it such that it's a closed system, | |
01:35.680 --> 01:37.200 | |
you can't get outside the system. | |
01:37.200 --> 01:38.760 | |
Well, maybe it's one of three worlds. | |
01:38.760 --> 01:41.400 | |
We're either in a simulation which can be exploited, | |
01:41.400 --> 01:44.200 | |
we're in a simulation which not only can't be exploited, | |
01:44.200 --> 01:46.440 | |
but like the same thing's true about VMs. | |
01:46.440 --> 01:48.160 | |
A really well designed VM, | |
01:48.160 --> 01:50.520 | |
you can't even detect if you're in a VM or not. | |
01:51.400 --> 01:52.520 | |
That's brilliant. | |
01:52.520 --> 01:55.160 | |
So we're, yeah, so the simulation is running | |
01:55.160 --> 01:56.800 | |
on a virtual machine. | |
01:56.800 --> 01:59.440 | |
But now in reality, all VMs have ways to detect. | |
01:59.440 --> 02:00.280 | |
That's the point. | |
02:00.280 --> 02:04.840 | |
I mean, is it, you've done quite a bit of hacking yourself. | |
02:04.840 --> 02:08.640 | |
So you should know that really any complicated system | |
02:08.640 --> 02:11.000 | |
will have ways in and out. | |
02:11.000 --> 02:14.240 | |
So this isn't necessarily true going forward. | |
02:15.280 --> 02:18.080 | |
I spent my time away from comma, | |
02:18.080 --> 02:21.240 | |
I learned a cock, it's a dependently typed, | |
02:21.240 --> 02:24.360 | |
like it's a language for writing math proofs. | |
02:24.360 --> 02:28.200 | |
And if you write code that compiles in a language like that, | |
02:28.200 --> 02:30.840 | |
it is correct by definition. | |
02:30.840 --> 02:33.560 | |
The types check it's correctance. | |
02:33.560 --> 02:35.000 | |
So it's possible that the simulation | |
02:35.000 --> 02:39.640 | |
is written in a language like this, in which case, yeah. | |
02:39.640 --> 02:42.680 | |
Yeah, but that can't be sufficiently expressive | |
02:42.680 --> 02:43.760 | |
of language like that. | |
02:43.760 --> 02:44.600 | |
Oh, it can. | |
02:44.600 --> 02:45.440 | |
It can be? | |
02:45.440 --> 02:46.280 | |
Oh, yeah. | |
02:46.280 --> 02:48.920 | |
Okay, well, so, all right, so. | |
02:48.920 --> 02:50.640 | |
The simulation doesn't have to be tearing complete | |
02:50.640 --> 02:52.320 | |
if it has a scheduled end date. | |
02:52.320 --> 02:54.600 | |
Looks like it does actually with entropy. | |
02:54.600 --> 02:58.520 | |
I mean, I don't think that a simulation | |
02:58.520 --> 03:02.200 | |
that results in something as complicated as the universe | |
03:03.080 --> 03:07.280 | |
would have a formal proof of correctness, right? | |
03:08.240 --> 03:09.880 | |
It's possible, of course. | |
03:09.880 --> 03:12.720 | |
We have no idea how good their tooling is. | |
03:12.720 --> 03:14.640 | |
And we have no idea how complicated | |
03:14.640 --> 03:16.280 | |
the universe computer really is. | |
03:16.280 --> 03:17.920 | |
It may be quite simple. | |
03:17.920 --> 03:19.680 | |
It's just very large, right? | |
03:19.680 --> 03:22.160 | |
It's very, it's definitely very large. | |
03:22.160 --> 03:24.480 | |
But the fundamental rules might be super simple. | |
03:24.480 --> 03:26.240 | |
Yeah, Conway's gonna like kinda stop. | |
03:26.240 --> 03:30.320 | |
Right, so if you could hack, | |
03:30.320 --> 03:32.400 | |
so imagine the simulation that is hackable, | |
03:32.400 --> 03:33.640 | |
if you could hack it, | |
03:35.040 --> 03:37.960 | |
what would you change about the universe? | |
03:37.960 --> 03:40.560 | |
Like how would you approach hacking a simulation? | |
03:41.640 --> 03:44.360 | |
The reason I gave that talk? | |
03:44.360 --> 03:46.680 | |
By the way, I'm not familiar with the talk you gave. | |
03:46.680 --> 03:50.160 | |
I just read that you talked about escaping the simulation | |
03:50.160 --> 03:51.280 | |
or something like that. | |
03:51.280 --> 03:52.640 | |
So maybe you can tell me a little bit | |
03:52.640 --> 03:55.360 | |
about the theme and the message there too. | |
03:55.360 --> 03:57.680 | |
It wasn't a very practical talk | |
03:57.680 --> 04:00.600 | |
about how to actually escape a simulation. | |
04:00.600 --> 04:03.320 | |
It was more about a way of restructuring | |
04:03.320 --> 04:05.120 | |
an us versus them narrative. | |
04:05.120 --> 04:10.120 | |
If we continue on the path we're going with technology, | |
04:12.360 --> 04:14.160 | |
I think we're in big trouble, | |
04:14.160 --> 04:16.760 | |
like as a species and not just as a species, | |
04:16.760 --> 04:19.480 | |
but even as me as an individual member of the species. | |
04:19.480 --> 04:23.680 | |
So if we could change rhetoric to be more like, | |
04:23.680 --> 04:24.920 | |
to think upwards, | |
04:26.240 --> 04:29.080 | |
like to think about that we're in a simulation | |
04:29.080 --> 04:30.360 | |
and how we could get out, | |
04:30.360 --> 04:32.640 | |
already we'd be on the right path. | |
04:32.640 --> 04:34.800 | |
What you actually do once you do that, | |
04:34.800 --> 04:37.360 | |
well, I assume I would have acquired way more intelligence | |
04:37.360 --> 04:39.760 | |
in the process of doing that, so I'll just ask that. | |
04:39.760 --> 04:42.080 | |
So the thinking upwards, | |
04:42.080 --> 04:43.760 | |
what kind of ideas, | |
04:43.760 --> 04:45.640 | |
what kind of breakthrough ideas do you think thinking | |
04:45.640 --> 04:47.280 | |
in that way could inspire? | |
04:47.280 --> 04:49.800 | |
And why did you say upwards? | |
04:49.800 --> 04:50.640 | |
Upwards. | |
04:50.640 --> 04:51.480 | |
Into space? | |
04:51.480 --> 04:54.120 | |
Are you thinking sort of exploration in all forms? | |
04:54.120 --> 04:59.120 | |
The space narrative that held for the modernist generation | |
04:59.880 --> 05:02.600 | |
doesn't hold as well for the postmodern generation. | |
05:04.560 --> 05:05.480 | |
What's the space narrative? | |
05:05.480 --> 05:06.520 | |
Are we talking about the same space? | |
05:06.520 --> 05:07.360 | |
The three dimensional space? | |
05:07.360 --> 05:08.840 | |
No, no, space, like going up space, | |
05:08.840 --> 05:10.040 | |
like building like Elon Musk, | |
05:10.040 --> 05:11.160 | |
like we're going to build rockets, | |
05:11.160 --> 05:12.080 | |
we're going to go to Mars, | |
05:12.080 --> 05:13.560 | |
we're going to colonize the universe. | |
05:13.560 --> 05:14.720 | |
And the narrative you're referring, | |
05:14.720 --> 05:16.040 | |
I was born in the Soviet Union, | |
05:16.040 --> 05:18.000 | |
you're referring to the race to space? | |
05:18.000 --> 05:18.840 | |
The race to space, yeah. | |
05:18.840 --> 05:19.680 | |
Yes, explore, okay. | |
05:19.680 --> 05:21.760 | |
That was a great modernist narrative. | |
05:21.760 --> 05:23.360 | |
Yeah. | |
05:23.360 --> 05:26.720 | |
It doesn't seem to hold the same weight in today's culture. | |
05:27.640 --> 05:32.160 | |
I'm hoping for good postmodern narratives that replace it. | |
05:32.160 --> 05:35.560 | |
So let's think, so you work a lot with AI. | |
05:35.560 --> 05:39.080 | |
So AI is one formulation of that narrative. | |
05:39.080 --> 05:40.080 | |
There could be also, | |
05:40.080 --> 05:42.320 | |
I don't know how much you do in VR and AR. | |
05:42.320 --> 05:43.160 | |
Yeah. | |
05:43.160 --> 05:45.160 | |
That's another, I know less about it, | |
05:45.160 --> 05:47.600 | |
but every time I play with it and our research, | |
05:47.600 --> 05:49.640 | |
it's fascinating, that virtual world. | |
05:49.640 --> 05:51.840 | |
Are you interested in the virtual world? | |
05:51.840 --> 05:54.200 | |
I would like to move to virtual reality. | |
05:55.360 --> 05:56.440 | |
In terms of your work? | |
05:56.440 --> 05:58.760 | |
No, I would like to physically move there. | |
05:58.760 --> 06:00.240 | |
The apartment I can rent in the cloud | |
06:00.240 --> 06:03.240 | |
is way better than the apartment I can rent in the real world. | |
06:03.240 --> 06:04.760 | |
Well, it's all relative, isn't it? | |
06:04.760 --> 06:07.280 | |
Because others will have very nice apartments too, | |
06:07.280 --> 06:09.200 | |
so you'll be inferior in the virtual world as well. | |
06:09.200 --> 06:11.320 | |
But that's not how I view the world, right? | |
06:11.320 --> 06:12.440 | |
I don't view the world. | |
06:12.440 --> 06:15.640 | |
I mean, that's a very like, almost zero summish way | |
06:15.640 --> 06:16.480 | |
to view the world. | |
06:16.480 --> 06:18.800 | |
Say like, my great apartment isn't great | |
06:18.800 --> 06:20.400 | |
because my neighbor has one too. | |
06:20.400 --> 06:21.640 | |
No, my great apartment is great | |
06:21.640 --> 06:24.320 | |
because like, look at this dishwasher, man. | |
06:24.320 --> 06:26.640 | |
You just touch the dish and it's washed, right? | |
06:26.640 --> 06:28.680 | |
And that is great in and of itself | |
06:28.680 --> 06:30.120 | |
if I had the only apartment | |
06:30.120 --> 06:31.520 | |
or if everybody had the apartment. | |
06:31.520 --> 06:32.400 | |
I don't care. | |
06:32.400 --> 06:34.760 | |
So you have fundamental gratitude. | |
06:34.760 --> 06:39.080 | |
The world first learned of Geohot, George Hots | |
06:39.080 --> 06:42.280 | |
in August 2007, maybe before then, | |
06:42.280 --> 06:44.080 | |
but certainly in August 2007 | |
06:44.080 --> 06:46.760 | |
when you were the first person to unlock, | |
06:46.760 --> 06:48.880 | |
carry on lock an iPhone. | |
06:48.880 --> 06:50.520 | |
How did you get into hacking? | |
06:50.520 --> 06:53.080 | |
What was the first system you discovered | |
06:53.080 --> 06:55.040 | |
vulnerabilities for and broke into? | |
06:56.240 --> 07:01.240 | |
So that was really kind of the first thing. | |
07:01.640 --> 07:06.640 | |
I had a book in 2006 called Gray Hat Hacking. | |
07:07.480 --> 07:11.000 | |
And I guess I realized that | |
07:11.000 --> 07:13.480 | |
if you acquired these sort of powers | |
07:13.480 --> 07:15.280 | |
you could control the world. | |
07:16.160 --> 07:18.920 | |
But I didn't really know that much | |
07:18.920 --> 07:20.560 | |
about computers back then. | |
07:20.560 --> 07:22.120 | |
I started with electronics. | |
07:22.120 --> 07:24.200 | |
The first iPhone hack was physical. | |
07:24.200 --> 07:25.040 | |
Cardware. | |
07:25.040 --> 07:28.160 | |
You had to open it up and pull an address line high. | |
07:28.160 --> 07:29.960 | |
And it was because I didn't really know | |
07:29.960 --> 07:31.320 | |
about software exploitation. | |
07:31.320 --> 07:32.960 | |
I learned that all in the next few years | |
07:32.960 --> 07:33.920 | |
and I got very good at it. | |
07:33.920 --> 07:36.560 | |
But back then I knew about like | |
07:36.560 --> 07:38.920 | |
how memory chips are connected to processors and stuff. | |
07:38.920 --> 07:41.040 | |
But you knew about software and programming. | |
07:41.040 --> 07:43.200 | |
You didn't know. | |
07:43.200 --> 07:46.160 | |
Oh really, so your view of the world | |
07:46.160 --> 07:49.320 | |
and computers was physical, was hardware. | |
07:49.320 --> 07:52.400 | |
Actually, if you read the code that I released with that | |
07:52.400 --> 07:55.760 | |
in August 2007, it's atrocious. | |
07:55.760 --> 07:56.760 | |
What language was it? | |
07:56.760 --> 07:57.600 | |
C. | |
07:57.600 --> 07:58.440 | |
C, nice. | |
07:58.440 --> 08:01.480 | |
And in a broken sort of state machine, ask C. | |
08:01.480 --> 08:02.960 | |
I didn't know how to program. | |
08:02.960 --> 08:04.160 | |
Yeah. | |
08:04.160 --> 08:06.600 | |
So how did you learn to program? | |
08:07.520 --> 08:08.440 | |
What was your journey? | |
08:08.440 --> 08:10.040 | |
I mean, we'll talk about it. | |
08:10.040 --> 08:12.680 | |
You've live streamed some of your programming. | |
08:12.680 --> 08:14.400 | |
This chaotic, beautiful mess. | |
08:14.400 --> 08:16.480 | |
How did you arrive at that? | |
08:16.480 --> 08:18.640 | |
Years and years of practice. | |
08:18.640 --> 08:22.240 | |
I interned at Google after, | |
08:22.240 --> 08:24.800 | |
the summer after the iPhone unlock. | |
08:24.800 --> 08:26.720 | |
And I did a contract for them | |
08:26.720 --> 08:29.040 | |
where I built a hardware for Street View | |
08:29.040 --> 08:31.760 | |
and I wrote a software library to interact with it. | |
08:32.680 --> 08:34.920 | |
And it was terrible code. | |
08:34.920 --> 08:36.560 | |
And for the first time I got feedback | |
08:36.560 --> 08:38.760 | |
from people who I respected saying, | |
08:38.760 --> 08:41.160 | |
no, like, don't write code like this. | |
08:42.680 --> 08:45.680 | |
Now, of course, just getting that feedback is not enough. | |
08:45.680 --> 08:50.680 | |
The way that I really got good was, | |
08:51.000 --> 08:54.800 | |
I wanted to write this thing that could emulate | |
08:54.800 --> 08:58.440 | |
and then visualize like arm binaries | |
08:58.440 --> 09:00.040 | |
because I wanted to hack the iPhone better. | |
09:00.040 --> 09:01.960 | |
And I didn't like that I couldn't see what the, | |
09:01.960 --> 09:03.800 | |
I couldn't single step through the processor | |
09:03.800 --> 09:05.200 | |
because I had no debugger on there, | |
09:05.200 --> 09:06.640 | |
especially for the low level things like the boot ROM | |
09:06.640 --> 09:07.480 | |
and the boot loader. | |
09:07.480 --> 09:09.440 | |
So I tried to build this tool to do it. | |
09:10.920 --> 09:13.440 | |
And I built the tool once and it was terrible. | |
09:13.440 --> 09:15.120 | |
I built the tool second times, it was terrible. | |
09:15.120 --> 09:16.320 | |
I built the tool third time. | |
09:16.320 --> 09:18.600 | |
This by the time I was at Facebook, it was kind of okay. | |
09:18.600 --> 09:20.560 | |
And then I built the tool fourth time | |
09:20.560 --> 09:22.560 | |
when I was a Google intern again in 2014. | |
09:22.560 --> 09:24.320 | |
And that was the first time I was like, | |
09:24.320 --> 09:25.880 | |
this is finally usable. | |
09:25.880 --> 09:27.120 | |
How do you pronounce this, Kira? | |
09:27.120 --> 09:28.360 | |
Kira, yeah. | |
09:28.360 --> 09:31.840 | |
So it's essentially the most efficient way | |
09:31.840 --> 09:35.720 | |
to visualize the change of state of the computer | |
09:35.720 --> 09:37.200 | |
as the program is running. | |
09:37.200 --> 09:38.920 | |
That's what you mean by debugger. | |
09:38.920 --> 09:41.760 | |
Yeah, it's a timeless debugger. | |
09:41.760 --> 09:45.080 | |
So you can rewind just as easily as going forward. | |
09:45.080 --> 09:46.280 | |
Think about, if you're using GDB, | |
09:46.280 --> 09:47.880 | |
you have to put a watch on a variable. | |
09:47.880 --> 09:49.680 | |
If you want to see if that variable changes. | |
09:49.680 --> 09:51.480 | |
In Kira, you can just click on that variable. | |
09:51.480 --> 09:53.880 | |
And then it shows every single time | |
09:53.880 --> 09:56.520 | |
when that variable was changed or accessed. | |
09:56.520 --> 09:59.760 | |
Think about it like get for your computer's, the run lock. | |
09:59.760 --> 10:04.760 | |
So there's like a deep log of the state of the computer | |
10:05.640 --> 10:07.840 | |
as the program runs and you can rewind. | |
10:07.840 --> 10:11.480 | |
Why isn't that, maybe it is, maybe you can educate me. | |
10:11.480 --> 10:14.640 | |
Why isn't that kind of debugging used more often? | |
10:14.640 --> 10:16.320 | |
Because the tooling's bad. | |
10:16.320 --> 10:17.160 | |
Well, two things. | |
10:17.160 --> 10:19.360 | |
One, if you're trying to debug Chrome, | |
10:19.360 --> 10:22.920 | |
Chrome is a 200 megabyte binary | |
10:22.920 --> 10:25.440 | |
that runs slowly on desktops. | |
10:25.440 --> 10:27.760 | |
So that's gonna be really hard to use for that. | |
10:27.760 --> 10:30.160 | |
But it's really good to use for like CTFs | |
10:30.160 --> 10:33.200 | |
and for boot ROMs and for small parts of code. | |
10:33.200 --> 10:36.360 | |
So it's hard if you're trying to debug like massive systems. | |
10:36.360 --> 10:38.200 | |
What's a CTF and what's a boot ROM? | |
10:38.200 --> 10:40.480 | |
A boot ROM is the first code that executes | |
10:40.480 --> 10:42.280 | |
the minute you give power to your iPhone. | |
10:42.280 --> 10:43.520 | |
Okay. | |
10:43.520 --> 10:46.040 | |
And CTF were these competitions that I played. | |
10:46.040 --> 10:46.880 | |
Capture the flag. | |
10:46.880 --> 10:47.720 | |
Capture the flag. | |
10:47.720 --> 10:48.560 | |
I was gonna ask you about that. | |
10:48.560 --> 10:49.920 | |
What are those, those look at, | |
10:49.920 --> 10:51.440 | |
I watched a couple of videos on YouTube. | |
10:51.440 --> 10:52.920 | |
Those look fascinating. | |
10:52.920 --> 10:55.560 | |
What have you learned about maybe at the high level | |
10:55.560 --> 10:58.040 | |
in the vulnerability of systems from these competitions? | |
11:00.840 --> 11:04.200 | |
I feel like in the heyday of CTFs, | |
11:04.200 --> 11:08.160 | |
you had all of the best security people in the world | |
11:08.160 --> 11:10.720 | |
challenging each other and coming up | |
11:10.720 --> 11:13.640 | |
with new toy exploitable things over here. | |
11:13.640 --> 11:15.400 | |
And then everybody, okay, who can break it? | |
11:15.400 --> 11:17.160 | |
And when you break it, you get like, | |
11:17.160 --> 11:19.360 | |
there's like a file in the server called flag. | |
11:19.360 --> 11:20.960 | |
And then there's a program running, | |
11:20.960 --> 11:22.680 | |
listening on a socket that's vulnerable. | |
11:22.680 --> 11:25.000 | |
So you write an exploit, you get a shell, | |
11:25.000 --> 11:27.160 | |
and then you cat flag, and then you type the flag | |
11:27.160 --> 11:29.480 | |
into like a web based scoreboard and you get points. | |
11:29.480 --> 11:33.000 | |
So the goal is essentially to find an exploit in the system | |
11:33.000 --> 11:35.280 | |
that allows you to run shell, | |
11:35.280 --> 11:38.040 | |
to run arbitrary code on that system. | |
11:38.040 --> 11:40.200 | |
That's one of the categories. | |
11:40.200 --> 11:41.960 | |
That's like the Poneable category. | |
11:43.560 --> 11:44.400 | |
Poneable? | |
11:44.400 --> 11:45.240 | |
Yeah, Poneable. | |
11:45.240 --> 11:47.600 | |
It's like, you know, you Pone the program. | |
11:47.600 --> 11:48.440 | |
It's a program. | |
11:48.440 --> 11:51.760 | |
Oh, yeah. | |
11:51.760 --> 11:55.360 | |
You know, first of all, I apologize, I'm gonna say, | |
11:55.360 --> 11:56.280 | |
it's because I'm Russian, | |
11:56.280 --> 11:59.120 | |
but maybe you can help educate me. | |
12:00.120 --> 12:01.680 | |
Some video game like misspelled | |
12:01.680 --> 12:02.840 | |
to own way back in the day. | |
12:02.840 --> 12:04.880 | |
Yeah, and it's just, | |
12:04.880 --> 12:06.280 | |
I wonder if there's a definition | |
12:06.280 --> 12:08.000 | |
and I'll have to go to Urban Dictionary for it. | |
12:08.000 --> 12:09.800 | |
Yeah, it'd be interesting to see what it says. | |
12:09.800 --> 12:12.760 | |
Okay, so what was the heyday of CTL, by the way, | |
12:12.760 --> 12:15.480 | |
but was it, what decade are we talking about? | |
12:15.480 --> 12:18.400 | |
I think like, I mean, maybe I'm biased | |
12:18.400 --> 12:21.120 | |
because it's the era that I played, | |
12:21.120 --> 12:25.800 | |
but like 2011 to 2015, | |
12:27.200 --> 12:30.320 | |
because the modern CTF scene | |
12:30.320 --> 12:32.640 | |
is similar to the modern competitive programming scene. | |
12:32.640 --> 12:34.280 | |
You have people who like do drills. | |
12:34.280 --> 12:35.880 | |
You have people who practice. | |
12:35.880 --> 12:37.040 | |
And then once you've done that, | |
12:37.040 --> 12:40.040 | |
you've turned it less into a game of generic computer skill | |
12:40.040 --> 12:42.440 | |
and more into a game of, okay, you memorize, | |
12:42.440 --> 12:44.620 | |
you drill on these five categories. | |
12:45.760 --> 12:48.920 | |
And then before that, it wasn't, | |
12:48.920 --> 12:51.560 | |
it didn't have like as much attention as it had. | |
12:52.800 --> 12:53.640 | |
I don't know, they were like, | |
12:53.640 --> 12:55.200 | |
I won $30,000 once in Korea | |
12:55.200 --> 12:56.120 | |
for one of these competitions. | |
12:56.120 --> 12:56.960 | |
Holy crap. | |
12:56.960 --> 12:57.920 | |
Yeah, they were, they were, that was... | |
12:57.920 --> 12:59.520 | |
So that means, I mean, money is money, | |
12:59.520 --> 13:02.320 | |
but that means there was probably good people there. | |
13:02.320 --> 13:03.600 | |
Exactly, yeah. | |
13:03.600 --> 13:06.800 | |
Are the challenges human constructed | |
13:06.800 --> 13:10.760 | |
or are they grounded in some real flaws in real systems? | |
13:10.760 --> 13:13.080 | |
Usually they're human constructed, | |
13:13.080 --> 13:15.760 | |
but they're usually inspired by real flaws. | |
13:15.760 --> 13:17.320 | |
What kind of systems are imagined | |
13:17.320 --> 13:19.080 | |
is really focused on mobile? | |
13:19.080 --> 13:20.920 | |
Like what has vulnerabilities these days? | |
13:20.920 --> 13:25.120 | |
Is it primarily mobile systems like Android? | |
13:25.120 --> 13:26.680 | |
Oh, everything does. | |
13:26.680 --> 13:28.120 | |
Yeah, of course. | |
13:28.120 --> 13:29.360 | |
The price has kind of gone up | |
13:29.360 --> 13:31.280 | |
because less and less people can find them. | |
13:31.280 --> 13:33.160 | |
And what's happened in security is now, | |
13:33.160 --> 13:34.560 | |
if you want to like jailbreak an iPhone, | |
13:34.560 --> 13:36.960 | |
you don't need one exploit anymore, you need nine. | |
13:37.960 --> 13:39.160 | |
Nine change together? | |
13:39.160 --> 13:40.000 | |
What would you mean? | |
13:40.000 --> 13:40.840 | |
Yeah, wow. | |
13:40.840 --> 13:44.800 | |
Okay, so it's really, what's the benefit? | |
13:44.800 --> 13:48.240 | |
Speaking higher level philosophically about hacking. | |
13:48.240 --> 13:50.400 | |
I mean, it sounds from everything I've seen about you, | |
13:50.400 --> 13:55.040 | |
you just love the challenge and you don't want to do anything. | |
13:55.040 --> 13:58.120 | |
You don't want to bring that exploit out into the world | |
13:58.120 --> 14:01.680 | |
and do any actual, let it run wild. | |
14:01.680 --> 14:02.760 | |
You just want to solve it | |
14:02.760 --> 14:05.400 | |
and then you go on to the next thing. | |
14:05.400 --> 14:08.440 | |
Oh yeah, I mean, doing criminal stuff's not really worth it. | |
14:08.440 --> 14:10.520 | |
And I'll actually use the same argument | |
14:10.520 --> 14:15.440 | |
for why I don't do defense for why I don't do crime. | |
14:15.440 --> 14:16.840 | |
If you want to defend a system, | |
14:16.840 --> 14:19.280 | |
say the system has 10 holes, right? | |
14:19.280 --> 14:22.240 | |
If you find nine of those holes as a defender, | |
14:22.240 --> 14:24.240 | |
you still lose because the attacker gets in | |
14:24.240 --> 14:25.520 | |
through the last one. | |
14:25.520 --> 14:26.360 | |
If you're an attacker, | |
14:26.360 --> 14:28.720 | |
you only have to find one out of the 10. | |
14:28.720 --> 14:30.760 | |
But if you're a criminal, | |
14:30.760 --> 14:34.800 | |
if you log on with a VPN nine out of the 10 times, | |
14:34.800 --> 14:37.760 | |
but one time you forget, you're done. | |
14:37.760 --> 14:39.400 | |
Because you're caught, okay. | |
14:39.400 --> 14:41.160 | |
Because you only have to mess up once | |
14:41.160 --> 14:42.920 | |
to be caught as a criminal. | |
14:42.920 --> 14:44.320 | |
That's why I'm not a criminal. | |
14:45.920 --> 14:47.080 | |
But okay, let me, | |
14:47.080 --> 14:49.520 | |
cause I was having a discussion with somebody | |
14:49.520 --> 14:52.440 | |
just at a high level about nuclear weapons, | |
14:52.440 --> 14:56.240 | |
actually why we're having blown ourselves up yet. | |
14:56.240 --> 14:59.840 | |
And my feeling is all the smart people in the world, | |
14:59.840 --> 15:04.120 | |
if you look at the distribution of smart people, | |
15:04.120 --> 15:06.760 | |
smart people are generally good. | |
15:06.760 --> 15:07.680 | |
And then the Southern person, | |
15:07.680 --> 15:09.480 | |
I was talking to Sean Carroll, the physicist, | |
15:09.480 --> 15:11.400 | |
and he was saying no good and bad people | |
15:11.400 --> 15:14.080 | |
are evenly distributed amongst everybody. | |
15:14.080 --> 15:18.080 | |
My sense was good hackers are in general good people | |
15:18.080 --> 15:20.400 | |
and they don't want to mess with the world. | |
15:20.400 --> 15:21.920 | |
What's your sense? | |
15:21.920 --> 15:24.720 | |
I'm not even sure about that. | |
15:25.920 --> 15:30.520 | |
Like, I have a nice life. | |
15:30.520 --> 15:32.120 | |
Crime wouldn't get me anything. | |
15:34.320 --> 15:36.520 | |
But if you're good and you have these skills, | |
15:36.520 --> 15:38.720 | |
you probably have a nice life too, right? | |
15:38.720 --> 15:40.160 | |
Right, you can use the father things. | |
15:40.160 --> 15:41.120 | |
But is there an ethical, | |
15:41.120 --> 15:44.200 | |
is there a little voice in your head that says, | |
15:46.120 --> 15:49.040 | |
well, yeah, if you could hack something | |
15:49.040 --> 15:50.720 | |
to where you could hurt people | |
15:52.840 --> 15:54.960 | |
and you could earn a lot of money doing it though, | |
15:54.960 --> 15:56.320 | |
not hurt physically perhaps, | |
15:56.320 --> 15:59.000 | |
but disrupt their life in some kind of way. | |
16:00.200 --> 16:02.360 | |
Isn't there a little voice that says, | |
16:03.360 --> 16:04.560 | |
Well, two things. | |
16:04.560 --> 16:06.800 | |
One, I don't really care about money. | |
16:06.800 --> 16:08.680 | |
So like the money wouldn't be an incentive. | |
16:08.680 --> 16:10.640 | |
The thrill might be an incentive. | |
16:10.640 --> 16:14.440 | |
But when I was 19, I read crime and punishment. | |
16:14.440 --> 16:16.120 | |
That was another great one | |
16:16.120 --> 16:18.440 | |
that talked me out of ever really doing crime. | |
16:19.400 --> 16:21.720 | |
Cause it's like, that's gonna be me. | |
16:21.720 --> 16:25.040 | |
I'd get away with it, but it would just run through my head. | |
16:25.040 --> 16:26.480 | |
Even if I got away with it, you know? | |
16:26.480 --> 16:27.640 | |
And then you do crime for long enough, | |
16:27.640 --> 16:28.960 | |
you'll never get away with it. | |
16:28.960 --> 16:30.360 | |
That's right, in the end. | |
16:30.360 --> 16:32.680 | |
That's a good reason to be good. | |
16:32.680 --> 16:34.880 | |
I wouldn't say I'm good, I would just say I'm not bad. | |
16:34.880 --> 16:38.080 | |
You're a talented programmer and a hacker | |
16:38.080 --> 16:40.920 | |
in a good positive sense of the word. | |
16:40.920 --> 16:43.360 | |
You've played around, found vulnerabilities | |
16:43.360 --> 16:44.720 | |
in various systems. | |
16:44.720 --> 16:46.120 | |
What have you learned broadly | |
16:46.120 --> 16:49.480 | |
about the design of systems and so on | |
16:49.480 --> 16:51.520 | |
from that whole process? | |
16:53.280 --> 16:58.280 | |
You learn to not take things | |
16:59.280 --> 17:02.160 | |
for what people say they are, | |
17:02.160 --> 17:05.320 | |
but you look at things for what they actually are. | |
17:07.040 --> 17:07.880 | |
Yeah. | |
17:07.880 --> 17:10.080 | |
I understand that's what you tell me it is, | |
17:10.080 --> 17:11.320 | |
but what does it do? | |
17:12.960 --> 17:14.600 | |
And you have nice visualization tools | |
17:14.600 --> 17:16.720 | |
to really know what it's really doing. | |
17:16.720 --> 17:20.080 | |
Oh, I wish I'm a better programmer now than I was in 2014. | |
17:20.080 --> 17:21.880 | |
I said, Kira, that was the first tool | |
17:21.880 --> 17:23.440 | |
that I wrote that was usable. | |
17:23.440 --> 17:25.360 | |
I wouldn't say the code was great. | |
17:25.360 --> 17:27.360 | |
I still wouldn't say my code is great. | |
17:28.840 --> 17:30.760 | |
So how was your evolution as a programmer? | |
17:30.760 --> 17:32.280 | |
Except practice. | |
17:32.280 --> 17:33.880 | |
You started with C, | |
17:33.880 --> 17:35.560 | |
what point did you pick up Python? | |
17:35.560 --> 17:37.080 | |
Because you're pretty big in Python now. | |
17:37.080 --> 17:39.960 | |
Now, yeah, in college, | |
17:39.960 --> 17:42.520 | |
I went to Carnegie Mellon when I was 22. | |
17:42.520 --> 17:44.200 | |
I went back, I'm like, | |
17:44.200 --> 17:46.640 | |
I'm gonna take all your hardest CS courses | |
17:46.640 --> 17:47.640 | |
and we'll see how I do, right? | |
17:47.640 --> 17:48.560 | |
Like, did I miss anything | |
17:48.560 --> 17:51.520 | |
by not having a real undergraduate education? | |
17:51.520 --> 17:54.240 | |
Took operating systems, compilers, AI, | |
17:54.240 --> 17:56.880 | |
and they're like a freshman Weeder math course. | |
17:56.880 --> 18:01.880 | |
And some of those classes you mentioned, | |
18:03.320 --> 18:04.240 | |
pretty tough, actually. | |
18:04.240 --> 18:05.640 | |
They're great. | |
18:05.640 --> 18:07.640 | |
At least when the 2012, | |
18:07.640 --> 18:10.240 | |
circa 2012 operating systems and compilers | |
18:11.240 --> 18:14.440 | |
were two of the best classes I've ever taken in my life. | |
18:14.440 --> 18:15.640 | |
Because you write an operating system | |
18:15.640 --> 18:16.840 | |
and you write a compiler. | |
18:18.080 --> 18:19.760 | |
I wrote my operating system in C | |
18:19.760 --> 18:21.400 | |
and I wrote my compiler in Haskell, | |
18:21.400 --> 18:26.400 | |
but somehow I picked up Python that semester as well. | |
18:26.400 --> 18:28.080 | |
I started using it for the CTFs, actually. | |
18:28.080 --> 18:30.320 | |
That's when I really started to get into CTFs | |
18:30.320 --> 18:33.360 | |
and CTFs, you're all to race against the clock. | |
18:33.360 --> 18:35.120 | |
So I can't write things and see. | |
18:35.120 --> 18:36.240 | |
Oh, there's a clock component. | |
18:36.240 --> 18:37.840 | |
So you really want to use the programming language | |
18:37.840 --> 18:38.960 | |
just so you can be fastest. | |
18:38.960 --> 18:40.080 | |
48 hours. | |
18:40.080 --> 18:41.440 | |
Pwn as many of these challenges as you can. | |
18:41.440 --> 18:42.280 | |
Pwn. | |
18:42.280 --> 18:43.120 | |
Yeah. | |
18:43.120 --> 18:43.960 | |
You got like 100 points of challenge, | |
18:43.960 --> 18:45.360 | |
whatever team gets the most. | |
18:46.360 --> 18:50.240 | |
You were both at Facebook and Google for a brief stint. | |
18:50.240 --> 18:51.080 | |
Yeah. | |
18:51.080 --> 18:54.920 | |
With Project Zero, actually, at Google for five months | |
18:54.920 --> 18:56.960 | |
where you develop Kira. | |
18:56.960 --> 18:59.280 | |
What was Project Zero about in general? | |
19:01.760 --> 19:05.160 | |
Just curious about the security efforts in these companies. | |
19:05.160 --> 19:08.840 | |
Well, Project Zero started the same time I went there. | |
19:08.840 --> 19:10.080 | |
What year is it there? | |
19:11.080 --> 19:12.320 | |
2015. | |
19:12.320 --> 19:13.160 | |
2015. | |
19:13.160 --> 19:15.040 | |
So that was right at the beginning of Project Zero. | |
19:15.040 --> 19:16.200 | |
It's small. | |
19:16.200 --> 19:18.840 | |
It's Google's offensive security team. | |
19:18.840 --> 19:23.840 | |
I'll try to give the best public facing explanation | |
19:25.680 --> 19:26.520 | |
that I can. | |
19:26.520 --> 19:30.960 | |
So the idea is basically, | |
19:30.960 --> 19:33.240 | |
these vulnerabilities exist in the world. | |
19:33.240 --> 19:35.240 | |
Nation states have them. | |
19:35.240 --> 19:37.440 | |
Some high powered bad actors have them. | |
19:39.840 --> 19:44.200 | |
Sometimes people will find these vulnerabilities | |
19:44.200 --> 19:47.960 | |
and submit them in bug bounties to the companies. | |
19:47.960 --> 19:49.440 | |
But a lot of the companies don't only care. | |
19:49.440 --> 19:50.520 | |
They don't even fix the bug. | |
19:50.520 --> 19:53.760 | |
It doesn't hurt for there to be a vulnerability. | |
19:53.760 --> 19:55.880 | |
So Project Zero is like, we're going to do it different. | |
19:55.880 --> 19:57.840 | |
We're going to announce a vulnerability | |
19:57.840 --> 19:59.640 | |
and we're going to give them 90 days to fix it. | |
19:59.640 --> 20:00.800 | |
And then whether they fix it or not, | |
20:00.800 --> 20:03.200 | |
we're going to drop the Zero Day. | |
20:03.200 --> 20:04.080 | |
Oh, wow. | |
20:04.080 --> 20:05.240 | |
We're going to drop the weapon on the textbook. | |
20:05.240 --> 20:06.080 | |
That's so cool. | |
20:06.080 --> 20:07.480 | |
That is so cool. | |
20:07.480 --> 20:09.200 | |
I love that deadlines. | |
20:09.200 --> 20:10.040 | |
Oh, that's so cool. | |
20:10.040 --> 20:10.880 | |
Give them real deadlines. | |
20:10.880 --> 20:12.320 | |
Yeah. | |
20:12.320 --> 20:15.800 | |
And I think it's done a lot for moving the industry forward. | |
20:15.800 --> 20:20.360 | |
I watched your coding sessions on the streamed online. | |
20:20.360 --> 20:25.280 | |
You code things up, the basic projects, usually from scratch. | |
20:25.280 --> 20:28.200 | |
I would say, sort of as a programmer myself, | |
20:28.200 --> 20:30.360 | |
just watching you, that you type really fast | |
20:30.360 --> 20:34.440 | |
and your brain works in both brilliant and chaotic ways. | |
20:34.440 --> 20:35.800 | |
I don't know if that's always true, | |
20:35.800 --> 20:37.600 | |
but certainly for the live streams. | |
20:37.600 --> 20:41.320 | |
So it's interesting to me because I'm much slower | |
20:41.320 --> 20:43.520 | |
and systematic and careful. | |
20:43.520 --> 20:48.040 | |
And you just move probably in order of magnitude faster. | |
20:48.040 --> 20:51.800 | |
So I'm curious, is there a method to your madness? | |
20:51.800 --> 20:53.040 | |
Or is it just who you are? | |
20:53.040 --> 20:54.720 | |
There's pros and cons. | |
20:54.720 --> 20:58.080 | |
There's pros and cons to my programming style. | |
20:58.080 --> 21:00.360 | |
And I'm aware of them. | |
21:00.360 --> 21:04.480 | |
If you ask me to get something up and working quickly | |
21:04.480 --> 21:06.800 | |
with an API that's kind of undocumented, | |
21:06.800 --> 21:08.880 | |
I will do this super fast because I will throw things | |
21:08.880 --> 21:10.200 | |
at it until it works. | |
21:10.200 --> 21:14.720 | |
If you ask me to take a vector and rotate it 90 degrees | |
21:14.720 --> 21:19.320 | |
and then flip it over the X, Y plane, | |
21:19.320 --> 21:22.280 | |
I'll spam program for two hours and won't get it. | |
21:22.280 --> 21:23.480 | |
Oh, because it's something that you | |
21:23.480 --> 21:26.240 | |
could do with a sheet of paper or think through design | |
21:26.240 --> 21:30.400 | |
and then just you really just throw stuff at the wall | |
21:30.400 --> 21:34.600 | |
and you get so good at it that it usually works. | |
21:34.600 --> 21:36.920 | |
I should become better at the other kind as well. | |
21:36.920 --> 21:39.440 | |
Sometimes I will do things methodically. | |
21:39.440 --> 21:41.200 | |
It's nowhere near as entertaining on the Twitch streams. | |
21:41.200 --> 21:43.520 | |
I do exaggerate it a bit on the Twitch streams as well. | |
21:43.520 --> 21:45.480 | |
The Twitch streams, I mean, what do you want to see a game | |
21:45.480 --> 21:46.840 | |
or you want to see actions permit, right? | |
21:46.840 --> 21:48.200 | |
I'll show you APM for programming too. | |
21:48.200 --> 21:50.280 | |
Yeah, I'd recommend people go to it. | |
21:50.280 --> 21:53.800 | |
I think I watched probably several hours that you put, | |
21:53.800 --> 21:57.480 | |
like I've actually left you programming in the background | |
21:57.480 --> 22:00.400 | |
while I was programming because you made me, | |
22:00.400 --> 22:03.120 | |
it was like watching a really good gamer. | |
22:03.120 --> 22:06.240 | |
It's like energizes you because you're like moving so fast | |
22:06.240 --> 22:08.840 | |
and so it's awesome, it's inspiring. | |
22:08.840 --> 22:11.200 | |
It made me jealous that like, | |
22:12.280 --> 22:14.280 | |
because my own programming is inadequate | |
22:14.280 --> 22:16.960 | |
in terms of speed, so I was like. | |
22:16.960 --> 22:20.520 | |
So I'm twice as frantic on the live streams | |
22:20.520 --> 22:22.680 | |
as I am when I code without, oh. | |
22:22.680 --> 22:23.720 | |
It's super entertaining. | |
22:23.720 --> 22:26.400 | |
So I wasn't even paying attention to what you were coding, | |
22:26.400 --> 22:29.760 | |
which is great, it's just watching you switch windows | |
22:29.760 --> 22:31.400 | |
and Vim, I guess is the most way. | |
22:31.400 --> 22:33.000 | |
Yeah, does Vim on screen? | |
22:33.000 --> 22:35.640 | |
I've developed a workload Facebook and stuck with it. | |
22:35.640 --> 22:37.320 | |
How do you learn new programming tools, | |
22:37.320 --> 22:39.440 | |
ideas, techniques these days? | |
22:39.440 --> 22:42.080 | |
What's your like methodology for learning new things? | |
22:42.080 --> 22:45.920 | |
So I wrote for comma, | |
22:47.200 --> 22:49.280 | |
the distributed file systems out in the world | |
22:49.280 --> 22:50.720 | |
are extremely complex. | |
22:50.720 --> 22:55.280 | |
Like if you want to install something like like like Ceph, | |
22:55.280 --> 22:58.760 | |
Ceph is I think the like open infrastructure | |
22:58.760 --> 23:03.040 | |
distributed file system or there's like newer ones | |
23:03.040 --> 23:05.880 | |
like seaweed FS, but these are all like 10,000 | |
23:05.880 --> 23:06.880 | |
plus line projects. | |
23:06.880 --> 23:09.520 | |
I think some of them are even 100,000 line | |
23:09.520 --> 23:11.120 | |
and just configuring them as a nightmare. | |
23:11.120 --> 23:16.120 | |
So I wrote, I wrote one, it's 200 lines | |
23:16.440 --> 23:18.880 | |
and it uses like engine X of the line servers | |
23:18.880 --> 23:21.600 | |
and has this little master server that I wrote and go. | |
23:21.600 --> 23:24.840 | |
And the way I go, this, if I would say | |
23:24.840 --> 23:27.240 | |
that I'm proud per line of any code I wrote, | |
23:27.240 --> 23:29.160 | |
maybe there's some exploits that I think are beautiful | |
23:29.160 --> 23:31.320 | |
and then this, this is 200 lines | |
23:31.320 --> 23:33.720 | |
and just the way that I thought about it, | |
23:33.720 --> 23:35.560 | |
I think was very good and the reason it's very good | |
23:35.560 --> 23:37.640 | |
is because that was the fourth version of it that I wrote | |
23:37.640 --> 23:39.320 | |
and I had three versions that I threw away. | |
23:39.320 --> 23:41.000 | |
You mentioned, did you say go? | |
23:41.000 --> 23:41.840 | |
I wrote a go, yeah. | |
23:41.840 --> 23:42.680 | |
And go. | |
23:42.680 --> 23:43.880 | |
Is that a functional language? | |
23:43.880 --> 23:45.280 | |
I forget what go is. | |
23:45.280 --> 23:47.160 | |
Go is Google's language. | |
23:47.160 --> 23:48.200 | |
Right. | |
23:48.200 --> 23:49.480 | |
It's not functional. | |
23:49.480 --> 23:54.480 | |
It's some, it's like, in a way it's C++, but easier. | |
23:56.160 --> 23:58.200 | |
It's strongly typed. | |
23:58.200 --> 23:59.760 | |
It has a nice ecosystem around it. | |
23:59.760 --> 24:01.680 | |
When I first looked at it, I was like, | |
24:01.680 --> 24:03.800 | |
this is like Python, but it takes twice as long | |
24:03.800 --> 24:05.600 | |
to do anything. | |
24:05.600 --> 24:09.600 | |
Now that I've open pilot is migrating to C, | |
24:09.600 --> 24:11.000 | |
but it still has large Python components, | |
24:11.000 --> 24:12.760 | |
I now understand why Python doesn't work | |
24:12.760 --> 24:15.840 | |
for large code bases and why you want something like go. | |
24:15.840 --> 24:16.680 | |
Interesting. | |
24:16.680 --> 24:18.680 | |
So why, why doesn't Python work for, | |
24:18.680 --> 24:21.720 | |
so even most, speaking for myself at least, | |
24:21.720 --> 24:24.960 | |
like we do a lot of stuff, basically demo level work | |
24:24.960 --> 24:29.240 | |
with autonomous vehicles and most of the work is Python. | |
24:29.240 --> 24:32.440 | |
Why doesn't Python work for large code bases? | |
24:32.440 --> 24:37.440 | |
Because, well, lack of type checking is a big one. | |
24:37.920 --> 24:39.360 | |
So errors creep in. | |
24:39.360 --> 24:41.920 | |
Yeah, and like you don't know, | |
24:41.920 --> 24:45.320 | |
the compiler can tell you like nothing, right? | |
24:45.320 --> 24:48.440 | |
So everything is either, you know, | |
24:48.440 --> 24:49.880 | |
like syntax errors, fine, | |
24:49.880 --> 24:51.800 | |
but if you misspell a variable in Python, | |
24:51.800 --> 24:53.000 | |
the compiler won't catch that. | |
24:53.000 --> 24:56.600 | |
There's like linters that can catch it some of the time. | |
24:56.600 --> 24:57.560 | |
There's no types. | |
24:57.560 --> 25:00.520 | |
This is really the biggest downside. | |
25:00.520 --> 25:02.640 | |
And then we'll Python slow, but that's not related to it. | |
25:02.640 --> 25:04.840 | |
Well, maybe it's kind of related to it, so it's lack of. | |
25:04.840 --> 25:06.600 | |
So what's in your toolbox these days? | |
25:06.600 --> 25:07.760 | |
Is it Python or what else? | |
25:07.760 --> 25:08.600 | |
Go. | |
25:08.600 --> 25:10.240 | |
I need to move to something else. | |
25:10.240 --> 25:12.880 | |
My adventure into dependently typed languages, | |
25:12.880 --> 25:14.240 | |
I love these languages. | |
25:14.240 --> 25:17.520 | |
They just have like syntax from the 80s. | |
25:18.520 --> 25:21.120 | |
What do you think about JavaScript? | |
25:21.120 --> 25:24.000 | |
ES6, like the modern type script? | |
25:24.000 --> 25:27.320 | |
JavaScript is, the whole ecosystem | |
25:27.320 --> 25:29.320 | |
is unbelievably confusing. | |
25:29.320 --> 25:32.840 | |
NPM updates a package from 022 to 025 | |
25:32.840 --> 25:34.560 | |
and that breaks your Babel linter, | |
25:34.560 --> 25:38.560 | |
which translates your ES5 into ES6, which doesn't run on. | |
25:38.560 --> 25:42.480 | |
So why do I have to compile my JavaScript again, huh? | |
25:42.480 --> 25:44.040 | |
It may be the future though. | |
25:44.040 --> 25:45.800 | |
You think about, I mean, | |
25:45.800 --> 25:47.400 | |
I've embraced JavaScript recently | |
25:47.400 --> 25:52.280 | |
just because just like I've continually embraced PHP, | |
25:52.280 --> 25:55.360 | |
it seems that these worst possible languages live on | |
25:55.360 --> 25:57.480 | |
for the longest, like cockroaches never die. | |
25:57.480 --> 26:00.760 | |
Yeah, well, it's in the browser and it's fast. | |
26:00.760 --> 26:01.680 | |
It's fast. | |
26:01.680 --> 26:02.520 | |
Yeah. | |
26:02.520 --> 26:05.480 | |
It's in the browser and compute might stay become, | |
26:05.480 --> 26:06.440 | |
you know, the browser, | |
26:06.440 --> 26:09.040 | |
it's unclear what the role of the browser is | |
26:09.040 --> 26:11.800 | |
in terms of distributed computation in the future. | |
26:11.800 --> 26:12.640 | |
So. | |
26:13.600 --> 26:15.240 | |
JavaScript is definitely here to stay. | |
26:15.240 --> 26:16.080 | |
Yeah. | |
26:16.080 --> 26:18.160 | |
It's interesting if autonomous vehicles | |
26:18.160 --> 26:19.480 | |
will run on JavaScript one day. | |
26:19.480 --> 26:21.760 | |
I mean, you have to consider these possibilities. | |
26:21.760 --> 26:24.280 | |
Well, all our debug tools are JavaScript. | |
26:24.280 --> 26:26.040 | |
We actually just open source them. | |
26:26.040 --> 26:28.160 | |
We have a tool explorer, which you can annotate | |
26:28.160 --> 26:30.080 | |
your disengagements and we have tool Kibana, | |
26:30.080 --> 26:32.920 | |
which lets you analyze the can traffic from the car. | |
26:32.920 --> 26:35.240 | |
So basically any time you're visualizing something | |
26:35.240 --> 26:37.720 | |
about the log using JavaScript. | |
26:37.720 --> 26:40.120 | |
Well, the web is the best UI toolkit by far. | |
26:40.120 --> 26:40.960 | |
Yeah. | |
26:40.960 --> 26:41.880 | |
So, and then, you know what? | |
26:41.880 --> 26:42.760 | |
You're coding in JavaScript. | |
26:42.760 --> 26:43.600 | |
We have a React guy. | |
26:43.600 --> 26:44.440 | |
He's good. | |
26:44.440 --> 26:46.080 | |
React, nice. | |
26:46.080 --> 26:46.920 | |
Let's get into it. | |
26:46.920 --> 26:49.120 | |
So let's talk autonomous vehicles. | |
26:49.120 --> 26:50.640 | |
You found a comma AI. | |
26:51.440 --> 26:54.920 | |
Let's, at a high level, | |
26:54.920 --> 26:57.880 | |
how did you get into the world of vehicle automation? | |
26:57.880 --> 26:59.920 | |
Can you also just, for people who don't know, | |
26:59.920 --> 27:01.400 | |
tell the story of comma AI? | |
27:01.400 --> 27:02.920 | |
Sure. | |
27:02.920 --> 27:06.120 | |
So I was working at this AI startup | |
27:06.120 --> 27:09.240 | |
and a friend approached me and he's like, | |
27:09.240 --> 27:12.080 | |
dude, I don't know where this is going, | |
27:12.080 --> 27:15.160 | |
but the coolest applied AI problem today | |
27:15.160 --> 27:16.480 | |
is self driving cars. | |
27:16.480 --> 27:17.720 | |
I'm like, well, absolutely. | |
27:18.800 --> 27:20.520 | |
You wanna meet with Elon Musk | |
27:20.520 --> 27:24.560 | |
and he's looking for somebody to build a vision system | |
27:24.560 --> 27:27.600 | |
for autopilot. | |
27:27.600 --> 27:29.320 | |
This is when they were still on AP one. | |
27:29.320 --> 27:30.840 | |
They were still using Mobileye. | |
27:30.840 --> 27:33.680 | |
Elon back then was looking for a replacement. | |
27:33.680 --> 27:37.320 | |
And he brought me in and we talked about a contract | |
27:37.320 --> 27:39.040 | |
where I would deliver something | |
27:39.040 --> 27:41.640 | |
that meets Mobileye level performance. | |
27:41.640 --> 27:43.920 | |
I would get paid $12 million if I could deliver it tomorrow | |
27:43.920 --> 27:46.720 | |
and I would lose $1 million for every month I didn't deliver. | |
27:47.720 --> 27:49.080 | |
So I was like, okay, this is a great deal. | |
27:49.080 --> 27:50.800 | |
This is a super exciting challenge. | |
27:52.360 --> 27:53.200 | |
You know what? | |
27:53.200 --> 27:55.840 | |
It takes me 10 months, I get $2 million, it's good. | |
27:55.840 --> 27:57.160 | |
Maybe I can finish up in five. | |
27:57.160 --> 27:58.880 | |
Maybe I don't finish it at all and I get paid nothing | |
27:58.880 --> 28:00.880 | |
and I'll work for 12 months for free. | |
28:00.880 --> 28:02.960 | |
So maybe just take a pause on that. | |
28:02.960 --> 28:04.280 | |
I'm also curious about this | |
28:04.280 --> 28:06.360 | |
because I've been working in robotics for a long time. | |
28:06.360 --> 28:08.320 | |
And I'm curious to see a person like you just step in | |
28:08.320 --> 28:12.000 | |
and sort of somewhat naive, but brilliant, right? | |
28:12.000 --> 28:14.000 | |
So that's the best place to be | |
28:14.000 --> 28:17.240 | |
because you basically full steam take on a problem. | |
28:17.240 --> 28:19.720 | |
How confident, from that time, | |
28:19.720 --> 28:21.320 | |
because you know a lot more now, | |
28:21.320 --> 28:23.440 | |
at that time, how hard do you think it is | |
28:23.440 --> 28:25.880 | |
to solve all of autonomous driving? | |
28:25.880 --> 28:30.440 | |
I remember I suggested to Elon in the meeting | |
28:30.440 --> 28:33.120 | |
on putting a GPU behind each camera | |
28:33.120 --> 28:35.120 | |
to keep the compute local. | |
28:35.120 --> 28:38.000 | |
This is an incredibly stupid idea. | |
28:38.000 --> 28:40.080 | |
I leave the meeting 10 minutes later and I'm like, | |
28:40.080 --> 28:41.560 | |
I could have spent a little bit of time | |
28:41.560 --> 28:42.880 | |
thinking about this problem before I went in. | |
28:42.880 --> 28:44.200 | |
Why is this a stupid idea? | |
28:44.200 --> 28:46.280 | |
Oh, just send all your cameras to one big GPU. | |
28:46.280 --> 28:48.240 | |
You're much better off doing that. | |
28:48.240 --> 28:50.160 | |
Oh, sorry, you said behind every camera. | |
28:50.160 --> 28:51.000 | |
Every camera. | |
28:51.000 --> 28:51.840 | |
Every small GPU. | |
28:51.840 --> 28:52.720 | |
I was like, oh, I'll put the first few layers | |
28:52.720 --> 28:54.520 | |
of my comms there. | |
28:54.520 --> 28:56.080 | |
Like why did I say that? | |
28:56.080 --> 28:56.920 | |
That's possible. | |
28:56.920 --> 28:59.000 | |
It's possible, but it's a bad idea. | |
28:59.000 --> 29:00.480 | |
It's not obviously a bad idea. | |
29:00.480 --> 29:01.320 | |
Pretty obviously bad. | |
29:01.320 --> 29:02.960 | |
But whether it's actually a bad idea or not, | |
29:02.960 --> 29:05.240 | |
I left that meeting with Elon, like beating myself up. | |
29:05.240 --> 29:07.080 | |
I'm like, why did I say something stupid? | |
29:07.080 --> 29:09.360 | |
Yeah, you haven't, like you haven't at least | |
29:09.360 --> 29:12.240 | |
like thought through every aspect fully. | |
29:12.240 --> 29:13.200 | |
He's very sharp too. | |
29:13.200 --> 29:15.760 | |
Like usually in life, I get away with saying stupid things | |
29:15.760 --> 29:16.960 | |
and then kind of course, | |
29:16.960 --> 29:18.560 | |
right away he called me out about it. | |
29:18.560 --> 29:19.800 | |
And like, usually in life, | |
29:19.800 --> 29:21.120 | |
I get away with saying stupid things. | |
29:21.120 --> 29:24.640 | |
And then like people will, you know, | |
29:24.640 --> 29:26.080 | |
a lot of times people don't even notice. | |
29:26.080 --> 29:28.200 | |
And I'll like correct it and bring the conversation back. | |
29:28.200 --> 29:30.600 | |
But with Elon, it was like, nope, like, okay. | |
29:30.600 --> 29:33.520 | |
Well, that's not at all why the contract fell through. | |
29:33.520 --> 29:35.520 | |
I was much more prepared the second time I met him. | |
29:35.520 --> 29:36.360 | |
Yeah. | |
29:36.360 --> 29:39.640 | |
But in general, how hard did you think it, | |
29:39.640 --> 29:43.680 | |
like 12 months is a tough timeline? | |
29:43.680 --> 29:45.720 | |
Oh, I just thought I'd clone Mobileye IQ three. | |
29:45.720 --> 29:47.560 | |
I didn't think I'd solve level five self driving | |
29:47.560 --> 29:48.400 | |
or anything. | |
29:48.400 --> 29:51.000 | |
So the goal there was to do lane keeping, | |
29:51.000 --> 29:52.840 | |
good lane keeping. | |
29:52.840 --> 29:55.560 | |
I saw my friend showed me the outputs from Mobileye. | |
29:55.560 --> 29:57.680 | |
And the outputs from Mobileye was just basically two lanes | |
29:57.680 --> 29:59.440 | |
and a position of a lead car. | |
29:59.440 --> 30:01.560 | |
I'm like, I can gather a data set | |
30:01.560 --> 30:03.440 | |
and train this net in weeks. | |
30:03.440 --> 30:04.840 | |
And I did. | |
30:04.840 --> 30:07.600 | |
Well, first time I tried the implementation of Mobileye | |
30:07.600 --> 30:11.240 | |
in a Tesla, I was really surprised how good it is. | |
30:11.240 --> 30:12.320 | |
It's quite incredibly good. | |
30:12.320 --> 30:14.080 | |
Cause I thought it's just cause I've done | |
30:14.080 --> 30:14.920 | |
a lot of computer vision. | |
30:14.920 --> 30:18.880 | |
I thought it'd be a lot harder to create a system | |
30:18.880 --> 30:20.040 | |
that that's stable. | |
30:21.000 --> 30:22.440 | |
So I was personally surprised. | |
30:22.440 --> 30:25.000 | |
Just, you know, have to admit it. | |
30:25.000 --> 30:27.840 | |
Cause I was kind of skeptical before trying it. | |
30:27.840 --> 30:31.200 | |
Cause I thought it would go in and out a lot more. | |
30:31.200 --> 30:33.160 | |
It would get disengaged a lot more. | |
30:33.160 --> 30:35.000 | |
And it's pretty robust. | |
30:36.200 --> 30:39.720 | |
So what, how, how, how hard is the problem | |
30:39.720 --> 30:42.080 | |
when you, when you tackled it? | |
30:42.080 --> 30:45.760 | |
So I think AP one was great. Like Elon talked | |
30:45.760 --> 30:49.040 | |
about disengagements on the 405 down in LA | |
30:49.040 --> 30:51.040 | |
with like the lane marks were kind of faded | |
30:51.040 --> 30:52.960 | |
and the Mobileye system would drop out. | |
30:53.960 --> 30:57.240 | |
Like I had something up and working | |
30:57.240 --> 31:01.440 | |
that I would say was like the same quality in three months. | |
31:02.480 --> 31:04.560 | |
Same quality, but how do you know? | |
31:04.560 --> 31:07.400 | |
You say stuff like that confidently, but you can't, | |
31:07.400 --> 31:12.120 | |
and I love it, but the question is you can't, | |
31:12.120 --> 31:13.880 | |
you're kind of going by feel cause you just, | |
31:13.880 --> 31:15.560 | |
You're going by feel, absolutely, absolutely. | |
31:15.560 --> 31:17.280 | |
Like, like I would take, I hadn't, | |
31:17.280 --> 31:18.480 | |
I borrowed my friend's Tesla. | |
31:18.480 --> 31:20.760 | |
I would take AP one out for a drive. | |
31:20.760 --> 31:22.320 | |
And then I would take my system out for a drive. | |
31:22.320 --> 31:24.440 | |
And seems reasonably like the same. | |
31:26.080 --> 31:30.480 | |
So the 405, how hard is it to create something | |
31:30.480 --> 31:34.200 | |
that could actually be a product that's deployed? | |
31:34.200 --> 31:39.200 | |
I mean, I've read an article where Elon, this respond, | |
31:39.520 --> 31:41.880 | |
it said something about you saying that | |
31:41.880 --> 31:46.880 | |
to build autopilot is more complicated | |
31:47.080 --> 31:51.880 | |
than a single George Hodds level job. | |
31:51.880 --> 31:55.520 | |
How hard is that job to create something | |
31:55.520 --> 31:57.480 | |
that would work across the globally? | |
31:58.960 --> 32:00.640 | |
Why don't the global is the challenge, | |
32:00.640 --> 32:02.240 | |
but Elon followed that up by saying | |
32:02.240 --> 32:04.920 | |
it's going to take two years and a company of 10 people. | |
32:04.920 --> 32:07.920 | |
And here I am four years later with a company of 12 people. | |
32:07.920 --> 32:09.960 | |
And I think we still have another two to go. | |
32:09.960 --> 32:10.800 | |
Two years. | |
32:10.800 --> 32:13.120 | |
So yeah, so what do you think, | |
32:13.120 --> 32:15.960 | |
what do you think about how Tesla's progressing | |
32:15.960 --> 32:19.200 | |
with autopilot of V2, V3? | |
32:19.200 --> 32:23.120 | |
I think we've kept pace with them pretty well. | |
32:24.080 --> 32:26.880 | |
I think navigating autopilot is terrible. | |
32:26.880 --> 32:31.120 | |
We had some demo features internally of the same stuff | |
32:31.120 --> 32:32.720 | |
and we would test it and I'm like, | |
32:32.720 --> 32:34.720 | |
I'm not shipping this even as like open source software | |
32:34.720 --> 32:35.560 | |
to people. | |
32:35.560 --> 32:37.400 | |
What do you think is terrible? | |
32:37.400 --> 32:39.600 | |
Consumer Reports does a great job of describing it. | |
32:39.600 --> 32:41.240 | |
Like when it makes a lane change, | |
32:41.240 --> 32:43.600 | |
it does it worse than a human. | |
32:43.600 --> 32:46.960 | |
You shouldn't ship things like autopilot, open pilot, | |
32:46.960 --> 32:49.760 | |
they lane keep better than a human. | |
32:49.760 --> 32:53.440 | |
If you turn it on for a stretch of highway, | |
32:53.440 --> 32:56.680 | |
like an hour long, it's never going to touch a lane line. | |
32:56.680 --> 32:59.040 | |
Human will touch probably a lane line twice. | |
32:59.040 --> 33:00.080 | |
You just inspired me. | |
33:00.080 --> 33:02.200 | |
I don't know if you're grounded in data on that. | |
33:02.200 --> 33:03.280 | |
I read your paper. | |
33:03.280 --> 33:05.400 | |
Okay, but no, but that's interesting. | |
33:06.720 --> 33:09.840 | |
I wonder actually how often we touch lane lines | |
33:11.200 --> 33:13.400 | |
a little bit because it is. | |
33:13.400 --> 33:14.960 | |
I could answer that question pretty easily | |
33:14.960 --> 33:15.800 | |
with the common data side. | |
33:15.800 --> 33:16.920 | |
Yeah, I'm curious. | |
33:16.920 --> 33:17.760 | |
I've never answered it. | |
33:17.760 --> 33:18.600 | |
I don't know. | |
33:18.600 --> 33:20.000 | |
I just too was like my personal. | |
33:20.000 --> 33:22.400 | |
It feels right, but that's interesting | |
33:22.400 --> 33:23.800 | |
because every time you touch a lane, | |
33:23.800 --> 33:26.760 | |
that's a source of a little bit of stress | |
33:26.760 --> 33:29.320 | |
and kind of lane keeping is removing that stress. | |
33:29.320 --> 33:31.840 | |
That's ultimately the biggest value add | |
33:31.840 --> 33:34.240 | |
honestly is just removing the stress | |
33:34.240 --> 33:35.480 | |
of having to stay in lane. | |
33:35.480 --> 33:39.040 | |
And I think I don't think people fully realize | |
33:39.040 --> 33:41.960 | |
first of all that that's a big value add, | |
33:41.960 --> 33:45.000 | |
but also that that's all it is. | |
33:45.000 --> 33:48.560 | |
And that not only I find it a huge value add. | |
33:48.560 --> 33:50.440 | |
I drove down when we moved to San Diego, | |
33:50.440 --> 33:52.640 | |
I drove down in an enterprise rental car | |
33:52.640 --> 33:53.480 | |
and I missed it. | |
33:53.480 --> 33:55.480 | |
So I missed having the system so much. | |
33:55.480 --> 33:59.200 | |
It's so much more tiring to drive | |
33:59.200 --> 34:00.320 | |
without it. | |
34:00.320 --> 34:02.960 | |
It's, it is that lane centering. | |
34:02.960 --> 34:04.840 | |
That's the key feature. | |
34:04.840 --> 34:05.680 | |
Yeah. | |
34:06.600 --> 34:08.960 | |
And in a way it's the only feature | |
34:08.960 --> 34:11.040 | |
that actually adds value to people's lives | |
34:11.040 --> 34:12.200 | |
in autonomous vehicles today. | |
34:12.200 --> 34:13.840 | |
Waymo does not add value to people's lives. | |
34:13.840 --> 34:15.880 | |
It's a more expensive, slower Uber. | |
34:15.880 --> 34:18.640 | |
Maybe someday it'll be this big cliff where it adds value, | |
34:18.640 --> 34:19.480 | |
but I don't usually. | |
34:19.480 --> 34:20.320 | |
It's fascinating. | |
34:20.320 --> 34:22.560 | |
I haven't talked to, this is good. | |
34:22.560 --> 34:25.840 | |
Cause I haven't, I have intuitively, | |
34:25.840 --> 34:28.320 | |
but I think we're making it explicit now. | |
34:28.320 --> 34:33.320 | |
I actually believe that really good lane keeping | |
34:35.480 --> 34:37.240 | |
is a reason to buy a car. | |
34:37.240 --> 34:38.440 | |
Will be a reason to buy a car. | |
34:38.440 --> 34:39.720 | |
It is a huge value add. | |
34:39.720 --> 34:41.760 | |
I've never, until we just started talking about it, | |
34:41.760 --> 34:43.880 | |
haven't really quite realized it, | |
34:43.880 --> 34:48.880 | |
that I've felt with Elon's chase of level four | |
34:49.440 --> 34:52.360 | |
is not the correct chase. | |
34:52.360 --> 34:56.000 | |
It was on, cause you should just say Tesla has the best | |
34:56.000 --> 34:58.320 | |
as if from a Tesla perspective say, | |
34:58.320 --> 35:00.600 | |
Tesla has the best lane keeping. | |
35:00.600 --> 35:04.160 | |
Kama AI should say Kama AI is the best lane keeping. | |
35:04.160 --> 35:05.640 | |
And that is it. | |
35:05.640 --> 35:06.480 | |
Yeah. | |
35:06.480 --> 35:07.320 | |
Yeah. | |
35:07.320 --> 35:08.160 | |
Do you think? | |
35:08.160 --> 35:09.920 | |
You have to do the longitudinal as well. | |
35:09.920 --> 35:10.960 | |
You can't just lane keep. | |
35:10.960 --> 35:12.920 | |
You have to do ACC, | |
35:12.920 --> 35:15.840 | |
but ACC is much more forgiving than lane keep, | |
35:15.840 --> 35:17.400 | |
especially on the highway. | |
35:17.400 --> 35:22.000 | |
By the way, are you Kama AI's camera only, correct? | |
35:22.000 --> 35:23.440 | |
No, we use the radar. | |
35:23.440 --> 35:26.960 | |
We, from the car, you're able to get to, okay. | |
35:26.960 --> 35:28.800 | |
We can do it camera only now. | |
35:28.800 --> 35:29.640 | |
It's gotten to the point, | |
35:29.640 --> 35:31.600 | |
but we leave the radar there as like a, | |
35:31.600 --> 35:33.440 | |
it's fusion now. | |
35:33.440 --> 35:35.440 | |
Okay, so let's maybe talk through | |
35:35.440 --> 35:37.920 | |
some of the system specs on the hardware. | |
35:37.920 --> 35:42.880 | |
What's the hardware side of what you're providing? | |
35:42.880 --> 35:44.720 | |
What's the capabilities on the software side | |
35:44.720 --> 35:46.800 | |
with OpenPilot and so on? | |
35:46.800 --> 35:51.800 | |
So OpenPilot as the box that we sell that it runs on, | |
35:51.800 --> 35:53.920 | |
it's a phone in a plastic case. | |
35:53.920 --> 35:54.840 | |
It's nothing special. | |
35:54.840 --> 35:56.200 | |
We sell it without the software. | |
35:56.200 --> 35:57.840 | |
So you're like, you know, you buy the phone, | |
35:57.840 --> 35:58.920 | |
it's just easy. | |
35:58.920 --> 36:00.240 | |
It'll be easy set up, | |
36:00.240 --> 36:01.720 | |
but it's sold with no software. | |
36:03.480 --> 36:06.600 | |
OpenPilot right now is about to be 0.6. | |
36:06.600 --> 36:07.880 | |
When it gets to 1.0, | |
36:07.880 --> 36:09.680 | |
I think we'll be ready for a consumer product. | |
36:09.680 --> 36:11.120 | |
We're not gonna add any new features. | |
36:11.120 --> 36:13.800 | |
We're just gonna make the lane keeping really, really good. | |
36:13.800 --> 36:15.120 | |
Okay, I got it. | |
36:15.120 --> 36:16.120 | |
So what do we have right now? | |
36:16.120 --> 36:18.200 | |
It's a Snapdragon 820. | |
36:18.200 --> 36:23.200 | |
It's a Sony IMX 298 forward facing camera, | |
36:23.680 --> 36:24.720 | |
driver monitoring camera. | |
36:24.720 --> 36:26.400 | |
It's just a selfie cam on the phone. | |
36:26.400 --> 36:30.000 | |
And a can transceiver, | |
36:30.000 --> 36:32.320 | |
maybe it's a little thing called pandas. | |
36:32.320 --> 36:35.040 | |
And they talk over USB to the phone | |
36:35.040 --> 36:36.400 | |
and then they have three can buses | |
36:36.400 --> 36:37.560 | |
that they talk to the car. | |
36:38.560 --> 36:40.920 | |
One of those can buses is the radar can bus. | |
36:40.920 --> 36:42.920 | |
One of them is the main car can bus. | |
36:42.920 --> 36:44.920 | |
And the other one is the proxy camera can bus. | |
36:44.920 --> 36:47.320 | |
We leave the existing camera in place. | |
36:47.320 --> 36:49.560 | |
So we don't turn AEB off. | |
36:49.560 --> 36:51.040 | |
Right now we still turn AEB off | |
36:51.040 --> 36:52.280 | |
if you're using our longitudinal, | |
36:52.280 --> 36:54.320 | |
but we're gonna fix that before 1.0. | |
36:54.320 --> 36:55.160 | |
Got it. | |
36:55.160 --> 36:56.000 | |
Wow, that's cool. | |
36:56.000 --> 36:57.960 | |
So in its can both ways. | |
36:57.960 --> 37:02.120 | |
So how are you able to control vehicles? | |
37:02.120 --> 37:05.520 | |
So we proxy the vehicles that we work with | |
37:05.520 --> 37:08.960 | |
already have a lane keeping assist system. | |
37:08.960 --> 37:12.520 | |
So lane keeping assist can mean a huge variety of things. | |
37:12.520 --> 37:16.120 | |
It can mean it will apply a small torque | |
37:16.120 --> 37:18.920 | |
to the wheel after you've already crossed a lane line | |
37:18.920 --> 37:22.720 | |
by a foot, which is the system in the older Toyotas. | |
37:22.720 --> 37:26.360 | |
Versus like, I think Tesla still calls it lane keeping assist | |
37:26.360 --> 37:28.920 | |
where it'll keep you perfectly in the center of the lane | |
37:28.920 --> 37:29.960 | |
on the highway. | |
37:31.240 --> 37:34.000 | |
You can control like you with the joystick, the cars. | |
37:34.000 --> 37:36.600 | |
So these cars already have the capability of drive by wire. | |
37:36.600 --> 37:41.600 | |
So is it, is it trivial to convert a car | |
37:41.600 --> 37:43.320 | |
that it operates with? | |
37:43.320 --> 37:47.480 | |
It open pilot is able to control the steering. | |
37:48.480 --> 37:49.720 | |
Oh, a new car or a car that we, | |
37:49.720 --> 37:52.800 | |
so we have support now for 45 different makes of cars. | |
37:52.800 --> 37:54.880 | |
What are the cars in general? | |
37:54.880 --> 37:56.360 | |
Mostly Honda's and Toyotas. | |
37:56.360 --> 38:00.640 | |
We support almost every Honda and Toyota made this year. | |
38:01.680 --> 38:04.480 | |
And then bunch of GM's, bunch of Subaru's. | |
38:04.480 --> 38:05.960 | |
But it doesn't have to be like a Prius. | |
38:05.960 --> 38:07.320 | |
It could be Corolla as well. | |
38:07.320 --> 38:10.760 | |
Oh, the 2020 Corolla is the best car with open pilot. | |
38:10.760 --> 38:11.720 | |
It just came out there. | |
38:11.720 --> 38:14.200 | |
The actuator has less lag than the older Corolla. | |
38:15.840 --> 38:18.240 | |
I think I started watching a video with you. | |
38:18.240 --> 38:21.480 | |
I mean, the way you make videos is awesome. | |
38:21.480 --> 38:24.320 | |
It's just literally at the dealership streaming. | |
38:25.320 --> 38:26.160 | |
I had my friend to follow him. | |
38:26.160 --> 38:27.560 | |
I probably want to stream for an hour. | |
38:27.560 --> 38:31.120 | |
Yeah, and basically like if stuff goes a little wrong, | |
38:31.120 --> 38:33.160 | |
you just like, you just go with it. | |
38:33.160 --> 38:34.000 | |
Yeah, I love it. | |
38:34.000 --> 38:34.840 | |
It's real. | |
38:34.840 --> 38:35.680 | |
Yeah, it's real. | |
38:35.680 --> 38:42.000 | |
That's so beautiful and it's so in contrast to the way | |
38:42.000 --> 38:44.600 | |
other companies would put together a video like that. | |
38:44.600 --> 38:46.000 | |
Kind of why I like to do it like that. | |
38:46.000 --> 38:46.840 | |
Good. | |
38:46.840 --> 38:49.720 | |
I mean, if you become super rich one day and successful, | |
38:49.720 --> 38:52.280 | |
I hope you keep it that way because I think that's actually | |
38:52.280 --> 38:54.600 | |
what people love, that kind of genuine. | |
38:54.600 --> 38:56.520 | |
Oh, it's all that has value to me. | |
38:56.520 --> 38:59.840 | |
Money has no, if I sell out to like make money, | |
38:59.840 --> 39:00.680 | |
I sold out. | |
39:00.680 --> 39:01.520 | |
It doesn't matter. | |
39:01.520 --> 39:02.360 | |
What do I get? | |
39:02.360 --> 39:04.440 | |
Yacht, I don't want a yacht. | |
39:04.440 --> 39:09.440 | |
And I think Tesla actually has a small inkling of that | |
39:09.440 --> 39:11.240 | |
as well with autonomy day. | |
39:11.240 --> 39:14.000 | |
They did reveal more than, I mean, of course, | |
39:14.000 --> 39:15.680 | |
there's marketing communications, you could tell, | |
39:15.680 --> 39:17.640 | |
but it's more than most companies would reveal, | |
39:17.640 --> 39:20.960 | |
which is I hope they go towards that direction | |
39:20.960 --> 39:23.000 | |
more other companies, GM, Ford. | |
39:23.000 --> 39:25.400 | |
Oh, Tesla's going to win level five. | |
39:25.400 --> 39:26.560 | |
They really are. | |
39:26.560 --> 39:27.800 | |
So let's talk about it. | |
39:27.800 --> 39:33.000 | |
You think, you're focused on level two currently, currently. | |
39:33.000 --> 39:36.160 | |
We're going to be one to two years behind Tesla | |
39:36.160 --> 39:37.160 | |
getting to level five. | |
39:37.160 --> 39:38.520 | |
OK. | |
39:38.520 --> 39:39.320 | |
We're Android, right? | |
39:39.320 --> 39:39.880 | |
We're Android. | |
39:39.880 --> 39:40.680 | |
You're Android. | |
39:40.680 --> 39:42.240 | |
I'm just saying once Tesla gets it, | |
39:42.240 --> 39:43.440 | |
we're one to two years behind. | |
39:43.440 --> 39:45.680 | |
I'm not making any timeline on when Tesla's going to get it. | |
39:45.680 --> 39:46.120 | |
That's right. | |
39:46.120 --> 39:46.360 | |
You did. | |
39:46.360 --> 39:46.960 | |
That's brilliant. | |
39:46.960 --> 39:48.560 | |
I'm sorry, Tesla investors, if you | |
39:48.560 --> 39:50.520 | |
think you're going to have an autonomous robot taxi | |
39:50.520 --> 39:54.920 | |
fleet by the end of the year, I'll bet against that. | |
39:54.920 --> 39:57.720 | |
So what do you think about this? | |
39:57.720 --> 40:03.280 | |
The most level four companies are kind of just | |
40:03.280 --> 40:08.360 | |
doing their usual safety driver, doing full autonomy kind | |
40:08.360 --> 40:08.800 | |
of testing. | |
40:08.800 --> 40:10.880 | |
And then Tesla does basically trying | |
40:10.880 --> 40:15.280 | |
to go from lane keeping to full autonomy. | |
40:15.280 --> 40:16.840 | |
What do you think about that approach? | |
40:16.840 --> 40:18.360 | |
How successful would it be? | |
40:18.360 --> 40:20.680 | |
It's a ton better approach. | |
40:20.680 --> 40:23.960 | |
Because Tesla is gathering data on a scale | |
40:23.960 --> 40:25.200 | |
that none of them are. | |
40:25.200 --> 40:29.560 | |
They're putting real users behind the wheel of the cars. | |
40:29.560 --> 40:34.440 | |
It's, I think, the only strategy that works, the incremental. | |
40:34.440 --> 40:37.000 | |
Well, so there's a few components to Tesla approach | |
40:37.000 --> 40:38.800 | |
that's more than just the incremental. | |
40:38.800 --> 40:41.400 | |
What you spoke with is the software, | |
40:41.400 --> 40:43.720 | |
so over the air software updates. | |
40:43.720 --> 40:44.800 | |
Necessity. | |
40:44.800 --> 40:46.440 | |
I mean, Waymo crews have those too. | |
40:46.440 --> 40:47.560 | |
Those aren't. | |
40:47.560 --> 40:48.080 | |
But no. | |
40:48.080 --> 40:49.800 | |
Those differentiate from the automakers. | |
40:49.800 --> 40:50.080 | |
Right. | |
40:50.080 --> 40:53.440 | |
No lane keeping systems have no cars with lane keeping system | |
40:53.440 --> 40:54.760 | |
have that except Tesla. | |
40:54.760 --> 40:55.720 | |
Yeah. | |
40:55.720 --> 40:59.760 | |
And the other one is the data, the other direction, | |
40:59.760 --> 41:01.840 | |
which is the ability to query the data. | |
41:01.840 --> 41:03.480 | |
I don't think they're actually collecting | |
41:03.480 --> 41:05.240 | |
as much data as people think, but the ability | |
41:05.240 --> 41:09.440 | |
to turn on collection and turn it off. | |
41:09.440 --> 41:13.400 | |
So I'm both in the robotics world, in the psychology, | |
41:13.400 --> 41:15.000 | |
human factors world. | |
41:15.000 --> 41:17.320 | |
Many people believe that level two autonomy | |
41:17.320 --> 41:20.040 | |
is problematic because of the human factor. | |
41:20.040 --> 41:23.280 | |
Like the more the task is automated, | |
41:23.280 --> 41:25.960 | |
the more there's a vigilance decrement. | |
41:25.960 --> 41:27.200 | |
You start to fall asleep. | |
41:27.200 --> 41:30.480 | |
You start to become complacent, start texting more and so on. | |
41:30.480 --> 41:32.200 | |
Do you worry about that? | |
41:32.200 --> 41:35.000 | |
Because if you're talking about transition from lane keeping | |
41:35.000 --> 41:40.960 | |
to full autonomy, if you're spending 80% of the time | |
41:40.960 --> 41:43.080 | |
not supervising the machine, do you | |
41:43.080 --> 41:47.080 | |
worry about what that means for the safety of the drivers? | |
41:47.080 --> 41:49.640 | |
One, we don't consider OpenPilot to be 1.0 | |
41:49.640 --> 41:52.880 | |
until we have 100% driver monitoring. | |
41:52.880 --> 41:55.000 | |
You can cheat right now, our driver monitoring system. | |
41:55.000 --> 41:56.080 | |
There's a few ways to cheat it. | |
41:56.080 --> 41:58.160 | |
They're pretty obvious. | |
41:58.160 --> 41:59.680 | |
We're working on making that better. | |
41:59.680 --> 42:02.520 | |
Before we ship a consumer product that can drive cars, | |
42:02.520 --> 42:04.240 | |
I want to make sure that I have driver monitoring | |
42:04.240 --> 42:05.440 | |
that you can't cheat. | |
42:05.440 --> 42:09.000 | |
What's a successful driver monitoring system look like? | |
42:09.000 --> 42:11.680 | |
Is it all about just keeping your eyes on the road? | |
42:11.680 --> 42:12.760 | |
Well, a few things. | |
42:12.760 --> 42:16.600 | |
So that's what we went with at first for driver monitoring. | |
42:16.600 --> 42:17.160 | |
I'm checking. | |
42:17.160 --> 42:19.000 | |
I'm actually looking at where your head is looking. | |
42:19.000 --> 42:19.880 | |
The camera's not that high. | |
42:19.880 --> 42:21.840 | |
Resolution eyes are a little bit hard to get. | |
42:21.840 --> 42:22.880 | |
Well, head is big. | |
42:22.880 --> 42:23.560 | |
I mean, that's just. | |
42:23.560 --> 42:24.640 | |
Head is good. | |
42:24.640 --> 42:28.720 | |
And actually, a lot of it, just psychology wise, | |
42:28.720 --> 42:30.720 | |
to have that monitor constantly there, | |
42:30.720 --> 42:33.400 | |
it reminds you that you have to be paying attention. | |
42:33.400 --> 42:35.080 | |
But we want to go further. | |
42:35.080 --> 42:36.760 | |
We just hired someone full time to come on | |
42:36.760 --> 42:37.960 | |
to do the driver monitoring. | |
42:37.960 --> 42:40.600 | |
I want to detect phone in frame, and I | |
42:40.600 --> 42:42.600 | |
want to make sure you're not sleeping. | |
42:42.600 --> 42:44.880 | |
How much does the camera see of the body? | |
42:44.880 --> 42:47.480 | |
This one, not enough. | |
42:47.480 --> 42:48.400 | |
Not enough. | |
42:48.400 --> 42:50.720 | |
The next one, everything. | |
42:50.720 --> 42:52.920 | |
What's interesting, FishEye, is we're | |
42:52.920 --> 42:55.200 | |
doing just data collection, not real time. | |
42:55.200 --> 42:59.200 | |
But FishEye is a beautiful being able to capture the body. | |
42:59.200 --> 43:03.280 | |
And the smartphone is really the biggest problem. | |
43:03.280 --> 43:03.880 | |
I'll show you. | |
43:03.880 --> 43:07.800 | |
I can show you one of the pictures from our new system. | |
43:07.800 --> 43:08.160 | |
Awesome. | |
43:08.160 --> 43:10.520 | |
So you're basically saying the driver monitoring | |
43:10.520 --> 43:13.080 | |
will be the answer to that. | |
43:13.080 --> 43:15.320 | |
I think the other point that you raised in your paper | |
43:15.320 --> 43:16.920 | |
is good as well. | |
43:16.920 --> 43:20.400 | |
You're not asking a human to supervise a machine | |
43:20.400 --> 43:23.920 | |
without giving them the they can take over at any time. | |
43:23.920 --> 43:25.760 | |
Our safety model, you can take over. | |
43:25.760 --> 43:27.720 | |
We disengage on both the gas or the brake. | |
43:27.720 --> 43:28.880 | |
We don't disengage on steering. | |
43:28.880 --> 43:29.920 | |
I don't feel you have to. | |
43:29.920 --> 43:31.720 | |
But we disengage on gas or brake. | |
43:31.720 --> 43:34.240 | |
So it's very easy for you to take over. | |
43:34.240 --> 43:36.400 | |
And it's very easy for you to reengage. | |
43:36.400 --> 43:39.320 | |
That switching should be super cheap. | |
43:39.320 --> 43:40.800 | |
The cars that require, even autopilot, | |
43:40.800 --> 43:42.400 | |
requires a double press. | |
43:42.400 --> 43:44.360 | |
That's almost, I see, I don't like that. | |
43:44.360 --> 43:46.440 | |
And then the cancel. | |
43:46.440 --> 43:48.320 | |
To cancel in autopilot, you either | |
43:48.320 --> 43:49.920 | |
have to press cancel, which no one knows where that is. | |
43:49.920 --> 43:51.000 | |
So they press the brake. | |
43:51.000 --> 43:53.360 | |
But a lot of times you don't want to press the brake. | |
43:53.360 --> 43:54.560 | |
You want to press the gas. | |
43:54.560 --> 43:56.880 | |
So you should cancel on gas or wiggle the steering wheel, | |
43:56.880 --> 43:57.960 | |
which is bad as well. | |
43:57.960 --> 43:58.920 | |
Wow, that's brilliant. | |
43:58.920 --> 44:01.440 | |
I haven't heard anyone articulate that point. | |
44:01.440 --> 44:04.960 | |
Oh, there's a lot I think about. | |
44:04.960 --> 44:09.800 | |
Because I think actually Tesla has done a better job | |
44:09.800 --> 44:12.920 | |
than most automakers at making that frictionless. | |
44:12.920 --> 44:16.600 | |
But you just described that it could be even better. | |
44:16.600 --> 44:19.320 | |
I love Super Cruise as an experience. | |
44:19.320 --> 44:21.120 | |
Once it's engaged. | |
44:21.120 --> 44:22.800 | |
I don't know if you've used it, but getting the thing | |
44:22.800 --> 44:25.040 | |
to try to engage. | |
44:25.040 --> 44:27.480 | |
Yeah, I've used the driven Super Cruise a lot. | |
44:27.480 --> 44:29.680 | |
So what's your thoughts on the Super Cruise system in general? | |
44:29.680 --> 44:32.640 | |
You disengage Super Cruise, and it falls back to ACC. | |
44:32.640 --> 44:34.600 | |
So my car is still accelerating. | |
44:34.600 --> 44:36.280 | |
It feels weird. | |
44:36.280 --> 44:39.000 | |
Otherwise, when you actually have Super Cruise engaged | |
44:39.000 --> 44:41.200 | |
on the highway, it is phenomenal. | |
44:41.200 --> 44:42.320 | |
We bought that Cadillac. | |
44:42.320 --> 44:43.240 | |
We just sold it. | |
44:43.240 --> 44:45.600 | |
But we bought it just to experience this. | |
44:45.600 --> 44:47.440 | |
And I wanted everyone in the office to be like, | |
44:47.440 --> 44:49.360 | |
this is what we're striving to build. | |
44:49.360 --> 44:52.800 | |
GM pioneering with the driver monitoring. | |
44:52.800 --> 44:55.040 | |
You like their driver monitoring system? | |
44:55.040 --> 44:56.440 | |
It has some bugs. | |
44:56.440 --> 45:01.960 | |
If there's a sun shining back here, it'll be blind to you. | |
45:01.960 --> 45:03.360 | |
But overall, mostly, yeah. | |
45:03.360 --> 45:05.960 | |
That's so cool that you know all this stuff. | |
45:05.960 --> 45:09.960 | |
I don't often talk to people that because it's such a rare car, | |
45:09.960 --> 45:10.960 | |
unfortunately, currently. | |
45:10.960 --> 45:12.760 | |
We bought one explicitly for that. | |
45:12.760 --> 45:15.040 | |
We lost like $25K in the deprecation, | |
45:15.040 --> 45:16.720 | |
but it feels worth it. | |
45:16.720 --> 45:21.280 | |
I was very pleasantly surprised that our GM system | |
45:21.280 --> 45:26.320 | |
was so innovative and really wasn't advertised much, | |
45:26.320 --> 45:28.480 | |
wasn't talked about much. | |
45:28.480 --> 45:31.840 | |
And I was nervous that it would die, that it would disappear. | |
45:31.840 --> 45:33.520 | |
Well, they put it on the wrong car. | |
45:33.520 --> 45:35.680 | |
They should have put it on the bolt and not some weird Cadillac | |
45:35.680 --> 45:36.640 | |
that nobody bought. | |
45:36.640 --> 45:39.520 | |
I think that's going to be into, they're saying at least | |
45:39.520 --> 45:41.840 | |
it's going to be into their entire fleet. | |
45:41.840 --> 45:44.320 | |
So what do you think about, as long as we're | |
45:44.320 --> 45:46.920 | |
on the driver monitoring, what do you think | |
45:46.920 --> 45:51.920 | |
about Elon Musk's claim that driver monitoring is not needed? | |
45:51.920 --> 45:53.680 | |
Normally, I love his claims. | |
45:53.680 --> 45:55.560 | |
That one is stupid. | |
45:55.560 --> 45:56.560 | |
That one is stupid. | |
45:56.560 --> 46:00.320 | |
And he's not going to have his level five fleet | |
46:00.320 --> 46:01.320 | |
by the end of the year. | |
46:01.320 --> 46:04.880 | |
Hopefully, he's like, OK, I was wrong. | |
46:04.880 --> 46:06.280 | |
I'm going to add driver monitoring. | |
46:06.280 --> 46:08.240 | |
Because when these systems get to the point | |
46:08.240 --> 46:10.320 | |
that they're only messing up once every 1,000 miles, | |
46:10.320 --> 46:14.080 | |
you absolutely need driver monitoring. | |
46:14.080 --> 46:15.880 | |
So let me play, because I agree with you, | |
46:15.880 --> 46:17.320 | |
but let me play devil's advocate. | |
46:17.320 --> 46:22.440 | |
One possibility is that without driver monitoring, | |
46:22.440 --> 46:29.400 | |
people are able to self regulate, monitor themselves. | |
46:29.400 --> 46:30.680 | |
So your idea is, I'm just. | |
46:30.680 --> 46:34.160 | |
You're seeing all the people sleeping in Teslas? | |
46:34.160 --> 46:35.280 | |
Yeah. | |
46:35.280 --> 46:38.320 | |
Well, I'm a little skeptical of all the people sleeping | |
46:38.320 --> 46:43.960 | |
in Teslas because I've stopped paying attention to that kind | |
46:43.960 --> 46:45.680 | |
of stuff because I want to see real data. | |
46:45.680 --> 46:47.240 | |
It's too much glorified. | |
46:47.240 --> 46:48.720 | |
It doesn't feel scientific to me. | |
46:48.720 --> 46:52.560 | |
So I want to know how many people are really sleeping | |
46:52.560 --> 46:55.080 | |
in Teslas versus sleeping. | |
46:55.080 --> 46:57.640 | |
I was driving here, sleep deprived, | |
46:57.640 --> 46:59.520 | |
in a car with no automation. | |
46:59.520 --> 47:01.040 | |
I was falling asleep. | |
47:01.040 --> 47:02.120 | |
I agree that it's hypey. | |
47:02.120 --> 47:04.840 | |
It's just like, you know what? | |
47:04.840 --> 47:08.480 | |
If Elon put driver monitoring, my last autopilot experience | |
47:08.480 --> 47:12.200 | |
was I rented a Model 3 in March and drove it around. | |
47:12.200 --> 47:13.640 | |
The wheel thing is annoying. | |
47:13.640 --> 47:15.440 | |
And the reason the wheel thing is annoying. | |
47:15.440 --> 47:17.080 | |
We use the wheel thing as well, but we | |
47:17.080 --> 47:18.720 | |
don't disengage on wheel. | |
47:18.720 --> 47:21.720 | |
For Tesla, you have to touch the wheel just enough | |
47:21.720 --> 47:25.320 | |
to trigger the torque sensor to tell it that you're there, | |
47:25.320 --> 47:29.720 | |
but not enough as to disengage it, which don't use it | |
47:29.720 --> 47:30.440 | |
for two things. | |
47:30.440 --> 47:31.360 | |
Don't disengage on wheel. | |
47:31.360 --> 47:32.400 | |
You don't have to. | |
47:32.400 --> 47:35.360 | |
That whole experience, wow, beautifully put. | |
47:35.360 --> 47:38.360 | |
All those elements, even if you don't have driver monitoring, | |
47:38.360 --> 47:41.080 | |
that whole experience needs to be better. | |
47:41.080 --> 47:43.760 | |
Driver monitoring, I think would make, | |
47:43.760 --> 47:46.200 | |
I mean, I think supercruise is a better experience | |
47:46.200 --> 47:48.440 | |
once it's engaged over autopilot. | |
47:48.440 --> 47:51.600 | |
I think supercruise is a transition to engagement | |
47:51.600 --> 47:55.200 | |
and disengagement are significantly worse. | |
47:55.200 --> 47:57.880 | |
There's a tricky thing, because if I were to criticize | |
47:57.880 --> 48:00.800 | |
supercruise, it's a little too crude. | |
48:00.800 --> 48:03.640 | |
And I think it's like six seconds or something. | |
48:03.640 --> 48:06.080 | |
If you look off road, it'll start warning you. | |
48:06.080 --> 48:09.120 | |
It's some ridiculously long period of time. | |
48:09.120 --> 48:14.120 | |
And just the way, I think it's basically, it's a binary. | |
48:15.840 --> 48:17.440 | |
It should be adapted. | |
48:17.440 --> 48:19.880 | |
Yeah, it needs to learn more about you. | |
48:19.880 --> 48:23.160 | |
It needs to communicate what it sees about you more. | |
48:23.160 --> 48:25.800 | |
I'm not, you know, Tesla shows what it sees | |
48:25.800 --> 48:27.160 | |
about the external world. | |
48:27.160 --> 48:29.120 | |
It would be nice if supercruise would tell us | |
48:29.120 --> 48:30.840 | |
what it sees about the internal world. | |
48:30.840 --> 48:31.960 | |
It's even worse than that. | |
48:31.960 --> 48:33.320 | |
You press the button to engage | |
48:33.320 --> 48:35.480 | |
and it just says supercruise unavailable. | |
48:35.480 --> 48:36.320 | |
Yeah, why? | |
48:36.320 --> 48:37.800 | |
Why? | |
48:37.800 --> 48:41.480 | |
Yeah, that transparency is good. | |
48:41.480 --> 48:43.520 | |
We've renamed the driver monitoring packet | |
48:43.520 --> 48:45.360 | |
to driver state. | |
48:45.360 --> 48:46.280 | |
Driver state. | |
48:46.280 --> 48:48.360 | |
We have car state packet, which has the state of the car | |
48:48.360 --> 48:51.040 | |
and driver state packet, which has state of the driver. | |
48:51.040 --> 48:52.240 | |
So what is it? | |
48:52.240 --> 48:54.080 | |
Estimate their BAC. | |
48:54.080 --> 48:54.920 | |
What's BAC? | |
48:54.920 --> 48:55.920 | |
Blood alcohol, kind of. | |
48:57.360 --> 48:59.240 | |
You think that's possible with computer vision? | |
48:59.240 --> 49:00.080 | |
Absolutely. | |
49:02.560 --> 49:04.520 | |
It's a, to me, it's an open question. | |
49:04.520 --> 49:06.600 | |
I haven't looked into too much. | |
49:06.600 --> 49:08.440 | |
Actually, I quite seriously looked at the literature. | |
49:08.440 --> 49:10.840 | |
It's not obvious to me that from the eyes and so on, | |
49:10.840 --> 49:11.680 | |
you can tell. | |
49:11.680 --> 49:13.440 | |
You might need stuff from the car as well. | |
49:13.440 --> 49:15.760 | |
You might need how they're controlling the car, right? | |
49:15.760 --> 49:17.360 | |
And that's fundamentally at the end of the day | |
49:17.360 --> 49:18.640 | |
what you care about. | |
49:18.640 --> 49:21.640 | |
But I think, especially when people are really drunk, | |
49:21.640 --> 49:23.640 | |
they're not controlling the car nearly as smoothly | |
49:23.640 --> 49:25.160 | |
as they would look at them walking, right? | |
49:25.160 --> 49:27.240 | |
They're, the car is like an extension of the body. | |
49:27.240 --> 49:29.360 | |
So I think you could totally detect. | |
49:29.360 --> 49:30.880 | |
And if you could fix people who are drunk, | |
49:30.880 --> 49:32.840 | |
distracted, asleep, if you fix those three. | |
49:32.840 --> 49:35.480 | |
Yeah, that's a huge, that's huge. | |
49:35.480 --> 49:38.240 | |
So what are the current limitations of OpenPilot? | |
49:38.240 --> 49:41.720 | |
What are the main problems that still need to be solved? | |
49:41.720 --> 49:45.440 | |
We're hopefully fixing a few of them in zero six. | |
49:45.440 --> 49:48.400 | |
We're not as good as autopilot at stop cars. | |
49:49.440 --> 49:54.240 | |
So if you're coming up to a red light at like 55, | |
49:55.200 --> 49:56.880 | |
so it's the radar stopped car problem, | |
49:56.880 --> 49:59.200 | |
which is responsible for two autopilot accidents, | |
49:59.200 --> 50:01.480 | |
it's hard to differentiate a stopped car | |
50:01.480 --> 50:03.640 | |
from a like signpost. | |
50:03.640 --> 50:05.320 | |
Yeah, static object. | |
50:05.320 --> 50:07.520 | |
So you have to fuse, you have to do this visually. | |
50:07.520 --> 50:09.600 | |
There's no way from the radar data to tell the difference. | |
50:09.600 --> 50:10.680 | |
Maybe you can make a map, | |
50:10.680 --> 50:13.840 | |
but I don't really believe in mapping at all anymore. | |
50:13.840 --> 50:14.920 | |
Wait, wait, wait, what? | |
50:14.920 --> 50:16.040 | |
You don't believe in mapping? | |
50:16.040 --> 50:16.880 | |
No. | |
50:16.880 --> 50:21.120 | |
So you're basically, the OpenPilot solution is saying, | |
50:21.120 --> 50:22.480 | |
react to the environment as you see it, | |
50:22.480 --> 50:24.480 | |
just like human doing beings do. | |
50:24.480 --> 50:26.200 | |
And then eventually when you want to do navigate | |
50:26.200 --> 50:30.400 | |
on OpenPilot, I'll train the net to look at ways. | |
50:30.400 --> 50:31.360 | |
I'll run ways in the background, | |
50:31.360 --> 50:32.200 | |
I'll train and come down a way. | |
50:32.200 --> 50:33.560 | |
Are you using GPS at all? | |
50:33.560 --> 50:34.840 | |
We use it to ground truth. | |
50:34.840 --> 50:37.440 | |
We use it to very carefully ground truth the paths. | |
50:37.440 --> 50:39.560 | |
We have a stack which can recover relative | |
50:39.560 --> 50:41.800 | |
to 10 centimeters over one minute. | |
50:41.800 --> 50:43.440 | |
And then we use that to ground truth | |
50:43.440 --> 50:45.880 | |
exactly where the car went in that local part | |
50:45.880 --> 50:47.800 | |
of the environment, but it's all local. | |
50:47.800 --> 50:49.160 | |
How are you testing in general? | |
50:49.160 --> 50:51.400 | |
Just for yourself, like experiments and stuff. | |
50:51.400 --> 50:54.000 | |
Where are you located? | |
50:54.000 --> 50:54.840 | |
San Diego. | |
50:54.840 --> 50:55.680 | |
San Diego. | |
50:55.680 --> 50:56.520 | |
Yeah. | |
50:56.520 --> 50:57.360 | |
Okay. | |
50:57.360 --> 50:59.760 | |
So you basically drive around there, | |
50:59.760 --> 51:02.200 | |
collect some data and watch the performance? | |
51:02.200 --> 51:04.800 | |
We have a simulator now and we have, | |
51:04.800 --> 51:06.440 | |
our simulator is really cool. | |
51:06.440 --> 51:08.120 | |
Our simulator is not, | |
51:08.120 --> 51:09.720 | |
it's not like a Unity based simulator. | |
51:09.720 --> 51:11.840 | |
Our simulator lets us load in real estate. | |
51:12.880 --> 51:13.720 | |
What do you mean? | |
51:13.720 --> 51:16.760 | |
We can load in a drive and simulate | |
51:16.760 --> 51:20.280 | |
what the system would have done on the historical data. | |
51:20.280 --> 51:21.480 | |
Ooh, nice. | |
51:22.520 --> 51:24.360 | |
Interesting. | |
51:24.360 --> 51:26.080 | |
Right now we're only using it for testing, | |
51:26.080 --> 51:28.640 | |
but as soon as we start using it for training. | |
51:28.640 --> 51:29.480 | |
That's it. | |
51:29.480 --> 51:30.840 | |
That's all set up for us. | |
51:30.840 --> 51:33.040 | |
What's your feeling about the real world versus simulation? | |
51:33.040 --> 51:34.320 | |
Do you like simulation for training? | |
51:34.320 --> 51:35.720 | |
If this moves to training? | |
51:35.720 --> 51:40.040 | |
So we have to distinguish two types of simulators, right? | |
51:40.040 --> 51:44.720 | |
There's a simulator that like is completely fake. | |
51:44.720 --> 51:46.720 | |
I could get my car to drive around in GTA. | |
51:47.800 --> 51:51.080 | |
I feel that this kind of simulator is useless. | |
51:51.880 --> 51:53.640 | |
You're never, there's so many. | |
51:54.640 --> 51:57.000 | |
My analogy here is like, okay, fine. | |
51:57.000 --> 51:59.920 | |
You're not solving the computer vision problem, | |
51:59.920 --> 52:02.440 | |
but you're solving the computer graphics problem. | |
52:02.440 --> 52:03.280 | |
Right. | |
52:03.280 --> 52:04.600 | |
And you don't think you can get very far | |
52:04.600 --> 52:08.040 | |
by creating ultra realistic graphics? | |
52:08.040 --> 52:10.360 | |
No, because you can create ultra realistic graphics | |
52:10.360 --> 52:13.160 | |
or the road, now create ultra realistic behavioral models | |
52:13.160 --> 52:14.600 | |
of the other cars. | |
52:14.600 --> 52:16.920 | |
Oh, well, I'll just use myself driving. | |
52:16.920 --> 52:18.280 | |
No, you won't. | |
52:18.280 --> 52:21.640 | |
You need real, you need actual human behavior | |
52:21.640 --> 52:23.320 | |
because that's what you're trying to learn. | |
52:23.320 --> 52:25.840 | |
The driving does not have a spec. | |
52:25.840 --> 52:29.920 | |
The definition of driving is what humans do when they drive. | |
52:29.920 --> 52:32.800 | |
Whatever Waymo does, I don't think it's driving. | |
52:32.800 --> 52:33.640 | |
Right. | |
52:33.640 --> 52:36.400 | |
Well, I think actually Waymo and others, | |
52:36.400 --> 52:38.920 | |
if there's any use for reinforcement learning, | |
52:38.920 --> 52:40.360 | |
I've seen it used quite well. | |
52:40.360 --> 52:41.640 | |
I studied pedestrians a lot too, | |
52:41.640 --> 52:44.360 | |
is try to train models from real data | |
52:44.360 --> 52:46.920 | |
of how pedestrians move and try to use reinforcement learning | |
52:46.920 --> 52:50.040 | |
models to make pedestrians move in human like ways. | |
52:50.040 --> 52:53.520 | |
By that point, you've already gone so many layers, | |
52:53.520 --> 52:55.680 | |
you detected a pedestrian. | |
52:55.680 --> 52:59.640 | |
Did you hand code the feature vector of their state? | |
52:59.640 --> 53:00.480 | |
Right. | |
53:00.480 --> 53:02.880 | |
Did you guys learn anything from computer vision | |
53:02.880 --> 53:04.600 | |
before deep learning? | |
53:04.600 --> 53:07.160 | |
Well, okay, I feel like this is... | |
53:07.160 --> 53:10.840 | |
So perception to you is the sticking point. | |
53:10.840 --> 53:13.760 | |
I mean, what's the hardest part of the stack here? | |
53:13.760 --> 53:18.760 | |
There is no human understandable feature vector | |
53:19.680 --> 53:22.000 | |
separating perception and planning. | |
53:23.040 --> 53:25.120 | |
That's the best way I can put that. | |
53:25.120 --> 53:25.960 | |
There is no... | |
53:25.960 --> 53:29.600 | |
So it's all together and it's a joint problem. | |
53:29.600 --> 53:31.480 | |
So you can take localization. | |
53:31.480 --> 53:32.960 | |
Localization and planning, | |
53:32.960 --> 53:34.760 | |
there is a human understandable feature vector | |
53:34.760 --> 53:36.000 | |
between these two things. | |
53:36.000 --> 53:38.720 | |
I mean, okay, so I have like three degrees position, | |
53:38.720 --> 53:40.560 | |
three degrees orientation and those derivatives, | |
53:40.560 --> 53:42.000 | |
maybe those second derivatives, right? | |
53:42.000 --> 53:44.520 | |
That's human understandable, that's physical. | |
53:44.520 --> 53:48.560 | |
The between perception and planning. | |
53:49.520 --> 53:53.600 | |
So like Waymo has a perception stack and then a planner. | |
53:53.600 --> 53:55.560 | |
And one of the things Waymo does right | |
53:55.560 --> 54:00.000 | |
is they have a simulator that can separate those two. | |
54:00.000 --> 54:02.920 | |
They can like replay their perception data | |
54:02.920 --> 54:03.920 | |
and test their system, | |
54:03.920 --> 54:04.880 | |
which is what I'm talking about | |
54:04.880 --> 54:06.520 | |
about like the two different kinds of simulators. | |
54:06.520 --> 54:08.240 | |
There's the kind that can work on real data | |
54:08.240 --> 54:10.920 | |
and there's the kind that can't work on real data. | |
54:10.920 --> 54:13.880 | |
Now, the problem is that I don't think | |
54:13.880 --> 54:16.160 | |
you can hand code a feature vector, right? | |
54:16.160 --> 54:17.360 | |
Like you have some list of like, | |
54:17.360 --> 54:19.040 | |
well, here's my list of cars in the scenes. | |
54:19.040 --> 54:21.280 | |
Here's my list of pedestrians in the scene. | |
54:21.280 --> 54:23.240 | |
This isn't what humans are doing. | |
54:23.240 --> 54:24.920 | |
What are humans doing? | |
54:24.920 --> 54:25.760 | |
Global. | |
54:27.200 --> 54:28.040 | |
Some, some. | |
54:28.040 --> 54:31.960 | |
You're saying that's too difficult to hand engineer. | |
54:31.960 --> 54:34.120 | |
I'm saying that there is no state vector. | |
54:34.120 --> 54:36.560 | |
Given a perfect, I could give you the best team | |
54:36.560 --> 54:38.520 | |
of engineers in the world to build a perception system | |
54:38.520 --> 54:40.640 | |
and the best team to build a planner. | |
54:40.640 --> 54:42.640 | |
All you have to do is define the state vector | |
54:42.640 --> 54:43.960 | |
that separates those two. | |
54:43.960 --> 54:48.560 | |
I'm missing the state vector that separates those two. | |
54:48.560 --> 54:49.400 | |
What do you mean? | |
54:49.400 --> 54:54.000 | |
So what is the output of your perception system? | |
54:54.000 --> 54:56.880 | |
Output of the perception system. | |
54:56.880 --> 55:01.560 | |
It's, there's, okay, well, there's several ways to do it. | |
55:01.560 --> 55:03.840 | |
One is the slam component is localization. | |
55:03.840 --> 55:05.920 | |
The other is drivable area, drivable space. | |
55:05.920 --> 55:06.760 | |
Drivable space, yep. | |
55:06.760 --> 55:09.000 | |
And then there's the different objects in the scene. | |
55:09.000 --> 55:09.840 | |
Yep. | |
55:11.000 --> 55:16.000 | |
And different objects in the scene over time maybe | |
55:16.000 --> 55:18.720 | |
to give you input to then try to start | |
55:18.720 --> 55:21.560 | |
modeling the trajectories of those objects. | |
55:21.560 --> 55:22.400 | |
Sure. | |
55:22.400 --> 55:23.240 | |
That's it. | |
55:23.240 --> 55:25.160 | |
I can give you a concrete example of something you missed. | |
55:25.160 --> 55:26.000 | |
What's that? | |
55:26.000 --> 55:28.640 | |
So say there's a bush in the scene. | |
55:28.640 --> 55:30.920 | |
Humans understand that when they see this bush | |
55:30.920 --> 55:34.680 | |
that there may or may not be a car behind that bush. | |
55:34.680 --> 55:37.280 | |
Drivable area and a list of objects does not include that. | |
55:37.280 --> 55:38.920 | |
Humans are doing this constantly | |
55:38.920 --> 55:40.920 | |
at the simplest intersections. | |
55:40.920 --> 55:43.880 | |
So now you have to talk about occluded area. | |
55:43.880 --> 55:44.720 | |
Right. | |
55:44.720 --> 55:47.800 | |
Right, but even that, what do you mean by occluded? | |
55:47.800 --> 55:49.640 | |
Okay, so I can't see it. | |
55:49.640 --> 55:51.840 | |
Well, if it's the other side of a house, I don't care. | |
55:51.840 --> 55:53.560 | |
What's the likelihood that there's a car | |
55:53.560 --> 55:55.280 | |
in that occluded area, right? | |
55:55.280 --> 55:58.080 | |
And if you say, okay, we'll add that, | |
55:58.080 --> 56:00.680 | |
I can come up with 10 more examples that you can't add. | |
56:01.680 --> 56:03.960 | |
Certainly occluded area would be something | |
56:03.960 --> 56:06.760 | |
that simulator would have because it's simulating | |
56:06.760 --> 56:11.320 | |
the entire, you know, occlusion is part of it. | |
56:11.320 --> 56:12.680 | |
Occlusion is part of a vision stack. | |
56:12.680 --> 56:13.520 | |
Vision stack. | |
56:13.520 --> 56:16.600 | |
But what I'm saying is if you have a hand engineered, | |
56:16.600 --> 56:20.040 | |
if your perception system output can be written | |
56:20.040 --> 56:22.240 | |
in a spec document, it is incomplete. | |
56:23.120 --> 56:27.800 | |
Yeah, I mean, I certainly, it's hard to argue with that | |
56:27.800 --> 56:30.120 | |
because in the end, that's going to be true. | |
56:30.120 --> 56:31.760 | |
Yeah, and I'll tell you what the output | |
56:31.760 --> 56:32.720 | |
of our perception system is. | |
56:32.720 --> 56:33.560 | |
What's that? | |
56:33.560 --> 56:37.120 | |
It's a 1024 dimensional vector. | |
56:37.120 --> 56:38.000 | |
Transparent neural net. | |
56:38.000 --> 56:39.000 | |
Oh, you know that. | |
56:39.000 --> 56:42.000 | |
No, that's the 1024 dimensions of who knows what. | |
56:43.520 --> 56:45.160 | |
Because it's operating on real data. | |
56:45.160 --> 56:47.000 | |
Yeah. | |
56:47.000 --> 56:48.320 | |
And that's the perception. | |
56:48.320 --> 56:50.360 | |
That's the perception state, right? | |
56:50.360 --> 56:53.520 | |
Think about an autoencoder for faces, right? | |
56:53.520 --> 56:54.720 | |
If you have an autoencoder for faces | |
56:54.720 --> 56:59.720 | |
and you say it has 256 dimensions in the middle, | |
56:59.720 --> 57:00.680 | |
and I'm taking a face over here | |
57:00.680 --> 57:02.800 | |
and projecting it to a face over here. | |
57:02.800 --> 57:05.360 | |
Can you hand label all 256 of those dimensions? | |
57:06.280 --> 57:09.240 | |
Well, no, but those are generated automatically. | |
57:09.240 --> 57:11.360 | |
But even if you tried to do it by hand, | |
57:11.360 --> 57:15.520 | |
could you come up with a spec between your encoder | |
57:15.520 --> 57:16.360 | |
and your decoder? | |
57:17.400 --> 57:20.720 | |
No, no, because it wasn't designed, but they're... | |
57:20.720 --> 57:22.600 | |
No, no, no, but if you could design it, | |
57:23.600 --> 57:26.480 | |
if you could design a face reconstructor system, | |
57:26.480 --> 57:28.080 | |
could you come up with a spec? | |
57:29.240 --> 57:32.320 | |
No, but I think we're missing here a little bit. | |
57:32.320 --> 57:35.080 | |
I think you're just being very poetic | |
57:35.080 --> 57:37.880 | |
about expressing a fundamental problem of simulators, | |
57:37.880 --> 57:41.640 | |
that they are going to be missing so much | |
57:42.480 --> 57:44.680 | |
that the feature of actually | |
57:44.680 --> 57:47.080 | |
would just look fundamentally different | |
57:47.080 --> 57:50.440 | |
from in the simulated world than the real world. | |
57:51.280 --> 57:53.800 | |
I'm not making a claim about simulators. | |
57:53.800 --> 57:57.120 | |
I'm making a claim about the spec division | |
57:57.120 --> 57:58.800 | |
between perception and planning. | |
57:58.800 --> 57:59.640 | |
And planning. | |
57:59.640 --> 58:00.840 | |
Even in your system. | |
58:00.840 --> 58:01.800 | |
Just in general. | |
58:01.800 --> 58:03.360 | |
Right, just in general. | |
58:03.360 --> 58:05.680 | |
If you're trying to build a car that drives, | |
58:05.680 --> 58:07.280 | |
if you're trying to hand code | |
58:07.280 --> 58:08.760 | |
the output of your perception system, | |
58:08.760 --> 58:10.960 | |
like saying, here's a list of all the cars in the scene. | |
58:10.960 --> 58:11.920 | |
Here's a list of all the people. | |
58:11.920 --> 58:13.120 | |
Here's a list of the occluded areas. | |
58:13.120 --> 58:14.920 | |
Here's a vector of drivable areas. | |
58:14.920 --> 58:16.600 | |
It's insufficient. | |
58:16.600 --> 58:18.000 | |
And if you start to believe that, | |
58:18.000 --> 58:20.840 | |
you realize that what Waymo and Cruz are doing is impossible. | |
58:20.840 --> 58:24.320 | |
Currently, what we're doing is the perception problem | |
58:24.320 --> 58:28.200 | |
is converting the scene into a chessboard. | |
58:29.200 --> 58:31.720 | |
And then you reason some basic reasoning | |
58:31.720 --> 58:33.400 | |
around that chessboard. | |
58:33.400 --> 58:38.080 | |
And you're saying that really there's a lot missing there. | |
58:38.080 --> 58:40.240 | |
First of all, why are we talking about this? | |
58:40.240 --> 58:42.840 | |
Because isn't this a full autonomy? | |
58:42.840 --> 58:44.720 | |
Is this something you think about? | |
58:44.720 --> 58:47.680 | |
Oh, I want to win self driving cars. | |
58:47.680 --> 58:52.680 | |
So your definition of win includes the full five. | |
58:53.680 --> 58:55.800 | |
I don't think level four is a real thing. | |
58:55.800 --> 58:59.720 | |
I want to build the AlphaGo of driving. | |
58:59.720 --> 59:04.720 | |
So AlphaGo is really end to end. | |
59:06.160 --> 59:07.000 | |
Yeah. | |
59:07.000 --> 59:09.840 | |
Is, yeah, it's end to end. | |
59:09.840 --> 59:12.480 | |
And do you think this whole problem, | |
59:12.480 --> 59:14.680 | |
is that also kind of what you're getting at | |
59:14.680 --> 59:16.640 | |
with the perception and the planning? | |
59:16.640 --> 59:19.440 | |
Is that this whole problem, the right way to do it, | |
59:19.440 --> 59:21.600 | |
is really to learn the entire thing? | |
59:21.600 --> 59:23.680 | |
I'll argue that not only is it the right way, | |
59:23.680 --> 59:27.640 | |
it's the only way that's going to exceed human performance. | |
59:27.640 --> 59:29.960 | |
Well, it's certainly true for Go. | |
59:29.960 --> 59:31.520 | |
Everyone who tried to hand code Go things | |
59:31.520 --> 59:33.440 | |
built human inferior things. | |
59:33.440 --> 59:36.200 | |
And then someone came along and wrote some 10,000 line thing | |
59:36.200 --> 59:39.800 | |
that doesn't know anything about Go that beat everybody. | |
59:39.800 --> 59:41.080 | |
It's 10,000 lines. | |
59:41.080 --> 59:43.360 | |
True, in that sense. | |
59:43.360 --> 59:47.520 | |
The open question then that maybe I can ask you | |
59:47.520 --> 59:52.520 | |
is driving is much harder than Go. | |
59:53.440 --> 59:56.240 | |
The open question is how much harder? | |
59:56.240 --> 59:59.480 | |
So how, because I think the Elon Musk approach here | |
59:59.480 --> 1:00:01.600 | |
with planning and perception is similar | |
1:00:01.600 --> 1:00:02.960 | |
to what you're describing, | |
1:00:02.960 --> 1:00:07.960 | |
which is really turning into not some kind of modular thing, | |
1:00:08.280 --> 1:00:11.120 | |
but really do formulate as a learning problem | |
1:00:11.120 --> 1:00:13.360 | |
and solve the learning problem with scale. | |
1:00:13.360 --> 1:00:17.120 | |
So how many years, put one, | |
1:00:17.120 --> 1:00:18.880 | |
how many years would it take to solve this problem | |
1:00:18.880 --> 1:00:21.680 | |
or just how hard is this freaking problem? | |
1:00:21.680 --> 1:00:24.560 | |
Well, the cool thing is, | |
1:00:24.560 --> 1:00:27.800 | |
I think there's a lot of value | |
1:00:27.800 --> 1:00:29.840 | |
that we can deliver along the way. | |
1:00:30.840 --> 1:00:35.840 | |
I think that you can build lame keeping assist | |
1:00:36.600 --> 1:00:41.440 | |
actually plus adaptive cruise control plus, okay, | |
1:00:41.440 --> 1:00:46.000 | |
looking at ways extends to like all of driving. | |
1:00:46.000 --> 1:00:47.920 | |
Yeah, most of driving, right? | |
1:00:47.920 --> 1:00:49.760 | |
Oh, your adaptive cruise control treats red lights | |
1:00:49.760 --> 1:00:51.200 | |
like cars, okay. | |
1:00:51.200 --> 1:00:53.480 | |
So let's jump around with you mentioned | |
1:00:53.480 --> 1:00:55.760 | |
that you didn't like navigate an autopilot. | |
1:00:55.760 --> 1:00:57.760 | |
What advice, how would you make it better? | |
1:00:57.760 --> 1:01:00.560 | |
Do you think as a feature that if it's done really well, | |
1:01:00.560 --> 1:01:02.360 | |
it's a good feature? | |
1:01:02.360 --> 1:01:07.360 | |
I think that it's too reliant on like hand coded hacks | |
1:01:07.520 --> 1:01:10.400 | |
for like, how does navigate an autopilot do a lane change? | |
1:01:10.400 --> 1:01:13.400 | |
It actually does the same lane change every time | |
1:01:13.400 --> 1:01:14.320 | |
and it feels mechanical. | |
1:01:14.320 --> 1:01:15.920 | |
Humans do different lane changes. | |
1:01:15.920 --> 1:01:17.360 | |
Humans, sometimes we'll do a slow one, | |
1:01:17.360 --> 1:01:18.920 | |
sometimes do a fast one. | |
1:01:18.920 --> 1:01:20.880 | |
Navigate an autopilot at least every time I use it | |
1:01:20.880 --> 1:01:23.040 | |
is it the identical lane change? | |
1:01:23.040 --> 1:01:24.280 | |
How do you learn? | |
1:01:24.280 --> 1:01:26.800 | |
I mean, this is a fundamental thing actually | |
1:01:26.800 --> 1:01:30.400 | |
is the breaking and accelerating, | |
1:01:30.400 --> 1:01:33.960 | |
something that still, Tesla probably does it better | |
1:01:33.960 --> 1:01:36.800 | |
than most cars, but it still doesn't do a great job | |
1:01:36.800 --> 1:01:39.960 | |
of creating a comfortable natural experience | |
1:01:39.960 --> 1:01:42.680 | |
and navigate an autopilot is just lane changes | |
1:01:42.680 --> 1:01:44.120 | |
and extension of that. | |
1:01:44.120 --> 1:01:49.120 | |
So how do you learn to do natural lane change? | |
1:01:49.120 --> 1:01:52.920 | |
So we have it and I can talk about how it works. | |
1:01:52.920 --> 1:01:57.920 | |
So I feel that we have the solution for lateral | |
1:01:58.720 --> 1:02:00.640 | |
but we don't yet have the solution for longitudinal. | |
1:02:00.640 --> 1:02:03.360 | |
There's a few reasons longitudinal is harder than lateral. | |
1:02:03.360 --> 1:02:06.920 | |
The lane change component, the way that we train on it | |
1:02:06.920 --> 1:02:10.840 | |
very simply is like our model has an input | |
1:02:10.840 --> 1:02:14.040 | |
for whether it's doing a lane change or not. | |
1:02:14.040 --> 1:02:16.360 | |
And then when we train the end to end model, | |
1:02:16.360 --> 1:02:19.560 | |
we hand label all the lane changes because you have to. | |
1:02:19.560 --> 1:02:22.440 | |
I've struggled a long time about not wanting to do that | |
1:02:22.440 --> 1:02:24.280 | |
but I think you have to. | |
1:02:24.280 --> 1:02:25.320 | |
Or the training data. | |
1:02:25.320 --> 1:02:26.520 | |
For the training data, right? | |
1:02:26.520 --> 1:02:28.280 | |
We actually have an automatic ground truth | |
1:02:28.280 --> 1:02:30.600 | |
or which automatically labels all the lane changes. | |
1:02:30.600 --> 1:02:31.680 | |
Was that possible? | |
1:02:31.680 --> 1:02:32.720 | |
To automatically label lane changes? | |
1:02:32.720 --> 1:02:33.560 | |
Yeah. | |
1:02:33.560 --> 1:02:34.800 | |
And detect the lane I see when it crosses it, right? | |
1:02:34.800 --> 1:02:36.680 | |
And I don't have to get that high percent accuracy | |
1:02:36.680 --> 1:02:38.080 | |
but it's like 95 good enough. | |
1:02:38.080 --> 1:02:38.960 | |
Okay. | |
1:02:38.960 --> 1:02:43.200 | |
Now I set the bit when it's doing the lane change | |
1:02:43.200 --> 1:02:44.840 | |
in the end to end learning. | |
1:02:44.840 --> 1:02:47.920 | |
And then I set it to zero when it's not doing a lane change. | |
1:02:47.920 --> 1:02:49.720 | |
So now if I want us to do a lane change a test time, | |
1:02:49.720 --> 1:02:52.360 | |
I just put the bit to a one and it'll do a lane change. | |
1:02:52.360 --> 1:02:54.640 | |
Yeah, but so if you look at the space of lane change, | |
1:02:54.640 --> 1:02:57.320 | |
you know some percentage, not a hundred percent, | |
1:02:57.320 --> 1:03:01.120 | |
that we make as humans is not a pleasant experience | |
1:03:01.120 --> 1:03:02.800 | |
because we messed some part of it up. | |
1:03:02.800 --> 1:03:04.320 | |
It's nerve wracking to change. | |
1:03:04.320 --> 1:03:05.760 | |
If you look, you have to see, | |
1:03:05.760 --> 1:03:06.920 | |
it has to accelerate. | |
1:03:06.920 --> 1:03:09.920 | |
How do we label the ones that are natural and feel good? | |
1:03:09.920 --> 1:03:11.560 | |
You know, that's the, | |
1:03:11.560 --> 1:03:13.360 | |
because that's your ultimate criticism, | |
1:03:13.360 --> 1:03:17.000 | |
the current navigate and autopilot just doesn't feel good. | |
1:03:17.000 --> 1:03:18.520 | |
Well, the current navigate and autopilot | |
1:03:18.520 --> 1:03:21.720 | |
is a hand coded policy written by an engineer in a room | |
1:03:21.720 --> 1:03:25.080 | |
who probably went out and tested it a few times on the 280. | |
1:03:25.080 --> 1:03:28.560 | |
Probably a more, a better version of that. | |
1:03:28.560 --> 1:03:29.400 | |
But yes. | |
1:03:29.400 --> 1:03:30.560 | |
That's how we would have written it. | |
1:03:30.560 --> 1:03:31.400 | |
Yeah. | |
1:03:31.400 --> 1:03:33.480 | |
Maybe Tesla did a Tesla, they tested it in. | |
1:03:33.480 --> 1:03:34.920 | |
That might have been two engineers. | |
1:03:34.920 --> 1:03:35.760 | |
Two engineers. | |
1:03:35.760 --> 1:03:37.400 | |
Yeah. | |
1:03:37.400 --> 1:03:40.120 | |
No, but so if you learn the lane change, | |
1:03:40.120 --> 1:03:42.480 | |
if you learn how to do a lane change from data, | |
1:03:42.480 --> 1:03:44.680 | |
just like you have a label that says lane change | |
1:03:44.680 --> 1:03:48.040 | |
and then you put it in when you want it to do the lane change, | |
1:03:48.040 --> 1:03:49.640 | |
it'll automatically do the lane change | |
1:03:49.640 --> 1:03:51.600 | |
that's appropriate for the situation. | |
1:03:51.600 --> 1:03:54.720 | |
Now, to get at the problem of some humans | |
1:03:54.720 --> 1:03:55.960 | |
do bad lane changes, | |
1:03:57.400 --> 1:03:59.920 | |
we haven't worked too much on this problem yet. | |
1:03:59.920 --> 1:04:03.120 | |
It's not that much of a problem in practice. | |
1:04:03.120 --> 1:04:06.160 | |
My theory is that all good drivers are good in the same way | |
1:04:06.160 --> 1:04:08.440 | |
and all bad drivers are bad in different ways. | |
1:04:09.360 --> 1:04:11.320 | |
And we've seen some data to back this up. | |
1:04:11.320 --> 1:04:12.400 | |
Well, beautifully put. | |
1:04:12.400 --> 1:04:16.560 | |
So you just basically, if that's true hypothesis, | |
1:04:16.560 --> 1:04:19.920 | |
then your task is to discover the good drivers. | |
1:04:19.920 --> 1:04:21.800 | |
The good drivers stand out | |
1:04:21.800 --> 1:04:23.360 | |
because they're in one cluster | |
1:04:23.360 --> 1:04:25.200 | |
and the bad drivers are scattered all over the place | |
1:04:25.200 --> 1:04:27.240 | |
and your net learns the cluster. | |
1:04:27.240 --> 1:04:28.080 | |
Yeah. | |
1:04:28.080 --> 1:04:30.800 | |
So you just learn from the good drivers | |
1:04:30.800 --> 1:04:32.200 | |
and they're easy to cluster. | |
1:04:33.200 --> 1:04:34.240 | |
In fact, we learned from all of them | |
1:04:34.240 --> 1:04:35.840 | |
and the net automatically learns the policy | |
1:04:35.840 --> 1:04:36.920 | |
that's like the majority. | |
1:04:36.920 --> 1:04:38.440 | |
But we'll eventually probably have to build some out. | |
1:04:38.440 --> 1:04:41.560 | |
So if that theory is true, I hope it's true | |
1:04:41.560 --> 1:04:46.440 | |
because the counter theory is there is many clusters, | |
1:04:49.480 --> 1:04:53.680 | |
maybe arbitrarily many clusters of good drivers. | |
1:04:53.680 --> 1:04:55.840 | |
Because if there's one cluster of good drivers, | |
1:04:55.840 --> 1:04:57.600 | |
you can at least discover a set of policies. | |
1:04:57.600 --> 1:04:59.000 | |
You can learn a set of policies | |
1:04:59.000 --> 1:05:00.640 | |
which would be good universally. | |
1:05:00.640 --> 1:05:01.640 | |
Yeah. | |
1:05:01.640 --> 1:05:04.560 | |
That would be nice if it's true. | |
1:05:04.560 --> 1:05:06.560 | |
And you're saying that there is some evidence that... | |
1:05:06.560 --> 1:05:09.720 | |
Let's say lane changes can be clustered into four clusters. | |
1:05:09.720 --> 1:05:10.560 | |
Right. | |
1:05:10.560 --> 1:05:12.040 | |
There's a finite level of... | |
1:05:12.040 --> 1:05:15.280 | |
I would argue that all four of those are good clusters. | |
1:05:15.280 --> 1:05:18.360 | |
All the things that are random are noise and probably bad. | |
1:05:18.360 --> 1:05:20.360 | |
And which one of the four you pick? | |
1:05:20.360 --> 1:05:21.920 | |
Or maybe it's 10 or maybe it's 20. | |
1:05:21.920 --> 1:05:22.760 | |
You can learn that. | |
1:05:22.760 --> 1:05:23.800 | |
It's context dependent. | |
1:05:23.800 --> 1:05:25.040 | |
It depends on the scene. | |
1:05:26.760 --> 1:05:30.440 | |
And the hope is it's not too dependent on the driver. | |
1:05:31.400 --> 1:05:34.240 | |
Yeah, the hope is that it all washes out. | |
1:05:34.240 --> 1:05:36.960 | |
The hope is that the distribution is not bimodal. | |
1:05:36.960 --> 1:05:39.080 | |
The hope is that it's a nice Gaussian. | |
1:05:39.080 --> 1:05:41.640 | |
So what advice would you give to Tesla? | |
1:05:41.640 --> 1:05:45.000 | |
How to fix, how to improve, navigate an autopilot? | |
1:05:45.000 --> 1:05:48.240 | |
That's the lessons that you've learned from Kamii. | |
1:05:48.240 --> 1:05:50.560 | |
The only real advice I would give to Tesla | |
1:05:50.560 --> 1:05:52.920 | |
is please put driver monitoring in your cars. | |
1:05:53.920 --> 1:05:55.160 | |
With respect to improving it. | |
1:05:55.160 --> 1:05:56.000 | |
You can't do that anymore. | |
1:05:56.000 --> 1:05:57.280 | |
I started to interrupt. | |
1:05:57.280 --> 1:06:01.760 | |
But there's a practical nature of many of hundreds of thousands | |
1:06:01.760 --> 1:06:05.760 | |
of cars being produced that don't have a good driver facing camera. | |
1:06:05.760 --> 1:06:07.520 | |
The Model 3 has a selfie cam. | |
1:06:07.520 --> 1:06:08.680 | |
Is it not good enough? | |
1:06:08.680 --> 1:06:10.800 | |
Did they not have put IR LEDs for night? | |
1:06:10.800 --> 1:06:11.640 | |
That's a good question. | |
1:06:11.640 --> 1:06:13.360 | |
But I do know that it's fish eye | |
1:06:13.360 --> 1:06:15.800 | |
and it's relatively low resolution. | |
1:06:15.800 --> 1:06:16.760 | |
So it's really not designed. | |
1:06:16.760 --> 1:06:18.760 | |
It wasn't designed for driver monitoring. | |
1:06:18.760 --> 1:06:21.760 | |
You can hope that you can kind of scrape up | |
1:06:21.760 --> 1:06:24.400 | |
and have something from it. | |
1:06:24.400 --> 1:06:27.520 | |
But why didn't they put it in today? | |
1:06:27.520 --> 1:06:28.280 | |
Put it in today. | |
1:06:28.280 --> 1:06:29.520 | |
Put it in today. | |
1:06:29.520 --> 1:06:31.520 | |
Every time I've heard Carpathian talk about the problem | |
1:06:31.520 --> 1:06:33.240 | |
and talking about like software 2.0 | |
1:06:33.240 --> 1:06:35.240 | |
and how the machine learning is gobbling up everything, | |
1:06:35.240 --> 1:06:37.440 | |
I think this is absolutely the right strategy. | |
1:06:37.440 --> 1:06:40.160 | |
I think that he didn't write and navigate on autopilot. | |
1:06:40.160 --> 1:06:43.240 | |
I think somebody else did and kind of hacked it on top of that stuff. | |
1:06:43.240 --> 1:06:45.680 | |
I think when Carpathian says, wait a second, | |
1:06:45.680 --> 1:06:47.440 | |
why did we hand code this lane change policy | |
1:06:47.440 --> 1:06:48.360 | |
with all these magic numbers? | |
1:06:48.360 --> 1:06:49.360 | |
We're going to learn it from data. | |
1:06:49.360 --> 1:06:49.840 | |
They'll fix it. | |
1:06:49.840 --> 1:06:51.040 | |
They already know what to do there. | |
1:06:51.040 --> 1:06:54.360 | |
Well, that's Andre's job is to turn everything | |
1:06:54.360 --> 1:06:57.480 | |
into a learning problem and collect a huge amount of data. | |
1:06:57.480 --> 1:07:01.120 | |
The reality is, though, not every problem | |
1:07:01.120 --> 1:07:04.080 | |
can be turned into a learning problem in the short term. | |
1:07:04.080 --> 1:07:07.280 | |
In the end, everything will be a learning problem. | |
1:07:07.280 --> 1:07:12.880 | |
The reality is, like if you want to build L5 vehicles today, | |
1:07:12.880 --> 1:07:15.600 | |
it will likely involve no learning. | |
1:07:15.600 --> 1:07:20.320 | |
And that's the reality is, so at which point does learning start? | |
1:07:20.320 --> 1:07:23.480 | |
It's the crutch statement that LiDAR is a crutch. | |
1:07:23.480 --> 1:07:27.240 | |
Which point will learning get up to part of human performance? | |
1:07:27.240 --> 1:07:31.960 | |
It's over human performance on ImageNet, classification, | |
1:07:31.960 --> 1:07:34.000 | |
on driving, it's a question still. | |
1:07:34.000 --> 1:07:35.760 | |
It is a question. | |
1:07:35.760 --> 1:07:39.160 | |
I'll say this, I'm here to play for 10 years. | |
1:07:39.160 --> 1:07:40.280 | |
I'm not here to try to. | |
1:07:40.280 --> 1:07:42.960 | |
I'm here to play for 10 years and make money along the way. | |
1:07:42.960 --> 1:07:45.040 | |
I'm not here to try to promise people | |
1:07:45.040 --> 1:07:47.600 | |
that I'm going to have my L5 taxi network up and working | |
1:07:47.600 --> 1:07:48.200 | |
in two years. | |
1:07:48.200 --> 1:07:49.400 | |
Do you think that was a mistake? | |
1:07:49.400 --> 1:07:50.520 | |
Yes. | |
1:07:50.520 --> 1:07:53.160 | |
What do you think was the motivation behind saying | |
1:07:53.160 --> 1:07:56.640 | |
that other companies are also promising L5 vehicles | |
1:07:56.640 --> 1:08:01.880 | |
with their different approaches in 2020, 2021, 2022? | |
1:08:01.880 --> 1:08:05.720 | |
If anybody would like to bet me that those things do not pan out, | |
1:08:05.720 --> 1:08:07.000 | |
I will bet you. | |
1:08:07.000 --> 1:08:10.800 | |
Even money, even money, I'll bet you as much as you want. | |
1:08:10.800 --> 1:08:13.600 | |
So are you worried about what's going to happen? | |
1:08:13.600 --> 1:08:16.040 | |
Because you're not in full agreement on that. | |
1:08:16.040 --> 1:08:19.160 | |
What's going to happen when 2022, 2021 come around | |
1:08:19.160 --> 1:08:22.800 | |
and nobody has fleets of autonomous vehicles? | |
1:08:22.800 --> 1:08:25.000 | |
Well, you can look at the history. | |
1:08:25.000 --> 1:08:26.880 | |
If you go back five years ago, they | |
1:08:26.880 --> 1:08:29.880 | |
were all promised by 2018 and 2017. | |
1:08:29.880 --> 1:08:32.200 | |
But they weren't that strong of promises. | |
1:08:32.200 --> 1:08:36.240 | |
I mean, Ford really declared. | |
1:08:36.240 --> 1:08:40.560 | |
I think not many have declared as definitively | |
1:08:40.560 --> 1:08:42.600 | |
as they have now these dates. | |
1:08:42.600 --> 1:08:43.320 | |
Well, OK. | |
1:08:43.320 --> 1:08:45.040 | |
So let's separate L4 and L5. | |
1:08:45.040 --> 1:08:46.800 | |
Do I think that it's possible for Waymo | |
1:08:46.800 --> 1:08:50.960 | |
to continue to hack on their system | |
1:08:50.960 --> 1:08:53.400 | |
until it gets to level four in Chandler, Arizona? | |
1:08:53.400 --> 1:08:55.040 | |
Yes. | |
1:08:55.040 --> 1:08:56.800 | |
No safety driver? | |
1:08:56.800 --> 1:08:57.600 | |
Chandler, Arizona? | |
1:08:57.600 --> 1:08:59.600 | |
Yeah. | |
1:08:59.600 --> 1:09:02.440 | |
By which year are we talking about? | |
1:09:02.440 --> 1:09:06.120 | |
Oh, I even think that's possible by like 2020, 2021. | |
1:09:06.120 --> 1:09:09.480 | |
But level four, Chandler, Arizona, not level five, | |
1:09:09.480 --> 1:09:11.480 | |
New York City. | |
1:09:11.480 --> 1:09:15.920 | |
Level four, meaning some very defined streets. | |
1:09:15.920 --> 1:09:17.400 | |
It works out really well. | |
1:09:17.400 --> 1:09:18.280 | |
Very defined streets. | |
1:09:18.280 --> 1:09:20.680 | |
And then practically, these streets are pretty empty. | |
1:09:20.680 --> 1:09:24.680 | |
If most of the streets are covered in Waymos, | |
1:09:24.680 --> 1:09:28.360 | |
Waymo can kind of change the definition of what driving is. | |
1:09:28.360 --> 1:09:28.920 | |
Right? | |
1:09:28.920 --> 1:09:31.720 | |
If your self driving network is the majority | |
1:09:31.720 --> 1:09:34.120 | |
of cars in an area, they only need | |
1:09:34.120 --> 1:09:35.720 | |
to be safe with respect to each other, | |
1:09:35.720 --> 1:09:38.640 | |
and all the humans will need to learn to adapt to them. | |
1:09:38.640 --> 1:09:41.120 | |
Now go drive in downtown New York. | |
1:09:41.120 --> 1:09:42.200 | |
Oh, yeah, that's. | |
1:09:42.200 --> 1:09:43.440 | |
I mean, already. | |
1:09:43.440 --> 1:09:46.040 | |
You can talk about autonomy and like on farms, | |
1:09:46.040 --> 1:09:48.520 | |
it already works great, because you can really just | |
1:09:48.520 --> 1:09:51.320 | |
follow the GPS line. | |
1:09:51.320 --> 1:09:56.800 | |
So what does success look like for Kama AI? | |
1:09:56.800 --> 1:09:58.200 | |
What are the milestones like where | |
1:09:58.200 --> 1:09:59.800 | |
you can sit back with some champagne | |
1:09:59.800 --> 1:10:04.120 | |
and say, we did it, boys and girls? | |
1:10:04.120 --> 1:10:06.320 | |
Well, it's never over. | |
1:10:06.320 --> 1:10:07.800 | |
Yeah, but don't be so. | |
1:10:07.800 --> 1:10:10.400 | |
You must drink champagne every time you celebrate. | |
1:10:10.400 --> 1:10:11.440 | |
So what is good? | |
1:10:11.440 --> 1:10:13.160 | |
What are some wins? | |
1:10:13.160 --> 1:10:19.480 | |
A big milestone that we're hoping for by mid next year | |
1:10:19.480 --> 1:10:20.680 | |
is profitability of the company. | |
1:10:20.680 --> 1:10:28.560 | |
And we're going to have to revisit the idea of selling | |
1:10:28.560 --> 1:10:30.280 | |
a consumer product. | |
1:10:30.280 --> 1:10:32.720 | |
But it's not going to be like the Kama One. | |
1:10:32.720 --> 1:10:36.240 | |
When we do it, it's going to be perfect. | |
1:10:36.240 --> 1:10:39.600 | |
OpenPilot has gotten so much better in the last two years. | |
1:10:39.600 --> 1:10:41.680 | |
We're going to have a few features. | |
1:10:41.680 --> 1:10:43.760 | |
We're going to have 100% driver monitoring. | |
1:10:43.760 --> 1:10:46.720 | |
We're going to disable no safety features in the car. | |
1:10:46.720 --> 1:10:48.760 | |
Actually, I think it'd be really cool what we're doing right | |
1:10:48.760 --> 1:10:51.600 | |
now, our project this week is we're analyzing the data set | |
1:10:51.600 --> 1:10:53.240 | |
and looking for all the AEB triggers | |
1:10:53.240 --> 1:10:55.640 | |
from the manufacturer systems. | |
1:10:55.640 --> 1:10:59.440 | |
We have better data set on that than the manufacturers. | |
1:10:59.440 --> 1:11:02.960 | |
How much does Toyota have 10 million miles of real world | |
1:11:02.960 --> 1:11:05.360 | |
driving to know how many times they're AEB triggered? | |
1:11:05.360 --> 1:11:10.880 | |
So let me give you, because you asked, financial advice. | |
1:11:10.880 --> 1:11:12.440 | |
Because I work with a lot of automakers | |
1:11:12.440 --> 1:11:15.840 | |
and one possible source of money for you, | |
1:11:15.840 --> 1:11:21.400 | |
which I'll be excited to see you take on, is basically | |
1:11:21.400 --> 1:11:29.120 | |
selling the data, which is something that most people, | |
1:11:29.120 --> 1:11:31.800 | |
and not selling in a way where here, here at Automaker, | |
1:11:31.800 --> 1:11:33.000 | |
but creating. | |
1:11:33.000 --> 1:11:35.480 | |
We've done this actually at MIT, not for money purposes, | |
1:11:35.480 --> 1:11:37.760 | |
but you could do it for significant money purposes | |
1:11:37.760 --> 1:11:39.440 | |
and make the world a better place | |
1:11:39.440 --> 1:11:44.240 | |
by creating a consortia where automakers would pay in | |
1:11:44.240 --> 1:11:46.960 | |
and then they get to have free access to the data. | |
1:11:46.960 --> 1:11:52.400 | |
And I think a lot of people are really hungry for that | |
1:11:52.400 --> 1:11:54.200 | |
and would pay significant amount of money for it. | |
1:11:54.200 --> 1:11:55.400 | |
Here's the problem with that. | |
1:11:55.400 --> 1:11:56.840 | |
I like this idea all in theory. | |
1:11:56.840 --> 1:11:59.640 | |
It'd be very easy for me to give them access to my servers. | |
1:11:59.640 --> 1:12:02.280 | |
And we already have all open source tools to access this data. | |
1:12:02.280 --> 1:12:03.400 | |
It's in a great format. | |
1:12:03.400 --> 1:12:05.560 | |
We have a great pipeline. | |
1:12:05.560 --> 1:12:07.120 | |
But they're going to put me in the room | |
1:12:07.120 --> 1:12:10.120 | |
with some business development guy. | |
1:12:10.120 --> 1:12:12.400 | |
And I'm going to have to talk to this guy. | |
1:12:12.400 --> 1:12:15.040 | |
And he's not going to know most of the words I'm saying. | |
1:12:15.040 --> 1:12:17.280 | |
I'm not willing to tolerate that. | |
1:12:17.280 --> 1:12:18.840 | |
OK, Mick Jagger. | |
1:12:18.840 --> 1:12:19.800 | |
No, no, no, no. | |
1:12:19.800 --> 1:12:21.040 | |
But I think I agree with you. | |
1:12:21.040 --> 1:12:21.720 | |
I'm the same way. | |
1:12:21.720 --> 1:12:22.960 | |
But you just tell them the terms | |
1:12:22.960 --> 1:12:24.640 | |
and there's no discussion needed. | |
1:12:24.640 --> 1:12:30.480 | |
If I could just tell them the terms, then like, all right. | |
1:12:30.480 --> 1:12:31.600 | |
Who wants access to my data? | |
1:12:31.600 --> 1:12:36.680 | |
I will sell it to you for, let's say, | |
1:12:36.680 --> 1:12:37.640 | |
you want a subscription? | |
1:12:37.640 --> 1:12:40.680 | |
I'll sell you for 100k a month. | |
1:12:40.680 --> 1:12:41.200 | |
Anyone? | |
1:12:41.200 --> 1:12:42.000 | |
100k a month? | |
1:12:42.000 --> 1:12:43.040 | |
100k a month? | |
1:12:43.040 --> 1:12:45.080 | |
I'll give you access to the data subscription? | |
1:12:45.080 --> 1:12:45.680 | |
Yeah. | |
1:12:45.680 --> 1:12:46.680 | |
Yeah, I think that's kind of fair. | |
1:12:46.680 --> 1:12:48.440 | |
Came up with that number off the top of my head. | |
1:12:48.440 --> 1:12:50.840 | |
If somebody sends me like a three line email where it's like, | |
1:12:50.840 --> 1:12:54.000 | |
we would like to pay 100k a month to get access to your data. | |
1:12:54.000 --> 1:12:56.160 | |
We would agree to like reasonable privacy terms | |
1:12:56.160 --> 1:12:58.360 | |
of the people who are in the data set. | |
1:12:58.360 --> 1:12:59.520 | |
I would be happy to do it. | |
1:12:59.520 --> 1:13:01.200 | |
But that's not going to be the email. | |
1:13:01.200 --> 1:13:03.120 | |
The email is going to be, hey, do you | |
1:13:03.120 --> 1:13:05.560 | |
have some time in the next month where we can sit down | |
1:13:05.560 --> 1:13:07.000 | |
and we can, I don't have time for that. | |
1:13:07.000 --> 1:13:08.360 | |
We're moving too fast. | |
1:13:08.360 --> 1:13:10.040 | |
You could politely respond to that email, | |
1:13:10.040 --> 1:13:13.240 | |
but not saying I don't have any time for your bullshit. | |
1:13:13.240 --> 1:13:15.440 | |
You say, oh, well, unfortunately, these are the terms. | |
1:13:15.440 --> 1:13:19.280 | |
And so this is what we try to, we brought the cost down | |
1:13:19.280 --> 1:13:22.320 | |
for you in order to minimize the friction, the communication. | |
1:13:22.320 --> 1:13:22.920 | |
Yeah, absolutely. | |
1:13:22.920 --> 1:13:26.720 | |
Here's the whatever it is, $1, $2 million a year. | |
1:13:26.720 --> 1:13:28.880 | |
And you have access. | |
1:13:28.880 --> 1:13:31.440 | |
And it's not like I get that email from like, | |
1:13:31.440 --> 1:13:32.720 | |
but OK, am I going to reach out? | |
1:13:32.720 --> 1:13:34.200 | |
Am I going to hire a business development person | |
1:13:34.200 --> 1:13:35.840 | |
who's going to reach out to the automakers? | |
1:13:35.840 --> 1:13:36.480 | |
No way. | |
1:13:36.480 --> 1:13:36.880 | |
Yeah. | |
1:13:36.880 --> 1:13:37.840 | |
OK, I got you. | |
1:13:37.840 --> 1:13:38.520 | |
I admire. | |
1:13:38.520 --> 1:13:39.680 | |
If they reached into me, I'm not | |
1:13:39.680 --> 1:13:40.600 | |
going to ignore the email. | |
1:13:40.600 --> 1:13:42.160 | |
I'll come back with something like, yeah, | |
1:13:42.160 --> 1:13:44.560 | |
if you're willing to pay $100,000 for access to the data, | |
1:13:44.560 --> 1:13:46.080 | |
I'm happy to set that up. | |
1:13:46.080 --> 1:13:48.200 | |
That's worth my engineering time. | |
1:13:48.200 --> 1:13:49.520 | |
That's actually quite insightful of you. | |
1:13:49.520 --> 1:13:50.440 | |
You're right. | |
1:13:50.440 --> 1:13:52.480 | |
Probably because many of the automakers | |
1:13:52.480 --> 1:13:54.480 | |
are quite a bit old school, there | |
1:13:54.480 --> 1:13:56.200 | |
will be a need to reach out. | |
1:13:56.200 --> 1:13:58.440 | |
And they want it, but there will need | |
1:13:58.440 --> 1:13:59.800 | |
to be some communication. | |
1:13:59.800 --> 1:14:00.160 | |
You're right. | |
1:14:00.160 --> 1:14:06.760 | |
Mobileye circa 2015 had the lowest R&D spend of any chipmaker. | |
1:14:06.760 --> 1:14:10.640 | |
Like per, and you look at all the people who work for them, | |
1:14:10.640 --> 1:14:12.120 | |
and it's all business development people | |
1:14:12.120 --> 1:14:15.320 | |
because the car companies are impossible to work with. | |
1:14:15.320 --> 1:14:17.880 | |
Yeah, so you have no patience for that, | |
1:14:17.880 --> 1:14:20.040 | |
and you're a legit Android, huh? | |
1:14:20.040 --> 1:14:21.440 | |
I have something to do, right? | |
1:14:21.440 --> 1:14:24.040 | |
Like, it's not like I don't mean to be a dick and say, | |
1:14:24.040 --> 1:14:25.920 | |
I don't have patience for that, but it's like, | |
1:14:25.920 --> 1:14:29.160 | |
that stuff doesn't help us with our goal of winning | |
1:14:29.160 --> 1:14:30.560 | |
self driving cars. | |
1:14:30.560 --> 1:14:33.800 | |
If I want money in the short term, | |
1:14:33.800 --> 1:14:38.040 | |
if I showed off the actual learning tech that we have, | |
1:14:38.040 --> 1:14:40.160 | |
it's somewhat sad. | |
1:14:40.160 --> 1:14:43.000 | |
It's years and years ahead of everybody else's. | |
1:14:43.000 --> 1:14:43.720 | |
Maybe not Tesla's. | |
1:14:43.720 --> 1:14:45.720 | |
I think Tesla has similar stuff to us, actually. | |
1:14:45.720 --> 1:14:47.640 | |
I think Tesla has similar stuff, but when you compare it | |
1:14:47.640 --> 1:14:50.920 | |
to what the Toyota Research Institute has, | |
1:14:50.920 --> 1:14:53.480 | |
you're not even close to what we have. | |
1:14:53.480 --> 1:14:55.840 | |
No comments, but I also can't. | |
1:14:55.840 --> 1:14:58.440 | |
I have to take your comments. | |
1:14:58.440 --> 1:15:01.960 | |
I intuitively believe you, but I have | |
1:15:01.960 --> 1:15:04.680 | |
to take it with a grain of salt because, | |
1:15:04.680 --> 1:15:07.440 | |
I mean, you are an inspiration because you basically | |
1:15:07.440 --> 1:15:10.000 | |
don't care about a lot of things that other companies care | |
1:15:10.000 --> 1:15:10.880 | |
about. | |
1:15:10.880 --> 1:15:16.600 | |
You don't try to bullshit, in a sense, like make up stuff, | |
1:15:16.600 --> 1:15:18.600 | |
so to drive up valuation. | |
1:15:18.600 --> 1:15:19.960 | |
You're really very real, and you're | |
1:15:19.960 --> 1:15:22.280 | |
trying to solve the problem, and I admire that a lot. | |
1:15:22.280 --> 1:15:26.520 | |
What I don't necessarily fully can't trust you on about your | |
1:15:26.520 --> 1:15:28.440 | |
respect is how good it is, right? | |
1:15:28.440 --> 1:15:33.320 | |
I can only, but I also know how bad others are. | |
1:15:33.320 --> 1:15:36.680 | |
I'll say two things about, trust, but verify, right? | |
1:15:36.680 --> 1:15:38.040 | |
I'll say two things about that. | |
1:15:38.040 --> 1:15:42.360 | |
One is try, get in a 2020 Corolla, | |
1:15:42.360 --> 1:15:46.680 | |
and try OpenPilot 0.6 when it comes out next month. | |
1:15:46.680 --> 1:15:48.400 | |
I think already, you'll look at this, | |
1:15:48.400 --> 1:15:51.400 | |
and you'll be like, this is already really good. | |
1:15:51.400 --> 1:15:54.240 | |
And then, I could be doing that all with hand labelers | |
1:15:54.240 --> 1:15:58.000 | |
and all with the same approach that Mobileye uses. | |
1:15:58.000 --> 1:16:00.040 | |
When we release a model that no longer | |
1:16:00.040 --> 1:16:05.000 | |
has the lanes in it, that only outputs a path, | |
1:16:05.000 --> 1:16:08.720 | |
then think about how we did that machine learning, | |
1:16:08.720 --> 1:16:10.080 | |
and then right away, when you see, | |
1:16:10.080 --> 1:16:11.240 | |
and that's going to be an OpenPilot, | |
1:16:11.240 --> 1:16:13.000 | |
that's going to be an OpenPilot before 1.0, | |
1:16:13.000 --> 1:16:14.400 | |
when you see that model, you'll know | |
1:16:14.400 --> 1:16:15.360 | |
that everything I'm saying is true, | |
1:16:15.360 --> 1:16:16.840 | |
because how else did I get that model? | |
1:16:16.840 --> 1:16:17.320 | |
Good. | |
1:16:17.320 --> 1:16:19.240 | |
You know what I'm saying is true about the simulator. | |
1:16:19.240 --> 1:16:20.600 | |
Yeah, yeah, yeah, this is super exciting. | |
1:16:20.600 --> 1:16:22.680 | |
That's super exciting. | |
1:16:22.680 --> 1:16:25.760 | |
But I listened to your talk with Kyle, | |
1:16:25.760 --> 1:16:30.480 | |
and Kyle was originally building the aftermarket system, | |
1:16:30.480 --> 1:16:34.920 | |
and he gave up on it because of technical challenges, | |
1:16:34.920 --> 1:16:37.360 | |
because of the fact that he's going | |
1:16:37.360 --> 1:16:39.160 | |
to have to support 20 to 50 cars. | |
1:16:39.160 --> 1:16:41.120 | |
We support 45, because what is he | |
1:16:41.120 --> 1:16:43.440 | |
going to do when the manufacturer ABS system triggers? | |
1:16:43.440 --> 1:16:45.480 | |
We have alerts and warnings to deal with all of that | |
1:16:45.480 --> 1:16:48.400 | |
and all the cars, and how is he going to formally verify it? | |
1:16:48.400 --> 1:16:49.800 | |
Well, I got 10 million miles of data. | |
1:16:49.800 --> 1:16:53.240 | |
It's probably better verified than the spec. | |
1:16:53.240 --> 1:16:57.720 | |
Yeah, I'm glad you're here talking to me. | |
1:16:57.720 --> 1:17:01.120 | |
I'll remember this day, because it's interesting. | |
1:17:01.120 --> 1:17:04.160 | |
If you look at Kyle's from Cruise, | |
1:17:04.160 --> 1:17:06.320 | |
I'm sure they have a large number of business development | |
1:17:06.320 --> 1:17:10.200 | |
folks, and he's working with GM. | |
1:17:10.200 --> 1:17:13.280 | |
He could work with Argo AI, worked with Ford. | |
1:17:13.280 --> 1:17:18.520 | |
It's interesting, because chances that you fail businesswise, | |
1:17:18.520 --> 1:17:21.120 | |
like bankrupt, are pretty high. | |
1:17:21.120 --> 1:17:23.880 | |
And yet, it's the Android model, | |
1:17:23.880 --> 1:17:26.440 | |
is you're actually taking on the problem. | |
1:17:26.440 --> 1:17:28.160 | |
So that's really inspiring. | |
1:17:28.160 --> 1:17:30.920 | |
Well, I have a long term way for comedy to make money, too. | |
1:17:30.920 --> 1:17:34.400 | |
And one of the nice things when you really take on the problem, | |
1:17:34.400 --> 1:17:36.760 | |
which is my hope for autopilot, for example, | |
1:17:36.760 --> 1:17:41.040 | |
is things you don't expect, ways to make money, | |
1:17:41.040 --> 1:17:44.160 | |
or create value that you don't expect will pop up. | |
1:17:44.160 --> 1:17:48.560 | |
I've known how to do it since 2017 is the first time I said it. | |
1:17:48.560 --> 1:17:50.440 | |
Which part to know how to do which part? | |
1:17:50.440 --> 1:17:52.520 | |
Our long term plan is to be a car insurance company. | |
1:17:52.520 --> 1:17:53.160 | |
Insurance. | |
1:17:53.160 --> 1:17:55.320 | |
Yeah, I love it. | |
1:17:55.320 --> 1:17:56.680 | |
I make driving twice as safe. | |
1:17:56.680 --> 1:17:57.680 | |
Not only that, I have the best data | |
1:17:57.680 --> 1:18:00.040 | |
such to know who statistically is the safest drivers. | |
1:18:00.040 --> 1:18:02.160 | |
And oh, oh, we see you. | |
1:18:02.160 --> 1:18:03.720 | |
We see you driving unsafely. | |
1:18:03.720 --> 1:18:05.360 | |
We're not going to insure you. | |
1:18:05.360 --> 1:18:08.960 | |
And that causes a bifurcation in the market, | |
1:18:08.960 --> 1:18:10.920 | |
because the only people who can't get common insurance | |
1:18:10.920 --> 1:18:12.760 | |
or the bad drivers, Geico can insure them. | |
1:18:12.760 --> 1:18:15.360 | |
Their premiums are crazy high, our premiums are crazy low. | |
1:18:15.360 --> 1:18:16.240 | |
We win car insurance. | |
1:18:16.240 --> 1:18:18.120 | |
Take over that whole market. | |
1:18:18.120 --> 1:18:21.560 | |
OK, so if we win, if we win, but that's | |
1:18:21.560 --> 1:18:23.800 | |
I'm saying like how do you turn comma into a $10 billion | |
1:18:23.800 --> 1:18:24.640 | |
company is that. | |
1:18:24.640 --> 1:18:25.600 | |
That's right. | |
1:18:25.600 --> 1:18:30.000 | |
So you Elon Musk, who else? | |
1:18:30.000 --> 1:18:32.720 | |
Who else is thinking like this and working like this | |
1:18:32.720 --> 1:18:33.160 | |
in your view? | |
1:18:33.160 --> 1:18:34.800 | |
Who are the competitors? | |
1:18:34.800 --> 1:18:36.160 | |
Are there people seriously? | |
1:18:36.160 --> 1:18:39.480 | |
I don't think anyone that I'm aware of is seriously | |
1:18:39.480 --> 1:18:45.280 | |
taking on lane keeping, like to where it's a huge business that | |
1:18:45.280 --> 1:18:51.400 | |
turns eventually to full autonomy that then creates | |
1:18:51.400 --> 1:18:53.440 | |
other businesses on top of it and so on. | |
1:18:53.440 --> 1:18:56.480 | |
Thinks insurance, thinks all kinds of ideas like that. | |
1:18:56.480 --> 1:19:00.480 | |
Do you know anyone else thinking like this? | |
1:19:00.480 --> 1:19:02.200 | |
Not really. | |
1:19:02.200 --> 1:19:02.960 | |
That's interesting. | |
1:19:02.960 --> 1:19:06.560 | |
I mean, my sense is everybody turns to that in like four | |
1:19:06.560 --> 1:19:07.800 | |
or five years. | |
1:19:07.800 --> 1:19:11.240 | |
Like Ford, once the autonomy doesn't fall through. | |
1:19:11.240 --> 1:19:12.600 | |
But at this time. | |
1:19:12.600 --> 1:19:14.120 | |
Elon's the iOS. | |
1:19:14.120 --> 1:19:16.720 | |
By the way, he paved the way for all of us. | |
1:19:16.720 --> 1:19:18.000 | |
It's not iOS, true. | |
1:19:18.000 --> 1:19:21.520 | |
I would not be doing comma AI today if it was not | |
1:19:21.520 --> 1:19:23.480 | |
for those conversations with Elon. | |
1:19:23.480 --> 1:19:26.840 | |
And if it were not for him saying like, | |
1:19:26.840 --> 1:19:28.600 | |
I think he said like, well, obviously we're not | |
1:19:28.600 --> 1:19:31.280 | |
going to use LiDAR, we use cameras, humans use cameras. | |
1:19:31.280 --> 1:19:32.600 | |
So what do you think about that? | |
1:19:32.600 --> 1:19:33.880 | |
How important is LiDAR? | |
1:19:33.880 --> 1:19:36.960 | |
Everybody else's on L5 is using LiDAR. | |
1:19:36.960 --> 1:19:39.160 | |
What are your thoughts on his provocative statement | |
1:19:39.160 --> 1:19:41.320 | |
that LiDAR is a crutch? | |
1:19:41.320 --> 1:19:43.520 | |
See, sometimes they'll say dumb things like the driver | |
1:19:43.520 --> 1:19:45.680 | |
monitoring thing, but sometimes they'll say absolutely | |
1:19:45.680 --> 1:19:48.400 | |
completely 100% obviously true things. | |
1:19:48.400 --> 1:19:50.840 | |
Of course LiDAR is a crutch. | |
1:19:50.840 --> 1:19:53.040 | |
It's not even a good crutch. | |
1:19:53.040 --> 1:19:54.200 | |
You're not even using it. | |
1:19:54.200 --> 1:19:56.920 | |
They're using it for localization, | |
1:19:56.920 --> 1:19:58.160 | |
which isn't good in the first place. | |
1:19:58.160 --> 1:20:00.480 | |
If you have to localize your car to centimeters | |
1:20:00.480 --> 1:20:04.280 | |
in order to drive, that's not driving. | |
1:20:04.280 --> 1:20:06.320 | |
Currently not doing much machine learning. | |
1:20:06.320 --> 1:20:09.280 | |
I thought LiDAR data, meaning like to help you | |
1:20:09.280 --> 1:20:12.840 | |
in the task of general task of perception. | |
1:20:12.840 --> 1:20:15.320 | |
The main goal of those LiDARs on those cars | |
1:20:15.320 --> 1:20:18.840 | |
I think is actually localization more than perception, | |
1:20:18.840 --> 1:20:20.080 | |
or at least that's what they use them for. | |
1:20:20.080 --> 1:20:20.920 | |
Yeah, that's true. | |
1:20:20.920 --> 1:20:22.480 | |
If you want to localize to centimeters, | |
1:20:22.480 --> 1:20:23.720 | |
you can't use GPS. | |
1:20:23.720 --> 1:20:25.120 | |
The fancies GPS in the world can't do it, | |
1:20:25.120 --> 1:20:26.960 | |
especially if you're under tree cover and stuff. | |
1:20:26.960 --> 1:20:28.480 | |
LiDAR you can do this pretty easily. | |
1:20:28.480 --> 1:20:30.240 | |
So really they're not taking on, | |
1:20:30.240 --> 1:20:33.200 | |
I mean in some research they're using it for perception, | |
1:20:33.200 --> 1:20:35.840 | |
but and they're certainly not, which is sad, | |
1:20:35.840 --> 1:20:38.680 | |
they're not fusing it well with vision. | |
1:20:38.680 --> 1:20:40.560 | |
They do use it for perception. | |
1:20:40.560 --> 1:20:42.400 | |
I'm not saying they don't use it for perception, | |
1:20:42.400 --> 1:20:45.480 | |
but the thing that they have vision based | |
1:20:45.480 --> 1:20:47.680 | |
and radar based perception systems as well. | |
1:20:47.680 --> 1:20:51.440 | |
You could remove the LiDAR and keep around | |
1:20:51.440 --> 1:20:54.040 | |
a lot of the dynamic object perception. | |
1:20:54.040 --> 1:20:56.320 | |
You want to get centimeter accurate localization. | |
1:20:56.320 --> 1:20:59.120 | |
Good luck doing that with anything else. | |
1:20:59.120 --> 1:21:02.880 | |
So what should a cruise Waymo do? | |
1:21:02.880 --> 1:21:05.360 | |
Like what would be your advice to them now? | |
1:21:06.400 --> 1:21:11.400 | |
I mean Waymo is actually, they're serious. | |
1:21:11.400 --> 1:21:13.120 | |
Waymo out of the ball of them, | |
1:21:13.120 --> 1:21:16.120 | |
are quite serious about the long game. | |
1:21:16.120 --> 1:21:20.680 | |
If L5 is a lot, is requires 50 years, | |
1:21:20.680 --> 1:21:24.000 | |
I think Waymo will be the only one left standing at the end | |
1:21:24.000 --> 1:21:26.560 | |
with a given the financial backing that they have. | |
1:21:26.560 --> 1:21:28.640 | |
They're boo Google box. | |
1:21:28.640 --> 1:21:31.040 | |
I'll say nice things about both Waymo and cruise. | |
1:21:32.320 --> 1:21:33.480 | |
Let's do it. | |
1:21:33.480 --> 1:21:34.320 | |
Nice is good. | |
1:21:35.720 --> 1:21:39.200 | |
Waymo is by far the furthest along with technology. | |
1:21:39.200 --> 1:21:41.160 | |
Waymo has a three to five year lead | |
1:21:41.160 --> 1:21:42.880 | |
on all the competitors. | |
1:21:43.960 --> 1:21:48.640 | |
If the Waymo looking stack works, | |
1:21:48.640 --> 1:21:49.720 | |
maybe three year lead. | |
1:21:49.720 --> 1:21:51.280 | |
If the Waymo looking stack works, | |
1:21:51.280 --> 1:21:52.800 | |
they have a three year lead. | |
1:21:52.800 --> 1:21:55.800 | |
Now, I argue that Waymo has spent too much money | |
1:21:55.800 --> 1:21:59.240 | |
to recapitalize, to gain back their losses | |
1:21:59.240 --> 1:22:00.160 | |
in those three years. | |
1:22:00.160 --> 1:22:03.600 | |
Also self driving cars have no network effect like that. | |
1:22:03.600 --> 1:22:04.800 | |
Uber has a network effect. | |
1:22:04.800 --> 1:22:07.120 | |
You have a market, you have drivers and you have riders. | |
1:22:07.120 --> 1:22:09.880 | |
Self driving cars, you have capital and you have riders. | |
1:22:09.880 --> 1:22:11.400 | |
There's no network effect. | |
1:22:11.400 --> 1:22:13.800 | |
If I want to blanket a new city in self driving cars, | |
1:22:13.800 --> 1:22:16.000 | |
I buy the off the shelf Chinese knockoff self driving cars | |
1:22:16.000 --> 1:22:17.160 | |
and I buy enough of them in the city. | |
1:22:17.160 --> 1:22:18.360 | |
I can't do that with drivers. | |
1:22:18.360 --> 1:22:20.840 | |
And that's why Uber has a first mover advantage | |
1:22:20.840 --> 1:22:22.640 | |
that no self driving car company will. | |
1:22:23.960 --> 1:22:26.520 | |
Can you just a thing, let a little bit. | |
1:22:26.520 --> 1:22:28.160 | |
Uber, you're not talking about Uber, | |
1:22:28.160 --> 1:22:29.240 | |
the autonomous vehicle Uber. | |
1:22:29.240 --> 1:22:30.960 | |
You're talking about the Uber cars. | |
1:22:30.960 --> 1:22:31.800 | |
Yeah. | |
1:22:31.800 --> 1:22:32.640 | |
I'm Uber. | |
1:22:32.640 --> 1:22:35.920 | |
I open for business in Austin, Texas, let's say. | |
1:22:35.920 --> 1:22:38.760 | |
I need to attract both sides of the market. | |
1:22:38.760 --> 1:22:41.200 | |
I need to both get drivers on my platform | |
1:22:41.200 --> 1:22:42.720 | |
and riders on my platform. | |
1:22:42.720 --> 1:22:45.320 | |
And I need to keep them both sufficiently happy, right? | |
1:22:45.320 --> 1:22:46.520 | |
Riders aren't going to use it | |
1:22:46.520 --> 1:22:48.960 | |
if it takes more than five minutes for an Uber to show up. | |
1:22:48.960 --> 1:22:50.120 | |
Drivers aren't going to use it | |
1:22:50.120 --> 1:22:52.120 | |
if they have to sit around all day and there's no riders. | |
1:22:52.120 --> 1:22:54.480 | |
So you have to carefully balance a market. | |
1:22:54.480 --> 1:22:56.240 | |
And whenever you have to carefully balance a market, | |
1:22:56.240 --> 1:22:58.280 | |
there's a great first mover advantage | |
1:22:58.280 --> 1:23:01.000 | |
because there's a switching cost for everybody, right? | |
1:23:01.000 --> 1:23:02.120 | |
The drivers and the riders | |
1:23:02.120 --> 1:23:04.080 | |
would have to switch at the same time. | |
1:23:04.080 --> 1:23:08.880 | |
Let's even say that, let's say, Uber shows up. | |
1:23:08.880 --> 1:23:13.880 | |
And Uber somehow agrees to do things at a bigger, | |
1:23:14.800 --> 1:23:17.440 | |
we've done it more efficiently, right? | |
1:23:17.440 --> 1:23:19.800 | |
Uber only takes 5% of a car | |
1:23:19.800 --> 1:23:21.600 | |
instead of the 10% that Uber takes. | |
1:23:21.600 --> 1:23:22.760 | |
No one is going to switch | |
1:23:22.760 --> 1:23:24.920 | |
because the switching cost is higher than that 5%. | |
1:23:24.920 --> 1:23:27.200 | |
So you actually can, in markets like that, | |
1:23:27.200 --> 1:23:28.520 | |
you have a first mover advantage. | |
1:23:28.520 --> 1:23:29.360 | |
Yeah. | |
1:23:30.160 --> 1:23:32.720 | |
Autonomous vehicles of the level five variety | |
1:23:32.720 --> 1:23:34.560 | |
have no first mover advantage. | |
1:23:34.560 --> 1:23:36.800 | |
If the technology becomes commoditized, | |
1:23:36.800 --> 1:23:39.520 | |
say I want to go to a new city, look at the scooters. | |
1:23:39.520 --> 1:23:41.480 | |
It's going to look a lot more like scooters. | |
1:23:41.480 --> 1:23:44.040 | |
Every person with a checkbook | |
1:23:44.040 --> 1:23:45.720 | |
can blanket a city in scooters | |
1:23:45.720 --> 1:23:47.920 | |
and that's why you have 10 different scooter companies. | |
1:23:47.920 --> 1:23:48.760 | |
Which one's going to win? | |
1:23:48.760 --> 1:23:49.600 | |
It's a race to the bottom. | |
1:23:49.600 --> 1:23:51.040 | |
It's a terrible market to be in | |
1:23:51.040 --> 1:23:53.160 | |
because there's no market for scooters. | |
1:23:54.960 --> 1:23:56.520 | |
And the scooters don't get a say | |
1:23:56.520 --> 1:23:57.480 | |
in whether they want to be bought | |
1:23:57.480 --> 1:23:58.440 | |
and deployed to a city or not. | |
1:23:58.440 --> 1:23:59.280 | |
Right. | |
1:23:59.280 --> 1:24:00.120 | |
So yeah. | |
1:24:00.120 --> 1:24:02.080 | |
We're going to entice the scooters with subsidies | |
1:24:02.080 --> 1:24:02.920 | |
and deals. | |
1:24:03.840 --> 1:24:05.480 | |
So whenever you have to invest that capital, | |
1:24:05.480 --> 1:24:06.720 | |
it doesn't... | |
1:24:06.720 --> 1:24:07.560 | |
It doesn't come back. | |
1:24:07.560 --> 1:24:08.600 | |
Yeah. | |
1:24:08.600 --> 1:24:12.320 | |
They can't be your main criticism of the Waymo approach. | |
1:24:12.320 --> 1:24:14.840 | |
Oh, I'm saying even if it does technically work. | |
1:24:14.840 --> 1:24:17.040 | |
Even if it does technically work, that's a problem. | |
1:24:17.040 --> 1:24:18.000 | |
Yeah. | |
1:24:18.000 --> 1:24:21.720 | |
I don't know if I were to say, I would say, | |
1:24:22.840 --> 1:24:23.520 | |
you're already there. | |
1:24:23.520 --> 1:24:24.560 | |
I haven't even thought about that. | |
1:24:24.560 --> 1:24:26.520 | |
But I would say the bigger challenge | |
1:24:26.520 --> 1:24:27.760 | |
is the technical approach. | |
1:24:29.760 --> 1:24:31.840 | |
So Waymo's cruise is... | |
1:24:31.840 --> 1:24:33.000 | |
And not just the technical approach, | |
1:24:33.000 --> 1:24:34.800 | |
but of creating value. | |
1:24:34.800 --> 1:24:39.800 | |
I still don't understand how you beat Uber, | |
1:24:40.760 --> 1:24:43.480 | |
the human driven cars. | |
1:24:43.480 --> 1:24:44.920 | |
In terms of financially, | |
1:24:44.920 --> 1:24:47.160 | |
it doesn't make sense to me | |
1:24:47.160 --> 1:24:50.080 | |
that people want to get an autonomous vehicle. | |
1:24:50.080 --> 1:24:52.800 | |
I don't understand how you make money. | |
1:24:52.800 --> 1:24:56.440 | |
In the long term, yes, like real long term, | |
1:24:56.440 --> 1:24:58.640 | |
but it just feels like there's too much | |
1:24:58.640 --> 1:24:59.960 | |
capital investment needed. | |
1:24:59.960 --> 1:25:01.200 | |
Oh, and they're going to be worse than Ubers | |
1:25:01.200 --> 1:25:02.440 | |
because they're going to stop | |
1:25:02.440 --> 1:25:04.760 | |
for every little thing everywhere. | |
1:25:06.320 --> 1:25:07.360 | |
I'll say a nice thing about cruise. | |
1:25:07.360 --> 1:25:08.440 | |
That was my nice thing about Waymo. | |
1:25:08.440 --> 1:25:09.280 | |
They're three years ahead of me. | |
1:25:09.280 --> 1:25:10.120 | |
It was a nice... | |
1:25:10.120 --> 1:25:10.960 | |
Oh, because they're three years. | |
1:25:10.960 --> 1:25:12.480 | |
They're three years technically ahead of everybody. | |
1:25:12.480 --> 1:25:13.960 | |
Their tech stack is great. | |
1:25:14.800 --> 1:25:17.920 | |
My nice thing about cruise is GM buying them | |
1:25:17.920 --> 1:25:19.160 | |
was a great move for GM. | |
1:25:20.600 --> 1:25:22.240 | |
For $1 billion, | |
1:25:22.240 --> 1:25:25.600 | |
GM bought an insurance policy against Waymo. | |
1:25:26.560 --> 1:25:30.000 | |
They put cruise is three years behind Waymo. | |
1:25:30.000 --> 1:25:32.600 | |
That means Google will get a monopoly | |
1:25:32.600 --> 1:25:35.160 | |
on the technology for at most three years. | |
1:25:36.840 --> 1:25:38.880 | |
And if technology works, | |
1:25:38.880 --> 1:25:40.840 | |
you might not even be right about the three years. | |
1:25:40.840 --> 1:25:41.840 | |
It might be less. | |
1:25:41.840 --> 1:25:42.680 | |
Might be less. | |
1:25:42.680 --> 1:25:44.320 | |
Cruise actually might not be that far behind. | |
1:25:44.320 --> 1:25:47.360 | |
I don't know how much Waymo has waffled around | |
1:25:47.360 --> 1:25:49.760 | |
or how much of it actually is just that long tail. | |
1:25:49.760 --> 1:25:50.600 | |
Yeah, okay. | |
1:25:50.600 --> 1:25:53.600 | |
If that's the best you could say in terms of nice things, | |
1:25:53.600 --> 1:25:55.200 | |
that's more of a nice thing for GM | |
1:25:55.200 --> 1:25:58.560 | |
that that's a smart insurance policy. | |
1:25:58.560 --> 1:25:59.680 | |
It's a smart insurance policy. | |
1:25:59.680 --> 1:26:01.880 | |
I mean, I think that's how... | |
1:26:01.880 --> 1:26:05.200 | |
I can't see cruise working out any other. | |
1:26:05.200 --> 1:26:07.840 | |
For cruise to leapfrog Waymo would really surprise me. | |
1:26:10.400 --> 1:26:13.000 | |
Yeah, so let's talk about the underlying assumptions | |
1:26:13.000 --> 1:26:13.840 | |
of everything is... | |
1:26:13.840 --> 1:26:15.440 | |
We're not gonna leapfrog Tesla. | |
1:26:17.560 --> 1:26:19.240 | |
Tesla would have to seriously mess up | |
1:26:19.240 --> 1:26:20.440 | |
for us to leapfrog them. | |
1:26:20.440 --> 1:26:23.240 | |
Okay, so the way you leapfrog, right, | |
1:26:23.240 --> 1:26:26.120 | |
is you come up with an idea | |
1:26:26.120 --> 1:26:28.560 | |
or you take a direction, perhaps secretly, | |
1:26:28.560 --> 1:26:30.640 | |
that the other people aren't taking. | |
1:26:31.640 --> 1:26:36.640 | |
And so cruise, Waymo, even Aurora... | |
1:26:38.080 --> 1:26:40.080 | |
I don't know, Aurora, Zooks is the same stack as well. | |
1:26:40.080 --> 1:26:41.720 | |
They're all the same code base even. | |
1:26:41.720 --> 1:26:44.120 | |
They're all the same DARPA Urban Challenge code base. | |
1:26:44.120 --> 1:26:45.360 | |
It's... | |
1:26:45.360 --> 1:26:47.760 | |
So the question is, do you think there's a room | |
1:26:47.760 --> 1:26:49.120 | |
for brilliance and innovation there | |
1:26:49.120 --> 1:26:50.560 | |
that will change everything? | |
1:26:51.560 --> 1:26:53.880 | |
Like say, okay, so I'll give you examples. | |
1:26:53.880 --> 1:26:58.880 | |
It could be if revolution and mapping, for example, | |
1:26:59.640 --> 1:27:03.040 | |
that allow you to map things, | |
1:27:03.040 --> 1:27:05.840 | |
do HD maps of the whole world, | |
1:27:05.840 --> 1:27:08.080 | |
all weather conditions somehow really well, | |
1:27:08.080 --> 1:27:13.080 | |
or revolution and simulation, | |
1:27:14.480 --> 1:27:18.840 | |
to where all the way you said before becomes incorrect. | |
1:27:20.480 --> 1:27:21.520 | |
That kind of thing. | |
1:27:21.520 --> 1:27:23.920 | |
Any room for breakthrough innovation? | |
1:27:24.920 --> 1:27:25.960 | |
What I said before about, | |
1:27:25.960 --> 1:27:28.280 | |
oh, they actually get the whole thing, well, | |
1:27:28.280 --> 1:27:32.600 | |
I'll say this about we divide driving into three problems. | |
1:27:32.600 --> 1:27:33.800 | |
And I actually haven't solved the third yet, | |
1:27:33.800 --> 1:27:34.800 | |
but I haven't had any idea how to do it. | |
1:27:34.800 --> 1:27:36.120 | |
So there's the static. | |
1:27:36.120 --> 1:27:38.000 | |
The static driving problem is assuming | |
1:27:38.000 --> 1:27:40.120 | |
you are the only car on the road, right? | |
1:27:40.120 --> 1:27:42.000 | |
And this problem can be solved 100% | |
1:27:42.000 --> 1:27:44.000 | |
with mapping and localization. | |
1:27:44.000 --> 1:27:45.760 | |
This is why farms work the way they do. | |
1:27:45.760 --> 1:27:48.440 | |
If all you have to deal with is the static problem, | |
1:27:48.440 --> 1:27:50.160 | |
and you can statically schedule your machines, right? | |
1:27:50.160 --> 1:27:52.680 | |
It's the same as like statically scheduling processes. | |
1:27:52.680 --> 1:27:54.040 | |
You can statically schedule your tractors | |
1:27:54.040 --> 1:27:56.160 | |
to never hit each other on their paths, right? | |
1:27:56.160 --> 1:27:57.520 | |
Because then you know the speed they go at. | |
1:27:57.520 --> 1:28:00.160 | |
So that's the static driving problem. | |
1:28:00.160 --> 1:28:03.160 | |
Maps only helps you with the static driving problem. | |
1:28:03.920 --> 1:28:06.960 | |
Yeah, the question about static driving, | |
1:28:06.960 --> 1:28:08.800 | |
you've just made it sound like it's really easy. | |
1:28:08.800 --> 1:28:10.160 | |
Static driving is really easy. | |
1:28:11.880 --> 1:28:13.040 | |
How easy? | |
1:28:13.040 --> 1:28:16.480 | |
How, well, because the whole drifting out of lane, | |
1:28:16.480 --> 1:28:18.760 | |
when Tesla drifts out of lane, | |
1:28:18.760 --> 1:28:21.960 | |
it's failing on the fundamental static driving problem. | |
1:28:21.960 --> 1:28:24.440 | |
Tesla is drifting out of lane? | |
1:28:24.440 --> 1:28:27.720 | |
The static driving problem is not easy for the world. | |
1:28:27.720 --> 1:28:30.320 | |
The static driving problem is easy for one route. | |
1:28:31.840 --> 1:28:33.920 | |
One route in one weather condition | |
1:28:33.920 --> 1:28:37.920 | |
with one state of lane markings | |
1:28:37.920 --> 1:28:40.920 | |
and like no deterioration, no cracks in the road. | |
1:28:40.920 --> 1:28:42.600 | |
Well, I'm assuming you have a perfect localizer. | |
1:28:42.600 --> 1:28:44.200 | |
So that's all for the weather condition | |
1:28:44.200 --> 1:28:45.560 | |
and the lane marking condition. | |
1:28:45.560 --> 1:28:46.640 | |
But that's the problem. | |
1:28:46.640 --> 1:28:47.680 | |
How do you have a perfect localizer? | |
1:28:47.680 --> 1:28:50.560 | |
You can build, perfect localizers are not that hard to build. | |
1:28:50.560 --> 1:28:53.360 | |
Okay, come on now, with LIDAR. | |
1:28:53.360 --> 1:28:54.200 | |
LIDAR, yeah. | |
1:28:54.200 --> 1:28:55.040 | |
With LIDAR, okay. | |
1:28:55.040 --> 1:28:56.440 | |
LIDAR, yeah, but you use LIDAR, right? | |
1:28:56.440 --> 1:28:58.640 | |
Like you use LIDAR, build a perfect localizer. | |
1:28:58.640 --> 1:29:00.960 | |
Building a perfect localizer without LIDAR, | |
1:29:03.000 --> 1:29:04.320 | |
it's gonna be hard. | |
1:29:04.320 --> 1:29:05.760 | |
You can get 10 centimeters without LIDAR, | |
1:29:05.760 --> 1:29:07.240 | |
you can get one centimeter with LIDAR. | |
1:29:07.240 --> 1:29:09.280 | |
I'm not even concerned about the one or 10 centimeters. | |
1:29:09.280 --> 1:29:12.680 | |
I'm concerned if every once in a while you just weigh off. | |
1:29:12.680 --> 1:29:17.480 | |
Yeah, so this is why you have to carefully | |
1:29:17.480 --> 1:29:20.040 | |
make sure you're always tracking your position. | |
1:29:20.040 --> 1:29:21.760 | |
You wanna use LIDAR camera fusion, | |
1:29:21.760 --> 1:29:24.480 | |
but you can get the reliability of that system | |
1:29:24.480 --> 1:29:28.000 | |
up to 100,000 miles | |
1:29:28.000 --> 1:29:29.720 | |
and then you write some fallback condition | |
1:29:29.720 --> 1:29:32.160 | |
where it's not that bad if you're way off, right? | |
1:29:32.160 --> 1:29:33.800 | |
I think that you can get it to the point, | |
1:29:33.800 --> 1:29:36.800 | |
it's like ASL D that you're never in a case | |
1:29:36.800 --> 1:29:38.480 | |
where you're way off and you don't know it. | |
1:29:38.480 --> 1:29:40.240 | |
Yeah, okay, so this is brilliant. | |
1:29:40.240 --> 1:29:41.160 | |
So that's the static. | |
1:29:41.160 --> 1:29:42.280 | |
Static. | |
1:29:42.280 --> 1:29:45.960 | |
We can, especially with LIDAR and good HD maps, | |
1:29:45.960 --> 1:29:47.080 | |
you can solve that problem. | |
1:29:47.080 --> 1:29:47.920 | |
It's easy. | |
1:29:47.920 --> 1:29:51.840 | |
The static, the static problem is so easy. | |
1:29:51.840 --> 1:29:54.000 | |
It's very typical for you to say something's easy. | |
1:29:54.000 --> 1:29:54.840 | |
I got it. | |
1:29:54.840 --> 1:29:56.920 | |
It's not as challenging as the other ones, okay. | |
1:29:56.920 --> 1:29:58.760 | |
Well, okay, maybe it's obvious how to solve it. | |
1:29:58.760 --> 1:29:59.760 | |
The third one's the hardest. | |
1:29:59.760 --> 1:30:01.920 | |
And a lot of people don't even think about the third one | |
1:30:01.920 --> 1:30:03.640 | |
and even see it as different from the second one. | |
1:30:03.640 --> 1:30:05.720 | |
So the second one is dynamic. | |
1:30:05.720 --> 1:30:08.560 | |
The second one is like, say there's an obvious example, | |
1:30:08.560 --> 1:30:10.360 | |
it's like a car stopped at a red light, right? | |
1:30:10.360 --> 1:30:12.520 | |
You can't have that car in your map | |
1:30:12.520 --> 1:30:13.720 | |
because you don't know whether that car | |
1:30:13.720 --> 1:30:14.880 | |
is gonna be there or not. | |
1:30:14.880 --> 1:30:17.960 | |
So you have to detect that car in real time | |
1:30:17.960 --> 1:30:21.600 | |
and then you have to do the appropriate action, right? | |
1:30:21.600 --> 1:30:24.800 | |
Also, that car is not a fixed object. | |
1:30:24.800 --> 1:30:26.600 | |
That car may move and you have to predict | |
1:30:26.600 --> 1:30:28.680 | |
what that car will do, right? | |
1:30:28.680 --> 1:30:30.840 | |
So this is the dynamic problem. | |
1:30:30.840 --> 1:30:31.680 | |
Yeah. | |
1:30:31.680 --> 1:30:32.800 | |
So you have to deal with this. | |
1:30:32.800 --> 1:30:36.640 | |
This involves, again, like you're gonna need models | |
1:30:36.640 --> 1:30:38.760 | |
of other people's behavior. | |
1:30:38.760 --> 1:30:40.160 | |
Do you, are you including in that? | |
1:30:40.160 --> 1:30:42.320 | |
I don't wanna step on the third one. | |
1:30:42.320 --> 1:30:46.600 | |
Oh, but are you including in that your influence | |
1:30:46.600 --> 1:30:47.440 | |
on people? | |
1:30:47.440 --> 1:30:48.280 | |
Ah, that's the third one. | |
1:30:48.280 --> 1:30:49.120 | |
Okay. | |
1:30:49.120 --> 1:30:49.960 | |
That's the third one. | |
1:30:49.960 --> 1:30:51.880 | |
We call it the counterfactual. | |
1:30:51.880 --> 1:30:52.720 | |
Yeah, brilliant. | |
1:30:52.720 --> 1:30:53.560 | |
And that. | |
1:30:53.560 --> 1:30:54.920 | |
I just talked to Judea Pro who's obsessed | |
1:30:54.920 --> 1:30:55.760 | |
with counterfactuals. | |
1:30:55.760 --> 1:30:58.640 | |
Counterfactual, oh yeah, yeah, I read his books. | |
1:30:58.640 --> 1:31:03.640 | |
So the static and the dynamic are our approach right now | |
1:31:03.960 --> 1:31:07.600 | |
for lateral will scale completely to the static and dynamic. | |
1:31:07.600 --> 1:31:10.760 | |
The counterfactual, the only way I have to do it yet, | |
1:31:10.760 --> 1:31:14.000 | |
the thing that I wanna do once we have all of these cars | |
1:31:14.000 --> 1:31:16.760 | |
is I wanna do reinforcement learning on the world. | |
1:31:16.760 --> 1:31:18.880 | |
I'm always gonna turn the exploiter up to max. | |
1:31:18.880 --> 1:31:20.440 | |
I'm not gonna have them explore. | |
1:31:20.440 --> 1:31:22.760 | |
But the only real way to get at the counterfactual | |
1:31:22.760 --> 1:31:24.080 | |
is to do reinforcement learning | |
1:31:24.080 --> 1:31:26.360 | |
because the other agents are humans. | |
1:31:27.760 --> 1:31:30.080 | |
So that's fascinating that you break it down like that. | |
1:31:30.080 --> 1:31:31.680 | |
I agree completely. | |
1:31:31.680 --> 1:31:33.600 | |
I've spent my life thinking about this problem. | |
1:31:33.600 --> 1:31:34.920 | |
This is beautiful. | |
1:31:34.920 --> 1:31:37.880 | |
And part of it, cause you're slightly insane, | |
1:31:37.880 --> 1:31:42.880 | |
because not my life, just the last four years. | |
1:31:43.120 --> 1:31:48.120 | |
No, no, you have some non zero percent of your brain | |
1:31:48.920 --> 1:31:52.360 | |
has a madman in it, which is a really good feature. | |
1:31:52.360 --> 1:31:55.920 | |
But there's a safety component to it | |
1:31:55.920 --> 1:31:57.320 | |
that I think when there's sort of | |
1:31:57.320 --> 1:31:59.040 | |
with counterfactuals and so on, | |
1:31:59.040 --> 1:32:00.280 | |
that would just freak people out. | |
1:32:00.280 --> 1:32:03.320 | |
How do you even start to think about this in general? | |
1:32:03.320 --> 1:32:07.600 | |
I mean, you've had some friction with NHTSA and so on. | |
1:32:07.600 --> 1:32:12.600 | |
I am frankly exhausted by safety engineers. | |
1:32:14.280 --> 1:32:19.280 | |
The prioritization on safety over innovation | |
1:32:21.360 --> 1:32:23.720 | |
to a degree where it kills, in my view, | |
1:32:23.720 --> 1:32:26.200 | |
kills safety in the longterm. | |
1:32:26.200 --> 1:32:28.080 | |
So the counterfactual thing, | |
1:32:28.080 --> 1:32:31.560 | |
they just actually exploring this world | |
1:32:31.560 --> 1:32:33.600 | |
of how do you interact with dynamic objects and so on? | |
1:32:33.600 --> 1:32:34.840 | |
How do you think about safety? | |
1:32:34.840 --> 1:32:38.120 | |
You can do reinforcement learning without ever exploring. | |
1:32:38.120 --> 1:32:39.200 | |
And I said that, like, | |
1:32:39.200 --> 1:32:41.560 | |
so you can think about your, in like reinforcement learning, | |
1:32:41.560 --> 1:32:44.320 | |
it's usually called like a temperature parameter. | |
1:32:44.320 --> 1:32:45.360 | |
And your temperature parameter | |
1:32:45.360 --> 1:32:48.080 | |
is how often you deviate from the argmax. | |
1:32:48.080 --> 1:32:50.720 | |
I could always set that to zero and still learn. | |
1:32:50.720 --> 1:32:52.840 | |
And I feel that you'd always want that set to zero | |
1:32:52.840 --> 1:32:54.080 | |
on your actual system. | |
1:32:54.080 --> 1:32:54.920 | |
Gotcha. | |
1:32:54.920 --> 1:32:58.160 | |
But the problem is you first don't know very much | |
1:32:58.160 --> 1:32:59.560 | |
and so you're going to make mistakes. | |
1:32:59.560 --> 1:33:01.680 | |
So the learning, the exploration happens through mistakes. | |
1:33:01.680 --> 1:33:03.240 | |
We're all ready, yeah, but. | |
1:33:03.240 --> 1:33:06.080 | |
Okay, so the consequences of a mistake. | |
1:33:06.080 --> 1:33:09.400 | |
OpenPilot and Autopilot are making mistakes left and right. | |
1:33:09.400 --> 1:33:12.560 | |
We have 700 daily active users, | |
1:33:12.560 --> 1:33:14.080 | |
1,000 weekly active users. | |
1:33:14.080 --> 1:33:18.920 | |
OpenPilot makes tens of thousands of mistakes a week. | |
1:33:18.920 --> 1:33:21.160 | |
These mistakes have zero consequences. | |
1:33:21.160 --> 1:33:22.520 | |
These mistakes are, | |
1:33:22.520 --> 1:33:26.800 | |
oh, I wanted to take this exit and it went straight. | |
1:33:26.800 --> 1:33:28.520 | |
So I'm just going to carefully touch the wheel. | |
1:33:28.520 --> 1:33:29.360 | |
The humans catch them. | |
1:33:29.360 --> 1:33:30.640 | |
The humans catch them. | |
1:33:30.640 --> 1:33:33.120 | |
And the human disengagement is labeling | |
1:33:33.120 --> 1:33:35.000 | |
that reinforcement learning in a completely | |
1:33:35.000 --> 1:33:36.200 | |
consequence free way. | |
1:33:37.240 --> 1:33:39.840 | |
So driver monitoring is the way you ensure they keep. | |
1:33:39.840 --> 1:33:40.680 | |
Yes. | |
1:33:40.680 --> 1:33:42.120 | |
They keep paying attention. | |
1:33:42.120 --> 1:33:43.240 | |
How's your messaging? | |
1:33:43.240 --> 1:33:45.200 | |
Say I gave you a billion dollars, | |
1:33:45.200 --> 1:33:46.960 | |
so you would be scaling it now. | |
1:33:47.800 --> 1:33:49.720 | |
Oh, if I could scale, I couldn't scale with any amount of money. | |
1:33:49.720 --> 1:33:51.640 | |
I'd raise money if I could, if I had a way to scale it. | |
1:33:51.640 --> 1:33:53.320 | |
Yeah, you're not, no, I'm not focused on scale. | |
1:33:53.320 --> 1:33:54.160 | |
I don't know how to do. | |
1:33:54.160 --> 1:33:55.760 | |
Oh, like, I guess I could sell it to more people, | |
1:33:55.760 --> 1:33:56.960 | |
but I want to make the system better. | |
1:33:56.960 --> 1:33:57.800 | |
Better, better. | |
1:33:57.800 --> 1:33:58.840 | |
And I don't know how to. | |
1:33:58.840 --> 1:34:01.080 | |
But what's the messaging here? | |
1:34:01.080 --> 1:34:02.560 | |
I got a chance to talk to Elon. | |
1:34:02.560 --> 1:34:07.560 | |
And he basically said that the human factor doesn't matter. | |
1:34:09.280 --> 1:34:10.360 | |
You know, the human doesn't matter | |
1:34:10.360 --> 1:34:12.280 | |
because the system will perform. | |
1:34:12.280 --> 1:34:14.760 | |
There'll be sort of a, sorry to use the term, | |
1:34:14.760 --> 1:34:16.120 | |
but like a singular, like a point | |
1:34:16.120 --> 1:34:17.920 | |
where it gets just much better. | |
1:34:17.920 --> 1:34:20.800 | |
And so the human, it won't really matter. | |
1:34:20.800 --> 1:34:25.000 | |
But it seems like that human catching the system | |
1:34:25.000 --> 1:34:29.360 | |
when it gets into trouble is like the thing | |
1:34:29.360 --> 1:34:32.720 | |
which will make something like reinforcement learning work. | |
1:34:32.720 --> 1:34:35.640 | |
So how do you, how do you think messaging for Tesla, | |
1:34:35.640 --> 1:34:39.080 | |
for you, for the industry in general, should change? | |
1:34:39.080 --> 1:34:40.840 | |
I think my messaging is pretty clear, | |
1:34:40.840 --> 1:34:43.080 | |
at least like our messaging wasn't that clear | |
1:34:43.080 --> 1:34:45.200 | |
in the beginning and I do kind of fault myself for that. | |
1:34:45.200 --> 1:34:48.480 | |
We are proud right now to be a level two system. | |
1:34:48.480 --> 1:34:50.360 | |
We are proud to be level two. | |
1:34:50.360 --> 1:34:51.640 | |
If we talk about level four, | |
1:34:51.640 --> 1:34:53.200 | |
it's not with the current hardware. | |
1:34:53.200 --> 1:34:55.920 | |
It's not going to be just a magical OTA upgrade. | |
1:34:55.920 --> 1:34:57.280 | |
It's going to be new hardware. | |
1:34:57.280 --> 1:35:00.000 | |
It's going to be very carefully thought out right now. | |
1:35:00.000 --> 1:35:01.560 | |
We are proud to be level two. | |
1:35:01.560 --> 1:35:03.320 | |
And we have a rigorous safety model. | |
1:35:03.320 --> 1:35:05.680 | |
I mean, not like, like, okay, rigorous. | |
1:35:05.680 --> 1:35:06.600 | |
Who knows what that means? | |
1:35:06.600 --> 1:35:08.600 | |
But we at least have a safety model | |
1:35:08.600 --> 1:35:09.560 | |
and we make it explicit. | |
1:35:09.560 --> 1:35:11.800 | |
It's in safety.md and open pilot. | |
1:35:11.800 --> 1:35:13.960 | |
And it says, seriously though. | |
1:35:13.960 --> 1:35:14.800 | |
Safety.md. | |
1:35:14.800 --> 1:35:15.840 | |
Safety.md. | |
1:35:16.840 --> 1:35:18.400 | |
This is really, this is so Android. | |
1:35:18.400 --> 1:35:21.800 | |
So, well, this is, this is the safety model | |
1:35:21.800 --> 1:35:25.520 | |
and I like to have conversations like if, like, you know, | |
1:35:25.520 --> 1:35:27.120 | |
sometimes people will come to you and they're like, | |
1:35:27.120 --> 1:35:29.240 | |
your system's not safe. | |
1:35:29.240 --> 1:35:30.080 | |
Okay. | |
1:35:30.080 --> 1:35:31.080 | |
Have you read my safety docs? | |
1:35:31.080 --> 1:35:32.720 | |
Would you like to have an intelligent conversation | |
1:35:32.720 --> 1:35:33.560 | |
about this? | |
1:35:33.560 --> 1:35:34.400 | |
And the answer is always no. | |
1:35:34.400 --> 1:35:36.880 | |
They just like scream about, it runs Python. | |
1:35:38.240 --> 1:35:39.080 | |
Okay. What? | |
1:35:39.080 --> 1:35:41.560 | |
So you're saying that, that because Python's not real time, | |
1:35:41.560 --> 1:35:44.240 | |
Python not being real time never causes disengagement. | |
1:35:44.240 --> 1:35:47.640 | |
Disengagement's are caused by, you know, the model is QM. | |
1:35:47.640 --> 1:35:49.760 | |
But safety.md says the following. | |
1:35:49.760 --> 1:35:50.600 | |
First and foremost, | |
1:35:50.600 --> 1:35:53.000 | |
the driver must be paying attention at all times. | |
1:35:54.240 --> 1:35:55.320 | |
I don't consider, I do, | |
1:35:55.320 --> 1:35:57.720 | |
I still consider the software to be alpha software | |
1:35:57.720 --> 1:36:00.080 | |
until we can actually enforce that statement. | |
1:36:00.080 --> 1:36:03.280 | |
But I feel it's very well communicated to our users. | |
1:36:03.280 --> 1:36:04.520 | |
Two more things. | |
1:36:04.520 --> 1:36:09.080 | |
One is the user must be able to easily take control | |
1:36:09.080 --> 1:36:10.880 | |
of the vehicle at all times. | |
1:36:10.880 --> 1:36:14.440 | |
So if you step on the gas or brake with open pilot, | |
1:36:14.440 --> 1:36:16.400 | |
it gives full manual control back to the user | |
1:36:16.400 --> 1:36:18.680 | |
or press the cancel button. | |
1:36:18.680 --> 1:36:23.240 | |
Step two, the car will never react so quickly. | |
1:36:23.240 --> 1:36:26.000 | |
We define so quickly to be about one second | |
1:36:26.000 --> 1:36:27.640 | |
that you can't react in time. | |
1:36:27.640 --> 1:36:29.480 | |
And we do this by enforcing torque limits, | |
1:36:29.480 --> 1:36:31.520 | |
braking limits and acceleration limits. | |
1:36:31.520 --> 1:36:36.520 | |
So we have like our torque limits way lower than Tesla's. | |
1:36:36.520 --> 1:36:39.080 | |
This is another potential. | |
1:36:39.080 --> 1:36:40.240 | |
If I could tweak autopilot, | |
1:36:40.240 --> 1:36:41.360 | |
I would lower their torque limit | |
1:36:41.360 --> 1:36:42.960 | |
and I would add driver monitoring. | |
1:36:42.960 --> 1:36:46.240 | |
Because autopilot can jerk the wheel hard. | |
1:36:46.240 --> 1:36:47.520 | |
Open pilot can't. | |
1:36:47.520 --> 1:36:52.080 | |
It's, we limit and all this code is open source, readable. | |
1:36:52.080 --> 1:36:54.880 | |
And I believe now it's all MISRA C compliant. | |
1:36:54.880 --> 1:36:55.800 | |
What's that mean? | |
1:36:57.080 --> 1:37:00.400 | |
MISRA is like the automotive coding standard. | |
1:37:00.400 --> 1:37:03.400 | |
At first, I've come to respect, | |
1:37:03.400 --> 1:37:04.960 | |
I've been reading like the standards lately | |
1:37:04.960 --> 1:37:05.920 | |
and I've come to respect them. | |
1:37:05.920 --> 1:37:07.800 | |
They're actually written by very smart people. | |
1:37:07.800 --> 1:37:09.920 | |
Yeah, they're brilliant people actually. | |
1:37:09.920 --> 1:37:11.320 | |
They have a lot of experience. | |
1:37:11.320 --> 1:37:13.360 | |
They're sometimes a little too cautious, | |
1:37:13.360 --> 1:37:16.800 | |
but in this case, it pays off. | |
1:37:16.800 --> 1:37:18.440 | |
MISRA is written by like computer scientists | |
1:37:18.440 --> 1:37:19.840 | |
and you can tell by the language they use. | |
1:37:19.840 --> 1:37:21.080 | |
You can tell by the language they use. | |
1:37:21.080 --> 1:37:24.440 | |
They talk about like whether certain conditions in MISRA | |
1:37:24.440 --> 1:37:26.520 | |
are decidable or undecidable. | |
1:37:26.520 --> 1:37:28.360 | |
And you mean like the halting problem? | |
1:37:28.360 --> 1:37:31.600 | |
And yes, all right, you've earned my respect. | |
1:37:31.600 --> 1:37:33.120 | |
I will read carefully what you have to say | |
1:37:33.120 --> 1:37:35.760 | |
and we want to make our code compliant with that. | |
1:37:35.760 --> 1:37:38.160 | |
All right, so you're proud level two, beautiful. | |
1:37:38.160 --> 1:37:42.320 | |
So you were the founder and I think CEO of Kama AI, | |
1:37:42.320 --> 1:37:44.320 | |
then you were the head of research. | |
1:37:44.320 --> 1:37:46.080 | |
What the heck are you now? | |
1:37:46.080 --> 1:37:47.480 | |
What's your connection to Kama AI? | |
1:37:47.480 --> 1:37:49.640 | |
I'm the president, but I'm one of those like | |
1:37:49.640 --> 1:37:53.440 | |
unelected presidents of like a small dictatorship country, | |
1:37:53.440 --> 1:37:55.200 | |
not one of those like elected presidents. | |
1:37:55.200 --> 1:37:57.640 | |
Oh, so you're like Putin when he was like the, yeah, | |
1:37:57.640 --> 1:37:58.980 | |
I got you. | |
1:37:58.980 --> 1:38:02.120 | |
So there's, what's the governance structure? | |
1:38:02.120 --> 1:38:04.800 | |
What's the future of Kama AI finance? | |
1:38:04.800 --> 1:38:08.120 | |
I mean, yeah, as a business, do you want, | |
1:38:08.120 --> 1:38:11.640 | |
are you just focused on getting things right now, | |
1:38:11.640 --> 1:38:14.920 | |
making some small amount of money in the meantime | |
1:38:14.920 --> 1:38:17.520 | |
and then when it works, it works a new scale. | |
1:38:17.520 --> 1:38:20.480 | |
Our burn rate is about 200 K a month | |
1:38:20.480 --> 1:38:23.040 | |
and our revenue is about 100 K a month. | |
1:38:23.040 --> 1:38:24.920 | |
So we need to forex our revenue, | |
1:38:24.920 --> 1:38:28.200 | |
but we haven't like tried very hard at that yet. | |
1:38:28.200 --> 1:38:30.160 | |
And the revenue is basically selling stuff online. | |
1:38:30.160 --> 1:38:32.360 | |
Yeah, we sell stuff shop.com.ai. | |
1:38:32.360 --> 1:38:33.920 | |
Is there other, well, okay. | |
1:38:33.920 --> 1:38:35.360 | |
So you'll have to figure out. | |
1:38:35.360 --> 1:38:37.880 | |
That's our only, see, but to me, | |
1:38:37.880 --> 1:38:40.400 | |
that's like respectable revenues. | |
1:38:40.400 --> 1:38:42.640 | |
We make it by selling products to consumers | |
1:38:42.640 --> 1:38:45.040 | |
for honest and transparent about what they are. | |
1:38:45.040 --> 1:38:49.000 | |
Most actually level four companies, right? | |
1:38:50.720 --> 1:38:54.320 | |
Cause you could easily start blowing up like smoke, | |
1:38:54.320 --> 1:38:57.080 | |
like overselling the hype and feeding into, | |
1:38:57.080 --> 1:38:59.080 | |
getting some fundraisers. | |
1:38:59.080 --> 1:39:00.520 | |
Oh, you're the guy, you're a genius | |
1:39:00.520 --> 1:39:01.800 | |
because you hacked the iPhone. | |
1:39:01.800 --> 1:39:02.920 | |
Oh, I hate that. | |
1:39:02.920 --> 1:39:03.760 | |
I hate that. | |
1:39:03.760 --> 1:39:06.360 | |
Yeah, I can trade my social capital for more money. | |
1:39:06.360 --> 1:39:07.320 | |
I did it once. | |
1:39:07.320 --> 1:39:10.320 | |
I almost regret it doing it the first time. | |
1:39:10.320 --> 1:39:11.640 | |
Well, on a small tangent, | |
1:39:11.640 --> 1:39:16.560 | |
what's your, you seem to not like fame | |
1:39:16.560 --> 1:39:18.840 | |
and yet you're also drawn to fame. | |
1:39:18.840 --> 1:39:23.840 | |
What's, where have you on, where are you on that currently? | |
1:39:24.560 --> 1:39:27.200 | |
Have you had some introspection, some soul searching? | |
1:39:27.200 --> 1:39:28.480 | |
Yeah. | |
1:39:28.480 --> 1:39:32.200 | |
I actually, I've come to a pretty stable position on that. | |
1:39:32.200 --> 1:39:33.880 | |
Like after the first time, | |
1:39:33.880 --> 1:39:36.840 | |
I realized that I don't want attention from the masses. | |
1:39:36.840 --> 1:39:39.160 | |
I want attention from people who I respect. | |
1:39:39.160 --> 1:39:41.960 | |
Who do you respect? | |
1:39:41.960 --> 1:39:43.960 | |
I can give a list of people. | |
1:39:43.960 --> 1:39:47.200 | |
So are these like Elon Musk type characters? | |
1:39:47.200 --> 1:39:49.040 | |
Yeah. | |
1:39:49.040 --> 1:39:50.000 | |
Actually, you know what? | |
1:39:50.000 --> 1:39:51.200 | |
I'll make it more broad than that. | |
1:39:51.200 --> 1:39:52.600 | |
I won't make it about a person. | |
1:39:52.600 --> 1:39:54.040 | |
I respect skill. | |
1:39:54.040 --> 1:39:56.880 | |
I respect people who have skills, right? | |
1:39:56.880 --> 1:40:00.280 | |
And I would like to like be, | |
1:40:00.280 --> 1:40:01.400 | |
I'm not gonna say famous, | |
1:40:01.400 --> 1:40:03.760 | |
but be like known among more people | |
1:40:03.760 --> 1:40:05.440 | |
who have like real skills. | |
1:40:05.440 --> 1:40:10.440 | |
Who in cars, do you think have skill? | |
1:40:12.560 --> 1:40:13.720 | |
Not do you respect? | |
1:40:15.000 --> 1:40:17.760 | |
Oh, Kyle Voat has skill. | |
1:40:17.760 --> 1:40:19.880 | |
A lot of people at Waymo have skill. | |
1:40:19.880 --> 1:40:20.840 | |
And I respect them. | |
1:40:20.840 --> 1:40:23.760 | |
I respect them as engineers. | |
1:40:23.760 --> 1:40:24.920 | |
Like I can think, I mean, | |
1:40:24.920 --> 1:40:26.280 | |
I think about all the times in my life | |
1:40:26.280 --> 1:40:27.960 | |
where I've been like dead set on approaches | |
1:40:27.960 --> 1:40:29.160 | |
and they turn out to be wrong. | |
1:40:29.160 --> 1:40:30.000 | |
Yeah. | |
1:40:30.000 --> 1:40:31.720 | |
So I mean, this might, I might be wrong. | |
1:40:31.720 --> 1:40:34.720 | |
I accept that, I accept that there's a decent chance | |
1:40:34.720 --> 1:40:36.600 | |
that I'm wrong. | |
1:40:36.600 --> 1:40:38.400 | |
And actually, I mean, having talked to Chris Armson, | |
1:40:38.400 --> 1:40:40.480 | |
Sterling Anderson, those guys, | |
1:40:40.480 --> 1:40:43.360 | |
I mean, I deeply respect Chris. | |
1:40:43.360 --> 1:40:44.640 | |
I just admire the guy. | |
1:40:46.040 --> 1:40:47.400 | |
He's legit. | |
1:40:47.400 --> 1:40:48.960 | |
When you drive a car through the desert | |
1:40:48.960 --> 1:40:52.400 | |
when everybody thinks it's impossible, that's legit. | |
1:40:52.400 --> 1:40:53.840 | |
And then I also really respect the people | |
1:40:53.840 --> 1:40:55.680 | |
who are like writing the infrastructure of the world, | |
1:40:55.680 --> 1:40:57.360 | |
like the Linus Torvalds and the Chris Ladin. | |
1:40:57.360 --> 1:40:59.080 | |
Oh yeah, they were doing the real work. | |
1:40:59.080 --> 1:41:00.800 | |
I know they're doing the real work. | |
1:41:02.000 --> 1:41:03.760 | |
Having talked to Chris Ladin, | |
1:41:03.760 --> 1:41:05.680 | |
you realize, especially when they're humble, | |
1:41:05.680 --> 1:41:07.680 | |
it's like, you realize, oh, you guys, | |
1:41:07.680 --> 1:41:09.640 | |
we're just using your... | |
1:41:09.640 --> 1:41:10.480 | |
Oh yeah. | |
1:41:10.480 --> 1:41:11.520 | |
All the hard work that you did. | |
1:41:11.520 --> 1:41:13.120 | |
Yeah, that's incredible. | |
1:41:13.120 --> 1:41:17.160 | |
What do you think, Mr. Anthony Lewandowski? | |
1:41:18.440 --> 1:41:21.640 | |
What do you, he's a, he's another mad genius. | |
1:41:21.640 --> 1:41:22.480 | |
Sharp guy. | |
1:41:22.480 --> 1:41:23.320 | |
Oh yeah. | |
1:41:23.320 --> 1:41:27.640 | |
What, do you think he might long term become a competitor? | |
1:41:27.640 --> 1:41:28.840 | |
Oh, to comma? | |
1:41:28.840 --> 1:41:32.400 | |
Well, so I think that he has the other right approach. | |
1:41:32.400 --> 1:41:35.280 | |
I think that right now, there's two right approaches. | |
1:41:35.280 --> 1:41:37.680 | |
One is what we're doing and one is what he's doing. | |
1:41:37.680 --> 1:41:39.800 | |
Can you describe, I think it's called Pronto AI, | |
1:41:39.800 --> 1:41:42.360 | |
he's starting using, do you know what the approach is? | |
1:41:42.360 --> 1:41:43.200 | |
I actually don't know. | |
1:41:43.200 --> 1:41:45.040 | |
Embark is also doing the same sort of thing. | |
1:41:45.040 --> 1:41:47.280 | |
The idea is almost that you want to, | |
1:41:47.280 --> 1:41:51.800 | |
so if you're, I can't partner with Honda and Toyota. | |
1:41:51.800 --> 1:41:56.800 | |
Honda and Toyota are like 400,000 person companies. | |
1:41:57.600 --> 1:41:59.400 | |
It's not even a company at that point. | |
1:41:59.400 --> 1:42:01.400 | |
Like I don't think of it like, I don't personify it. | |
1:42:01.400 --> 1:42:06.400 | |
I think of it like an object, but a trucker drives for a fleet. | |
1:42:07.120 --> 1:42:10.280 | |
Maybe that has like, some truckers are independent. | |
1:42:10.280 --> 1:42:12.080 | |
Some truckers drive for fleets with a hundred trucks. | |
1:42:12.080 --> 1:42:14.960 | |
There are tons of independent trucking companies out there. | |
1:42:14.960 --> 1:42:18.120 | |
Start a trucking company and drive your costs down | |
1:42:18.120 --> 1:42:23.120 | |
or figure out how to drive down the cost of trucking. | |
1:42:23.760 --> 1:42:26.560 | |
Another company that I really respect is Nauto. | |
1:42:26.560 --> 1:42:28.320 | |
Actually, I respect their business model. | |
1:42:28.320 --> 1:42:31.560 | |
Nauto sells a driver monitoring camera | |
1:42:31.560 --> 1:42:33.920 | |
and they sell it to fleet owners. | |
1:42:33.920 --> 1:42:38.920 | |
If I owned a fleet of cars and I could pay 40 bucks a month | |
1:42:39.120 --> 1:42:41.280 | |
to monitor my employees, | |
1:42:42.400 --> 1:42:45.520 | |
this is gonna like reduces accidents 18%. | |
1:42:45.520 --> 1:42:48.960 | |
It's so like that in the space, | |
1:42:48.960 --> 1:42:52.000 | |
that is like the business model that I like most respect | |
1:42:53.400 --> 1:42:55.360 | |
because they're creating value today. | |
1:42:55.360 --> 1:42:57.840 | |
Yeah, which is, that's a huge one. | |
1:42:57.840 --> 1:42:59.800 | |
How do we create value today with some of this? | |
1:42:59.800 --> 1:43:01.680 | |
And the length keeping thing is huge. | |
1:43:01.680 --> 1:43:03.800 | |
And it sounds like you're creeping in | |
1:43:03.800 --> 1:43:06.680 | |
or full steam ahead on the drive of monitoring too. | |
1:43:06.680 --> 1:43:09.240 | |
Which I think actually where the short term value, | |
1:43:09.240 --> 1:43:10.480 | |
if you can get right. | |
1:43:10.480 --> 1:43:12.800 | |
I still, I'm not a huge fan of the statement | |
1:43:12.800 --> 1:43:15.120 | |
that everything is to have drive of monitoring. | |
1:43:15.120 --> 1:43:16.120 | |
I agree with that completely, | |
1:43:16.120 --> 1:43:18.680 | |
but that statement usually misses the point | |
1:43:18.680 --> 1:43:21.920 | |
that to get the experience of it right is not trivial. | |
1:43:21.920 --> 1:43:22.840 | |
Oh, no, not at all. | |
1:43:22.840 --> 1:43:25.280 | |
In fact, like, so right now we have, | |
1:43:25.280 --> 1:43:28.480 | |
I think the timeout depends on speed of the car, | |
1:43:29.560 --> 1:43:32.520 | |
but we want to depend on like the scene state. | |
1:43:32.520 --> 1:43:35.440 | |
If you're on like an empty highway, | |
1:43:35.440 --> 1:43:37.680 | |
it's very different if you don't pay attention | |
1:43:37.680 --> 1:43:40.600 | |
than if like you're like coming up to a traffic light. | |
1:43:42.040 --> 1:43:45.720 | |
And long term, it should probably learn from the driver | |
1:43:45.720 --> 1:43:48.120 | |
because that's to do, I watched a lot of video. | |
1:43:48.120 --> 1:43:49.480 | |
We've built a smartphone detector | |
1:43:49.480 --> 1:43:51.520 | |
just to analyze how people are using smartphones | |
1:43:51.520 --> 1:43:53.400 | |
and people are using it very differently. | |
1:43:53.400 --> 1:43:57.760 | |
And there's this, it's a texting styles. | |
1:43:57.760 --> 1:44:00.320 | |
We haven't watched nearly enough of the videos. | |
1:44:00.320 --> 1:44:01.800 | |
We haven't, I got millions of miles | |
1:44:01.800 --> 1:44:02.960 | |
of people driving cars. | |
1:44:02.960 --> 1:44:05.960 | |
In this moment, I spent a large fraction of my time | |
1:44:05.960 --> 1:44:10.880 | |
just watching videos because it's never fails to learn. | |
1:44:10.880 --> 1:44:13.480 | |
Like I've never failed from a video watching session | |
1:44:13.480 --> 1:44:15.400 | |
to learn something I didn't know before. | |
1:44:15.400 --> 1:44:19.640 | |
In fact, I usually, like when I eat lunch, I'll sit, | |
1:44:19.640 --> 1:44:20.680 | |
especially when the weather is good | |
1:44:20.680 --> 1:44:22.080 | |
and just watch pedestrians. | |
1:44:22.080 --> 1:44:26.400 | |
With an eye to understand like from a computer vision eye, | |
1:44:26.400 --> 1:44:29.280 | |
just to see, can this model, can you predict | |
1:44:29.280 --> 1:44:30.480 | |
what are the decisions made? | |
1:44:30.480 --> 1:44:33.040 | |
And there's so many things that we don't understand. | |
1:44:33.040 --> 1:44:34.760 | |
This is what I mean about state vector. | |
1:44:34.760 --> 1:44:37.880 | |
Yeah, it's, I'm trying to always think like, | |
1:44:37.880 --> 1:44:40.280 | |
because I'm understanding in my human brain, | |
1:44:40.280 --> 1:44:42.000 | |
how do we convert that into, | |
1:44:43.000 --> 1:44:44.960 | |
how hard is the learning problem here? | |
1:44:44.960 --> 1:44:46.960 | |
I guess is the fundamental question. | |
1:44:46.960 --> 1:44:51.800 | |
So something that's from a hacking perspective, | |
1:44:51.800 --> 1:44:54.200 | |
this is always comes up, especially with folks. | |
1:44:54.200 --> 1:44:56.480 | |
Well, first, the most popular question is | |
1:44:56.480 --> 1:44:58.440 | |
the trolley problem, right? | |
1:44:58.440 --> 1:45:01.960 | |
So that's not a sort of a serious problem. | |
1:45:01.960 --> 1:45:05.000 | |
There are some ethical questions, I think that arise. | |
1:45:06.080 --> 1:45:09.600 | |
Maybe you wanna, do you think there's any ethical, | |
1:45:09.600 --> 1:45:11.280 | |
serious ethical questions? | |
1:45:11.280 --> 1:45:14.080 | |
We have a solution to the trolley problem at com.ai. | |
1:45:14.080 --> 1:45:15.920 | |
Well, so there is actually an alert | |
1:45:15.920 --> 1:45:18.000 | |
in our code, ethical dilemma detected. | |
1:45:18.000 --> 1:45:18.960 | |
It's not triggered yet. | |
1:45:18.960 --> 1:45:21.040 | |
We don't know how yet to detect the ethical dilemmas, | |
1:45:21.040 --> 1:45:22.360 | |
but we're a level two system. | |
1:45:22.360 --> 1:45:23.760 | |
So we're going to disengage and leave | |
1:45:23.760 --> 1:45:25.320 | |
that decision to the human. | |
1:45:25.320 --> 1:45:26.680 | |
You're such a troll. | |
1:45:26.680 --> 1:45:28.760 | |
No, but the trolley problem deserves to be trolled. | |
1:45:28.760 --> 1:45:32.040 | |
Yeah, that's a beautiful answer actually. | |
1:45:32.040 --> 1:45:34.440 | |
I know, I gave it to someone who was like, | |
1:45:34.440 --> 1:45:36.600 | |
sometimes people ask like you asked about the trolley problem. | |
1:45:36.600 --> 1:45:38.080 | |
Like you can have a kind of discussion about it. | |
1:45:38.080 --> 1:45:39.720 | |
Like when you get someone who's like really like | |
1:45:39.720 --> 1:45:43.600 | |
earnest about it, because it's the kind of thing where | |
1:45:43.600 --> 1:45:45.600 | |
if you ask a bunch of people in an office, | |
1:45:45.600 --> 1:45:48.360 | |
whether we should use a SQL stack or no SQL stack, | |
1:45:48.360 --> 1:45:50.600 | |
if they're not that technical, they have no opinion. | |
1:45:50.600 --> 1:45:52.360 | |
But if you ask them what color they want to paint the office, | |
1:45:52.360 --> 1:45:54.040 | |
everyone has an opinion on that. | |
1:45:54.040 --> 1:45:56.040 | |
And that's why the trolley problem is. | |
1:45:56.040 --> 1:45:57.280 | |
I mean, that's a beautiful answer. | |
1:45:57.280 --> 1:45:59.240 | |
Yeah, we're able to detect the problem | |
1:45:59.240 --> 1:46:01.960 | |
and we're able to pass it on to the human. | |
1:46:01.960 --> 1:46:03.760 | |
Wow, I've never heard anyone say it. | |
1:46:03.760 --> 1:46:06.160 | |
This is your nice escape route. | |
1:46:06.160 --> 1:46:07.320 | |
Okay, but... | |
1:46:07.320 --> 1:46:08.680 | |
Proud level two. | |
1:46:08.680 --> 1:46:09.760 | |
I'm proud level two. | |
1:46:09.760 --> 1:46:10.600 | |
I love it. | |
1:46:10.600 --> 1:46:14.400 | |
So the other thing that people have some concern about | |
1:46:14.400 --> 1:46:17.800 | |
with AI in general is hacking. | |
1:46:17.800 --> 1:46:21.400 | |
So how hard is it, do you think, to hack an autonomous vehicle | |
1:46:21.400 --> 1:46:25.000 | |
either through physical access or through the more sort of | |
1:46:25.000 --> 1:46:28.240 | |
popular now, these adversarial examples on the sensors? | |
1:46:28.240 --> 1:46:30.720 | |
Okay, the adversarial examples one. | |
1:46:30.720 --> 1:46:32.320 | |
You want to see some adversarial examples | |
1:46:32.320 --> 1:46:34.880 | |
that affect humans, right? | |
1:46:34.880 --> 1:46:38.040 | |
Oh, well, there used to be a stop sign here, | |
1:46:38.040 --> 1:46:40.000 | |
but I put a black bag over the stop sign | |
1:46:40.000 --> 1:46:43.520 | |
and then people ran it, adversarial, right? | |
1:46:43.520 --> 1:46:48.360 | |
Like, there's tons of human adversarial examples too. | |
1:46:48.360 --> 1:46:52.240 | |
The question in general about security, if you saw, | |
1:46:52.240 --> 1:46:54.040 | |
something just came out today and there are always | |
1:46:54.040 --> 1:46:57.560 | |
such hypey headlines about how navigate on autopilot | |
1:46:57.560 --> 1:47:00.960 | |
was fooled by a GPS spoof to take an exit. | |
1:47:00.960 --> 1:47:01.800 | |
Right. | |
1:47:01.800 --> 1:47:03.920 | |
At least that's all they could do was take an exit. | |
1:47:03.920 --> 1:47:06.720 | |
If your car is relying on GPS in order | |
1:47:06.720 --> 1:47:10.240 | |
to have a safe driving policy, you're doing something wrong. | |
1:47:10.240 --> 1:47:12.680 | |
If you're relying, and this is why V2V | |
1:47:12.680 --> 1:47:17.680 | |
is such a terrible idea, V2V now relies on both parties | |
1:47:18.160 --> 1:47:19.800 | |
getting communication right. | |
1:47:19.800 --> 1:47:24.800 | |
This is not even, so I think of safety, | |
1:47:26.080 --> 1:47:28.480 | |
security is like a special case of safety, right? | |
1:47:28.480 --> 1:47:31.880 | |
Safety is like we put a little, you know, | |
1:47:31.880 --> 1:47:33.360 | |
piece of caution tape around the hole | |
1:47:33.360 --> 1:47:35.560 | |
so that people won't walk into it by accident. | |
1:47:35.560 --> 1:47:38.240 | |
Security is like put a 10 foot fence around the hole | |
1:47:38.240 --> 1:47:40.120 | |
so you actually physically cannot climb into it | |
1:47:40.120 --> 1:47:42.360 | |
with barbed wire on the top and stuff, right? | |
1:47:42.360 --> 1:47:44.560 | |
So like if you're designing systems | |
1:47:44.560 --> 1:47:47.440 | |
that are like unreliable, they're definitely not secure. | |
1:47:48.440 --> 1:47:51.240 | |
Your car should always do something safe | |
1:47:51.240 --> 1:47:53.400 | |
using its local sensors. | |
1:47:53.400 --> 1:47:55.240 | |
And then the local sensor should be hardwired. | |
1:47:55.240 --> 1:47:57.400 | |
And then could somebody hack into your can boss | |
1:47:57.400 --> 1:47:58.640 | |
and turn your steering wheel on your brakes? | |
1:47:58.640 --> 1:48:01.240 | |
Yes, but they could do it before comma AI too, so. | |
1:48:02.800 --> 1:48:04.680 | |
Let's think out of the box and some things. | |
1:48:04.680 --> 1:48:09.400 | |
So do you think teleoperation has a role in any of this? | |
1:48:09.400 --> 1:48:13.880 | |
So remotely stepping in and controlling the cars? | |
1:48:13.880 --> 1:48:18.880 | |
No, I think that if the safety operation | |
1:48:21.320 --> 1:48:26.160 | |
by design requires a constant link to the cars, | |
1:48:26.160 --> 1:48:27.560 | |
I think it doesn't work. | |
1:48:27.560 --> 1:48:31.080 | |
So that's the same argument used for V2I, V2V. | |
1:48:31.080 --> 1:48:34.280 | |
Well, there's a lot of non safety critical stuff | |
1:48:34.280 --> 1:48:35.120 | |
you can do with V2I. | |
1:48:35.120 --> 1:48:37.440 | |
I like V2I, I like V2I way more than V2V | |
1:48:37.440 --> 1:48:39.280 | |
because V2I is already like, | |
1:48:39.280 --> 1:48:40.880 | |
I already have internet in the car, right? | |
1:48:40.880 --> 1:48:43.280 | |
There's a lot of great stuff you can do with V2I. | |
1:48:44.280 --> 1:48:46.320 | |
Like for example, you can, | |
1:48:46.320 --> 1:48:48.880 | |
well, where I already have V2V, Waze is V2I, right? | |
1:48:48.880 --> 1:48:50.520 | |
Waze can route me around traffic jams. | |
1:48:50.520 --> 1:48:52.760 | |
That's a great example of V2I. | |
1:48:52.760 --> 1:48:54.440 | |
And then, okay, the car automatically talks | |
1:48:54.440 --> 1:48:55.800 | |
to that same service, like it works. | |
1:48:55.800 --> 1:48:56.800 | |
So it's improving the experience, | |
1:48:56.800 --> 1:48:59.480 | |
but it's not a fundamental fallback for safety. | |
1:48:59.480 --> 1:49:04.160 | |
No, if any of your things that require | |
1:49:04.160 --> 1:49:07.480 | |
wireless communication are more than QM, | |
1:49:07.480 --> 1:49:10.640 | |
like have an ASL rating, you shouldn't. | |
1:49:10.640 --> 1:49:14.200 | |
You previously said that life is work | |
1:49:15.440 --> 1:49:17.480 | |
and that you don't do anything to relax. | |
1:49:17.480 --> 1:49:20.800 | |
So how do you think about hard work? | |
1:49:20.800 --> 1:49:22.200 | |
Well, what is it? | |
1:49:22.200 --> 1:49:24.720 | |
What do you think it takes to accomplish great things? | |
1:49:24.720 --> 1:49:25.840 | |
And there's a lot of people saying | |
1:49:25.840 --> 1:49:28.280 | |
that there needs to be some balance. | |
1:49:28.280 --> 1:49:29.600 | |
You know, you need to, | |
1:49:29.600 --> 1:49:31.120 | |
in order to accomplish great things, | |
1:49:31.120 --> 1:49:32.200 | |
you need to take some time off, | |
1:49:32.200 --> 1:49:34.640 | |
you need to reflect and so on. | |
1:49:34.640 --> 1:49:37.840 | |
And then some people are just insanely working, | |
1:49:37.840 --> 1:49:39.640 | |
burning the candle on both ends. | |
1:49:39.640 --> 1:49:41.360 | |
How do you think about that? | |
1:49:41.360 --> 1:49:43.400 | |
I think I was trolling in the Suraj interview | |
1:49:43.400 --> 1:49:45.600 | |
when I said that off camera, | |
1:49:45.600 --> 1:49:47.240 | |
but right before I spoke to a little bit of weed, | |
1:49:47.240 --> 1:49:49.800 | |
like, you know, come on, this is a joke, right? | |
1:49:49.800 --> 1:49:50.880 | |
Like I do nothing to relax. | |
1:49:50.880 --> 1:49:52.560 | |
Look where I am, I'm at a party, right? | |
1:49:52.560 --> 1:49:53.960 | |
Yeah, yeah, yeah, sure. | |
1:49:53.960 --> 1:49:55.200 | |
That's true. | |
1:49:55.200 --> 1:49:58.040 | |
So no, no, of course I don't. | |
1:49:58.040 --> 1:49:59.800 | |
When I say that life is work though, | |
1:49:59.800 --> 1:50:01.960 | |
I mean that like, I think that | |
1:50:01.960 --> 1:50:04.200 | |
what gives my life meaning is work. | |
1:50:04.200 --> 1:50:05.720 | |
I don't mean that every minute of the day | |
1:50:05.720 --> 1:50:06.560 | |
you should be working. | |
1:50:06.560 --> 1:50:08.000 | |
I actually think this is not the best way | |
1:50:08.000 --> 1:50:09.800 | |
to maximize results. | |
1:50:09.800 --> 1:50:12.040 | |
I think that if you're working 12 hours a day, | |
1:50:12.040 --> 1:50:14.920 | |
you should be working smarter and not harder. | |
1:50:14.920 --> 1:50:17.880 | |
Well, so it gives work gives you meaning | |
1:50:17.880 --> 1:50:20.520 | |
for some people, other sorts of meaning | |
1:50:20.520 --> 1:50:24.560 | |
is personal relationships, like family and so on. | |
1:50:24.560 --> 1:50:27.200 | |
You've also in that interview with Suraj | |
1:50:27.200 --> 1:50:30.720 | |
or the trolling mentioned that one of the things | |
1:50:30.720 --> 1:50:33.400 | |
you look forward to in the future is AI girlfriends. | |
1:50:33.400 --> 1:50:34.360 | |
Yes. | |
1:50:34.360 --> 1:50:38.800 | |
So that's a topic that I'm very much fascinated by, | |
1:50:38.800 --> 1:50:39.840 | |
not necessarily girlfriends, | |
1:50:39.840 --> 1:50:41.880 | |
but just forming a deep connection with AI. | |
1:50:41.880 --> 1:50:42.960 | |
Yeah. | |
1:50:42.960 --> 1:50:44.400 | |
What kind of system do you imagine | |
1:50:44.400 --> 1:50:46.240 | |
when you say AI girlfriend, | |
1:50:46.240 --> 1:50:47.800 | |
whether you were trolling or not? | |
1:50:47.800 --> 1:50:49.720 | |
No, that one I'm very serious about. | |
1:50:49.720 --> 1:50:52.360 | |
And I'm serious about that on both a shallow level | |
1:50:52.360 --> 1:50:53.680 | |
and a deep level. | |
1:50:53.680 --> 1:50:55.720 | |
I think that VR brothels are coming soon | |
1:50:55.720 --> 1:50:57.800 | |
and are gonna be really cool. | |
1:50:57.800 --> 1:50:59.760 | |
It's not cheating if it's a robot. | |
1:50:59.760 --> 1:51:01.080 | |
I see the slogan already. | |
1:51:01.080 --> 1:51:04.320 | |
Um, but... | |
1:51:04.320 --> 1:51:06.200 | |
There's a, I don't know if you've watched | |
1:51:06.200 --> 1:51:08.320 | |
or just watched the Black Mirror episode. | |
1:51:08.320 --> 1:51:09.320 | |
I watched the latest one, yeah. | |
1:51:09.320 --> 1:51:11.320 | |
Yeah, yeah. | |
1:51:11.320 --> 1:51:13.160 | |
Oh, the Ashley 2 one? | |
1:51:13.160 --> 1:51:15.120 | |
Or the... | |
1:51:15.120 --> 1:51:16.920 | |
No, where there's two friends | |
1:51:16.920 --> 1:51:20.160 | |
who were having sex with each other in... | |
1:51:20.160 --> 1:51:21.240 | |
Oh, in the VR game. | |
1:51:21.240 --> 1:51:23.560 | |
In the VR game, it's the two guys, | |
1:51:23.560 --> 1:51:26.720 | |
but one of them was a female, yeah. | |
1:51:26.720 --> 1:51:27.560 | |
Yeah, the... | |
1:51:27.560 --> 1:51:29.560 | |
Which is another mind blowing concept. | |
1:51:29.560 --> 1:51:33.320 | |
That in VR, you don't have to be the form. | |
1:51:33.320 --> 1:51:37.720 | |
You can be two animals having sex, it's weird. | |
1:51:37.720 --> 1:51:38.560 | |
I mean, I'll see how nice | |
1:51:38.560 --> 1:51:40.280 | |
that the software maps the nerve endings, right? | |
1:51:40.280 --> 1:51:41.600 | |
Yeah, it's weird. | |
1:51:41.600 --> 1:51:44.480 | |
I mean, yeah, they sweep a lot of the fascinating, | |
1:51:44.480 --> 1:51:46.440 | |
really difficult technical challenges under the rug, | |
1:51:46.440 --> 1:51:48.360 | |
like assuming it's possible | |
1:51:48.360 --> 1:51:51.160 | |
to do the mapping of the nerve endings, then... | |
1:51:51.160 --> 1:51:52.000 | |
I wish, yeah, I saw that. | |
1:51:52.000 --> 1:51:53.800 | |
The way they did it with the little like stim unit | |
1:51:53.800 --> 1:51:55.400 | |
on the head, that'd be amazing. | |
1:51:56.800 --> 1:51:58.760 | |
So, well, no, no, on a shallow level, | |
1:51:58.760 --> 1:52:01.640 | |
like you could set up like almost a brothel | |
1:52:01.640 --> 1:52:05.160 | |
with like real dolls and Oculus quests, | |
1:52:05.160 --> 1:52:06.200 | |
write some good software. | |
1:52:06.200 --> 1:52:08.280 | |
I think it'd be a cool novelty experience. | |
1:52:09.280 --> 1:52:11.400 | |
But no, on a deeper, like emotional level. | |
1:52:12.800 --> 1:52:16.960 | |
I mean, yeah, I would really like to fall in love | |
1:52:16.960 --> 1:52:18.120 | |
with the machine. | |
1:52:18.120 --> 1:52:23.120 | |
Do you see yourself having a long term relationship | |
1:52:23.120 --> 1:52:27.520 | |
of the kind monogamous relationship that we have now | |
1:52:27.520 --> 1:52:31.360 | |
with the robot, with the AI system, even? | |
1:52:31.360 --> 1:52:32.680 | |
Not even just the robot. | |
1:52:32.680 --> 1:52:37.680 | |
So, I think about maybe my ideal future. | |
1:52:38.200 --> 1:52:43.200 | |
When I was 15, I read Eliezer Yudkowsky's early writings | |
1:52:44.320 --> 1:52:49.120 | |
on the singularity and like that AI | |
1:52:49.120 --> 1:52:53.040 | |
is going to surpass human intelligence massively. | |
1:52:53.040 --> 1:52:55.480 | |
He made some Moore's law based predictions | |
1:52:55.480 --> 1:52:57.400 | |
that I mostly agree with. | |
1:52:57.400 --> 1:52:59.360 | |
And then I really struggled | |
1:52:59.360 --> 1:53:01.360 | |
for the next couple of years of my life. | |
1:53:01.360 --> 1:53:03.360 | |
Like, why should I even bother to learn anything? | |
1:53:03.360 --> 1:53:06.160 | |
It's all gonna be meaningless when the machine show up. | |
1:53:06.160 --> 1:53:07.000 | |
Right. | |
1:53:07.000 --> 1:53:10.520 | |
Well, maybe when I was that young, | |
1:53:10.520 --> 1:53:12.040 | |
I was still a little bit more pure | |
1:53:12.040 --> 1:53:13.160 | |
and really like clung to that. | |
1:53:13.160 --> 1:53:14.720 | |
And then I'm like, well, the machine's ain't here yet. | |
1:53:14.720 --> 1:53:16.800 | |
You know, and I seem to be pretty good at this stuff. | |
1:53:16.800 --> 1:53:18.520 | |
Let's try my best, you know, | |
1:53:18.520 --> 1:53:20.320 | |
like what's the worst that happens? | |
1:53:20.320 --> 1:53:23.440 | |
But the best possible future I see | |
1:53:23.440 --> 1:53:26.120 | |
is me sort of merging with the machine. | |
1:53:26.120 --> 1:53:28.120 | |
And the way that I personify this | |
1:53:28.120 --> 1:53:30.800 | |
is in a longterm and augments relationship with the machine. | |
1:53:32.160 --> 1:53:33.320 | |
Oh, you don't think there's room | |
1:53:33.320 --> 1:53:35.040 | |
for another human in your life | |
1:53:35.040 --> 1:53:37.440 | |
if you really truly merge with another machine? | |
1:53:38.440 --> 1:53:40.240 | |
I mean, I see merging. | |
1:53:40.240 --> 1:53:44.240 | |
I see like the best interface to my brain | |
1:53:45.520 --> 1:53:48.000 | |
is like the same relationship interface | |
1:53:48.000 --> 1:53:49.320 | |
to merge with an AI, right? | |
1:53:49.320 --> 1:53:51.440 | |
What does that merging feel like? | |
1:53:52.440 --> 1:53:55.320 | |
I've seen couples who've been together for a long time | |
1:53:55.320 --> 1:53:57.840 | |
and like, I almost think of them as one person. | |
1:53:57.840 --> 1:54:01.280 | |
Like couples who spend all their time together and... | |
1:54:01.280 --> 1:54:02.120 | |
That's fascinating. | |
1:54:02.120 --> 1:54:03.320 | |
You're actually putting, | |
1:54:03.320 --> 1:54:05.520 | |
what does that merging actually looks like? | |
1:54:05.520 --> 1:54:07.600 | |
It's not just a nice channel. | |
1:54:07.600 --> 1:54:11.640 | |
Like a lot of people imagine it's just an efficient link, | |
1:54:11.640 --> 1:54:13.800 | |
search link to Wikipedia or something. | |
1:54:13.800 --> 1:54:14.640 | |
I don't believe in that. | |
1:54:14.640 --> 1:54:17.120 | |
But it's more, you're saying that there's the same kind of, | |
1:54:17.120 --> 1:54:19.520 | |
the same kind of relationship you have with another human | |
1:54:19.520 --> 1:54:22.960 | |
as a deep relationship is that's what merging looks like. | |
1:54:22.960 --> 1:54:24.480 | |
That's pretty... | |
1:54:24.480 --> 1:54:26.680 | |
I don't believe that link is possible. | |
1:54:26.680 --> 1:54:28.120 | |
I think that that link, so you're like, | |
1:54:28.120 --> 1:54:30.160 | |
oh, I'm gonna download Wikipedia right to my brain. | |
1:54:30.160 --> 1:54:33.360 | |
My reading speed is not limited by my eyes. | |
1:54:33.360 --> 1:54:36.800 | |
My reading speed is limited by my inner processing loop. | |
1:54:36.800 --> 1:54:38.680 | |
And to like bootstrap that | |
1:54:38.680 --> 1:54:42.440 | |
sounds kind of unclear how to do it and horrifying. | |
1:54:42.440 --> 1:54:46.560 | |
But if I am with somebody, and I'll use somebody | |
1:54:46.560 --> 1:54:51.400 | |
who is making a super sophisticated model of me | |
1:54:51.400 --> 1:54:53.200 | |
and then running simulations on that model, | |
1:54:53.200 --> 1:54:54.120 | |
I'm not gonna get into the question | |
1:54:54.120 --> 1:54:55.880 | |
whether the simulations are conscious or not. | |
1:54:55.880 --> 1:54:58.240 | |
I don't really wanna know what it's doing. | |
1:54:58.240 --> 1:55:01.600 | |
But using those simulations to play out hypothetical futures | |
1:55:01.600 --> 1:55:04.880 | |
for me, deciding what things to say to me | |
1:55:04.880 --> 1:55:08.720 | |
to guide me along a path and that's how I envision it. | |
1:55:08.720 --> 1:55:13.720 | |
So on that path to AI of super human level intelligence, | |
1:55:13.720 --> 1:55:15.680 | |
you've mentioned that you believe in the singularity, | |
1:55:15.680 --> 1:55:17.280 | |
that singularity is coming. | |
1:55:17.280 --> 1:55:20.440 | |
Again, could be trolling, could be not, could be part... | |
1:55:20.440 --> 1:55:21.760 | |
All trolling has truth in it. | |
1:55:21.760 --> 1:55:22.840 | |
I don't know what that means anymore. | |
1:55:22.840 --> 1:55:24.520 | |
What is the singularity? | |
1:55:24.520 --> 1:55:26.720 | |
So yeah, so that's really the question. | |
1:55:26.720 --> 1:55:29.280 | |
How many years do you think before the singularity | |
1:55:29.280 --> 1:55:30.920 | |
of what form do you think it will take? | |
1:55:30.920 --> 1:55:34.200 | |
Does that mean fundamental shifts in capabilities of AI? | |
1:55:34.200 --> 1:55:36.960 | |
Does it mean some other kind of ideas? | |
1:55:36.960 --> 1:55:40.120 | |
Maybe that's just my roots, but... | |
1:55:40.120 --> 1:55:42.920 | |
So I can buy a human being's worth of computers | |
1:55:42.920 --> 1:55:46.000 | |
for things worth of compute for like a million bucks a day. | |
1:55:46.000 --> 1:55:47.800 | |
It's about one TPU pod V3. | |
1:55:47.800 --> 1:55:50.240 | |
I want like, I think they claim a hundred pay to flops. | |
1:55:50.240 --> 1:55:51.080 | |
That's being generous. | |
1:55:51.080 --> 1:55:52.320 | |
I think humans are actually more like 20. | |
1:55:52.320 --> 1:55:53.160 | |
So that's like five humans. | |
1:55:53.160 --> 1:55:54.040 | |
That's pretty good. | |
1:55:54.040 --> 1:55:55.560 | |
Google needs to sell their TPUs. | |
1:55:56.840 --> 1:55:58.640 | |
But no, I could buy GPUs. | |
1:55:58.640 --> 1:56:02.280 | |
I could buy a stack of like, I buy 1080TIs, | |
1:56:02.280 --> 1:56:03.880 | |
build data center full of them. | |
1:56:03.880 --> 1:56:07.280 | |
And for a million bucks, I can get a human worth of compute. | |
1:56:08.160 --> 1:56:12.280 | |
But when you look at the total number of flops in the world, | |
1:56:12.280 --> 1:56:14.400 | |
when you look at human flops, | |
1:56:14.400 --> 1:56:17.040 | |
which goes up very, very slowly with the population, | |
1:56:17.040 --> 1:56:19.760 | |
and machine flops, which goes up exponentially, | |
1:56:19.760 --> 1:56:22.360 | |
but it's still nowhere near. | |
1:56:22.360 --> 1:56:24.040 | |
I think that's the key thing | |
1:56:24.040 --> 1:56:25.880 | |
to talk about when the singularity happened. | |
1:56:25.880 --> 1:56:28.560 | |
When most flops in the world are silicon | |
1:56:28.560 --> 1:56:32.280 | |
and not biological, that's kind of the crossing point. | |
1:56:32.280 --> 1:56:35.480 | |
Like they are now the dominant species on the planet. | |
1:56:35.480 --> 1:56:38.720 | |
And just looking at how technology is progressing, | |
1:56:38.720 --> 1:56:40.360 | |
when do you think that could possibly happen? | |
1:56:40.360 --> 1:56:41.680 | |
Do you think it would happen in your lifetime? | |
1:56:41.680 --> 1:56:43.640 | |
Oh yeah, definitely in my lifetime. | |
1:56:43.640 --> 1:56:44.480 | |
I've done the math. | |
1:56:44.480 --> 1:56:47.560 | |
I like 2038 because it's the UNIX timestamp roll over. | |
1:56:49.920 --> 1:56:51.840 | |
Yeah, beautifully put. | |
1:56:52.680 --> 1:56:57.680 | |
So you've said that the meaning of life is to win. | |
1:56:58.000 --> 1:56:59.560 | |
If you look five years into the future, | |
1:56:59.560 --> 1:57:01.000 | |
what does winning look like? | |
1:57:02.640 --> 1:57:03.720 | |
So... | |
1:57:03.720 --> 1:57:08.720 | |
I can go into technical depth to what I mean by that, to win. | |
1:57:11.720 --> 1:57:12.720 | |
It may not mean... | |
1:57:12.720 --> 1:57:14.400 | |
I was criticized for that in the comments. | |
1:57:14.400 --> 1:57:17.720 | |
Like, doesn't this guy want to save the penguins in Antarctica? | |
1:57:17.720 --> 1:57:20.960 | |
Or like, oh man, listen to what I'm saying. | |
1:57:20.960 --> 1:57:23.720 | |
I'm not talking about like I have a yacht or something. | |
1:57:24.720 --> 1:57:26.720 | |
I am an agent. | |
1:57:26.720 --> 1:57:28.720 | |
I am put into this world. | |
1:57:28.720 --> 1:57:32.720 | |
And I don't really know what my purpose is. | |
1:57:33.720 --> 1:57:36.720 | |
But if you're a reinforcement, if you're an intelligent agent | |
1:57:36.720 --> 1:57:39.720 | |
and you're put into a world, what is the ideal thing to do? | |
1:57:39.720 --> 1:57:41.720 | |
Well, the ideal thing, mathematically, | |
1:57:41.720 --> 1:57:43.720 | |
you can go back to like Schmidt Hoover theories about this, | |
1:57:43.720 --> 1:57:46.720 | |
is to build a compressive model of the world. | |
1:57:46.720 --> 1:57:49.720 | |
To build a maximally compressive to explore the world | |
1:57:49.720 --> 1:57:52.720 | |
such that your exploration function maximizes | |
1:57:52.720 --> 1:57:55.720 | |
the derivative of compression of the past. | |
1:57:55.720 --> 1:57:58.720 | |
Schmidt Hoover has a paper about this. | |
1:57:58.720 --> 1:58:01.720 | |
And like, I took that kind of as like a personal goal function. | |
1:58:02.720 --> 1:58:04.720 | |
So what I mean to win, I mean like, | |
1:58:04.720 --> 1:58:08.720 | |
maybe this is religious, but like I think that in the future | |
1:58:08.720 --> 1:58:10.720 | |
I might be given a real purpose. | |
1:58:10.720 --> 1:58:12.720 | |
Or I may decide this purpose myself. | |
1:58:12.720 --> 1:58:14.720 | |
And then at that point, now I know what the game is | |
1:58:14.720 --> 1:58:15.720 | |
and I know how to win. | |
1:58:15.720 --> 1:58:18.720 | |
I think right now I'm still just trying to figure out what the game is. | |
1:58:18.720 --> 1:58:19.720 | |
But once I know... | |
1:58:20.720 --> 1:58:22.720 | |
So you have... | |
1:58:22.720 --> 1:58:25.720 | |
You have imperfect information. | |
1:58:25.720 --> 1:58:27.720 | |
You have a lot of uncertainty about the reward function | |
1:58:27.720 --> 1:58:28.720 | |
and you're discovering it. | |
1:58:28.720 --> 1:58:29.720 | |
Exactly. | |
1:58:29.720 --> 1:58:30.720 | |
But the purpose is... | |
1:58:30.720 --> 1:58:31.720 | |
That's a better way to put it. | |
1:58:31.720 --> 1:58:33.720 | |
The purpose is to maximize it | |
1:58:33.720 --> 1:58:36.720 | |
while you have a lot of uncertainty around it. | |
1:58:36.720 --> 1:58:38.720 | |
And you're both reducing the uncertainty | |
1:58:38.720 --> 1:58:40.720 | |
and maximizing at the same time. | |
1:58:40.720 --> 1:58:43.720 | |
And so that's at the technical level. | |
1:58:43.720 --> 1:58:44.720 | |
What is the... | |
1:58:44.720 --> 1:58:46.720 | |
If you believe in the universal prior, | |
1:58:46.720 --> 1:58:48.720 | |
what is the universal reward function? | |
1:58:48.720 --> 1:58:50.720 | |
That's the better way to put it. | |
1:58:50.720 --> 1:58:53.720 | |
So that win is interesting. | |
1:58:53.720 --> 1:58:56.720 | |
I think I speak for everyone in saying that | |
1:58:56.720 --> 1:59:01.720 | |
I wonder what that reward function is for you. | |
1:59:01.720 --> 1:59:06.720 | |
And I look forward to seeing that in five years and ten years. | |
1:59:06.720 --> 1:59:09.720 | |
I think a lot of people including myself are cheering you on, man. | |
1:59:09.720 --> 1:59:11.720 | |
So I'm happy you exist. | |
1:59:11.720 --> 1:59:13.720 | |
And I wish you the best of luck. | |
1:59:13.720 --> 1:59:14.720 | |
Thanks for talking today, man. | |
1:59:14.720 --> 1:59:15.720 | |
Thank you. | |
1:59:15.720 --> 1:59:20.720 | |
This was a lot of fun. | |