|
WEBVTT |
|
|
|
00:00.000 --> 00:02.520 |
|
The following is a conversation with George Hotz. |
|
|
|
00:02.520 --> 00:04.480 |
|
He's the founder of Comma AI, |
|
|
|
00:04.480 --> 00:07.400 |
|
a machine learning based vehicle automation company. |
|
|
|
00:07.400 --> 00:10.200 |
|
He is most certainly an outspoken personality |
|
|
|
00:10.200 --> 00:13.160 |
|
in the field of AI and technology in general. |
|
|
|
00:13.160 --> 00:16.240 |
|
He first gained recognition for being the first person |
|
|
|
00:16.240 --> 00:18.400 |
|
to carry on lock and iPhone. |
|
|
|
00:18.400 --> 00:21.280 |
|
And since then, he's done quite a few interesting things |
|
|
|
00:21.280 --> 00:24.400 |
|
at the intersection of hardware and software. |
|
|
|
00:24.400 --> 00:27.440 |
|
This is the artificial intelligence podcast. |
|
|
|
00:27.440 --> 00:29.560 |
|
If you enjoy it, subscribe on YouTube, |
|
|
|
00:29.560 --> 00:32.920 |
|
give it five stars on iTunes, support it on Patreon, |
|
|
|
00:32.920 --> 00:34.920 |
|
or simply connect with me on Twitter. |
|
|
|
00:34.920 --> 00:39.120 |
|
Alex Friedman, spelled F R I D M A N. |
|
|
|
00:39.120 --> 00:42.000 |
|
And I'd like to give a special thank you to Jennifer |
|
|
|
00:42.000 --> 00:45.880 |
|
from Canada for her support of the podcast on Patreon. |
|
|
|
00:45.880 --> 00:47.720 |
|
Merci beaucoup, Jennifer. |
|
|
|
00:47.720 --> 00:50.600 |
|
She's been a friend and an engineering colleague |
|
|
|
00:50.600 --> 00:52.800 |
|
for many years since I was in grad school. |
|
|
|
00:52.800 --> 00:55.520 |
|
Your support means a lot and inspires me |
|
|
|
00:55.520 --> 00:57.920 |
|
to keep this series going. |
|
|
|
00:57.920 --> 01:01.600 |
|
And now here's my conversation with George Hotz. |
|
|
|
01:02.720 --> 01:04.720 |
|
Do you think we're living in a simulation? |
|
|
|
01:06.480 --> 01:10.080 |
|
Yes, but it may be unfalsifiable. |
|
|
|
01:10.080 --> 01:12.440 |
|
What do you mean by unfalsifiable? |
|
|
|
01:12.440 --> 01:16.840 |
|
So if the simulation is designed in such a way |
|
|
|
01:16.840 --> 01:19.640 |
|
that they did like a formal proof |
|
|
|
01:19.640 --> 01:22.320 |
|
to show that no information can get in and out. |
|
|
|
01:22.320 --> 01:25.200 |
|
And if their hardware is designed for the anything |
|
|
|
01:25.200 --> 01:27.880 |
|
in the simulation to always keep the hardware in spec, |
|
|
|
01:27.880 --> 01:29.480 |
|
it may be impossible to prove |
|
|
|
01:29.480 --> 01:31.280 |
|
whether we're in a simulation or not. |
|
|
|
01:32.600 --> 01:35.680 |
|
So they've designed it such that it's a closed system, |
|
|
|
01:35.680 --> 01:37.200 |
|
you can't get outside the system. |
|
|
|
01:37.200 --> 01:38.760 |
|
Well, maybe it's one of three worlds. |
|
|
|
01:38.760 --> 01:41.400 |
|
We're either in a simulation which can be exploited, |
|
|
|
01:41.400 --> 01:44.200 |
|
we're in a simulation which not only can't be exploited, |
|
|
|
01:44.200 --> 01:46.440 |
|
but like the same thing's true about VMs. |
|
|
|
01:46.440 --> 01:48.160 |
|
A really well designed VM, |
|
|
|
01:48.160 --> 01:50.520 |
|
you can't even detect if you're in a VM or not. |
|
|
|
01:51.400 --> 01:52.520 |
|
That's brilliant. |
|
|
|
01:52.520 --> 01:55.160 |
|
So we're, yeah, so the simulation is running |
|
|
|
01:55.160 --> 01:56.800 |
|
on a virtual machine. |
|
|
|
01:56.800 --> 01:59.440 |
|
But now in reality, all VMs have ways to detect. |
|
|
|
01:59.440 --> 02:00.280 |
|
That's the point. |
|
|
|
02:00.280 --> 02:04.840 |
|
I mean, is it, you've done quite a bit of hacking yourself. |
|
|
|
02:04.840 --> 02:08.640 |
|
So you should know that really any complicated system |
|
|
|
02:08.640 --> 02:11.000 |
|
will have ways in and out. |
|
|
|
02:11.000 --> 02:14.240 |
|
So this isn't necessarily true going forward. |
|
|
|
02:15.280 --> 02:18.080 |
|
I spent my time away from comma, |
|
|
|
02:18.080 --> 02:21.240 |
|
I learned a cock, it's a dependently typed, |
|
|
|
02:21.240 --> 02:24.360 |
|
like it's a language for writing math proofs. |
|
|
|
02:24.360 --> 02:28.200 |
|
And if you write code that compiles in a language like that, |
|
|
|
02:28.200 --> 02:30.840 |
|
it is correct by definition. |
|
|
|
02:30.840 --> 02:33.560 |
|
The types check it's correctance. |
|
|
|
02:33.560 --> 02:35.000 |
|
So it's possible that the simulation |
|
|
|
02:35.000 --> 02:39.640 |
|
is written in a language like this, in which case, yeah. |
|
|
|
02:39.640 --> 02:42.680 |
|
Yeah, but that can't be sufficiently expressive |
|
|
|
02:42.680 --> 02:43.760 |
|
of language like that. |
|
|
|
02:43.760 --> 02:44.600 |
|
Oh, it can. |
|
|
|
02:44.600 --> 02:45.440 |
|
It can be? |
|
|
|
02:45.440 --> 02:46.280 |
|
Oh, yeah. |
|
|
|
02:46.280 --> 02:48.920 |
|
Okay, well, so, all right, so. |
|
|
|
02:48.920 --> 02:50.640 |
|
The simulation doesn't have to be tearing complete |
|
|
|
02:50.640 --> 02:52.320 |
|
if it has a scheduled end date. |
|
|
|
02:52.320 --> 02:54.600 |
|
Looks like it does actually with entropy. |
|
|
|
02:54.600 --> 02:58.520 |
|
I mean, I don't think that a simulation |
|
|
|
02:58.520 --> 03:02.200 |
|
that results in something as complicated as the universe |
|
|
|
03:03.080 --> 03:07.280 |
|
would have a formal proof of correctness, right? |
|
|
|
03:08.240 --> 03:09.880 |
|
It's possible, of course. |
|
|
|
03:09.880 --> 03:12.720 |
|
We have no idea how good their tooling is. |
|
|
|
03:12.720 --> 03:14.640 |
|
And we have no idea how complicated |
|
|
|
03:14.640 --> 03:16.280 |
|
the universe computer really is. |
|
|
|
03:16.280 --> 03:17.920 |
|
It may be quite simple. |
|
|
|
03:17.920 --> 03:19.680 |
|
It's just very large, right? |
|
|
|
03:19.680 --> 03:22.160 |
|
It's very, it's definitely very large. |
|
|
|
03:22.160 --> 03:24.480 |
|
But the fundamental rules might be super simple. |
|
|
|
03:24.480 --> 03:26.240 |
|
Yeah, Conway's gonna like kinda stop. |
|
|
|
03:26.240 --> 03:30.320 |
|
Right, so if you could hack, |
|
|
|
03:30.320 --> 03:32.400 |
|
so imagine the simulation that is hackable, |
|
|
|
03:32.400 --> 03:33.640 |
|
if you could hack it, |
|
|
|
03:35.040 --> 03:37.960 |
|
what would you change about the universe? |
|
|
|
03:37.960 --> 03:40.560 |
|
Like how would you approach hacking a simulation? |
|
|
|
03:41.640 --> 03:44.360 |
|
The reason I gave that talk? |
|
|
|
03:44.360 --> 03:46.680 |
|
By the way, I'm not familiar with the talk you gave. |
|
|
|
03:46.680 --> 03:50.160 |
|
I just read that you talked about escaping the simulation |
|
|
|
03:50.160 --> 03:51.280 |
|
or something like that. |
|
|
|
03:51.280 --> 03:52.640 |
|
So maybe you can tell me a little bit |
|
|
|
03:52.640 --> 03:55.360 |
|
about the theme and the message there too. |
|
|
|
03:55.360 --> 03:57.680 |
|
It wasn't a very practical talk |
|
|
|
03:57.680 --> 04:00.600 |
|
about how to actually escape a simulation. |
|
|
|
04:00.600 --> 04:03.320 |
|
It was more about a way of restructuring |
|
|
|
04:03.320 --> 04:05.120 |
|
an us versus them narrative. |
|
|
|
04:05.120 --> 04:10.120 |
|
If we continue on the path we're going with technology, |
|
|
|
04:12.360 --> 04:14.160 |
|
I think we're in big trouble, |
|
|
|
04:14.160 --> 04:16.760 |
|
like as a species and not just as a species, |
|
|
|
04:16.760 --> 04:19.480 |
|
but even as me as an individual member of the species. |
|
|
|
04:19.480 --> 04:23.680 |
|
So if we could change rhetoric to be more like, |
|
|
|
04:23.680 --> 04:24.920 |
|
to think upwards, |
|
|
|
04:26.240 --> 04:29.080 |
|
like to think about that we're in a simulation |
|
|
|
04:29.080 --> 04:30.360 |
|
and how we could get out, |
|
|
|
04:30.360 --> 04:32.640 |
|
already we'd be on the right path. |
|
|
|
04:32.640 --> 04:34.800 |
|
What you actually do once you do that, |
|
|
|
04:34.800 --> 04:37.360 |
|
well, I assume I would have acquired way more intelligence |
|
|
|
04:37.360 --> 04:39.760 |
|
in the process of doing that, so I'll just ask that. |
|
|
|
04:39.760 --> 04:42.080 |
|
So the thinking upwards, |
|
|
|
04:42.080 --> 04:43.760 |
|
what kind of ideas, |
|
|
|
04:43.760 --> 04:45.640 |
|
what kind of breakthrough ideas do you think thinking |
|
|
|
04:45.640 --> 04:47.280 |
|
in that way could inspire? |
|
|
|
04:47.280 --> 04:49.800 |
|
And why did you say upwards? |
|
|
|
04:49.800 --> 04:50.640 |
|
Upwards. |
|
|
|
04:50.640 --> 04:51.480 |
|
Into space? |
|
|
|
04:51.480 --> 04:54.120 |
|
Are you thinking sort of exploration in all forms? |
|
|
|
04:54.120 --> 04:59.120 |
|
The space narrative that held for the modernist generation |
|
|
|
04:59.880 --> 05:02.600 |
|
doesn't hold as well for the postmodern generation. |
|
|
|
05:04.560 --> 05:05.480 |
|
What's the space narrative? |
|
|
|
05:05.480 --> 05:06.520 |
|
Are we talking about the same space? |
|
|
|
05:06.520 --> 05:07.360 |
|
The three dimensional space? |
|
|
|
05:07.360 --> 05:08.840 |
|
No, no, space, like going up space, |
|
|
|
05:08.840 --> 05:10.040 |
|
like building like Elon Musk, |
|
|
|
05:10.040 --> 05:11.160 |
|
like we're going to build rockets, |
|
|
|
05:11.160 --> 05:12.080 |
|
we're going to go to Mars, |
|
|
|
05:12.080 --> 05:13.560 |
|
we're going to colonize the universe. |
|
|
|
05:13.560 --> 05:14.720 |
|
And the narrative you're referring, |
|
|
|
05:14.720 --> 05:16.040 |
|
I was born in the Soviet Union, |
|
|
|
05:16.040 --> 05:18.000 |
|
you're referring to the race to space? |
|
|
|
05:18.000 --> 05:18.840 |
|
The race to space, yeah. |
|
|
|
05:18.840 --> 05:19.680 |
|
Yes, explore, okay. |
|
|
|
05:19.680 --> 05:21.760 |
|
That was a great modernist narrative. |
|
|
|
05:21.760 --> 05:23.360 |
|
Yeah. |
|
|
|
05:23.360 --> 05:26.720 |
|
It doesn't seem to hold the same weight in today's culture. |
|
|
|
05:27.640 --> 05:32.160 |
|
I'm hoping for good postmodern narratives that replace it. |
|
|
|
05:32.160 --> 05:35.560 |
|
So let's think, so you work a lot with AI. |
|
|
|
05:35.560 --> 05:39.080 |
|
So AI is one formulation of that narrative. |
|
|
|
05:39.080 --> 05:40.080 |
|
There could be also, |
|
|
|
05:40.080 --> 05:42.320 |
|
I don't know how much you do in VR and AR. |
|
|
|
05:42.320 --> 05:43.160 |
|
Yeah. |
|
|
|
05:43.160 --> 05:45.160 |
|
That's another, I know less about it, |
|
|
|
05:45.160 --> 05:47.600 |
|
but every time I play with it and our research, |
|
|
|
05:47.600 --> 05:49.640 |
|
it's fascinating, that virtual world. |
|
|
|
05:49.640 --> 05:51.840 |
|
Are you interested in the virtual world? |
|
|
|
05:51.840 --> 05:54.200 |
|
I would like to move to virtual reality. |
|
|
|
05:55.360 --> 05:56.440 |
|
In terms of your work? |
|
|
|
05:56.440 --> 05:58.760 |
|
No, I would like to physically move there. |
|
|
|
05:58.760 --> 06:00.240 |
|
The apartment I can rent in the cloud |
|
|
|
06:00.240 --> 06:03.240 |
|
is way better than the apartment I can rent in the real world. |
|
|
|
06:03.240 --> 06:04.760 |
|
Well, it's all relative, isn't it? |
|
|
|
06:04.760 --> 06:07.280 |
|
Because others will have very nice apartments too, |
|
|
|
06:07.280 --> 06:09.200 |
|
so you'll be inferior in the virtual world as well. |
|
|
|
06:09.200 --> 06:11.320 |
|
But that's not how I view the world, right? |
|
|
|
06:11.320 --> 06:12.440 |
|
I don't view the world. |
|
|
|
06:12.440 --> 06:15.640 |
|
I mean, that's a very like, almost zero summish way |
|
|
|
06:15.640 --> 06:16.480 |
|
to view the world. |
|
|
|
06:16.480 --> 06:18.800 |
|
Say like, my great apartment isn't great |
|
|
|
06:18.800 --> 06:20.400 |
|
because my neighbor has one too. |
|
|
|
06:20.400 --> 06:21.640 |
|
No, my great apartment is great |
|
|
|
06:21.640 --> 06:24.320 |
|
because like, look at this dishwasher, man. |
|
|
|
06:24.320 --> 06:26.640 |
|
You just touch the dish and it's washed, right? |
|
|
|
06:26.640 --> 06:28.680 |
|
And that is great in and of itself |
|
|
|
06:28.680 --> 06:30.120 |
|
if I had the only apartment |
|
|
|
06:30.120 --> 06:31.520 |
|
or if everybody had the apartment. |
|
|
|
06:31.520 --> 06:32.400 |
|
I don't care. |
|
|
|
06:32.400 --> 06:34.760 |
|
So you have fundamental gratitude. |
|
|
|
06:34.760 --> 06:39.080 |
|
The world first learned of Geohot, George Hots |
|
|
|
06:39.080 --> 06:42.280 |
|
in August 2007, maybe before then, |
|
|
|
06:42.280 --> 06:44.080 |
|
but certainly in August 2007 |
|
|
|
06:44.080 --> 06:46.760 |
|
when you were the first person to unlock, |
|
|
|
06:46.760 --> 06:48.880 |
|
carry on lock an iPhone. |
|
|
|
06:48.880 --> 06:50.520 |
|
How did you get into hacking? |
|
|
|
06:50.520 --> 06:53.080 |
|
What was the first system you discovered |
|
|
|
06:53.080 --> 06:55.040 |
|
vulnerabilities for and broke into? |
|
|
|
06:56.240 --> 07:01.240 |
|
So that was really kind of the first thing. |
|
|
|
07:01.640 --> 07:06.640 |
|
I had a book in 2006 called Gray Hat Hacking. |
|
|
|
07:07.480 --> 07:11.000 |
|
And I guess I realized that |
|
|
|
07:11.000 --> 07:13.480 |
|
if you acquired these sort of powers |
|
|
|
07:13.480 --> 07:15.280 |
|
you could control the world. |
|
|
|
07:16.160 --> 07:18.920 |
|
But I didn't really know that much |
|
|
|
07:18.920 --> 07:20.560 |
|
about computers back then. |
|
|
|
07:20.560 --> 07:22.120 |
|
I started with electronics. |
|
|
|
07:22.120 --> 07:24.200 |
|
The first iPhone hack was physical. |
|
|
|
07:24.200 --> 07:25.040 |
|
Cardware. |
|
|
|
07:25.040 --> 07:28.160 |
|
You had to open it up and pull an address line high. |
|
|
|
07:28.160 --> 07:29.960 |
|
And it was because I didn't really know |
|
|
|
07:29.960 --> 07:31.320 |
|
about software exploitation. |
|
|
|
07:31.320 --> 07:32.960 |
|
I learned that all in the next few years |
|
|
|
07:32.960 --> 07:33.920 |
|
and I got very good at it. |
|
|
|
07:33.920 --> 07:36.560 |
|
But back then I knew about like |
|
|
|
07:36.560 --> 07:38.920 |
|
how memory chips are connected to processors and stuff. |
|
|
|
07:38.920 --> 07:41.040 |
|
But you knew about software and programming. |
|
|
|
07:41.040 --> 07:43.200 |
|
You didn't know. |
|
|
|
07:43.200 --> 07:46.160 |
|
Oh really, so your view of the world |
|
|
|
07:46.160 --> 07:49.320 |
|
and computers was physical, was hardware. |
|
|
|
07:49.320 --> 07:52.400 |
|
Actually, if you read the code that I released with that |
|
|
|
07:52.400 --> 07:55.760 |
|
in August 2007, it's atrocious. |
|
|
|
07:55.760 --> 07:56.760 |
|
What language was it? |
|
|
|
07:56.760 --> 07:57.600 |
|
C. |
|
|
|
07:57.600 --> 07:58.440 |
|
C, nice. |
|
|
|
07:58.440 --> 08:01.480 |
|
And in a broken sort of state machine, ask C. |
|
|
|
08:01.480 --> 08:02.960 |
|
I didn't know how to program. |
|
|
|
08:02.960 --> 08:04.160 |
|
Yeah. |
|
|
|
08:04.160 --> 08:06.600 |
|
So how did you learn to program? |
|
|
|
08:07.520 --> 08:08.440 |
|
What was your journey? |
|
|
|
08:08.440 --> 08:10.040 |
|
I mean, we'll talk about it. |
|
|
|
08:10.040 --> 08:12.680 |
|
You've live streamed some of your programming. |
|
|
|
08:12.680 --> 08:14.400 |
|
This chaotic, beautiful mess. |
|
|
|
08:14.400 --> 08:16.480 |
|
How did you arrive at that? |
|
|
|
08:16.480 --> 08:18.640 |
|
Years and years of practice. |
|
|
|
08:18.640 --> 08:22.240 |
|
I interned at Google after, |
|
|
|
08:22.240 --> 08:24.800 |
|
the summer after the iPhone unlock. |
|
|
|
08:24.800 --> 08:26.720 |
|
And I did a contract for them |
|
|
|
08:26.720 --> 08:29.040 |
|
where I built a hardware for Street View |
|
|
|
08:29.040 --> 08:31.760 |
|
and I wrote a software library to interact with it. |
|
|
|
08:32.680 --> 08:34.920 |
|
And it was terrible code. |
|
|
|
08:34.920 --> 08:36.560 |
|
And for the first time I got feedback |
|
|
|
08:36.560 --> 08:38.760 |
|
from people who I respected saying, |
|
|
|
08:38.760 --> 08:41.160 |
|
no, like, don't write code like this. |
|
|
|
08:42.680 --> 08:45.680 |
|
Now, of course, just getting that feedback is not enough. |
|
|
|
08:45.680 --> 08:50.680 |
|
The way that I really got good was, |
|
|
|
08:51.000 --> 08:54.800 |
|
I wanted to write this thing that could emulate |
|
|
|
08:54.800 --> 08:58.440 |
|
and then visualize like arm binaries |
|
|
|
08:58.440 --> 09:00.040 |
|
because I wanted to hack the iPhone better. |
|
|
|
09:00.040 --> 09:01.960 |
|
And I didn't like that I couldn't see what the, |
|
|
|
09:01.960 --> 09:03.800 |
|
I couldn't single step through the processor |
|
|
|
09:03.800 --> 09:05.200 |
|
because I had no debugger on there, |
|
|
|
09:05.200 --> 09:06.640 |
|
especially for the low level things like the boot ROM |
|
|
|
09:06.640 --> 09:07.480 |
|
and the boot loader. |
|
|
|
09:07.480 --> 09:09.440 |
|
So I tried to build this tool to do it. |
|
|
|
09:10.920 --> 09:13.440 |
|
And I built the tool once and it was terrible. |
|
|
|
09:13.440 --> 09:15.120 |
|
I built the tool second times, it was terrible. |
|
|
|
09:15.120 --> 09:16.320 |
|
I built the tool third time. |
|
|
|
09:16.320 --> 09:18.600 |
|
This by the time I was at Facebook, it was kind of okay. |
|
|
|
09:18.600 --> 09:20.560 |
|
And then I built the tool fourth time |
|
|
|
09:20.560 --> 09:22.560 |
|
when I was a Google intern again in 2014. |
|
|
|
09:22.560 --> 09:24.320 |
|
And that was the first time I was like, |
|
|
|
09:24.320 --> 09:25.880 |
|
this is finally usable. |
|
|
|
09:25.880 --> 09:27.120 |
|
How do you pronounce this, Kira? |
|
|
|
09:27.120 --> 09:28.360 |
|
Kira, yeah. |
|
|
|
09:28.360 --> 09:31.840 |
|
So it's essentially the most efficient way |
|
|
|
09:31.840 --> 09:35.720 |
|
to visualize the change of state of the computer |
|
|
|
09:35.720 --> 09:37.200 |
|
as the program is running. |
|
|
|
09:37.200 --> 09:38.920 |
|
That's what you mean by debugger. |
|
|
|
09:38.920 --> 09:41.760 |
|
Yeah, it's a timeless debugger. |
|
|
|
09:41.760 --> 09:45.080 |
|
So you can rewind just as easily as going forward. |
|
|
|
09:45.080 --> 09:46.280 |
|
Think about, if you're using GDB, |
|
|
|
09:46.280 --> 09:47.880 |
|
you have to put a watch on a variable. |
|
|
|
09:47.880 --> 09:49.680 |
|
If you want to see if that variable changes. |
|
|
|
09:49.680 --> 09:51.480 |
|
In Kira, you can just click on that variable. |
|
|
|
09:51.480 --> 09:53.880 |
|
And then it shows every single time |
|
|
|
09:53.880 --> 09:56.520 |
|
when that variable was changed or accessed. |
|
|
|
09:56.520 --> 09:59.760 |
|
Think about it like get for your computer's, the run lock. |
|
|
|
09:59.760 --> 10:04.760 |
|
So there's like a deep log of the state of the computer |
|
|
|
10:05.640 --> 10:07.840 |
|
as the program runs and you can rewind. |
|
|
|
10:07.840 --> 10:11.480 |
|
Why isn't that, maybe it is, maybe you can educate me. |
|
|
|
10:11.480 --> 10:14.640 |
|
Why isn't that kind of debugging used more often? |
|
|
|
10:14.640 --> 10:16.320 |
|
Because the tooling's bad. |
|
|
|
10:16.320 --> 10:17.160 |
|
Well, two things. |
|
|
|
10:17.160 --> 10:19.360 |
|
One, if you're trying to debug Chrome, |
|
|
|
10:19.360 --> 10:22.920 |
|
Chrome is a 200 megabyte binary |
|
|
|
10:22.920 --> 10:25.440 |
|
that runs slowly on desktops. |
|
|
|
10:25.440 --> 10:27.760 |
|
So that's gonna be really hard to use for that. |
|
|
|
10:27.760 --> 10:30.160 |
|
But it's really good to use for like CTFs |
|
|
|
10:30.160 --> 10:33.200 |
|
and for boot ROMs and for small parts of code. |
|
|
|
10:33.200 --> 10:36.360 |
|
So it's hard if you're trying to debug like massive systems. |
|
|
|
10:36.360 --> 10:38.200 |
|
What's a CTF and what's a boot ROM? |
|
|
|
10:38.200 --> 10:40.480 |
|
A boot ROM is the first code that executes |
|
|
|
10:40.480 --> 10:42.280 |
|
the minute you give power to your iPhone. |
|
|
|
10:42.280 --> 10:43.520 |
|
Okay. |
|
|
|
10:43.520 --> 10:46.040 |
|
And CTF were these competitions that I played. |
|
|
|
10:46.040 --> 10:46.880 |
|
Capture the flag. |
|
|
|
10:46.880 --> 10:47.720 |
|
Capture the flag. |
|
|
|
10:47.720 --> 10:48.560 |
|
I was gonna ask you about that. |
|
|
|
10:48.560 --> 10:49.920 |
|
What are those, those look at, |
|
|
|
10:49.920 --> 10:51.440 |
|
I watched a couple of videos on YouTube. |
|
|
|
10:51.440 --> 10:52.920 |
|
Those look fascinating. |
|
|
|
10:52.920 --> 10:55.560 |
|
What have you learned about maybe at the high level |
|
|
|
10:55.560 --> 10:58.040 |
|
in the vulnerability of systems from these competitions? |
|
|
|
11:00.840 --> 11:04.200 |
|
I feel like in the heyday of CTFs, |
|
|
|
11:04.200 --> 11:08.160 |
|
you had all of the best security people in the world |
|
|
|
11:08.160 --> 11:10.720 |
|
challenging each other and coming up |
|
|
|
11:10.720 --> 11:13.640 |
|
with new toy exploitable things over here. |
|
|
|
11:13.640 --> 11:15.400 |
|
And then everybody, okay, who can break it? |
|
|
|
11:15.400 --> 11:17.160 |
|
And when you break it, you get like, |
|
|
|
11:17.160 --> 11:19.360 |
|
there's like a file in the server called flag. |
|
|
|
11:19.360 --> 11:20.960 |
|
And then there's a program running, |
|
|
|
11:20.960 --> 11:22.680 |
|
listening on a socket that's vulnerable. |
|
|
|
11:22.680 --> 11:25.000 |
|
So you write an exploit, you get a shell, |
|
|
|
11:25.000 --> 11:27.160 |
|
and then you cat flag, and then you type the flag |
|
|
|
11:27.160 --> 11:29.480 |
|
into like a web based scoreboard and you get points. |
|
|
|
11:29.480 --> 11:33.000 |
|
So the goal is essentially to find an exploit in the system |
|
|
|
11:33.000 --> 11:35.280 |
|
that allows you to run shell, |
|
|
|
11:35.280 --> 11:38.040 |
|
to run arbitrary code on that system. |
|
|
|
11:38.040 --> 11:40.200 |
|
That's one of the categories. |
|
|
|
11:40.200 --> 11:41.960 |
|
That's like the Poneable category. |
|
|
|
11:43.560 --> 11:44.400 |
|
Poneable? |
|
|
|
11:44.400 --> 11:45.240 |
|
Yeah, Poneable. |
|
|
|
11:45.240 --> 11:47.600 |
|
It's like, you know, you Pone the program. |
|
|
|
11:47.600 --> 11:48.440 |
|
It's a program. |
|
|
|
11:48.440 --> 11:51.760 |
|
Oh, yeah. |
|
|
|
11:51.760 --> 11:55.360 |
|
You know, first of all, I apologize, I'm gonna say, |
|
|
|
11:55.360 --> 11:56.280 |
|
it's because I'm Russian, |
|
|
|
11:56.280 --> 11:59.120 |
|
but maybe you can help educate me. |
|
|
|
12:00.120 --> 12:01.680 |
|
Some video game like misspelled |
|
|
|
12:01.680 --> 12:02.840 |
|
to own way back in the day. |
|
|
|
12:02.840 --> 12:04.880 |
|
Yeah, and it's just, |
|
|
|
12:04.880 --> 12:06.280 |
|
I wonder if there's a definition |
|
|
|
12:06.280 --> 12:08.000 |
|
and I'll have to go to Urban Dictionary for it. |
|
|
|
12:08.000 --> 12:09.800 |
|
Yeah, it'd be interesting to see what it says. |
|
|
|
12:09.800 --> 12:12.760 |
|
Okay, so what was the heyday of CTL, by the way, |
|
|
|
12:12.760 --> 12:15.480 |
|
but was it, what decade are we talking about? |
|
|
|
12:15.480 --> 12:18.400 |
|
I think like, I mean, maybe I'm biased |
|
|
|
12:18.400 --> 12:21.120 |
|
because it's the era that I played, |
|
|
|
12:21.120 --> 12:25.800 |
|
but like 2011 to 2015, |
|
|
|
12:27.200 --> 12:30.320 |
|
because the modern CTF scene |
|
|
|
12:30.320 --> 12:32.640 |
|
is similar to the modern competitive programming scene. |
|
|
|
12:32.640 --> 12:34.280 |
|
You have people who like do drills. |
|
|
|
12:34.280 --> 12:35.880 |
|
You have people who practice. |
|
|
|
12:35.880 --> 12:37.040 |
|
And then once you've done that, |
|
|
|
12:37.040 --> 12:40.040 |
|
you've turned it less into a game of generic computer skill |
|
|
|
12:40.040 --> 12:42.440 |
|
and more into a game of, okay, you memorize, |
|
|
|
12:42.440 --> 12:44.620 |
|
you drill on these five categories. |
|
|
|
12:45.760 --> 12:48.920 |
|
And then before that, it wasn't, |
|
|
|
12:48.920 --> 12:51.560 |
|
it didn't have like as much attention as it had. |
|
|
|
12:52.800 --> 12:53.640 |
|
I don't know, they were like, |
|
|
|
12:53.640 --> 12:55.200 |
|
I won $30,000 once in Korea |
|
|
|
12:55.200 --> 12:56.120 |
|
for one of these competitions. |
|
|
|
12:56.120 --> 12:56.960 |
|
Holy crap. |
|
|
|
12:56.960 --> 12:57.920 |
|
Yeah, they were, they were, that was... |
|
|
|
12:57.920 --> 12:59.520 |
|
So that means, I mean, money is money, |
|
|
|
12:59.520 --> 13:02.320 |
|
but that means there was probably good people there. |
|
|
|
13:02.320 --> 13:03.600 |
|
Exactly, yeah. |
|
|
|
13:03.600 --> 13:06.800 |
|
Are the challenges human constructed |
|
|
|
13:06.800 --> 13:10.760 |
|
or are they grounded in some real flaws in real systems? |
|
|
|
13:10.760 --> 13:13.080 |
|
Usually they're human constructed, |
|
|
|
13:13.080 --> 13:15.760 |
|
but they're usually inspired by real flaws. |
|
|
|
13:15.760 --> 13:17.320 |
|
What kind of systems are imagined |
|
|
|
13:17.320 --> 13:19.080 |
|
is really focused on mobile? |
|
|
|
13:19.080 --> 13:20.920 |
|
Like what has vulnerabilities these days? |
|
|
|
13:20.920 --> 13:25.120 |
|
Is it primarily mobile systems like Android? |
|
|
|
13:25.120 --> 13:26.680 |
|
Oh, everything does. |
|
|
|
13:26.680 --> 13:28.120 |
|
Yeah, of course. |
|
|
|
13:28.120 --> 13:29.360 |
|
The price has kind of gone up |
|
|
|
13:29.360 --> 13:31.280 |
|
because less and less people can find them. |
|
|
|
13:31.280 --> 13:33.160 |
|
And what's happened in security is now, |
|
|
|
13:33.160 --> 13:34.560 |
|
if you want to like jailbreak an iPhone, |
|
|
|
13:34.560 --> 13:36.960 |
|
you don't need one exploit anymore, you need nine. |
|
|
|
13:37.960 --> 13:39.160 |
|
Nine change together? |
|
|
|
13:39.160 --> 13:40.000 |
|
What would you mean? |
|
|
|
13:40.000 --> 13:40.840 |
|
Yeah, wow. |
|
|
|
13:40.840 --> 13:44.800 |
|
Okay, so it's really, what's the benefit? |
|
|
|
13:44.800 --> 13:48.240 |
|
Speaking higher level philosophically about hacking. |
|
|
|
13:48.240 --> 13:50.400 |
|
I mean, it sounds from everything I've seen about you, |
|
|
|
13:50.400 --> 13:55.040 |
|
you just love the challenge and you don't want to do anything. |
|
|
|
13:55.040 --> 13:58.120 |
|
You don't want to bring that exploit out into the world |
|
|
|
13:58.120 --> 14:01.680 |
|
and do any actual, let it run wild. |
|
|
|
14:01.680 --> 14:02.760 |
|
You just want to solve it |
|
|
|
14:02.760 --> 14:05.400 |
|
and then you go on to the next thing. |
|
|
|
14:05.400 --> 14:08.440 |
|
Oh yeah, I mean, doing criminal stuff's not really worth it. |
|
|
|
14:08.440 --> 14:10.520 |
|
And I'll actually use the same argument |
|
|
|
14:10.520 --> 14:15.440 |
|
for why I don't do defense for why I don't do crime. |
|
|
|
14:15.440 --> 14:16.840 |
|
If you want to defend a system, |
|
|
|
14:16.840 --> 14:19.280 |
|
say the system has 10 holes, right? |
|
|
|
14:19.280 --> 14:22.240 |
|
If you find nine of those holes as a defender, |
|
|
|
14:22.240 --> 14:24.240 |
|
you still lose because the attacker gets in |
|
|
|
14:24.240 --> 14:25.520 |
|
through the last one. |
|
|
|
14:25.520 --> 14:26.360 |
|
If you're an attacker, |
|
|
|
14:26.360 --> 14:28.720 |
|
you only have to find one out of the 10. |
|
|
|
14:28.720 --> 14:30.760 |
|
But if you're a criminal, |
|
|
|
14:30.760 --> 14:34.800 |
|
if you log on with a VPN nine out of the 10 times, |
|
|
|
14:34.800 --> 14:37.760 |
|
but one time you forget, you're done. |
|
|
|
14:37.760 --> 14:39.400 |
|
Because you're caught, okay. |
|
|
|
14:39.400 --> 14:41.160 |
|
Because you only have to mess up once |
|
|
|
14:41.160 --> 14:42.920 |
|
to be caught as a criminal. |
|
|
|
14:42.920 --> 14:44.320 |
|
That's why I'm not a criminal. |
|
|
|
14:45.920 --> 14:47.080 |
|
But okay, let me, |
|
|
|
14:47.080 --> 14:49.520 |
|
cause I was having a discussion with somebody |
|
|
|
14:49.520 --> 14:52.440 |
|
just at a high level about nuclear weapons, |
|
|
|
14:52.440 --> 14:56.240 |
|
actually why we're having blown ourselves up yet. |
|
|
|
14:56.240 --> 14:59.840 |
|
And my feeling is all the smart people in the world, |
|
|
|
14:59.840 --> 15:04.120 |
|
if you look at the distribution of smart people, |
|
|
|
15:04.120 --> 15:06.760 |
|
smart people are generally good. |
|
|
|
15:06.760 --> 15:07.680 |
|
And then the Southern person, |
|
|
|
15:07.680 --> 15:09.480 |
|
I was talking to Sean Carroll, the physicist, |
|
|
|
15:09.480 --> 15:11.400 |
|
and he was saying no good and bad people |
|
|
|
15:11.400 --> 15:14.080 |
|
are evenly distributed amongst everybody. |
|
|
|
15:14.080 --> 15:18.080 |
|
My sense was good hackers are in general good people |
|
|
|
15:18.080 --> 15:20.400 |
|
and they don't want to mess with the world. |
|
|
|
15:20.400 --> 15:21.920 |
|
What's your sense? |
|
|
|
15:21.920 --> 15:24.720 |
|
I'm not even sure about that. |
|
|
|
15:25.920 --> 15:30.520 |
|
Like, I have a nice life. |
|
|
|
15:30.520 --> 15:32.120 |
|
Crime wouldn't get me anything. |
|
|
|
15:34.320 --> 15:36.520 |
|
But if you're good and you have these skills, |
|
|
|
15:36.520 --> 15:38.720 |
|
you probably have a nice life too, right? |
|
|
|
15:38.720 --> 15:40.160 |
|
Right, you can use the father things. |
|
|
|
15:40.160 --> 15:41.120 |
|
But is there an ethical, |
|
|
|
15:41.120 --> 15:44.200 |
|
is there a little voice in your head that says, |
|
|
|
15:46.120 --> 15:49.040 |
|
well, yeah, if you could hack something |
|
|
|
15:49.040 --> 15:50.720 |
|
to where you could hurt people |
|
|
|
15:52.840 --> 15:54.960 |
|
and you could earn a lot of money doing it though, |
|
|
|
15:54.960 --> 15:56.320 |
|
not hurt physically perhaps, |
|
|
|
15:56.320 --> 15:59.000 |
|
but disrupt their life in some kind of way. |
|
|
|
16:00.200 --> 16:02.360 |
|
Isn't there a little voice that says, |
|
|
|
16:03.360 --> 16:04.560 |
|
Well, two things. |
|
|
|
16:04.560 --> 16:06.800 |
|
One, I don't really care about money. |
|
|
|
16:06.800 --> 16:08.680 |
|
So like the money wouldn't be an incentive. |
|
|
|
16:08.680 --> 16:10.640 |
|
The thrill might be an incentive. |
|
|
|
16:10.640 --> 16:14.440 |
|
But when I was 19, I read crime and punishment. |
|
|
|
16:14.440 --> 16:16.120 |
|
That was another great one |
|
|
|
16:16.120 --> 16:18.440 |
|
that talked me out of ever really doing crime. |
|
|
|
16:19.400 --> 16:21.720 |
|
Cause it's like, that's gonna be me. |
|
|
|
16:21.720 --> 16:25.040 |
|
I'd get away with it, but it would just run through my head. |
|
|
|
16:25.040 --> 16:26.480 |
|
Even if I got away with it, you know? |
|
|
|
16:26.480 --> 16:27.640 |
|
And then you do crime for long enough, |
|
|
|
16:27.640 --> 16:28.960 |
|
you'll never get away with it. |
|
|
|
16:28.960 --> 16:30.360 |
|
That's right, in the end. |
|
|
|
16:30.360 --> 16:32.680 |
|
That's a good reason to be good. |
|
|
|
16:32.680 --> 16:34.880 |
|
I wouldn't say I'm good, I would just say I'm not bad. |
|
|
|
16:34.880 --> 16:38.080 |
|
You're a talented programmer and a hacker |
|
|
|
16:38.080 --> 16:40.920 |
|
in a good positive sense of the word. |
|
|
|
16:40.920 --> 16:43.360 |
|
You've played around, found vulnerabilities |
|
|
|
16:43.360 --> 16:44.720 |
|
in various systems. |
|
|
|
16:44.720 --> 16:46.120 |
|
What have you learned broadly |
|
|
|
16:46.120 --> 16:49.480 |
|
about the design of systems and so on |
|
|
|
16:49.480 --> 16:51.520 |
|
from that whole process? |
|
|
|
16:53.280 --> 16:58.280 |
|
You learn to not take things |
|
|
|
16:59.280 --> 17:02.160 |
|
for what people say they are, |
|
|
|
17:02.160 --> 17:05.320 |
|
but you look at things for what they actually are. |
|
|
|
17:07.040 --> 17:07.880 |
|
Yeah. |
|
|
|
17:07.880 --> 17:10.080 |
|
I understand that's what you tell me it is, |
|
|
|
17:10.080 --> 17:11.320 |
|
but what does it do? |
|
|
|
17:12.960 --> 17:14.600 |
|
And you have nice visualization tools |
|
|
|
17:14.600 --> 17:16.720 |
|
to really know what it's really doing. |
|
|
|
17:16.720 --> 17:20.080 |
|
Oh, I wish I'm a better programmer now than I was in 2014. |
|
|
|
17:20.080 --> 17:21.880 |
|
I said, Kira, that was the first tool |
|
|
|
17:21.880 --> 17:23.440 |
|
that I wrote that was usable. |
|
|
|
17:23.440 --> 17:25.360 |
|
I wouldn't say the code was great. |
|
|
|
17:25.360 --> 17:27.360 |
|
I still wouldn't say my code is great. |
|
|
|
17:28.840 --> 17:30.760 |
|
So how was your evolution as a programmer? |
|
|
|
17:30.760 --> 17:32.280 |
|
Except practice. |
|
|
|
17:32.280 --> 17:33.880 |
|
You started with C, |
|
|
|
17:33.880 --> 17:35.560 |
|
what point did you pick up Python? |
|
|
|
17:35.560 --> 17:37.080 |
|
Because you're pretty big in Python now. |
|
|
|
17:37.080 --> 17:39.960 |
|
Now, yeah, in college, |
|
|
|
17:39.960 --> 17:42.520 |
|
I went to Carnegie Mellon when I was 22. |
|
|
|
17:42.520 --> 17:44.200 |
|
I went back, I'm like, |
|
|
|
17:44.200 --> 17:46.640 |
|
I'm gonna take all your hardest CS courses |
|
|
|
17:46.640 --> 17:47.640 |
|
and we'll see how I do, right? |
|
|
|
17:47.640 --> 17:48.560 |
|
Like, did I miss anything |
|
|
|
17:48.560 --> 17:51.520 |
|
by not having a real undergraduate education? |
|
|
|
17:51.520 --> 17:54.240 |
|
Took operating systems, compilers, AI, |
|
|
|
17:54.240 --> 17:56.880 |
|
and they're like a freshman Weeder math course. |
|
|
|
17:56.880 --> 18:01.880 |
|
And some of those classes you mentioned, |
|
|
|
18:03.320 --> 18:04.240 |
|
pretty tough, actually. |
|
|
|
18:04.240 --> 18:05.640 |
|
They're great. |
|
|
|
18:05.640 --> 18:07.640 |
|
At least when the 2012, |
|
|
|
18:07.640 --> 18:10.240 |
|
circa 2012 operating systems and compilers |
|
|
|
18:11.240 --> 18:14.440 |
|
were two of the best classes I've ever taken in my life. |
|
|
|
18:14.440 --> 18:15.640 |
|
Because you write an operating system |
|
|
|
18:15.640 --> 18:16.840 |
|
and you write a compiler. |
|
|
|
18:18.080 --> 18:19.760 |
|
I wrote my operating system in C |
|
|
|
18:19.760 --> 18:21.400 |
|
and I wrote my compiler in Haskell, |
|
|
|
18:21.400 --> 18:26.400 |
|
but somehow I picked up Python that semester as well. |
|
|
|
18:26.400 --> 18:28.080 |
|
I started using it for the CTFs, actually. |
|
|
|
18:28.080 --> 18:30.320 |
|
That's when I really started to get into CTFs |
|
|
|
18:30.320 --> 18:33.360 |
|
and CTFs, you're all to race against the clock. |
|
|
|
18:33.360 --> 18:35.120 |
|
So I can't write things and see. |
|
|
|
18:35.120 --> 18:36.240 |
|
Oh, there's a clock component. |
|
|
|
18:36.240 --> 18:37.840 |
|
So you really want to use the programming language |
|
|
|
18:37.840 --> 18:38.960 |
|
just so you can be fastest. |
|
|
|
18:38.960 --> 18:40.080 |
|
48 hours. |
|
|
|
18:40.080 --> 18:41.440 |
|
Pwn as many of these challenges as you can. |
|
|
|
18:41.440 --> 18:42.280 |
|
Pwn. |
|
|
|
18:42.280 --> 18:43.120 |
|
Yeah. |
|
|
|
18:43.120 --> 18:43.960 |
|
You got like 100 points of challenge, |
|
|
|
18:43.960 --> 18:45.360 |
|
whatever team gets the most. |
|
|
|
18:46.360 --> 18:50.240 |
|
You were both at Facebook and Google for a brief stint. |
|
|
|
18:50.240 --> 18:51.080 |
|
Yeah. |
|
|
|
18:51.080 --> 18:54.920 |
|
With Project Zero, actually, at Google for five months |
|
|
|
18:54.920 --> 18:56.960 |
|
where you develop Kira. |
|
|
|
18:56.960 --> 18:59.280 |
|
What was Project Zero about in general? |
|
|
|
19:01.760 --> 19:05.160 |
|
Just curious about the security efforts in these companies. |
|
|
|
19:05.160 --> 19:08.840 |
|
Well, Project Zero started the same time I went there. |
|
|
|
19:08.840 --> 19:10.080 |
|
What year is it there? |
|
|
|
19:11.080 --> 19:12.320 |
|
2015. |
|
|
|
19:12.320 --> 19:13.160 |
|
2015. |
|
|
|
19:13.160 --> 19:15.040 |
|
So that was right at the beginning of Project Zero. |
|
|
|
19:15.040 --> 19:16.200 |
|
It's small. |
|
|
|
19:16.200 --> 19:18.840 |
|
It's Google's offensive security team. |
|
|
|
19:18.840 --> 19:23.840 |
|
I'll try to give the best public facing explanation |
|
|
|
19:25.680 --> 19:26.520 |
|
that I can. |
|
|
|
19:26.520 --> 19:30.960 |
|
So the idea is basically, |
|
|
|
19:30.960 --> 19:33.240 |
|
these vulnerabilities exist in the world. |
|
|
|
19:33.240 --> 19:35.240 |
|
Nation states have them. |
|
|
|
19:35.240 --> 19:37.440 |
|
Some high powered bad actors have them. |
|
|
|
19:39.840 --> 19:44.200 |
|
Sometimes people will find these vulnerabilities |
|
|
|
19:44.200 --> 19:47.960 |
|
and submit them in bug bounties to the companies. |
|
|
|
19:47.960 --> 19:49.440 |
|
But a lot of the companies don't only care. |
|
|
|
19:49.440 --> 19:50.520 |
|
They don't even fix the bug. |
|
|
|
19:50.520 --> 19:53.760 |
|
It doesn't hurt for there to be a vulnerability. |
|
|
|
19:53.760 --> 19:55.880 |
|
So Project Zero is like, we're going to do it different. |
|
|
|
19:55.880 --> 19:57.840 |
|
We're going to announce a vulnerability |
|
|
|
19:57.840 --> 19:59.640 |
|
and we're going to give them 90 days to fix it. |
|
|
|
19:59.640 --> 20:00.800 |
|
And then whether they fix it or not, |
|
|
|
20:00.800 --> 20:03.200 |
|
we're going to drop the Zero Day. |
|
|
|
20:03.200 --> 20:04.080 |
|
Oh, wow. |
|
|
|
20:04.080 --> 20:05.240 |
|
We're going to drop the weapon on the textbook. |
|
|
|
20:05.240 --> 20:06.080 |
|
That's so cool. |
|
|
|
20:06.080 --> 20:07.480 |
|
That is so cool. |
|
|
|
20:07.480 --> 20:09.200 |
|
I love that deadlines. |
|
|
|
20:09.200 --> 20:10.040 |
|
Oh, that's so cool. |
|
|
|
20:10.040 --> 20:10.880 |
|
Give them real deadlines. |
|
|
|
20:10.880 --> 20:12.320 |
|
Yeah. |
|
|
|
20:12.320 --> 20:15.800 |
|
And I think it's done a lot for moving the industry forward. |
|
|
|
20:15.800 --> 20:20.360 |
|
I watched your coding sessions on the streamed online. |
|
|
|
20:20.360 --> 20:25.280 |
|
You code things up, the basic projects, usually from scratch. |
|
|
|
20:25.280 --> 20:28.200 |
|
I would say, sort of as a programmer myself, |
|
|
|
20:28.200 --> 20:30.360 |
|
just watching you, that you type really fast |
|
|
|
20:30.360 --> 20:34.440 |
|
and your brain works in both brilliant and chaotic ways. |
|
|
|
20:34.440 --> 20:35.800 |
|
I don't know if that's always true, |
|
|
|
20:35.800 --> 20:37.600 |
|
but certainly for the live streams. |
|
|
|
20:37.600 --> 20:41.320 |
|
So it's interesting to me because I'm much slower |
|
|
|
20:41.320 --> 20:43.520 |
|
and systematic and careful. |
|
|
|
20:43.520 --> 20:48.040 |
|
And you just move probably in order of magnitude faster. |
|
|
|
20:48.040 --> 20:51.800 |
|
So I'm curious, is there a method to your madness? |
|
|
|
20:51.800 --> 20:53.040 |
|
Or is it just who you are? |
|
|
|
20:53.040 --> 20:54.720 |
|
There's pros and cons. |
|
|
|
20:54.720 --> 20:58.080 |
|
There's pros and cons to my programming style. |
|
|
|
20:58.080 --> 21:00.360 |
|
And I'm aware of them. |
|
|
|
21:00.360 --> 21:04.480 |
|
If you ask me to get something up and working quickly |
|
|
|
21:04.480 --> 21:06.800 |
|
with an API that's kind of undocumented, |
|
|
|
21:06.800 --> 21:08.880 |
|
I will do this super fast because I will throw things |
|
|
|
21:08.880 --> 21:10.200 |
|
at it until it works. |
|
|
|
21:10.200 --> 21:14.720 |
|
If you ask me to take a vector and rotate it 90 degrees |
|
|
|
21:14.720 --> 21:19.320 |
|
and then flip it over the X, Y plane, |
|
|
|
21:19.320 --> 21:22.280 |
|
I'll spam program for two hours and won't get it. |
|
|
|
21:22.280 --> 21:23.480 |
|
Oh, because it's something that you |
|
|
|
21:23.480 --> 21:26.240 |
|
could do with a sheet of paper or think through design |
|
|
|
21:26.240 --> 21:30.400 |
|
and then just you really just throw stuff at the wall |
|
|
|
21:30.400 --> 21:34.600 |
|
and you get so good at it that it usually works. |
|
|
|
21:34.600 --> 21:36.920 |
|
I should become better at the other kind as well. |
|
|
|
21:36.920 --> 21:39.440 |
|
Sometimes I will do things methodically. |
|
|
|
21:39.440 --> 21:41.200 |
|
It's nowhere near as entertaining on the Twitch streams. |
|
|
|
21:41.200 --> 21:43.520 |
|
I do exaggerate it a bit on the Twitch streams as well. |
|
|
|
21:43.520 --> 21:45.480 |
|
The Twitch streams, I mean, what do you want to see a game |
|
|
|
21:45.480 --> 21:46.840 |
|
or you want to see actions permit, right? |
|
|
|
21:46.840 --> 21:48.200 |
|
I'll show you APM for programming too. |
|
|
|
21:48.200 --> 21:50.280 |
|
Yeah, I'd recommend people go to it. |
|
|
|
21:50.280 --> 21:53.800 |
|
I think I watched probably several hours that you put, |
|
|
|
21:53.800 --> 21:57.480 |
|
like I've actually left you programming in the background |
|
|
|
21:57.480 --> 22:00.400 |
|
while I was programming because you made me, |
|
|
|
22:00.400 --> 22:03.120 |
|
it was like watching a really good gamer. |
|
|
|
22:03.120 --> 22:06.240 |
|
It's like energizes you because you're like moving so fast |
|
|
|
22:06.240 --> 22:08.840 |
|
and so it's awesome, it's inspiring. |
|
|
|
22:08.840 --> 22:11.200 |
|
It made me jealous that like, |
|
|
|
22:12.280 --> 22:14.280 |
|
because my own programming is inadequate |
|
|
|
22:14.280 --> 22:16.960 |
|
in terms of speed, so I was like. |
|
|
|
22:16.960 --> 22:20.520 |
|
So I'm twice as frantic on the live streams |
|
|
|
22:20.520 --> 22:22.680 |
|
as I am when I code without, oh. |
|
|
|
22:22.680 --> 22:23.720 |
|
It's super entertaining. |
|
|
|
22:23.720 --> 22:26.400 |
|
So I wasn't even paying attention to what you were coding, |
|
|
|
22:26.400 --> 22:29.760 |
|
which is great, it's just watching you switch windows |
|
|
|
22:29.760 --> 22:31.400 |
|
and Vim, I guess is the most way. |
|
|
|
22:31.400 --> 22:33.000 |
|
Yeah, does Vim on screen? |
|
|
|
22:33.000 --> 22:35.640 |
|
I've developed a workload Facebook and stuck with it. |
|
|
|
22:35.640 --> 22:37.320 |
|
How do you learn new programming tools, |
|
|
|
22:37.320 --> 22:39.440 |
|
ideas, techniques these days? |
|
|
|
22:39.440 --> 22:42.080 |
|
What's your like methodology for learning new things? |
|
|
|
22:42.080 --> 22:45.920 |
|
So I wrote for comma, |
|
|
|
22:47.200 --> 22:49.280 |
|
the distributed file systems out in the world |
|
|
|
22:49.280 --> 22:50.720 |
|
are extremely complex. |
|
|
|
22:50.720 --> 22:55.280 |
|
Like if you want to install something like like like Ceph, |
|
|
|
22:55.280 --> 22:58.760 |
|
Ceph is I think the like open infrastructure |
|
|
|
22:58.760 --> 23:03.040 |
|
distributed file system or there's like newer ones |
|
|
|
23:03.040 --> 23:05.880 |
|
like seaweed FS, but these are all like 10,000 |
|
|
|
23:05.880 --> 23:06.880 |
|
plus line projects. |
|
|
|
23:06.880 --> 23:09.520 |
|
I think some of them are even 100,000 line |
|
|
|
23:09.520 --> 23:11.120 |
|
and just configuring them as a nightmare. |
|
|
|
23:11.120 --> 23:16.120 |
|
So I wrote, I wrote one, it's 200 lines |
|
|
|
23:16.440 --> 23:18.880 |
|
and it uses like engine X of the line servers |
|
|
|
23:18.880 --> 23:21.600 |
|
and has this little master server that I wrote and go. |
|
|
|
23:21.600 --> 23:24.840 |
|
And the way I go, this, if I would say |
|
|
|
23:24.840 --> 23:27.240 |
|
that I'm proud per line of any code I wrote, |
|
|
|
23:27.240 --> 23:29.160 |
|
maybe there's some exploits that I think are beautiful |
|
|
|
23:29.160 --> 23:31.320 |
|
and then this, this is 200 lines |
|
|
|
23:31.320 --> 23:33.720 |
|
and just the way that I thought about it, |
|
|
|
23:33.720 --> 23:35.560 |
|
I think was very good and the reason it's very good |
|
|
|
23:35.560 --> 23:37.640 |
|
is because that was the fourth version of it that I wrote |
|
|
|
23:37.640 --> 23:39.320 |
|
and I had three versions that I threw away. |
|
|
|
23:39.320 --> 23:41.000 |
|
You mentioned, did you say go? |
|
|
|
23:41.000 --> 23:41.840 |
|
I wrote a go, yeah. |
|
|
|
23:41.840 --> 23:42.680 |
|
And go. |
|
|
|
23:42.680 --> 23:43.880 |
|
Is that a functional language? |
|
|
|
23:43.880 --> 23:45.280 |
|
I forget what go is. |
|
|
|
23:45.280 --> 23:47.160 |
|
Go is Google's language. |
|
|
|
23:47.160 --> 23:48.200 |
|
Right. |
|
|
|
23:48.200 --> 23:49.480 |
|
It's not functional. |
|
|
|
23:49.480 --> 23:54.480 |
|
It's some, it's like, in a way it's C++, but easier. |
|
|
|
23:56.160 --> 23:58.200 |
|
It's strongly typed. |
|
|
|
23:58.200 --> 23:59.760 |
|
It has a nice ecosystem around it. |
|
|
|
23:59.760 --> 24:01.680 |
|
When I first looked at it, I was like, |
|
|
|
24:01.680 --> 24:03.800 |
|
this is like Python, but it takes twice as long |
|
|
|
24:03.800 --> 24:05.600 |
|
to do anything. |
|
|
|
24:05.600 --> 24:09.600 |
|
Now that I've open pilot is migrating to C, |
|
|
|
24:09.600 --> 24:11.000 |
|
but it still has large Python components, |
|
|
|
24:11.000 --> 24:12.760 |
|
I now understand why Python doesn't work |
|
|
|
24:12.760 --> 24:15.840 |
|
for large code bases and why you want something like go. |
|
|
|
24:15.840 --> 24:16.680 |
|
Interesting. |
|
|
|
24:16.680 --> 24:18.680 |
|
So why, why doesn't Python work for, |
|
|
|
24:18.680 --> 24:21.720 |
|
so even most, speaking for myself at least, |
|
|
|
24:21.720 --> 24:24.960 |
|
like we do a lot of stuff, basically demo level work |
|
|
|
24:24.960 --> 24:29.240 |
|
with autonomous vehicles and most of the work is Python. |
|
|
|
24:29.240 --> 24:32.440 |
|
Why doesn't Python work for large code bases? |
|
|
|
24:32.440 --> 24:37.440 |
|
Because, well, lack of type checking is a big one. |
|
|
|
24:37.920 --> 24:39.360 |
|
So errors creep in. |
|
|
|
24:39.360 --> 24:41.920 |
|
Yeah, and like you don't know, |
|
|
|
24:41.920 --> 24:45.320 |
|
the compiler can tell you like nothing, right? |
|
|
|
24:45.320 --> 24:48.440 |
|
So everything is either, you know, |
|
|
|
24:48.440 --> 24:49.880 |
|
like syntax errors, fine, |
|
|
|
24:49.880 --> 24:51.800 |
|
but if you misspell a variable in Python, |
|
|
|
24:51.800 --> 24:53.000 |
|
the compiler won't catch that. |
|
|
|
24:53.000 --> 24:56.600 |
|
There's like linters that can catch it some of the time. |
|
|
|
24:56.600 --> 24:57.560 |
|
There's no types. |
|
|
|
24:57.560 --> 25:00.520 |
|
This is really the biggest downside. |
|
|
|
25:00.520 --> 25:02.640 |
|
And then we'll Python slow, but that's not related to it. |
|
|
|
25:02.640 --> 25:04.840 |
|
Well, maybe it's kind of related to it, so it's lack of. |
|
|
|
25:04.840 --> 25:06.600 |
|
So what's in your toolbox these days? |
|
|
|
25:06.600 --> 25:07.760 |
|
Is it Python or what else? |
|
|
|
25:07.760 --> 25:08.600 |
|
Go. |
|
|
|
25:08.600 --> 25:10.240 |
|
I need to move to something else. |
|
|
|
25:10.240 --> 25:12.880 |
|
My adventure into dependently typed languages, |
|
|
|
25:12.880 --> 25:14.240 |
|
I love these languages. |
|
|
|
25:14.240 --> 25:17.520 |
|
They just have like syntax from the 80s. |
|
|
|
25:18.520 --> 25:21.120 |
|
What do you think about JavaScript? |
|
|
|
25:21.120 --> 25:24.000 |
|
ES6, like the modern type script? |
|
|
|
25:24.000 --> 25:27.320 |
|
JavaScript is, the whole ecosystem |
|
|
|
25:27.320 --> 25:29.320 |
|
is unbelievably confusing. |
|
|
|
25:29.320 --> 25:32.840 |
|
NPM updates a package from 022 to 025 |
|
|
|
25:32.840 --> 25:34.560 |
|
and that breaks your Babel linter, |
|
|
|
25:34.560 --> 25:38.560 |
|
which translates your ES5 into ES6, which doesn't run on. |
|
|
|
25:38.560 --> 25:42.480 |
|
So why do I have to compile my JavaScript again, huh? |
|
|
|
25:42.480 --> 25:44.040 |
|
It may be the future though. |
|
|
|
25:44.040 --> 25:45.800 |
|
You think about, I mean, |
|
|
|
25:45.800 --> 25:47.400 |
|
I've embraced JavaScript recently |
|
|
|
25:47.400 --> 25:52.280 |
|
just because just like I've continually embraced PHP, |
|
|
|
25:52.280 --> 25:55.360 |
|
it seems that these worst possible languages live on |
|
|
|
25:55.360 --> 25:57.480 |
|
for the longest, like cockroaches never die. |
|
|
|
25:57.480 --> 26:00.760 |
|
Yeah, well, it's in the browser and it's fast. |
|
|
|
26:00.760 --> 26:01.680 |
|
It's fast. |
|
|
|
26:01.680 --> 26:02.520 |
|
Yeah. |
|
|
|
26:02.520 --> 26:05.480 |
|
It's in the browser and compute might stay become, |
|
|
|
26:05.480 --> 26:06.440 |
|
you know, the browser, |
|
|
|
26:06.440 --> 26:09.040 |
|
it's unclear what the role of the browser is |
|
|
|
26:09.040 --> 26:11.800 |
|
in terms of distributed computation in the future. |
|
|
|
26:11.800 --> 26:12.640 |
|
So. |
|
|
|
26:13.600 --> 26:15.240 |
|
JavaScript is definitely here to stay. |
|
|
|
26:15.240 --> 26:16.080 |
|
Yeah. |
|
|
|
26:16.080 --> 26:18.160 |
|
It's interesting if autonomous vehicles |
|
|
|
26:18.160 --> 26:19.480 |
|
will run on JavaScript one day. |
|
|
|
26:19.480 --> 26:21.760 |
|
I mean, you have to consider these possibilities. |
|
|
|
26:21.760 --> 26:24.280 |
|
Well, all our debug tools are JavaScript. |
|
|
|
26:24.280 --> 26:26.040 |
|
We actually just open source them. |
|
|
|
26:26.040 --> 26:28.160 |
|
We have a tool explorer, which you can annotate |
|
|
|
26:28.160 --> 26:30.080 |
|
your disengagements and we have tool Kibana, |
|
|
|
26:30.080 --> 26:32.920 |
|
which lets you analyze the can traffic from the car. |
|
|
|
26:32.920 --> 26:35.240 |
|
So basically any time you're visualizing something |
|
|
|
26:35.240 --> 26:37.720 |
|
about the log using JavaScript. |
|
|
|
26:37.720 --> 26:40.120 |
|
Well, the web is the best UI toolkit by far. |
|
|
|
26:40.120 --> 26:40.960 |
|
Yeah. |
|
|
|
26:40.960 --> 26:41.880 |
|
So, and then, you know what? |
|
|
|
26:41.880 --> 26:42.760 |
|
You're coding in JavaScript. |
|
|
|
26:42.760 --> 26:43.600 |
|
We have a React guy. |
|
|
|
26:43.600 --> 26:44.440 |
|
He's good. |
|
|
|
26:44.440 --> 26:46.080 |
|
React, nice. |
|
|
|
26:46.080 --> 26:46.920 |
|
Let's get into it. |
|
|
|
26:46.920 --> 26:49.120 |
|
So let's talk autonomous vehicles. |
|
|
|
26:49.120 --> 26:50.640 |
|
You found a comma AI. |
|
|
|
26:51.440 --> 26:54.920 |
|
Let's, at a high level, |
|
|
|
26:54.920 --> 26:57.880 |
|
how did you get into the world of vehicle automation? |
|
|
|
26:57.880 --> 26:59.920 |
|
Can you also just, for people who don't know, |
|
|
|
26:59.920 --> 27:01.400 |
|
tell the story of comma AI? |
|
|
|
27:01.400 --> 27:02.920 |
|
Sure. |
|
|
|
27:02.920 --> 27:06.120 |
|
So I was working at this AI startup |
|
|
|
27:06.120 --> 27:09.240 |
|
and a friend approached me and he's like, |
|
|
|
27:09.240 --> 27:12.080 |
|
dude, I don't know where this is going, |
|
|
|
27:12.080 --> 27:15.160 |
|
but the coolest applied AI problem today |
|
|
|
27:15.160 --> 27:16.480 |
|
is self driving cars. |
|
|
|
27:16.480 --> 27:17.720 |
|
I'm like, well, absolutely. |
|
|
|
27:18.800 --> 27:20.520 |
|
You wanna meet with Elon Musk |
|
|
|
27:20.520 --> 27:24.560 |
|
and he's looking for somebody to build a vision system |
|
|
|
27:24.560 --> 27:27.600 |
|
for autopilot. |
|
|
|
27:27.600 --> 27:29.320 |
|
This is when they were still on AP one. |
|
|
|
27:29.320 --> 27:30.840 |
|
They were still using Mobileye. |
|
|
|
27:30.840 --> 27:33.680 |
|
Elon back then was looking for a replacement. |
|
|
|
27:33.680 --> 27:37.320 |
|
And he brought me in and we talked about a contract |
|
|
|
27:37.320 --> 27:39.040 |
|
where I would deliver something |
|
|
|
27:39.040 --> 27:41.640 |
|
that meets Mobileye level performance. |
|
|
|
27:41.640 --> 27:43.920 |
|
I would get paid $12 million if I could deliver it tomorrow |
|
|
|
27:43.920 --> 27:46.720 |
|
and I would lose $1 million for every month I didn't deliver. |
|
|
|
27:47.720 --> 27:49.080 |
|
So I was like, okay, this is a great deal. |
|
|
|
27:49.080 --> 27:50.800 |
|
This is a super exciting challenge. |
|
|
|
27:52.360 --> 27:53.200 |
|
You know what? |
|
|
|
27:53.200 --> 27:55.840 |
|
It takes me 10 months, I get $2 million, it's good. |
|
|
|
27:55.840 --> 27:57.160 |
|
Maybe I can finish up in five. |
|
|
|
27:57.160 --> 27:58.880 |
|
Maybe I don't finish it at all and I get paid nothing |
|
|
|
27:58.880 --> 28:00.880 |
|
and I'll work for 12 months for free. |
|
|
|
28:00.880 --> 28:02.960 |
|
So maybe just take a pause on that. |
|
|
|
28:02.960 --> 28:04.280 |
|
I'm also curious about this |
|
|
|
28:04.280 --> 28:06.360 |
|
because I've been working in robotics for a long time. |
|
|
|
28:06.360 --> 28:08.320 |
|
And I'm curious to see a person like you just step in |
|
|
|
28:08.320 --> 28:12.000 |
|
and sort of somewhat naive, but brilliant, right? |
|
|
|
28:12.000 --> 28:14.000 |
|
So that's the best place to be |
|
|
|
28:14.000 --> 28:17.240 |
|
because you basically full steam take on a problem. |
|
|
|
28:17.240 --> 28:19.720 |
|
How confident, from that time, |
|
|
|
28:19.720 --> 28:21.320 |
|
because you know a lot more now, |
|
|
|
28:21.320 --> 28:23.440 |
|
at that time, how hard do you think it is |
|
|
|
28:23.440 --> 28:25.880 |
|
to solve all of autonomous driving? |
|
|
|
28:25.880 --> 28:30.440 |
|
I remember I suggested to Elon in the meeting |
|
|
|
28:30.440 --> 28:33.120 |
|
on putting a GPU behind each camera |
|
|
|
28:33.120 --> 28:35.120 |
|
to keep the compute local. |
|
|
|
28:35.120 --> 28:38.000 |
|
This is an incredibly stupid idea. |
|
|
|
28:38.000 --> 28:40.080 |
|
I leave the meeting 10 minutes later and I'm like, |
|
|
|
28:40.080 --> 28:41.560 |
|
I could have spent a little bit of time |
|
|
|
28:41.560 --> 28:42.880 |
|
thinking about this problem before I went in. |
|
|
|
28:42.880 --> 28:44.200 |
|
Why is this a stupid idea? |
|
|
|
28:44.200 --> 28:46.280 |
|
Oh, just send all your cameras to one big GPU. |
|
|
|
28:46.280 --> 28:48.240 |
|
You're much better off doing that. |
|
|
|
28:48.240 --> 28:50.160 |
|
Oh, sorry, you said behind every camera. |
|
|
|
28:50.160 --> 28:51.000 |
|
Every camera. |
|
|
|
28:51.000 --> 28:51.840 |
|
Every small GPU. |
|
|
|
28:51.840 --> 28:52.720 |
|
I was like, oh, I'll put the first few layers |
|
|
|
28:52.720 --> 28:54.520 |
|
of my comms there. |
|
|
|
28:54.520 --> 28:56.080 |
|
Like why did I say that? |
|
|
|
28:56.080 --> 28:56.920 |
|
That's possible. |
|
|
|
28:56.920 --> 28:59.000 |
|
It's possible, but it's a bad idea. |
|
|
|
28:59.000 --> 29:00.480 |
|
It's not obviously a bad idea. |
|
|
|
29:00.480 --> 29:01.320 |
|
Pretty obviously bad. |
|
|
|
29:01.320 --> 29:02.960 |
|
But whether it's actually a bad idea or not, |
|
|
|
29:02.960 --> 29:05.240 |
|
I left that meeting with Elon, like beating myself up. |
|
|
|
29:05.240 --> 29:07.080 |
|
I'm like, why did I say something stupid? |
|
|
|
29:07.080 --> 29:09.360 |
|
Yeah, you haven't, like you haven't at least |
|
|
|
29:09.360 --> 29:12.240 |
|
like thought through every aspect fully. |
|
|
|
29:12.240 --> 29:13.200 |
|
He's very sharp too. |
|
|
|
29:13.200 --> 29:15.760 |
|
Like usually in life, I get away with saying stupid things |
|
|
|
29:15.760 --> 29:16.960 |
|
and then kind of course, |
|
|
|
29:16.960 --> 29:18.560 |
|
right away he called me out about it. |
|
|
|
29:18.560 --> 29:19.800 |
|
And like, usually in life, |
|
|
|
29:19.800 --> 29:21.120 |
|
I get away with saying stupid things. |
|
|
|
29:21.120 --> 29:24.640 |
|
And then like people will, you know, |
|
|
|
29:24.640 --> 29:26.080 |
|
a lot of times people don't even notice. |
|
|
|
29:26.080 --> 29:28.200 |
|
And I'll like correct it and bring the conversation back. |
|
|
|
29:28.200 --> 29:30.600 |
|
But with Elon, it was like, nope, like, okay. |
|
|
|
29:30.600 --> 29:33.520 |
|
Well, that's not at all why the contract fell through. |
|
|
|
29:33.520 --> 29:35.520 |
|
I was much more prepared the second time I met him. |
|
|
|
29:35.520 --> 29:36.360 |
|
Yeah. |
|
|
|
29:36.360 --> 29:39.640 |
|
But in general, how hard did you think it, |
|
|
|
29:39.640 --> 29:43.680 |
|
like 12 months is a tough timeline? |
|
|
|
29:43.680 --> 29:45.720 |
|
Oh, I just thought I'd clone Mobileye IQ three. |
|
|
|
29:45.720 --> 29:47.560 |
|
I didn't think I'd solve level five self driving |
|
|
|
29:47.560 --> 29:48.400 |
|
or anything. |
|
|
|
29:48.400 --> 29:51.000 |
|
So the goal there was to do lane keeping, |
|
|
|
29:51.000 --> 29:52.840 |
|
good lane keeping. |
|
|
|
29:52.840 --> 29:55.560 |
|
I saw my friend showed me the outputs from Mobileye. |
|
|
|
29:55.560 --> 29:57.680 |
|
And the outputs from Mobileye was just basically two lanes |
|
|
|
29:57.680 --> 29:59.440 |
|
and a position of a lead car. |
|
|
|
29:59.440 --> 30:01.560 |
|
I'm like, I can gather a data set |
|
|
|
30:01.560 --> 30:03.440 |
|
and train this net in weeks. |
|
|
|
30:03.440 --> 30:04.840 |
|
And I did. |
|
|
|
30:04.840 --> 30:07.600 |
|
Well, first time I tried the implementation of Mobileye |
|
|
|
30:07.600 --> 30:11.240 |
|
in a Tesla, I was really surprised how good it is. |
|
|
|
30:11.240 --> 30:12.320 |
|
It's quite incredibly good. |
|
|
|
30:12.320 --> 30:14.080 |
|
Cause I thought it's just cause I've done |
|
|
|
30:14.080 --> 30:14.920 |
|
a lot of computer vision. |
|
|
|
30:14.920 --> 30:18.880 |
|
I thought it'd be a lot harder to create a system |
|
|
|
30:18.880 --> 30:20.040 |
|
that that's stable. |
|
|
|
30:21.000 --> 30:22.440 |
|
So I was personally surprised. |
|
|
|
30:22.440 --> 30:25.000 |
|
Just, you know, have to admit it. |
|
|
|
30:25.000 --> 30:27.840 |
|
Cause I was kind of skeptical before trying it. |
|
|
|
30:27.840 --> 30:31.200 |
|
Cause I thought it would go in and out a lot more. |
|
|
|
30:31.200 --> 30:33.160 |
|
It would get disengaged a lot more. |
|
|
|
30:33.160 --> 30:35.000 |
|
And it's pretty robust. |
|
|
|
30:36.200 --> 30:39.720 |
|
So what, how, how, how hard is the problem |
|
|
|
30:39.720 --> 30:42.080 |
|
when you, when you tackled it? |
|
|
|
30:42.080 --> 30:45.760 |
|
So I think AP one was great. Like Elon talked |
|
|
|
30:45.760 --> 30:49.040 |
|
about disengagements on the 405 down in LA |
|
|
|
30:49.040 --> 30:51.040 |
|
with like the lane marks were kind of faded |
|
|
|
30:51.040 --> 30:52.960 |
|
and the Mobileye system would drop out. |
|
|
|
30:53.960 --> 30:57.240 |
|
Like I had something up and working |
|
|
|
30:57.240 --> 31:01.440 |
|
that I would say was like the same quality in three months. |
|
|
|
31:02.480 --> 31:04.560 |
|
Same quality, but how do you know? |
|
|
|
31:04.560 --> 31:07.400 |
|
You say stuff like that confidently, but you can't, |
|
|
|
31:07.400 --> 31:12.120 |
|
and I love it, but the question is you can't, |
|
|
|
31:12.120 --> 31:13.880 |
|
you're kind of going by feel cause you just, |
|
|
|
31:13.880 --> 31:15.560 |
|
You're going by feel, absolutely, absolutely. |
|
|
|
31:15.560 --> 31:17.280 |
|
Like, like I would take, I hadn't, |
|
|
|
31:17.280 --> 31:18.480 |
|
I borrowed my friend's Tesla. |
|
|
|
31:18.480 --> 31:20.760 |
|
I would take AP one out for a drive. |
|
|
|
31:20.760 --> 31:22.320 |
|
And then I would take my system out for a drive. |
|
|
|
31:22.320 --> 31:24.440 |
|
And seems reasonably like the same. |
|
|
|
31:26.080 --> 31:30.480 |
|
So the 405, how hard is it to create something |
|
|
|
31:30.480 --> 31:34.200 |
|
that could actually be a product that's deployed? |
|
|
|
31:34.200 --> 31:39.200 |
|
I mean, I've read an article where Elon, this respond, |
|
|
|
31:39.520 --> 31:41.880 |
|
it said something about you saying that |
|
|
|
31:41.880 --> 31:46.880 |
|
to build autopilot is more complicated |
|
|
|
31:47.080 --> 31:51.880 |
|
than a single George Hodds level job. |
|
|
|
31:51.880 --> 31:55.520 |
|
How hard is that job to create something |
|
|
|
31:55.520 --> 31:57.480 |
|
that would work across the globally? |
|
|
|
31:58.960 --> 32:00.640 |
|
Why don't the global is the challenge, |
|
|
|
32:00.640 --> 32:02.240 |
|
but Elon followed that up by saying |
|
|
|
32:02.240 --> 32:04.920 |
|
it's going to take two years and a company of 10 people. |
|
|
|
32:04.920 --> 32:07.920 |
|
And here I am four years later with a company of 12 people. |
|
|
|
32:07.920 --> 32:09.960 |
|
And I think we still have another two to go. |
|
|
|
32:09.960 --> 32:10.800 |
|
Two years. |
|
|
|
32:10.800 --> 32:13.120 |
|
So yeah, so what do you think, |
|
|
|
32:13.120 --> 32:15.960 |
|
what do you think about how Tesla's progressing |
|
|
|
32:15.960 --> 32:19.200 |
|
with autopilot of V2, V3? |
|
|
|
32:19.200 --> 32:23.120 |
|
I think we've kept pace with them pretty well. |
|
|
|
32:24.080 --> 32:26.880 |
|
I think navigating autopilot is terrible. |
|
|
|
32:26.880 --> 32:31.120 |
|
We had some demo features internally of the same stuff |
|
|
|
32:31.120 --> 32:32.720 |
|
and we would test it and I'm like, |
|
|
|
32:32.720 --> 32:34.720 |
|
I'm not shipping this even as like open source software |
|
|
|
32:34.720 --> 32:35.560 |
|
to people. |
|
|
|
32:35.560 --> 32:37.400 |
|
What do you think is terrible? |
|
|
|
32:37.400 --> 32:39.600 |
|
Consumer Reports does a great job of describing it. |
|
|
|
32:39.600 --> 32:41.240 |
|
Like when it makes a lane change, |
|
|
|
32:41.240 --> 32:43.600 |
|
it does it worse than a human. |
|
|
|
32:43.600 --> 32:46.960 |
|
You shouldn't ship things like autopilot, open pilot, |
|
|
|
32:46.960 --> 32:49.760 |
|
they lane keep better than a human. |
|
|
|
32:49.760 --> 32:53.440 |
|
If you turn it on for a stretch of highway, |
|
|
|
32:53.440 --> 32:56.680 |
|
like an hour long, it's never going to touch a lane line. |
|
|
|
32:56.680 --> 32:59.040 |
|
Human will touch probably a lane line twice. |
|
|
|
32:59.040 --> 33:00.080 |
|
You just inspired me. |
|
|
|
33:00.080 --> 33:02.200 |
|
I don't know if you're grounded in data on that. |
|
|
|
33:02.200 --> 33:03.280 |
|
I read your paper. |
|
|
|
33:03.280 --> 33:05.400 |
|
Okay, but no, but that's interesting. |
|
|
|
33:06.720 --> 33:09.840 |
|
I wonder actually how often we touch lane lines |
|
|
|
33:11.200 --> 33:13.400 |
|
a little bit because it is. |
|
|
|
33:13.400 --> 33:14.960 |
|
I could answer that question pretty easily |
|
|
|
33:14.960 --> 33:15.800 |
|
with the common data side. |
|
|
|
33:15.800 --> 33:16.920 |
|
Yeah, I'm curious. |
|
|
|
33:16.920 --> 33:17.760 |
|
I've never answered it. |
|
|
|
33:17.760 --> 33:18.600 |
|
I don't know. |
|
|
|
33:18.600 --> 33:20.000 |
|
I just too was like my personal. |
|
|
|
33:20.000 --> 33:22.400 |
|
It feels right, but that's interesting |
|
|
|
33:22.400 --> 33:23.800 |
|
because every time you touch a lane, |
|
|
|
33:23.800 --> 33:26.760 |
|
that's a source of a little bit of stress |
|
|
|
33:26.760 --> 33:29.320 |
|
and kind of lane keeping is removing that stress. |
|
|
|
33:29.320 --> 33:31.840 |
|
That's ultimately the biggest value add |
|
|
|
33:31.840 --> 33:34.240 |
|
honestly is just removing the stress |
|
|
|
33:34.240 --> 33:35.480 |
|
of having to stay in lane. |
|
|
|
33:35.480 --> 33:39.040 |
|
And I think I don't think people fully realize |
|
|
|
33:39.040 --> 33:41.960 |
|
first of all that that's a big value add, |
|
|
|
33:41.960 --> 33:45.000 |
|
but also that that's all it is. |
|
|
|
33:45.000 --> 33:48.560 |
|
And that not only I find it a huge value add. |
|
|
|
33:48.560 --> 33:50.440 |
|
I drove down when we moved to San Diego, |
|
|
|
33:50.440 --> 33:52.640 |
|
I drove down in an enterprise rental car |
|
|
|
33:52.640 --> 33:53.480 |
|
and I missed it. |
|
|
|
33:53.480 --> 33:55.480 |
|
So I missed having the system so much. |
|
|
|
33:55.480 --> 33:59.200 |
|
It's so much more tiring to drive |
|
|
|
33:59.200 --> 34:00.320 |
|
without it. |
|
|
|
34:00.320 --> 34:02.960 |
|
It's, it is that lane centering. |
|
|
|
34:02.960 --> 34:04.840 |
|
That's the key feature. |
|
|
|
34:04.840 --> 34:05.680 |
|
Yeah. |
|
|
|
34:06.600 --> 34:08.960 |
|
And in a way it's the only feature |
|
|
|
34:08.960 --> 34:11.040 |
|
that actually adds value to people's lives |
|
|
|
34:11.040 --> 34:12.200 |
|
in autonomous vehicles today. |
|
|
|
34:12.200 --> 34:13.840 |
|
Waymo does not add value to people's lives. |
|
|
|
34:13.840 --> 34:15.880 |
|
It's a more expensive, slower Uber. |
|
|
|
34:15.880 --> 34:18.640 |
|
Maybe someday it'll be this big cliff where it adds value, |
|
|
|
34:18.640 --> 34:19.480 |
|
but I don't usually. |
|
|
|
34:19.480 --> 34:20.320 |
|
It's fascinating. |
|
|
|
34:20.320 --> 34:22.560 |
|
I haven't talked to, this is good. |
|
|
|
34:22.560 --> 34:25.840 |
|
Cause I haven't, I have intuitively, |
|
|
|
34:25.840 --> 34:28.320 |
|
but I think we're making it explicit now. |
|
|
|
34:28.320 --> 34:33.320 |
|
I actually believe that really good lane keeping |
|
|
|
34:35.480 --> 34:37.240 |
|
is a reason to buy a car. |
|
|
|
34:37.240 --> 34:38.440 |
|
Will be a reason to buy a car. |
|
|
|
34:38.440 --> 34:39.720 |
|
It is a huge value add. |
|
|
|
34:39.720 --> 34:41.760 |
|
I've never, until we just started talking about it, |
|
|
|
34:41.760 --> 34:43.880 |
|
haven't really quite realized it, |
|
|
|
34:43.880 --> 34:48.880 |
|
that I've felt with Elon's chase of level four |
|
|
|
34:49.440 --> 34:52.360 |
|
is not the correct chase. |
|
|
|
34:52.360 --> 34:56.000 |
|
It was on, cause you should just say Tesla has the best |
|
|
|
34:56.000 --> 34:58.320 |
|
as if from a Tesla perspective say, |
|
|
|
34:58.320 --> 35:00.600 |
|
Tesla has the best lane keeping. |
|
|
|
35:00.600 --> 35:04.160 |
|
Kama AI should say Kama AI is the best lane keeping. |
|
|
|
35:04.160 --> 35:05.640 |
|
And that is it. |
|
|
|
35:05.640 --> 35:06.480 |
|
Yeah. |
|
|
|
35:06.480 --> 35:07.320 |
|
Yeah. |
|
|
|
35:07.320 --> 35:08.160 |
|
Do you think? |
|
|
|
35:08.160 --> 35:09.920 |
|
You have to do the longitudinal as well. |
|
|
|
35:09.920 --> 35:10.960 |
|
You can't just lane keep. |
|
|
|
35:10.960 --> 35:12.920 |
|
You have to do ACC, |
|
|
|
35:12.920 --> 35:15.840 |
|
but ACC is much more forgiving than lane keep, |
|
|
|
35:15.840 --> 35:17.400 |
|
especially on the highway. |
|
|
|
35:17.400 --> 35:22.000 |
|
By the way, are you Kama AI's camera only, correct? |
|
|
|
35:22.000 --> 35:23.440 |
|
No, we use the radar. |
|
|
|
35:23.440 --> 35:26.960 |
|
We, from the car, you're able to get to, okay. |
|
|
|
35:26.960 --> 35:28.800 |
|
We can do it camera only now. |
|
|
|
35:28.800 --> 35:29.640 |
|
It's gotten to the point, |
|
|
|
35:29.640 --> 35:31.600 |
|
but we leave the radar there as like a, |
|
|
|
35:31.600 --> 35:33.440 |
|
it's fusion now. |
|
|
|
35:33.440 --> 35:35.440 |
|
Okay, so let's maybe talk through |
|
|
|
35:35.440 --> 35:37.920 |
|
some of the system specs on the hardware. |
|
|
|
35:37.920 --> 35:42.880 |
|
What's the hardware side of what you're providing? |
|
|
|
35:42.880 --> 35:44.720 |
|
What's the capabilities on the software side |
|
|
|
35:44.720 --> 35:46.800 |
|
with OpenPilot and so on? |
|
|
|
35:46.800 --> 35:51.800 |
|
So OpenPilot as the box that we sell that it runs on, |
|
|
|
35:51.800 --> 35:53.920 |
|
it's a phone in a plastic case. |
|
|
|
35:53.920 --> 35:54.840 |
|
It's nothing special. |
|
|
|
35:54.840 --> 35:56.200 |
|
We sell it without the software. |
|
|
|
35:56.200 --> 35:57.840 |
|
So you're like, you know, you buy the phone, |
|
|
|
35:57.840 --> 35:58.920 |
|
it's just easy. |
|
|
|
35:58.920 --> 36:00.240 |
|
It'll be easy set up, |
|
|
|
36:00.240 --> 36:01.720 |
|
but it's sold with no software. |
|
|
|
36:03.480 --> 36:06.600 |
|
OpenPilot right now is about to be 0.6. |
|
|
|
36:06.600 --> 36:07.880 |
|
When it gets to 1.0, |
|
|
|
36:07.880 --> 36:09.680 |
|
I think we'll be ready for a consumer product. |
|
|
|
36:09.680 --> 36:11.120 |
|
We're not gonna add any new features. |
|
|
|
36:11.120 --> 36:13.800 |
|
We're just gonna make the lane keeping really, really good. |
|
|
|
36:13.800 --> 36:15.120 |
|
Okay, I got it. |
|
|
|
36:15.120 --> 36:16.120 |
|
So what do we have right now? |
|
|
|
36:16.120 --> 36:18.200 |
|
It's a Snapdragon 820. |
|
|
|
36:18.200 --> 36:23.200 |
|
It's a Sony IMX 298 forward facing camera, |
|
|
|
36:23.680 --> 36:24.720 |
|
driver monitoring camera. |
|
|
|
36:24.720 --> 36:26.400 |
|
It's just a selfie cam on the phone. |
|
|
|
36:26.400 --> 36:30.000 |
|
And a can transceiver, |
|
|
|
36:30.000 --> 36:32.320 |
|
maybe it's a little thing called pandas. |
|
|
|
36:32.320 --> 36:35.040 |
|
And they talk over USB to the phone |
|
|
|
36:35.040 --> 36:36.400 |
|
and then they have three can buses |
|
|
|
36:36.400 --> 36:37.560 |
|
that they talk to the car. |
|
|
|
36:38.560 --> 36:40.920 |
|
One of those can buses is the radar can bus. |
|
|
|
36:40.920 --> 36:42.920 |
|
One of them is the main car can bus. |
|
|
|
36:42.920 --> 36:44.920 |
|
And the other one is the proxy camera can bus. |
|
|
|
36:44.920 --> 36:47.320 |
|
We leave the existing camera in place. |
|
|
|
36:47.320 --> 36:49.560 |
|
So we don't turn AEB off. |
|
|
|
36:49.560 --> 36:51.040 |
|
Right now we still turn AEB off |
|
|
|
36:51.040 --> 36:52.280 |
|
if you're using our longitudinal, |
|
|
|
36:52.280 --> 36:54.320 |
|
but we're gonna fix that before 1.0. |
|
|
|
36:54.320 --> 36:55.160 |
|
Got it. |
|
|
|
36:55.160 --> 36:56.000 |
|
Wow, that's cool. |
|
|
|
36:56.000 --> 36:57.960 |
|
So in its can both ways. |
|
|
|
36:57.960 --> 37:02.120 |
|
So how are you able to control vehicles? |
|
|
|
37:02.120 --> 37:05.520 |
|
So we proxy the vehicles that we work with |
|
|
|
37:05.520 --> 37:08.960 |
|
already have a lane keeping assist system. |
|
|
|
37:08.960 --> 37:12.520 |
|
So lane keeping assist can mean a huge variety of things. |
|
|
|
37:12.520 --> 37:16.120 |
|
It can mean it will apply a small torque |
|
|
|
37:16.120 --> 37:18.920 |
|
to the wheel after you've already crossed a lane line |
|
|
|
37:18.920 --> 37:22.720 |
|
by a foot, which is the system in the older Toyotas. |
|
|
|
37:22.720 --> 37:26.360 |
|
Versus like, I think Tesla still calls it lane keeping assist |
|
|
|
37:26.360 --> 37:28.920 |
|
where it'll keep you perfectly in the center of the lane |
|
|
|
37:28.920 --> 37:29.960 |
|
on the highway. |
|
|
|
37:31.240 --> 37:34.000 |
|
You can control like you with the joystick, the cars. |
|
|
|
37:34.000 --> 37:36.600 |
|
So these cars already have the capability of drive by wire. |
|
|
|
37:36.600 --> 37:41.600 |
|
So is it, is it trivial to convert a car |
|
|
|
37:41.600 --> 37:43.320 |
|
that it operates with? |
|
|
|
37:43.320 --> 37:47.480 |
|
It open pilot is able to control the steering. |
|
|
|
37:48.480 --> 37:49.720 |
|
Oh, a new car or a car that we, |
|
|
|
37:49.720 --> 37:52.800 |
|
so we have support now for 45 different makes of cars. |
|
|
|
37:52.800 --> 37:54.880 |
|
What are the cars in general? |
|
|
|
37:54.880 --> 37:56.360 |
|
Mostly Honda's and Toyotas. |
|
|
|
37:56.360 --> 38:00.640 |
|
We support almost every Honda and Toyota made this year. |
|
|
|
38:01.680 --> 38:04.480 |
|
And then bunch of GM's, bunch of Subaru's. |
|
|
|
38:04.480 --> 38:05.960 |
|
But it doesn't have to be like a Prius. |
|
|
|
38:05.960 --> 38:07.320 |
|
It could be Corolla as well. |
|
|
|
38:07.320 --> 38:10.760 |
|
Oh, the 2020 Corolla is the best car with open pilot. |
|
|
|
38:10.760 --> 38:11.720 |
|
It just came out there. |
|
|
|
38:11.720 --> 38:14.200 |
|
The actuator has less lag than the older Corolla. |
|
|
|
38:15.840 --> 38:18.240 |
|
I think I started watching a video with you. |
|
|
|
38:18.240 --> 38:21.480 |
|
I mean, the way you make videos is awesome. |
|
|
|
38:21.480 --> 38:24.320 |
|
It's just literally at the dealership streaming. |
|
|
|
38:25.320 --> 38:26.160 |
|
I had my friend to follow him. |
|
|
|
38:26.160 --> 38:27.560 |
|
I probably want to stream for an hour. |
|
|
|
38:27.560 --> 38:31.120 |
|
Yeah, and basically like if stuff goes a little wrong, |
|
|
|
38:31.120 --> 38:33.160 |
|
you just like, you just go with it. |
|
|
|
38:33.160 --> 38:34.000 |
|
Yeah, I love it. |
|
|
|
38:34.000 --> 38:34.840 |
|
It's real. |
|
|
|
38:34.840 --> 38:35.680 |
|
Yeah, it's real. |
|
|
|
38:35.680 --> 38:42.000 |
|
That's so beautiful and it's so in contrast to the way |
|
|
|
38:42.000 --> 38:44.600 |
|
other companies would put together a video like that. |
|
|
|
38:44.600 --> 38:46.000 |
|
Kind of why I like to do it like that. |
|
|
|
38:46.000 --> 38:46.840 |
|
Good. |
|
|
|
38:46.840 --> 38:49.720 |
|
I mean, if you become super rich one day and successful, |
|
|
|
38:49.720 --> 38:52.280 |
|
I hope you keep it that way because I think that's actually |
|
|
|
38:52.280 --> 38:54.600 |
|
what people love, that kind of genuine. |
|
|
|
38:54.600 --> 38:56.520 |
|
Oh, it's all that has value to me. |
|
|
|
38:56.520 --> 38:59.840 |
|
Money has no, if I sell out to like make money, |
|
|
|
38:59.840 --> 39:00.680 |
|
I sold out. |
|
|
|
39:00.680 --> 39:01.520 |
|
It doesn't matter. |
|
|
|
39:01.520 --> 39:02.360 |
|
What do I get? |
|
|
|
39:02.360 --> 39:04.440 |
|
Yacht, I don't want a yacht. |
|
|
|
39:04.440 --> 39:09.440 |
|
And I think Tesla actually has a small inkling of that |
|
|
|
39:09.440 --> 39:11.240 |
|
as well with autonomy day. |
|
|
|
39:11.240 --> 39:14.000 |
|
They did reveal more than, I mean, of course, |
|
|
|
39:14.000 --> 39:15.680 |
|
there's marketing communications, you could tell, |
|
|
|
39:15.680 --> 39:17.640 |
|
but it's more than most companies would reveal, |
|
|
|
39:17.640 --> 39:20.960 |
|
which is I hope they go towards that direction |
|
|
|
39:20.960 --> 39:23.000 |
|
more other companies, GM, Ford. |
|
|
|
39:23.000 --> 39:25.400 |
|
Oh, Tesla's going to win level five. |
|
|
|
39:25.400 --> 39:26.560 |
|
They really are. |
|
|
|
39:26.560 --> 39:27.800 |
|
So let's talk about it. |
|
|
|
39:27.800 --> 39:33.000 |
|
You think, you're focused on level two currently, currently. |
|
|
|
39:33.000 --> 39:36.160 |
|
We're going to be one to two years behind Tesla |
|
|
|
39:36.160 --> 39:37.160 |
|
getting to level five. |
|
|
|
39:37.160 --> 39:38.520 |
|
OK. |
|
|
|
39:38.520 --> 39:39.320 |
|
We're Android, right? |
|
|
|
39:39.320 --> 39:39.880 |
|
We're Android. |
|
|
|
39:39.880 --> 39:40.680 |
|
You're Android. |
|
|
|
39:40.680 --> 39:42.240 |
|
I'm just saying once Tesla gets it, |
|
|
|
39:42.240 --> 39:43.440 |
|
we're one to two years behind. |
|
|
|
39:43.440 --> 39:45.680 |
|
I'm not making any timeline on when Tesla's going to get it. |
|
|
|
39:45.680 --> 39:46.120 |
|
That's right. |
|
|
|
39:46.120 --> 39:46.360 |
|
You did. |
|
|
|
39:46.360 --> 39:46.960 |
|
That's brilliant. |
|
|
|
39:46.960 --> 39:48.560 |
|
I'm sorry, Tesla investors, if you |
|
|
|
39:48.560 --> 39:50.520 |
|
think you're going to have an autonomous robot taxi |
|
|
|
39:50.520 --> 39:54.920 |
|
fleet by the end of the year, I'll bet against that. |
|
|
|
39:54.920 --> 39:57.720 |
|
So what do you think about this? |
|
|
|
39:57.720 --> 40:03.280 |
|
The most level four companies are kind of just |
|
|
|
40:03.280 --> 40:08.360 |
|
doing their usual safety driver, doing full autonomy kind |
|
|
|
40:08.360 --> 40:08.800 |
|
of testing. |
|
|
|
40:08.800 --> 40:10.880 |
|
And then Tesla does basically trying |
|
|
|
40:10.880 --> 40:15.280 |
|
to go from lane keeping to full autonomy. |
|
|
|
40:15.280 --> 40:16.840 |
|
What do you think about that approach? |
|
|
|
40:16.840 --> 40:18.360 |
|
How successful would it be? |
|
|
|
40:18.360 --> 40:20.680 |
|
It's a ton better approach. |
|
|
|
40:20.680 --> 40:23.960 |
|
Because Tesla is gathering data on a scale |
|
|
|
40:23.960 --> 40:25.200 |
|
that none of them are. |
|
|
|
40:25.200 --> 40:29.560 |
|
They're putting real users behind the wheel of the cars. |
|
|
|
40:29.560 --> 40:34.440 |
|
It's, I think, the only strategy that works, the incremental. |
|
|
|
40:34.440 --> 40:37.000 |
|
Well, so there's a few components to Tesla approach |
|
|
|
40:37.000 --> 40:38.800 |
|
that's more than just the incremental. |
|
|
|
40:38.800 --> 40:41.400 |
|
What you spoke with is the software, |
|
|
|
40:41.400 --> 40:43.720 |
|
so over the air software updates. |
|
|
|
40:43.720 --> 40:44.800 |
|
Necessity. |
|
|
|
40:44.800 --> 40:46.440 |
|
I mean, Waymo crews have those too. |
|
|
|
40:46.440 --> 40:47.560 |
|
Those aren't. |
|
|
|
40:47.560 --> 40:48.080 |
|
But no. |
|
|
|
40:48.080 --> 40:49.800 |
|
Those differentiate from the automakers. |
|
|
|
40:49.800 --> 40:50.080 |
|
Right. |
|
|
|
40:50.080 --> 40:53.440 |
|
No lane keeping systems have no cars with lane keeping system |
|
|
|
40:53.440 --> 40:54.760 |
|
have that except Tesla. |
|
|
|
40:54.760 --> 40:55.720 |
|
Yeah. |
|
|
|
40:55.720 --> 40:59.760 |
|
And the other one is the data, the other direction, |
|
|
|
40:59.760 --> 41:01.840 |
|
which is the ability to query the data. |
|
|
|
41:01.840 --> 41:03.480 |
|
I don't think they're actually collecting |
|
|
|
41:03.480 --> 41:05.240 |
|
as much data as people think, but the ability |
|
|
|
41:05.240 --> 41:09.440 |
|
to turn on collection and turn it off. |
|
|
|
41:09.440 --> 41:13.400 |
|
So I'm both in the robotics world, in the psychology, |
|
|
|
41:13.400 --> 41:15.000 |
|
human factors world. |
|
|
|
41:15.000 --> 41:17.320 |
|
Many people believe that level two autonomy |
|
|
|
41:17.320 --> 41:20.040 |
|
is problematic because of the human factor. |
|
|
|
41:20.040 --> 41:23.280 |
|
Like the more the task is automated, |
|
|
|
41:23.280 --> 41:25.960 |
|
the more there's a vigilance decrement. |
|
|
|
41:25.960 --> 41:27.200 |
|
You start to fall asleep. |
|
|
|
41:27.200 --> 41:30.480 |
|
You start to become complacent, start texting more and so on. |
|
|
|
41:30.480 --> 41:32.200 |
|
Do you worry about that? |
|
|
|
41:32.200 --> 41:35.000 |
|
Because if you're talking about transition from lane keeping |
|
|
|
41:35.000 --> 41:40.960 |
|
to full autonomy, if you're spending 80% of the time |
|
|
|
41:40.960 --> 41:43.080 |
|
not supervising the machine, do you |
|
|
|
41:43.080 --> 41:47.080 |
|
worry about what that means for the safety of the drivers? |
|
|
|
41:47.080 --> 41:49.640 |
|
One, we don't consider OpenPilot to be 1.0 |
|
|
|
41:49.640 --> 41:52.880 |
|
until we have 100% driver monitoring. |
|
|
|
41:52.880 --> 41:55.000 |
|
You can cheat right now, our driver monitoring system. |
|
|
|
41:55.000 --> 41:56.080 |
|
There's a few ways to cheat it. |
|
|
|
41:56.080 --> 41:58.160 |
|
They're pretty obvious. |
|
|
|
41:58.160 --> 41:59.680 |
|
We're working on making that better. |
|
|
|
41:59.680 --> 42:02.520 |
|
Before we ship a consumer product that can drive cars, |
|
|
|
42:02.520 --> 42:04.240 |
|
I want to make sure that I have driver monitoring |
|
|
|
42:04.240 --> 42:05.440 |
|
that you can't cheat. |
|
|
|
42:05.440 --> 42:09.000 |
|
What's a successful driver monitoring system look like? |
|
|
|
42:09.000 --> 42:11.680 |
|
Is it all about just keeping your eyes on the road? |
|
|
|
42:11.680 --> 42:12.760 |
|
Well, a few things. |
|
|
|
42:12.760 --> 42:16.600 |
|
So that's what we went with at first for driver monitoring. |
|
|
|
42:16.600 --> 42:17.160 |
|
I'm checking. |
|
|
|
42:17.160 --> 42:19.000 |
|
I'm actually looking at where your head is looking. |
|
|
|
42:19.000 --> 42:19.880 |
|
The camera's not that high. |
|
|
|
42:19.880 --> 42:21.840 |
|
Resolution eyes are a little bit hard to get. |
|
|
|
42:21.840 --> 42:22.880 |
|
Well, head is big. |
|
|
|
42:22.880 --> 42:23.560 |
|
I mean, that's just. |
|
|
|
42:23.560 --> 42:24.640 |
|
Head is good. |
|
|
|
42:24.640 --> 42:28.720 |
|
And actually, a lot of it, just psychology wise, |
|
|
|
42:28.720 --> 42:30.720 |
|
to have that monitor constantly there, |
|
|
|
42:30.720 --> 42:33.400 |
|
it reminds you that you have to be paying attention. |
|
|
|
42:33.400 --> 42:35.080 |
|
But we want to go further. |
|
|
|
42:35.080 --> 42:36.760 |
|
We just hired someone full time to come on |
|
|
|
42:36.760 --> 42:37.960 |
|
to do the driver monitoring. |
|
|
|
42:37.960 --> 42:40.600 |
|
I want to detect phone in frame, and I |
|
|
|
42:40.600 --> 42:42.600 |
|
want to make sure you're not sleeping. |
|
|
|
42:42.600 --> 42:44.880 |
|
How much does the camera see of the body? |
|
|
|
42:44.880 --> 42:47.480 |
|
This one, not enough. |
|
|
|
42:47.480 --> 42:48.400 |
|
Not enough. |
|
|
|
42:48.400 --> 42:50.720 |
|
The next one, everything. |
|
|
|
42:50.720 --> 42:52.920 |
|
What's interesting, FishEye, is we're |
|
|
|
42:52.920 --> 42:55.200 |
|
doing just data collection, not real time. |
|
|
|
42:55.200 --> 42:59.200 |
|
But FishEye is a beautiful being able to capture the body. |
|
|
|
42:59.200 --> 43:03.280 |
|
And the smartphone is really the biggest problem. |
|
|
|
43:03.280 --> 43:03.880 |
|
I'll show you. |
|
|
|
43:03.880 --> 43:07.800 |
|
I can show you one of the pictures from our new system. |
|
|
|
43:07.800 --> 43:08.160 |
|
Awesome. |
|
|
|
43:08.160 --> 43:10.520 |
|
So you're basically saying the driver monitoring |
|
|
|
43:10.520 --> 43:13.080 |
|
will be the answer to that. |
|
|
|
43:13.080 --> 43:15.320 |
|
I think the other point that you raised in your paper |
|
|
|
43:15.320 --> 43:16.920 |
|
is good as well. |
|
|
|
43:16.920 --> 43:20.400 |
|
You're not asking a human to supervise a machine |
|
|
|
43:20.400 --> 43:23.920 |
|
without giving them the they can take over at any time. |
|
|
|
43:23.920 --> 43:25.760 |
|
Our safety model, you can take over. |
|
|
|
43:25.760 --> 43:27.720 |
|
We disengage on both the gas or the brake. |
|
|
|
43:27.720 --> 43:28.880 |
|
We don't disengage on steering. |
|
|
|
43:28.880 --> 43:29.920 |
|
I don't feel you have to. |
|
|
|
43:29.920 --> 43:31.720 |
|
But we disengage on gas or brake. |
|
|
|
43:31.720 --> 43:34.240 |
|
So it's very easy for you to take over. |
|
|
|
43:34.240 --> 43:36.400 |
|
And it's very easy for you to reengage. |
|
|
|
43:36.400 --> 43:39.320 |
|
That switching should be super cheap. |
|
|
|
43:39.320 --> 43:40.800 |
|
The cars that require, even autopilot, |
|
|
|
43:40.800 --> 43:42.400 |
|
requires a double press. |
|
|
|
43:42.400 --> 43:44.360 |
|
That's almost, I see, I don't like that. |
|
|
|
43:44.360 --> 43:46.440 |
|
And then the cancel. |
|
|
|
43:46.440 --> 43:48.320 |
|
To cancel in autopilot, you either |
|
|
|
43:48.320 --> 43:49.920 |
|
have to press cancel, which no one knows where that is. |
|
|
|
43:49.920 --> 43:51.000 |
|
So they press the brake. |
|
|
|
43:51.000 --> 43:53.360 |
|
But a lot of times you don't want to press the brake. |
|
|
|
43:53.360 --> 43:54.560 |
|
You want to press the gas. |
|
|
|
43:54.560 --> 43:56.880 |
|
So you should cancel on gas or wiggle the steering wheel, |
|
|
|
43:56.880 --> 43:57.960 |
|
which is bad as well. |
|
|
|
43:57.960 --> 43:58.920 |
|
Wow, that's brilliant. |
|
|
|
43:58.920 --> 44:01.440 |
|
I haven't heard anyone articulate that point. |
|
|
|
44:01.440 --> 44:04.960 |
|
Oh, there's a lot I think about. |
|
|
|
44:04.960 --> 44:09.800 |
|
Because I think actually Tesla has done a better job |
|
|
|
44:09.800 --> 44:12.920 |
|
than most automakers at making that frictionless. |
|
|
|
44:12.920 --> 44:16.600 |
|
But you just described that it could be even better. |
|
|
|
44:16.600 --> 44:19.320 |
|
I love Super Cruise as an experience. |
|
|
|
44:19.320 --> 44:21.120 |
|
Once it's engaged. |
|
|
|
44:21.120 --> 44:22.800 |
|
I don't know if you've used it, but getting the thing |
|
|
|
44:22.800 --> 44:25.040 |
|
to try to engage. |
|
|
|
44:25.040 --> 44:27.480 |
|
Yeah, I've used the driven Super Cruise a lot. |
|
|
|
44:27.480 --> 44:29.680 |
|
So what's your thoughts on the Super Cruise system in general? |
|
|
|
44:29.680 --> 44:32.640 |
|
You disengage Super Cruise, and it falls back to ACC. |
|
|
|
44:32.640 --> 44:34.600 |
|
So my car is still accelerating. |
|
|
|
44:34.600 --> 44:36.280 |
|
It feels weird. |
|
|
|
44:36.280 --> 44:39.000 |
|
Otherwise, when you actually have Super Cruise engaged |
|
|
|
44:39.000 --> 44:41.200 |
|
on the highway, it is phenomenal. |
|
|
|
44:41.200 --> 44:42.320 |
|
We bought that Cadillac. |
|
|
|
44:42.320 --> 44:43.240 |
|
We just sold it. |
|
|
|
44:43.240 --> 44:45.600 |
|
But we bought it just to experience this. |
|
|
|
44:45.600 --> 44:47.440 |
|
And I wanted everyone in the office to be like, |
|
|
|
44:47.440 --> 44:49.360 |
|
this is what we're striving to build. |
|
|
|
44:49.360 --> 44:52.800 |
|
GM pioneering with the driver monitoring. |
|
|
|
44:52.800 --> 44:55.040 |
|
You like their driver monitoring system? |
|
|
|
44:55.040 --> 44:56.440 |
|
It has some bugs. |
|
|
|
44:56.440 --> 45:01.960 |
|
If there's a sun shining back here, it'll be blind to you. |
|
|
|
45:01.960 --> 45:03.360 |
|
But overall, mostly, yeah. |
|
|
|
45:03.360 --> 45:05.960 |
|
That's so cool that you know all this stuff. |
|
|
|
45:05.960 --> 45:09.960 |
|
I don't often talk to people that because it's such a rare car, |
|
|
|
45:09.960 --> 45:10.960 |
|
unfortunately, currently. |
|
|
|
45:10.960 --> 45:12.760 |
|
We bought one explicitly for that. |
|
|
|
45:12.760 --> 45:15.040 |
|
We lost like $25K in the deprecation, |
|
|
|
45:15.040 --> 45:16.720 |
|
but it feels worth it. |
|
|
|
45:16.720 --> 45:21.280 |
|
I was very pleasantly surprised that our GM system |
|
|
|
45:21.280 --> 45:26.320 |
|
was so innovative and really wasn't advertised much, |
|
|
|
45:26.320 --> 45:28.480 |
|
wasn't talked about much. |
|
|
|
45:28.480 --> 45:31.840 |
|
And I was nervous that it would die, that it would disappear. |
|
|
|
45:31.840 --> 45:33.520 |
|
Well, they put it on the wrong car. |
|
|
|
45:33.520 --> 45:35.680 |
|
They should have put it on the bolt and not some weird Cadillac |
|
|
|
45:35.680 --> 45:36.640 |
|
that nobody bought. |
|
|
|
45:36.640 --> 45:39.520 |
|
I think that's going to be into, they're saying at least |
|
|
|
45:39.520 --> 45:41.840 |
|
it's going to be into their entire fleet. |
|
|
|
45:41.840 --> 45:44.320 |
|
So what do you think about, as long as we're |
|
|
|
45:44.320 --> 45:46.920 |
|
on the driver monitoring, what do you think |
|
|
|
45:46.920 --> 45:51.920 |
|
about Elon Musk's claim that driver monitoring is not needed? |
|
|
|
45:51.920 --> 45:53.680 |
|
Normally, I love his claims. |
|
|
|
45:53.680 --> 45:55.560 |
|
That one is stupid. |
|
|
|
45:55.560 --> 45:56.560 |
|
That one is stupid. |
|
|
|
45:56.560 --> 46:00.320 |
|
And he's not going to have his level five fleet |
|
|
|
46:00.320 --> 46:01.320 |
|
by the end of the year. |
|
|
|
46:01.320 --> 46:04.880 |
|
Hopefully, he's like, OK, I was wrong. |
|
|
|
46:04.880 --> 46:06.280 |
|
I'm going to add driver monitoring. |
|
|
|
46:06.280 --> 46:08.240 |
|
Because when these systems get to the point |
|
|
|
46:08.240 --> 46:10.320 |
|
that they're only messing up once every 1,000 miles, |
|
|
|
46:10.320 --> 46:14.080 |
|
you absolutely need driver monitoring. |
|
|
|
46:14.080 --> 46:15.880 |
|
So let me play, because I agree with you, |
|
|
|
46:15.880 --> 46:17.320 |
|
but let me play devil's advocate. |
|
|
|
46:17.320 --> 46:22.440 |
|
One possibility is that without driver monitoring, |
|
|
|
46:22.440 --> 46:29.400 |
|
people are able to self regulate, monitor themselves. |
|
|
|
46:29.400 --> 46:30.680 |
|
So your idea is, I'm just. |
|
|
|
46:30.680 --> 46:34.160 |
|
You're seeing all the people sleeping in Teslas? |
|
|
|
46:34.160 --> 46:35.280 |
|
Yeah. |
|
|
|
46:35.280 --> 46:38.320 |
|
Well, I'm a little skeptical of all the people sleeping |
|
|
|
46:38.320 --> 46:43.960 |
|
in Teslas because I've stopped paying attention to that kind |
|
|
|
46:43.960 --> 46:45.680 |
|
of stuff because I want to see real data. |
|
|
|
46:45.680 --> 46:47.240 |
|
It's too much glorified. |
|
|
|
46:47.240 --> 46:48.720 |
|
It doesn't feel scientific to me. |
|
|
|
46:48.720 --> 46:52.560 |
|
So I want to know how many people are really sleeping |
|
|
|
46:52.560 --> 46:55.080 |
|
in Teslas versus sleeping. |
|
|
|
46:55.080 --> 46:57.640 |
|
I was driving here, sleep deprived, |
|
|
|
46:57.640 --> 46:59.520 |
|
in a car with no automation. |
|
|
|
46:59.520 --> 47:01.040 |
|
I was falling asleep. |
|
|
|
47:01.040 --> 47:02.120 |
|
I agree that it's hypey. |
|
|
|
47:02.120 --> 47:04.840 |
|
It's just like, you know what? |
|
|
|
47:04.840 --> 47:08.480 |
|
If Elon put driver monitoring, my last autopilot experience |
|
|
|
47:08.480 --> 47:12.200 |
|
was I rented a Model 3 in March and drove it around. |
|
|
|
47:12.200 --> 47:13.640 |
|
The wheel thing is annoying. |
|
|
|
47:13.640 --> 47:15.440 |
|
And the reason the wheel thing is annoying. |
|
|
|
47:15.440 --> 47:17.080 |
|
We use the wheel thing as well, but we |
|
|
|
47:17.080 --> 47:18.720 |
|
don't disengage on wheel. |
|
|
|
47:18.720 --> 47:21.720 |
|
For Tesla, you have to touch the wheel just enough |
|
|
|
47:21.720 --> 47:25.320 |
|
to trigger the torque sensor to tell it that you're there, |
|
|
|
47:25.320 --> 47:29.720 |
|
but not enough as to disengage it, which don't use it |
|
|
|
47:29.720 --> 47:30.440 |
|
for two things. |
|
|
|
47:30.440 --> 47:31.360 |
|
Don't disengage on wheel. |
|
|
|
47:31.360 --> 47:32.400 |
|
You don't have to. |
|
|
|
47:32.400 --> 47:35.360 |
|
That whole experience, wow, beautifully put. |
|
|
|
47:35.360 --> 47:38.360 |
|
All those elements, even if you don't have driver monitoring, |
|
|
|
47:38.360 --> 47:41.080 |
|
that whole experience needs to be better. |
|
|
|
47:41.080 --> 47:43.760 |
|
Driver monitoring, I think would make, |
|
|
|
47:43.760 --> 47:46.200 |
|
I mean, I think supercruise is a better experience |
|
|
|
47:46.200 --> 47:48.440 |
|
once it's engaged over autopilot. |
|
|
|
47:48.440 --> 47:51.600 |
|
I think supercruise is a transition to engagement |
|
|
|
47:51.600 --> 47:55.200 |
|
and disengagement are significantly worse. |
|
|
|
47:55.200 --> 47:57.880 |
|
There's a tricky thing, because if I were to criticize |
|
|
|
47:57.880 --> 48:00.800 |
|
supercruise, it's a little too crude. |
|
|
|
48:00.800 --> 48:03.640 |
|
And I think it's like six seconds or something. |
|
|
|
48:03.640 --> 48:06.080 |
|
If you look off road, it'll start warning you. |
|
|
|
48:06.080 --> 48:09.120 |
|
It's some ridiculously long period of time. |
|
|
|
48:09.120 --> 48:14.120 |
|
And just the way, I think it's basically, it's a binary. |
|
|
|
48:15.840 --> 48:17.440 |
|
It should be adapted. |
|
|
|
48:17.440 --> 48:19.880 |
|
Yeah, it needs to learn more about you. |
|
|
|
48:19.880 --> 48:23.160 |
|
It needs to communicate what it sees about you more. |
|
|
|
48:23.160 --> 48:25.800 |
|
I'm not, you know, Tesla shows what it sees |
|
|
|
48:25.800 --> 48:27.160 |
|
about the external world. |
|
|
|
48:27.160 --> 48:29.120 |
|
It would be nice if supercruise would tell us |
|
|
|
48:29.120 --> 48:30.840 |
|
what it sees about the internal world. |
|
|
|
48:30.840 --> 48:31.960 |
|
It's even worse than that. |
|
|
|
48:31.960 --> 48:33.320 |
|
You press the button to engage |
|
|
|
48:33.320 --> 48:35.480 |
|
and it just says supercruise unavailable. |
|
|
|
48:35.480 --> 48:36.320 |
|
Yeah, why? |
|
|
|
48:36.320 --> 48:37.800 |
|
Why? |
|
|
|
48:37.800 --> 48:41.480 |
|
Yeah, that transparency is good. |
|
|
|
48:41.480 --> 48:43.520 |
|
We've renamed the driver monitoring packet |
|
|
|
48:43.520 --> 48:45.360 |
|
to driver state. |
|
|
|
48:45.360 --> 48:46.280 |
|
Driver state. |
|
|
|
48:46.280 --> 48:48.360 |
|
We have car state packet, which has the state of the car |
|
|
|
48:48.360 --> 48:51.040 |
|
and driver state packet, which has state of the driver. |
|
|
|
48:51.040 --> 48:52.240 |
|
So what is it? |
|
|
|
48:52.240 --> 48:54.080 |
|
Estimate their BAC. |
|
|
|
48:54.080 --> 48:54.920 |
|
What's BAC? |
|
|
|
48:54.920 --> 48:55.920 |
|
Blood alcohol, kind of. |
|
|
|
48:57.360 --> 48:59.240 |
|
You think that's possible with computer vision? |
|
|
|
48:59.240 --> 49:00.080 |
|
Absolutely. |
|
|
|
49:02.560 --> 49:04.520 |
|
It's a, to me, it's an open question. |
|
|
|
49:04.520 --> 49:06.600 |
|
I haven't looked into too much. |
|
|
|
49:06.600 --> 49:08.440 |
|
Actually, I quite seriously looked at the literature. |
|
|
|
49:08.440 --> 49:10.840 |
|
It's not obvious to me that from the eyes and so on, |
|
|
|
49:10.840 --> 49:11.680 |
|
you can tell. |
|
|
|
49:11.680 --> 49:13.440 |
|
You might need stuff from the car as well. |
|
|
|
49:13.440 --> 49:15.760 |
|
You might need how they're controlling the car, right? |
|
|
|
49:15.760 --> 49:17.360 |
|
And that's fundamentally at the end of the day |
|
|
|
49:17.360 --> 49:18.640 |
|
what you care about. |
|
|
|
49:18.640 --> 49:21.640 |
|
But I think, especially when people are really drunk, |
|
|
|
49:21.640 --> 49:23.640 |
|
they're not controlling the car nearly as smoothly |
|
|
|
49:23.640 --> 49:25.160 |
|
as they would look at them walking, right? |
|
|
|
49:25.160 --> 49:27.240 |
|
They're, the car is like an extension of the body. |
|
|
|
49:27.240 --> 49:29.360 |
|
So I think you could totally detect. |
|
|
|
49:29.360 --> 49:30.880 |
|
And if you could fix people who are drunk, |
|
|
|
49:30.880 --> 49:32.840 |
|
distracted, asleep, if you fix those three. |
|
|
|
49:32.840 --> 49:35.480 |
|
Yeah, that's a huge, that's huge. |
|
|
|
49:35.480 --> 49:38.240 |
|
So what are the current limitations of OpenPilot? |
|
|
|
49:38.240 --> 49:41.720 |
|
What are the main problems that still need to be solved? |
|
|
|
49:41.720 --> 49:45.440 |
|
We're hopefully fixing a few of them in zero six. |
|
|
|
49:45.440 --> 49:48.400 |
|
We're not as good as autopilot at stop cars. |
|
|
|
49:49.440 --> 49:54.240 |
|
So if you're coming up to a red light at like 55, |
|
|
|
49:55.200 --> 49:56.880 |
|
so it's the radar stopped car problem, |
|
|
|
49:56.880 --> 49:59.200 |
|
which is responsible for two autopilot accidents, |
|
|
|
49:59.200 --> 50:01.480 |
|
it's hard to differentiate a stopped car |
|
|
|
50:01.480 --> 50:03.640 |
|
from a like signpost. |
|
|
|
50:03.640 --> 50:05.320 |
|
Yeah, static object. |
|
|
|
50:05.320 --> 50:07.520 |
|
So you have to fuse, you have to do this visually. |
|
|
|
50:07.520 --> 50:09.600 |
|
There's no way from the radar data to tell the difference. |
|
|
|
50:09.600 --> 50:10.680 |
|
Maybe you can make a map, |
|
|
|
50:10.680 --> 50:13.840 |
|
but I don't really believe in mapping at all anymore. |
|
|
|
50:13.840 --> 50:14.920 |
|
Wait, wait, wait, what? |
|
|
|
50:14.920 --> 50:16.040 |
|
You don't believe in mapping? |
|
|
|
50:16.040 --> 50:16.880 |
|
No. |
|
|
|
50:16.880 --> 50:21.120 |
|
So you're basically, the OpenPilot solution is saying, |
|
|
|
50:21.120 --> 50:22.480 |
|
react to the environment as you see it, |
|
|
|
50:22.480 --> 50:24.480 |
|
just like human doing beings do. |
|
|
|
50:24.480 --> 50:26.200 |
|
And then eventually when you want to do navigate |
|
|
|
50:26.200 --> 50:30.400 |
|
on OpenPilot, I'll train the net to look at ways. |
|
|
|
50:30.400 --> 50:31.360 |
|
I'll run ways in the background, |
|
|
|
50:31.360 --> 50:32.200 |
|
I'll train and come down a way. |
|
|
|
50:32.200 --> 50:33.560 |
|
Are you using GPS at all? |
|
|
|
50:33.560 --> 50:34.840 |
|
We use it to ground truth. |
|
|
|
50:34.840 --> 50:37.440 |
|
We use it to very carefully ground truth the paths. |
|
|
|
50:37.440 --> 50:39.560 |
|
We have a stack which can recover relative |
|
|
|
50:39.560 --> 50:41.800 |
|
to 10 centimeters over one minute. |
|
|
|
50:41.800 --> 50:43.440 |
|
And then we use that to ground truth |
|
|
|
50:43.440 --> 50:45.880 |
|
exactly where the car went in that local part |
|
|
|
50:45.880 --> 50:47.800 |
|
of the environment, but it's all local. |
|
|
|
50:47.800 --> 50:49.160 |
|
How are you testing in general? |
|
|
|
50:49.160 --> 50:51.400 |
|
Just for yourself, like experiments and stuff. |
|
|
|
50:51.400 --> 50:54.000 |
|
Where are you located? |
|
|
|
50:54.000 --> 50:54.840 |
|
San Diego. |
|
|
|
50:54.840 --> 50:55.680 |
|
San Diego. |
|
|
|
50:55.680 --> 50:56.520 |
|
Yeah. |
|
|
|
50:56.520 --> 50:57.360 |
|
Okay. |
|
|
|
50:57.360 --> 50:59.760 |
|
So you basically drive around there, |
|
|
|
50:59.760 --> 51:02.200 |
|
collect some data and watch the performance? |
|
|
|
51:02.200 --> 51:04.800 |
|
We have a simulator now and we have, |
|
|
|
51:04.800 --> 51:06.440 |
|
our simulator is really cool. |
|
|
|
51:06.440 --> 51:08.120 |
|
Our simulator is not, |
|
|
|
51:08.120 --> 51:09.720 |
|
it's not like a Unity based simulator. |
|
|
|
51:09.720 --> 51:11.840 |
|
Our simulator lets us load in real estate. |
|
|
|
51:12.880 --> 51:13.720 |
|
What do you mean? |
|
|
|
51:13.720 --> 51:16.760 |
|
We can load in a drive and simulate |
|
|
|
51:16.760 --> 51:20.280 |
|
what the system would have done on the historical data. |
|
|
|
51:20.280 --> 51:21.480 |
|
Ooh, nice. |
|
|
|
51:22.520 --> 51:24.360 |
|
Interesting. |
|
|
|
51:24.360 --> 51:26.080 |
|
Right now we're only using it for testing, |
|
|
|
51:26.080 --> 51:28.640 |
|
but as soon as we start using it for training. |
|
|
|
51:28.640 --> 51:29.480 |
|
That's it. |
|
|
|
51:29.480 --> 51:30.840 |
|
That's all set up for us. |
|
|
|
51:30.840 --> 51:33.040 |
|
What's your feeling about the real world versus simulation? |
|
|
|
51:33.040 --> 51:34.320 |
|
Do you like simulation for training? |
|
|
|
51:34.320 --> 51:35.720 |
|
If this moves to training? |
|
|
|
51:35.720 --> 51:40.040 |
|
So we have to distinguish two types of simulators, right? |
|
|
|
51:40.040 --> 51:44.720 |
|
There's a simulator that like is completely fake. |
|
|
|
51:44.720 --> 51:46.720 |
|
I could get my car to drive around in GTA. |
|
|
|
51:47.800 --> 51:51.080 |
|
I feel that this kind of simulator is useless. |
|
|
|
51:51.880 --> 51:53.640 |
|
You're never, there's so many. |
|
|
|
51:54.640 --> 51:57.000 |
|
My analogy here is like, okay, fine. |
|
|
|
51:57.000 --> 51:59.920 |
|
You're not solving the computer vision problem, |
|
|
|
51:59.920 --> 52:02.440 |
|
but you're solving the computer graphics problem. |
|
|
|
52:02.440 --> 52:03.280 |
|
Right. |
|
|
|
52:03.280 --> 52:04.600 |
|
And you don't think you can get very far |
|
|
|
52:04.600 --> 52:08.040 |
|
by creating ultra realistic graphics? |
|
|
|
52:08.040 --> 52:10.360 |
|
No, because you can create ultra realistic graphics |
|
|
|
52:10.360 --> 52:13.160 |
|
or the road, now create ultra realistic behavioral models |
|
|
|
52:13.160 --> 52:14.600 |
|
of the other cars. |
|
|
|
52:14.600 --> 52:16.920 |
|
Oh, well, I'll just use myself driving. |
|
|
|
52:16.920 --> 52:18.280 |
|
No, you won't. |
|
|
|
52:18.280 --> 52:21.640 |
|
You need real, you need actual human behavior |
|
|
|
52:21.640 --> 52:23.320 |
|
because that's what you're trying to learn. |
|
|
|
52:23.320 --> 52:25.840 |
|
The driving does not have a spec. |
|
|
|
52:25.840 --> 52:29.920 |
|
The definition of driving is what humans do when they drive. |
|
|
|
52:29.920 --> 52:32.800 |
|
Whatever Waymo does, I don't think it's driving. |
|
|
|
52:32.800 --> 52:33.640 |
|
Right. |
|
|
|
52:33.640 --> 52:36.400 |
|
Well, I think actually Waymo and others, |
|
|
|
52:36.400 --> 52:38.920 |
|
if there's any use for reinforcement learning, |
|
|
|
52:38.920 --> 52:40.360 |
|
I've seen it used quite well. |
|
|
|
52:40.360 --> 52:41.640 |
|
I studied pedestrians a lot too, |
|
|
|
52:41.640 --> 52:44.360 |
|
is try to train models from real data |
|
|
|
52:44.360 --> 52:46.920 |
|
of how pedestrians move and try to use reinforcement learning |
|
|
|
52:46.920 --> 52:50.040 |
|
models to make pedestrians move in human like ways. |
|
|
|
52:50.040 --> 52:53.520 |
|
By that point, you've already gone so many layers, |
|
|
|
52:53.520 --> 52:55.680 |
|
you detected a pedestrian. |
|
|
|
52:55.680 --> 52:59.640 |
|
Did you hand code the feature vector of their state? |
|
|
|
52:59.640 --> 53:00.480 |
|
Right. |
|
|
|
53:00.480 --> 53:02.880 |
|
Did you guys learn anything from computer vision |
|
|
|
53:02.880 --> 53:04.600 |
|
before deep learning? |
|
|
|
53:04.600 --> 53:07.160 |
|
Well, okay, I feel like this is... |
|
|
|
53:07.160 --> 53:10.840 |
|
So perception to you is the sticking point. |
|
|
|
53:10.840 --> 53:13.760 |
|
I mean, what's the hardest part of the stack here? |
|
|
|
53:13.760 --> 53:18.760 |
|
There is no human understandable feature vector |
|
|
|
53:19.680 --> 53:22.000 |
|
separating perception and planning. |
|
|
|
53:23.040 --> 53:25.120 |
|
That's the best way I can put that. |
|
|
|
53:25.120 --> 53:25.960 |
|
There is no... |
|
|
|
53:25.960 --> 53:29.600 |
|
So it's all together and it's a joint problem. |
|
|
|
53:29.600 --> 53:31.480 |
|
So you can take localization. |
|
|
|
53:31.480 --> 53:32.960 |
|
Localization and planning, |
|
|
|
53:32.960 --> 53:34.760 |
|
there is a human understandable feature vector |
|
|
|
53:34.760 --> 53:36.000 |
|
between these two things. |
|
|
|
53:36.000 --> 53:38.720 |
|
I mean, okay, so I have like three degrees position, |
|
|
|
53:38.720 --> 53:40.560 |
|
three degrees orientation and those derivatives, |
|
|
|
53:40.560 --> 53:42.000 |
|
maybe those second derivatives, right? |
|
|
|
53:42.000 --> 53:44.520 |
|
That's human understandable, that's physical. |
|
|
|
53:44.520 --> 53:48.560 |
|
The between perception and planning. |
|
|
|
53:49.520 --> 53:53.600 |
|
So like Waymo has a perception stack and then a planner. |
|
|
|
53:53.600 --> 53:55.560 |
|
And one of the things Waymo does right |
|
|
|
53:55.560 --> 54:00.000 |
|
is they have a simulator that can separate those two. |
|
|
|
54:00.000 --> 54:02.920 |
|
They can like replay their perception data |
|
|
|
54:02.920 --> 54:03.920 |
|
and test their system, |
|
|
|
54:03.920 --> 54:04.880 |
|
which is what I'm talking about |
|
|
|
54:04.880 --> 54:06.520 |
|
about like the two different kinds of simulators. |
|
|
|
54:06.520 --> 54:08.240 |
|
There's the kind that can work on real data |
|
|
|
54:08.240 --> 54:10.920 |
|
and there's the kind that can't work on real data. |
|
|
|
54:10.920 --> 54:13.880 |
|
Now, the problem is that I don't think |
|
|
|
54:13.880 --> 54:16.160 |
|
you can hand code a feature vector, right? |
|
|
|
54:16.160 --> 54:17.360 |
|
Like you have some list of like, |
|
|
|
54:17.360 --> 54:19.040 |
|
well, here's my list of cars in the scenes. |
|
|
|
54:19.040 --> 54:21.280 |
|
Here's my list of pedestrians in the scene. |
|
|
|
54:21.280 --> 54:23.240 |
|
This isn't what humans are doing. |
|
|
|
54:23.240 --> 54:24.920 |
|
What are humans doing? |
|
|
|
54:24.920 --> 54:25.760 |
|
Global. |
|
|
|
54:27.200 --> 54:28.040 |
|
Some, some. |
|
|
|
54:28.040 --> 54:31.960 |
|
You're saying that's too difficult to hand engineer. |
|
|
|
54:31.960 --> 54:34.120 |
|
I'm saying that there is no state vector. |
|
|
|
54:34.120 --> 54:36.560 |
|
Given a perfect, I could give you the best team |
|
|
|
54:36.560 --> 54:38.520 |
|
of engineers in the world to build a perception system |
|
|
|
54:38.520 --> 54:40.640 |
|
and the best team to build a planner. |
|
|
|
54:40.640 --> 54:42.640 |
|
All you have to do is define the state vector |
|
|
|
54:42.640 --> 54:43.960 |
|
that separates those two. |
|
|
|
54:43.960 --> 54:48.560 |
|
I'm missing the state vector that separates those two. |
|
|
|
54:48.560 --> 54:49.400 |
|
What do you mean? |
|
|
|
54:49.400 --> 54:54.000 |
|
So what is the output of your perception system? |
|
|
|
54:54.000 --> 54:56.880 |
|
Output of the perception system. |
|
|
|
54:56.880 --> 55:01.560 |
|
It's, there's, okay, well, there's several ways to do it. |
|
|
|
55:01.560 --> 55:03.840 |
|
One is the slam component is localization. |
|
|
|
55:03.840 --> 55:05.920 |
|
The other is drivable area, drivable space. |
|
|
|
55:05.920 --> 55:06.760 |
|
Drivable space, yep. |
|
|
|
55:06.760 --> 55:09.000 |
|
And then there's the different objects in the scene. |
|
|
|
55:09.000 --> 55:09.840 |
|
Yep. |
|
|
|
55:11.000 --> 55:16.000 |
|
And different objects in the scene over time maybe |
|
|
|
55:16.000 --> 55:18.720 |
|
to give you input to then try to start |
|
|
|
55:18.720 --> 55:21.560 |
|
modeling the trajectories of those objects. |
|
|
|
55:21.560 --> 55:22.400 |
|
Sure. |
|
|
|
55:22.400 --> 55:23.240 |
|
That's it. |
|
|
|
55:23.240 --> 55:25.160 |
|
I can give you a concrete example of something you missed. |
|
|
|
55:25.160 --> 55:26.000 |
|
What's that? |
|
|
|
55:26.000 --> 55:28.640 |
|
So say there's a bush in the scene. |
|
|
|
55:28.640 --> 55:30.920 |
|
Humans understand that when they see this bush |
|
|
|
55:30.920 --> 55:34.680 |
|
that there may or may not be a car behind that bush. |
|
|
|
55:34.680 --> 55:37.280 |
|
Drivable area and a list of objects does not include that. |
|
|
|
55:37.280 --> 55:38.920 |
|
Humans are doing this constantly |
|
|
|
55:38.920 --> 55:40.920 |
|
at the simplest intersections. |
|
|
|
55:40.920 --> 55:43.880 |
|
So now you have to talk about occluded area. |
|
|
|
55:43.880 --> 55:44.720 |
|
Right. |
|
|
|
55:44.720 --> 55:47.800 |
|
Right, but even that, what do you mean by occluded? |
|
|
|
55:47.800 --> 55:49.640 |
|
Okay, so I can't see it. |
|
|
|
55:49.640 --> 55:51.840 |
|
Well, if it's the other side of a house, I don't care. |
|
|
|
55:51.840 --> 55:53.560 |
|
What's the likelihood that there's a car |
|
|
|
55:53.560 --> 55:55.280 |
|
in that occluded area, right? |
|
|
|
55:55.280 --> 55:58.080 |
|
And if you say, okay, we'll add that, |
|
|
|
55:58.080 --> 56:00.680 |
|
I can come up with 10 more examples that you can't add. |
|
|
|
56:01.680 --> 56:03.960 |
|
Certainly occluded area would be something |
|
|
|
56:03.960 --> 56:06.760 |
|
that simulator would have because it's simulating |
|
|
|
56:06.760 --> 56:11.320 |
|
the entire, you know, occlusion is part of it. |
|
|
|
56:11.320 --> 56:12.680 |
|
Occlusion is part of a vision stack. |
|
|
|
56:12.680 --> 56:13.520 |
|
Vision stack. |
|
|
|
56:13.520 --> 56:16.600 |
|
But what I'm saying is if you have a hand engineered, |
|
|
|
56:16.600 --> 56:20.040 |
|
if your perception system output can be written |
|
|
|
56:20.040 --> 56:22.240 |
|
in a spec document, it is incomplete. |
|
|
|
56:23.120 --> 56:27.800 |
|
Yeah, I mean, I certainly, it's hard to argue with that |
|
|
|
56:27.800 --> 56:30.120 |
|
because in the end, that's going to be true. |
|
|
|
56:30.120 --> 56:31.760 |
|
Yeah, and I'll tell you what the output |
|
|
|
56:31.760 --> 56:32.720 |
|
of our perception system is. |
|
|
|
56:32.720 --> 56:33.560 |
|
What's that? |
|
|
|
56:33.560 --> 56:37.120 |
|
It's a 1024 dimensional vector. |
|
|
|
56:37.120 --> 56:38.000 |
|
Transparent neural net. |
|
|
|
56:38.000 --> 56:39.000 |
|
Oh, you know that. |
|
|
|
56:39.000 --> 56:42.000 |
|
No, that's the 1024 dimensions of who knows what. |
|
|
|
56:43.520 --> 56:45.160 |
|
Because it's operating on real data. |
|
|
|
56:45.160 --> 56:47.000 |
|
Yeah. |
|
|
|
56:47.000 --> 56:48.320 |
|
And that's the perception. |
|
|
|
56:48.320 --> 56:50.360 |
|
That's the perception state, right? |
|
|
|
56:50.360 --> 56:53.520 |
|
Think about an autoencoder for faces, right? |
|
|
|
56:53.520 --> 56:54.720 |
|
If you have an autoencoder for faces |
|
|
|
56:54.720 --> 56:59.720 |
|
and you say it has 256 dimensions in the middle, |
|
|
|
56:59.720 --> 57:00.680 |
|
and I'm taking a face over here |
|
|
|
57:00.680 --> 57:02.800 |
|
and projecting it to a face over here. |
|
|
|
57:02.800 --> 57:05.360 |
|
Can you hand label all 256 of those dimensions? |
|
|
|
57:06.280 --> 57:09.240 |
|
Well, no, but those are generated automatically. |
|
|
|
57:09.240 --> 57:11.360 |
|
But even if you tried to do it by hand, |
|
|
|
57:11.360 --> 57:15.520 |
|
could you come up with a spec between your encoder |
|
|
|
57:15.520 --> 57:16.360 |
|
and your decoder? |
|
|
|
57:17.400 --> 57:20.720 |
|
No, no, because it wasn't designed, but they're... |
|
|
|
57:20.720 --> 57:22.600 |
|
No, no, no, but if you could design it, |
|
|
|
57:23.600 --> 57:26.480 |
|
if you could design a face reconstructor system, |
|
|
|
57:26.480 --> 57:28.080 |
|
could you come up with a spec? |
|
|
|
57:29.240 --> 57:32.320 |
|
No, but I think we're missing here a little bit. |
|
|
|
57:32.320 --> 57:35.080 |
|
I think you're just being very poetic |
|
|
|
57:35.080 --> 57:37.880 |
|
about expressing a fundamental problem of simulators, |
|
|
|
57:37.880 --> 57:41.640 |
|
that they are going to be missing so much |
|
|
|
57:42.480 --> 57:44.680 |
|
that the feature of actually |
|
|
|
57:44.680 --> 57:47.080 |
|
would just look fundamentally different |
|
|
|
57:47.080 --> 57:50.440 |
|
from in the simulated world than the real world. |
|
|
|
57:51.280 --> 57:53.800 |
|
I'm not making a claim about simulators. |
|
|
|
57:53.800 --> 57:57.120 |
|
I'm making a claim about the spec division |
|
|
|
57:57.120 --> 57:58.800 |
|
between perception and planning. |
|
|
|
57:58.800 --> 57:59.640 |
|
And planning. |
|
|
|
57:59.640 --> 58:00.840 |
|
Even in your system. |
|
|
|
58:00.840 --> 58:01.800 |
|
Just in general. |
|
|
|
58:01.800 --> 58:03.360 |
|
Right, just in general. |
|
|
|
58:03.360 --> 58:05.680 |
|
If you're trying to build a car that drives, |
|
|
|
58:05.680 --> 58:07.280 |
|
if you're trying to hand code |
|
|
|
58:07.280 --> 58:08.760 |
|
the output of your perception system, |
|
|
|
58:08.760 --> 58:10.960 |
|
like saying, here's a list of all the cars in the scene. |
|
|
|
58:10.960 --> 58:11.920 |
|
Here's a list of all the people. |
|
|
|
58:11.920 --> 58:13.120 |
|
Here's a list of the occluded areas. |
|
|
|
58:13.120 --> 58:14.920 |
|
Here's a vector of drivable areas. |
|
|
|
58:14.920 --> 58:16.600 |
|
It's insufficient. |
|
|
|
58:16.600 --> 58:18.000 |
|
And if you start to believe that, |
|
|
|
58:18.000 --> 58:20.840 |
|
you realize that what Waymo and Cruz are doing is impossible. |
|
|
|
58:20.840 --> 58:24.320 |
|
Currently, what we're doing is the perception problem |
|
|
|
58:24.320 --> 58:28.200 |
|
is converting the scene into a chessboard. |
|
|
|
58:29.200 --> 58:31.720 |
|
And then you reason some basic reasoning |
|
|
|
58:31.720 --> 58:33.400 |
|
around that chessboard. |
|
|
|
58:33.400 --> 58:38.080 |
|
And you're saying that really there's a lot missing there. |
|
|
|
58:38.080 --> 58:40.240 |
|
First of all, why are we talking about this? |
|
|
|
58:40.240 --> 58:42.840 |
|
Because isn't this a full autonomy? |
|
|
|
58:42.840 --> 58:44.720 |
|
Is this something you think about? |
|
|
|
58:44.720 --> 58:47.680 |
|
Oh, I want to win self driving cars. |
|
|
|
58:47.680 --> 58:52.680 |
|
So your definition of win includes the full five. |
|
|
|
58:53.680 --> 58:55.800 |
|
I don't think level four is a real thing. |
|
|
|
58:55.800 --> 58:59.720 |
|
I want to build the AlphaGo of driving. |
|
|
|
58:59.720 --> 59:04.720 |
|
So AlphaGo is really end to end. |
|
|
|
59:06.160 --> 59:07.000 |
|
Yeah. |
|
|
|
59:07.000 --> 59:09.840 |
|
Is, yeah, it's end to end. |
|
|
|
59:09.840 --> 59:12.480 |
|
And do you think this whole problem, |
|
|
|
59:12.480 --> 59:14.680 |
|
is that also kind of what you're getting at |
|
|
|
59:14.680 --> 59:16.640 |
|
with the perception and the planning? |
|
|
|
59:16.640 --> 59:19.440 |
|
Is that this whole problem, the right way to do it, |
|
|
|
59:19.440 --> 59:21.600 |
|
is really to learn the entire thing? |
|
|
|
59:21.600 --> 59:23.680 |
|
I'll argue that not only is it the right way, |
|
|
|
59:23.680 --> 59:27.640 |
|
it's the only way that's going to exceed human performance. |
|
|
|
59:27.640 --> 59:29.960 |
|
Well, it's certainly true for Go. |
|
|
|
59:29.960 --> 59:31.520 |
|
Everyone who tried to hand code Go things |
|
|
|
59:31.520 --> 59:33.440 |
|
built human inferior things. |
|
|
|
59:33.440 --> 59:36.200 |
|
And then someone came along and wrote some 10,000 line thing |
|
|
|
59:36.200 --> 59:39.800 |
|
that doesn't know anything about Go that beat everybody. |
|
|
|
59:39.800 --> 59:41.080 |
|
It's 10,000 lines. |
|
|
|
59:41.080 --> 59:43.360 |
|
True, in that sense. |
|
|
|
59:43.360 --> 59:47.520 |
|
The open question then that maybe I can ask you |
|
|
|
59:47.520 --> 59:52.520 |
|
is driving is much harder than Go. |
|
|
|
59:53.440 --> 59:56.240 |
|
The open question is how much harder? |
|
|
|
59:56.240 --> 59:59.480 |
|
So how, because I think the Elon Musk approach here |
|
|
|
59:59.480 --> 1:00:01.600 |
|
with planning and perception is similar |
|
|
|
1:00:01.600 --> 1:00:02.960 |
|
to what you're describing, |
|
|
|
1:00:02.960 --> 1:00:07.960 |
|
which is really turning into not some kind of modular thing, |
|
|
|
1:00:08.280 --> 1:00:11.120 |
|
but really do formulate as a learning problem |
|
|
|
1:00:11.120 --> 1:00:13.360 |
|
and solve the learning problem with scale. |
|
|
|
1:00:13.360 --> 1:00:17.120 |
|
So how many years, put one, |
|
|
|
1:00:17.120 --> 1:00:18.880 |
|
how many years would it take to solve this problem |
|
|
|
1:00:18.880 --> 1:00:21.680 |
|
or just how hard is this freaking problem? |
|
|
|
1:00:21.680 --> 1:00:24.560 |
|
Well, the cool thing is, |
|
|
|
1:00:24.560 --> 1:00:27.800 |
|
I think there's a lot of value |
|
|
|
1:00:27.800 --> 1:00:29.840 |
|
that we can deliver along the way. |
|
|
|
1:00:30.840 --> 1:00:35.840 |
|
I think that you can build lame keeping assist |
|
|
|
1:00:36.600 --> 1:00:41.440 |
|
actually plus adaptive cruise control plus, okay, |
|
|
|
1:00:41.440 --> 1:00:46.000 |
|
looking at ways extends to like all of driving. |
|
|
|
1:00:46.000 --> 1:00:47.920 |
|
Yeah, most of driving, right? |
|
|
|
1:00:47.920 --> 1:00:49.760 |
|
Oh, your adaptive cruise control treats red lights |
|
|
|
1:00:49.760 --> 1:00:51.200 |
|
like cars, okay. |
|
|
|
1:00:51.200 --> 1:00:53.480 |
|
So let's jump around with you mentioned |
|
|
|
1:00:53.480 --> 1:00:55.760 |
|
that you didn't like navigate an autopilot. |
|
|
|
1:00:55.760 --> 1:00:57.760 |
|
What advice, how would you make it better? |
|
|
|
1:00:57.760 --> 1:01:00.560 |
|
Do you think as a feature that if it's done really well, |
|
|
|
1:01:00.560 --> 1:01:02.360 |
|
it's a good feature? |
|
|
|
1:01:02.360 --> 1:01:07.360 |
|
I think that it's too reliant on like hand coded hacks |
|
|
|
1:01:07.520 --> 1:01:10.400 |
|
for like, how does navigate an autopilot do a lane change? |
|
|
|
1:01:10.400 --> 1:01:13.400 |
|
It actually does the same lane change every time |
|
|
|
1:01:13.400 --> 1:01:14.320 |
|
and it feels mechanical. |
|
|
|
1:01:14.320 --> 1:01:15.920 |
|
Humans do different lane changes. |
|
|
|
1:01:15.920 --> 1:01:17.360 |
|
Humans, sometimes we'll do a slow one, |
|
|
|
1:01:17.360 --> 1:01:18.920 |
|
sometimes do a fast one. |
|
|
|
1:01:18.920 --> 1:01:20.880 |
|
Navigate an autopilot at least every time I use it |
|
|
|
1:01:20.880 --> 1:01:23.040 |
|
is it the identical lane change? |
|
|
|
1:01:23.040 --> 1:01:24.280 |
|
How do you learn? |
|
|
|
1:01:24.280 --> 1:01:26.800 |
|
I mean, this is a fundamental thing actually |
|
|
|
1:01:26.800 --> 1:01:30.400 |
|
is the breaking and accelerating, |
|
|
|
1:01:30.400 --> 1:01:33.960 |
|
something that still, Tesla probably does it better |
|
|
|
1:01:33.960 --> 1:01:36.800 |
|
than most cars, but it still doesn't do a great job |
|
|
|
1:01:36.800 --> 1:01:39.960 |
|
of creating a comfortable natural experience |
|
|
|
1:01:39.960 --> 1:01:42.680 |
|
and navigate an autopilot is just lane changes |
|
|
|
1:01:42.680 --> 1:01:44.120 |
|
and extension of that. |
|
|
|
1:01:44.120 --> 1:01:49.120 |
|
So how do you learn to do natural lane change? |
|
|
|
1:01:49.120 --> 1:01:52.920 |
|
So we have it and I can talk about how it works. |
|
|
|
1:01:52.920 --> 1:01:57.920 |
|
So I feel that we have the solution for lateral |
|
|
|
1:01:58.720 --> 1:02:00.640 |
|
but we don't yet have the solution for longitudinal. |
|
|
|
1:02:00.640 --> 1:02:03.360 |
|
There's a few reasons longitudinal is harder than lateral. |
|
|
|
1:02:03.360 --> 1:02:06.920 |
|
The lane change component, the way that we train on it |
|
|
|
1:02:06.920 --> 1:02:10.840 |
|
very simply is like our model has an input |
|
|
|
1:02:10.840 --> 1:02:14.040 |
|
for whether it's doing a lane change or not. |
|
|
|
1:02:14.040 --> 1:02:16.360 |
|
And then when we train the end to end model, |
|
|
|
1:02:16.360 --> 1:02:19.560 |
|
we hand label all the lane changes because you have to. |
|
|
|
1:02:19.560 --> 1:02:22.440 |
|
I've struggled a long time about not wanting to do that |
|
|
|
1:02:22.440 --> 1:02:24.280 |
|
but I think you have to. |
|
|
|
1:02:24.280 --> 1:02:25.320 |
|
Or the training data. |
|
|
|
1:02:25.320 --> 1:02:26.520 |
|
For the training data, right? |
|
|
|
1:02:26.520 --> 1:02:28.280 |
|
We actually have an automatic ground truth |
|
|
|
1:02:28.280 --> 1:02:30.600 |
|
or which automatically labels all the lane changes. |
|
|
|
1:02:30.600 --> 1:02:31.680 |
|
Was that possible? |
|
|
|
1:02:31.680 --> 1:02:32.720 |
|
To automatically label lane changes? |
|
|
|
1:02:32.720 --> 1:02:33.560 |
|
Yeah. |
|
|
|
1:02:33.560 --> 1:02:34.800 |
|
And detect the lane I see when it crosses it, right? |
|
|
|
1:02:34.800 --> 1:02:36.680 |
|
And I don't have to get that high percent accuracy |
|
|
|
1:02:36.680 --> 1:02:38.080 |
|
but it's like 95 good enough. |
|
|
|
1:02:38.080 --> 1:02:38.960 |
|
Okay. |
|
|
|
1:02:38.960 --> 1:02:43.200 |
|
Now I set the bit when it's doing the lane change |
|
|
|
1:02:43.200 --> 1:02:44.840 |
|
in the end to end learning. |
|
|
|
1:02:44.840 --> 1:02:47.920 |
|
And then I set it to zero when it's not doing a lane change. |
|
|
|
1:02:47.920 --> 1:02:49.720 |
|
So now if I want us to do a lane change a test time, |
|
|
|
1:02:49.720 --> 1:02:52.360 |
|
I just put the bit to a one and it'll do a lane change. |
|
|
|
1:02:52.360 --> 1:02:54.640 |
|
Yeah, but so if you look at the space of lane change, |
|
|
|
1:02:54.640 --> 1:02:57.320 |
|
you know some percentage, not a hundred percent, |
|
|
|
1:02:57.320 --> 1:03:01.120 |
|
that we make as humans is not a pleasant experience |
|
|
|
1:03:01.120 --> 1:03:02.800 |
|
because we messed some part of it up. |
|
|
|
1:03:02.800 --> 1:03:04.320 |
|
It's nerve wracking to change. |
|
|
|
1:03:04.320 --> 1:03:05.760 |
|
If you look, you have to see, |
|
|
|
1:03:05.760 --> 1:03:06.920 |
|
it has to accelerate. |
|
|
|
1:03:06.920 --> 1:03:09.920 |
|
How do we label the ones that are natural and feel good? |
|
|
|
1:03:09.920 --> 1:03:11.560 |
|
You know, that's the, |
|
|
|
1:03:11.560 --> 1:03:13.360 |
|
because that's your ultimate criticism, |
|
|
|
1:03:13.360 --> 1:03:17.000 |
|
the current navigate and autopilot just doesn't feel good. |
|
|
|
1:03:17.000 --> 1:03:18.520 |
|
Well, the current navigate and autopilot |
|
|
|
1:03:18.520 --> 1:03:21.720 |
|
is a hand coded policy written by an engineer in a room |
|
|
|
1:03:21.720 --> 1:03:25.080 |
|
who probably went out and tested it a few times on the 280. |
|
|
|
1:03:25.080 --> 1:03:28.560 |
|
Probably a more, a better version of that. |
|
|
|
1:03:28.560 --> 1:03:29.400 |
|
But yes. |
|
|
|
1:03:29.400 --> 1:03:30.560 |
|
That's how we would have written it. |
|
|
|
1:03:30.560 --> 1:03:31.400 |
|
Yeah. |
|
|
|
1:03:31.400 --> 1:03:33.480 |
|
Maybe Tesla did a Tesla, they tested it in. |
|
|
|
1:03:33.480 --> 1:03:34.920 |
|
That might have been two engineers. |
|
|
|
1:03:34.920 --> 1:03:35.760 |
|
Two engineers. |
|
|
|
1:03:35.760 --> 1:03:37.400 |
|
Yeah. |
|
|
|
1:03:37.400 --> 1:03:40.120 |
|
No, but so if you learn the lane change, |
|
|
|
1:03:40.120 --> 1:03:42.480 |
|
if you learn how to do a lane change from data, |
|
|
|
1:03:42.480 --> 1:03:44.680 |
|
just like you have a label that says lane change |
|
|
|
1:03:44.680 --> 1:03:48.040 |
|
and then you put it in when you want it to do the lane change, |
|
|
|
1:03:48.040 --> 1:03:49.640 |
|
it'll automatically do the lane change |
|
|
|
1:03:49.640 --> 1:03:51.600 |
|
that's appropriate for the situation. |
|
|
|
1:03:51.600 --> 1:03:54.720 |
|
Now, to get at the problem of some humans |
|
|
|
1:03:54.720 --> 1:03:55.960 |
|
do bad lane changes, |
|
|
|
1:03:57.400 --> 1:03:59.920 |
|
we haven't worked too much on this problem yet. |
|
|
|
1:03:59.920 --> 1:04:03.120 |
|
It's not that much of a problem in practice. |
|
|
|
1:04:03.120 --> 1:04:06.160 |
|
My theory is that all good drivers are good in the same way |
|
|
|
1:04:06.160 --> 1:04:08.440 |
|
and all bad drivers are bad in different ways. |
|
|
|
1:04:09.360 --> 1:04:11.320 |
|
And we've seen some data to back this up. |
|
|
|
1:04:11.320 --> 1:04:12.400 |
|
Well, beautifully put. |
|
|
|
1:04:12.400 --> 1:04:16.560 |
|
So you just basically, if that's true hypothesis, |
|
|
|
1:04:16.560 --> 1:04:19.920 |
|
then your task is to discover the good drivers. |
|
|
|
1:04:19.920 --> 1:04:21.800 |
|
The good drivers stand out |
|
|
|
1:04:21.800 --> 1:04:23.360 |
|
because they're in one cluster |
|
|
|
1:04:23.360 --> 1:04:25.200 |
|
and the bad drivers are scattered all over the place |
|
|
|
1:04:25.200 --> 1:04:27.240 |
|
and your net learns the cluster. |
|
|
|
1:04:27.240 --> 1:04:28.080 |
|
Yeah. |
|
|
|
1:04:28.080 --> 1:04:30.800 |
|
So you just learn from the good drivers |
|
|
|
1:04:30.800 --> 1:04:32.200 |
|
and they're easy to cluster. |
|
|
|
1:04:33.200 --> 1:04:34.240 |
|
In fact, we learned from all of them |
|
|
|
1:04:34.240 --> 1:04:35.840 |
|
and the net automatically learns the policy |
|
|
|
1:04:35.840 --> 1:04:36.920 |
|
that's like the majority. |
|
|
|
1:04:36.920 --> 1:04:38.440 |
|
But we'll eventually probably have to build some out. |
|
|
|
1:04:38.440 --> 1:04:41.560 |
|
So if that theory is true, I hope it's true |
|
|
|
1:04:41.560 --> 1:04:46.440 |
|
because the counter theory is there is many clusters, |
|
|
|
1:04:49.480 --> 1:04:53.680 |
|
maybe arbitrarily many clusters of good drivers. |
|
|
|
1:04:53.680 --> 1:04:55.840 |
|
Because if there's one cluster of good drivers, |
|
|
|
1:04:55.840 --> 1:04:57.600 |
|
you can at least discover a set of policies. |
|
|
|
1:04:57.600 --> 1:04:59.000 |
|
You can learn a set of policies |
|
|
|
1:04:59.000 --> 1:05:00.640 |
|
which would be good universally. |
|
|
|
1:05:00.640 --> 1:05:01.640 |
|
Yeah. |
|
|
|
1:05:01.640 --> 1:05:04.560 |
|
That would be nice if it's true. |
|
|
|
1:05:04.560 --> 1:05:06.560 |
|
And you're saying that there is some evidence that... |
|
|
|
1:05:06.560 --> 1:05:09.720 |
|
Let's say lane changes can be clustered into four clusters. |
|
|
|
1:05:09.720 --> 1:05:10.560 |
|
Right. |
|
|
|
1:05:10.560 --> 1:05:12.040 |
|
There's a finite level of... |
|
|
|
1:05:12.040 --> 1:05:15.280 |
|
I would argue that all four of those are good clusters. |
|
|
|
1:05:15.280 --> 1:05:18.360 |
|
All the things that are random are noise and probably bad. |
|
|
|
1:05:18.360 --> 1:05:20.360 |
|
And which one of the four you pick? |
|
|
|
1:05:20.360 --> 1:05:21.920 |
|
Or maybe it's 10 or maybe it's 20. |
|
|
|
1:05:21.920 --> 1:05:22.760 |
|
You can learn that. |
|
|
|
1:05:22.760 --> 1:05:23.800 |
|
It's context dependent. |
|
|
|
1:05:23.800 --> 1:05:25.040 |
|
It depends on the scene. |
|
|
|
1:05:26.760 --> 1:05:30.440 |
|
And the hope is it's not too dependent on the driver. |
|
|
|
1:05:31.400 --> 1:05:34.240 |
|
Yeah, the hope is that it all washes out. |
|
|
|
1:05:34.240 --> 1:05:36.960 |
|
The hope is that the distribution is not bimodal. |
|
|
|
1:05:36.960 --> 1:05:39.080 |
|
The hope is that it's a nice Gaussian. |
|
|
|
1:05:39.080 --> 1:05:41.640 |
|
So what advice would you give to Tesla? |
|
|
|
1:05:41.640 --> 1:05:45.000 |
|
How to fix, how to improve, navigate an autopilot? |
|
|
|
1:05:45.000 --> 1:05:48.240 |
|
That's the lessons that you've learned from Kamii. |
|
|
|
1:05:48.240 --> 1:05:50.560 |
|
The only real advice I would give to Tesla |
|
|
|
1:05:50.560 --> 1:05:52.920 |
|
is please put driver monitoring in your cars. |
|
|
|
1:05:53.920 --> 1:05:55.160 |
|
With respect to improving it. |
|
|
|
1:05:55.160 --> 1:05:56.000 |
|
You can't do that anymore. |
|
|
|
1:05:56.000 --> 1:05:57.280 |
|
I started to interrupt. |
|
|
|
1:05:57.280 --> 1:06:01.760 |
|
But there's a practical nature of many of hundreds of thousands |
|
|
|
1:06:01.760 --> 1:06:05.760 |
|
of cars being produced that don't have a good driver facing camera. |
|
|
|
1:06:05.760 --> 1:06:07.520 |
|
The Model 3 has a selfie cam. |
|
|
|
1:06:07.520 --> 1:06:08.680 |
|
Is it not good enough? |
|
|
|
1:06:08.680 --> 1:06:10.800 |
|
Did they not have put IR LEDs for night? |
|
|
|
1:06:10.800 --> 1:06:11.640 |
|
That's a good question. |
|
|
|
1:06:11.640 --> 1:06:13.360 |
|
But I do know that it's fish eye |
|
|
|
1:06:13.360 --> 1:06:15.800 |
|
and it's relatively low resolution. |
|
|
|
1:06:15.800 --> 1:06:16.760 |
|
So it's really not designed. |
|
|
|
1:06:16.760 --> 1:06:18.760 |
|
It wasn't designed for driver monitoring. |
|
|
|
1:06:18.760 --> 1:06:21.760 |
|
You can hope that you can kind of scrape up |
|
|
|
1:06:21.760 --> 1:06:24.400 |
|
and have something from it. |
|
|
|
1:06:24.400 --> 1:06:27.520 |
|
But why didn't they put it in today? |
|
|
|
1:06:27.520 --> 1:06:28.280 |
|
Put it in today. |
|
|
|
1:06:28.280 --> 1:06:29.520 |
|
Put it in today. |
|
|
|
1:06:29.520 --> 1:06:31.520 |
|
Every time I've heard Carpathian talk about the problem |
|
|
|
1:06:31.520 --> 1:06:33.240 |
|
and talking about like software 2.0 |
|
|
|
1:06:33.240 --> 1:06:35.240 |
|
and how the machine learning is gobbling up everything, |
|
|
|
1:06:35.240 --> 1:06:37.440 |
|
I think this is absolutely the right strategy. |
|
|
|
1:06:37.440 --> 1:06:40.160 |
|
I think that he didn't write and navigate on autopilot. |
|
|
|
1:06:40.160 --> 1:06:43.240 |
|
I think somebody else did and kind of hacked it on top of that stuff. |
|
|
|
1:06:43.240 --> 1:06:45.680 |
|
I think when Carpathian says, wait a second, |
|
|
|
1:06:45.680 --> 1:06:47.440 |
|
why did we hand code this lane change policy |
|
|
|
1:06:47.440 --> 1:06:48.360 |
|
with all these magic numbers? |
|
|
|
1:06:48.360 --> 1:06:49.360 |
|
We're going to learn it from data. |
|
|
|
1:06:49.360 --> 1:06:49.840 |
|
They'll fix it. |
|
|
|
1:06:49.840 --> 1:06:51.040 |
|
They already know what to do there. |
|
|
|
1:06:51.040 --> 1:06:54.360 |
|
Well, that's Andre's job is to turn everything |
|
|
|
1:06:54.360 --> 1:06:57.480 |
|
into a learning problem and collect a huge amount of data. |
|
|
|
1:06:57.480 --> 1:07:01.120 |
|
The reality is, though, not every problem |
|
|
|
1:07:01.120 --> 1:07:04.080 |
|
can be turned into a learning problem in the short term. |
|
|
|
1:07:04.080 --> 1:07:07.280 |
|
In the end, everything will be a learning problem. |
|
|
|
1:07:07.280 --> 1:07:12.880 |
|
The reality is, like if you want to build L5 vehicles today, |
|
|
|
1:07:12.880 --> 1:07:15.600 |
|
it will likely involve no learning. |
|
|
|
1:07:15.600 --> 1:07:20.320 |
|
And that's the reality is, so at which point does learning start? |
|
|
|
1:07:20.320 --> 1:07:23.480 |
|
It's the crutch statement that LiDAR is a crutch. |
|
|
|
1:07:23.480 --> 1:07:27.240 |
|
Which point will learning get up to part of human performance? |
|
|
|
1:07:27.240 --> 1:07:31.960 |
|
It's over human performance on ImageNet, classification, |
|
|
|
1:07:31.960 --> 1:07:34.000 |
|
on driving, it's a question still. |
|
|
|
1:07:34.000 --> 1:07:35.760 |
|
It is a question. |
|
|
|
1:07:35.760 --> 1:07:39.160 |
|
I'll say this, I'm here to play for 10 years. |
|
|
|
1:07:39.160 --> 1:07:40.280 |
|
I'm not here to try to. |
|
|
|
1:07:40.280 --> 1:07:42.960 |
|
I'm here to play for 10 years and make money along the way. |
|
|
|
1:07:42.960 --> 1:07:45.040 |
|
I'm not here to try to promise people |
|
|
|
1:07:45.040 --> 1:07:47.600 |
|
that I'm going to have my L5 taxi network up and working |
|
|
|
1:07:47.600 --> 1:07:48.200 |
|
in two years. |
|
|
|
1:07:48.200 --> 1:07:49.400 |
|
Do you think that was a mistake? |
|
|
|
1:07:49.400 --> 1:07:50.520 |
|
Yes. |
|
|
|
1:07:50.520 --> 1:07:53.160 |
|
What do you think was the motivation behind saying |
|
|
|
1:07:53.160 --> 1:07:56.640 |
|
that other companies are also promising L5 vehicles |
|
|
|
1:07:56.640 --> 1:08:01.880 |
|
with their different approaches in 2020, 2021, 2022? |
|
|
|
1:08:01.880 --> 1:08:05.720 |
|
If anybody would like to bet me that those things do not pan out, |
|
|
|
1:08:05.720 --> 1:08:07.000 |
|
I will bet you. |
|
|
|
1:08:07.000 --> 1:08:10.800 |
|
Even money, even money, I'll bet you as much as you want. |
|
|
|
1:08:10.800 --> 1:08:13.600 |
|
So are you worried about what's going to happen? |
|
|
|
1:08:13.600 --> 1:08:16.040 |
|
Because you're not in full agreement on that. |
|
|
|
1:08:16.040 --> 1:08:19.160 |
|
What's going to happen when 2022, 2021 come around |
|
|
|
1:08:19.160 --> 1:08:22.800 |
|
and nobody has fleets of autonomous vehicles? |
|
|
|
1:08:22.800 --> 1:08:25.000 |
|
Well, you can look at the history. |
|
|
|
1:08:25.000 --> 1:08:26.880 |
|
If you go back five years ago, they |
|
|
|
1:08:26.880 --> 1:08:29.880 |
|
were all promised by 2018 and 2017. |
|
|
|
1:08:29.880 --> 1:08:32.200 |
|
But they weren't that strong of promises. |
|
|
|
1:08:32.200 --> 1:08:36.240 |
|
I mean, Ford really declared. |
|
|
|
1:08:36.240 --> 1:08:40.560 |
|
I think not many have declared as definitively |
|
|
|
1:08:40.560 --> 1:08:42.600 |
|
as they have now these dates. |
|
|
|
1:08:42.600 --> 1:08:43.320 |
|
Well, OK. |
|
|
|
1:08:43.320 --> 1:08:45.040 |
|
So let's separate L4 and L5. |
|
|
|
1:08:45.040 --> 1:08:46.800 |
|
Do I think that it's possible for Waymo |
|
|
|
1:08:46.800 --> 1:08:50.960 |
|
to continue to hack on their system |
|
|
|
1:08:50.960 --> 1:08:53.400 |
|
until it gets to level four in Chandler, Arizona? |
|
|
|
1:08:53.400 --> 1:08:55.040 |
|
Yes. |
|
|
|
1:08:55.040 --> 1:08:56.800 |
|
No safety driver? |
|
|
|
1:08:56.800 --> 1:08:57.600 |
|
Chandler, Arizona? |
|
|
|
1:08:57.600 --> 1:08:59.600 |
|
Yeah. |
|
|
|
1:08:59.600 --> 1:09:02.440 |
|
By which year are we talking about? |
|
|
|
1:09:02.440 --> 1:09:06.120 |
|
Oh, I even think that's possible by like 2020, 2021. |
|
|
|
1:09:06.120 --> 1:09:09.480 |
|
But level four, Chandler, Arizona, not level five, |
|
|
|
1:09:09.480 --> 1:09:11.480 |
|
New York City. |
|
|
|
1:09:11.480 --> 1:09:15.920 |
|
Level four, meaning some very defined streets. |
|
|
|
1:09:15.920 --> 1:09:17.400 |
|
It works out really well. |
|
|
|
1:09:17.400 --> 1:09:18.280 |
|
Very defined streets. |
|
|
|
1:09:18.280 --> 1:09:20.680 |
|
And then practically, these streets are pretty empty. |
|
|
|
1:09:20.680 --> 1:09:24.680 |
|
If most of the streets are covered in Waymos, |
|
|
|
1:09:24.680 --> 1:09:28.360 |
|
Waymo can kind of change the definition of what driving is. |
|
|
|
1:09:28.360 --> 1:09:28.920 |
|
Right? |
|
|
|
1:09:28.920 --> 1:09:31.720 |
|
If your self driving network is the majority |
|
|
|
1:09:31.720 --> 1:09:34.120 |
|
of cars in an area, they only need |
|
|
|
1:09:34.120 --> 1:09:35.720 |
|
to be safe with respect to each other, |
|
|
|
1:09:35.720 --> 1:09:38.640 |
|
and all the humans will need to learn to adapt to them. |
|
|
|
1:09:38.640 --> 1:09:41.120 |
|
Now go drive in downtown New York. |
|
|
|
1:09:41.120 --> 1:09:42.200 |
|
Oh, yeah, that's. |
|
|
|
1:09:42.200 --> 1:09:43.440 |
|
I mean, already. |
|
|
|
1:09:43.440 --> 1:09:46.040 |
|
You can talk about autonomy and like on farms, |
|
|
|
1:09:46.040 --> 1:09:48.520 |
|
it already works great, because you can really just |
|
|
|
1:09:48.520 --> 1:09:51.320 |
|
follow the GPS line. |
|
|
|
1:09:51.320 --> 1:09:56.800 |
|
So what does success look like for Kama AI? |
|
|
|
1:09:56.800 --> 1:09:58.200 |
|
What are the milestones like where |
|
|
|
1:09:58.200 --> 1:09:59.800 |
|
you can sit back with some champagne |
|
|
|
1:09:59.800 --> 1:10:04.120 |
|
and say, we did it, boys and girls? |
|
|
|
1:10:04.120 --> 1:10:06.320 |
|
Well, it's never over. |
|
|
|
1:10:06.320 --> 1:10:07.800 |
|
Yeah, but don't be so. |
|
|
|
1:10:07.800 --> 1:10:10.400 |
|
You must drink champagne every time you celebrate. |
|
|
|
1:10:10.400 --> 1:10:11.440 |
|
So what is good? |
|
|
|
1:10:11.440 --> 1:10:13.160 |
|
What are some wins? |
|
|
|
1:10:13.160 --> 1:10:19.480 |
|
A big milestone that we're hoping for by mid next year |
|
|
|
1:10:19.480 --> 1:10:20.680 |
|
is profitability of the company. |
|
|
|
1:10:20.680 --> 1:10:28.560 |
|
And we're going to have to revisit the idea of selling |
|
|
|
1:10:28.560 --> 1:10:30.280 |
|
a consumer product. |
|
|
|
1:10:30.280 --> 1:10:32.720 |
|
But it's not going to be like the Kama One. |
|
|
|
1:10:32.720 --> 1:10:36.240 |
|
When we do it, it's going to be perfect. |
|
|
|
1:10:36.240 --> 1:10:39.600 |
|
OpenPilot has gotten so much better in the last two years. |
|
|
|
1:10:39.600 --> 1:10:41.680 |
|
We're going to have a few features. |
|
|
|
1:10:41.680 --> 1:10:43.760 |
|
We're going to have 100% driver monitoring. |
|
|
|
1:10:43.760 --> 1:10:46.720 |
|
We're going to disable no safety features in the car. |
|
|
|
1:10:46.720 --> 1:10:48.760 |
|
Actually, I think it'd be really cool what we're doing right |
|
|
|
1:10:48.760 --> 1:10:51.600 |
|
now, our project this week is we're analyzing the data set |
|
|
|
1:10:51.600 --> 1:10:53.240 |
|
and looking for all the AEB triggers |
|
|
|
1:10:53.240 --> 1:10:55.640 |
|
from the manufacturer systems. |
|
|
|
1:10:55.640 --> 1:10:59.440 |
|
We have better data set on that than the manufacturers. |
|
|
|
1:10:59.440 --> 1:11:02.960 |
|
How much does Toyota have 10 million miles of real world |
|
|
|
1:11:02.960 --> 1:11:05.360 |
|
driving to know how many times they're AEB triggered? |
|
|
|
1:11:05.360 --> 1:11:10.880 |
|
So let me give you, because you asked, financial advice. |
|
|
|
1:11:10.880 --> 1:11:12.440 |
|
Because I work with a lot of automakers |
|
|
|
1:11:12.440 --> 1:11:15.840 |
|
and one possible source of money for you, |
|
|
|
1:11:15.840 --> 1:11:21.400 |
|
which I'll be excited to see you take on, is basically |
|
|
|
1:11:21.400 --> 1:11:29.120 |
|
selling the data, which is something that most people, |
|
|
|
1:11:29.120 --> 1:11:31.800 |
|
and not selling in a way where here, here at Automaker, |
|
|
|
1:11:31.800 --> 1:11:33.000 |
|
but creating. |
|
|
|
1:11:33.000 --> 1:11:35.480 |
|
We've done this actually at MIT, not for money purposes, |
|
|
|
1:11:35.480 --> 1:11:37.760 |
|
but you could do it for significant money purposes |
|
|
|
1:11:37.760 --> 1:11:39.440 |
|
and make the world a better place |
|
|
|
1:11:39.440 --> 1:11:44.240 |
|
by creating a consortia where automakers would pay in |
|
|
|
1:11:44.240 --> 1:11:46.960 |
|
and then they get to have free access to the data. |
|
|
|
1:11:46.960 --> 1:11:52.400 |
|
And I think a lot of people are really hungry for that |
|
|
|
1:11:52.400 --> 1:11:54.200 |
|
and would pay significant amount of money for it. |
|
|
|
1:11:54.200 --> 1:11:55.400 |
|
Here's the problem with that. |
|
|
|
1:11:55.400 --> 1:11:56.840 |
|
I like this idea all in theory. |
|
|
|
1:11:56.840 --> 1:11:59.640 |
|
It'd be very easy for me to give them access to my servers. |
|
|
|
1:11:59.640 --> 1:12:02.280 |
|
And we already have all open source tools to access this data. |
|
|
|
1:12:02.280 --> 1:12:03.400 |
|
It's in a great format. |
|
|
|
1:12:03.400 --> 1:12:05.560 |
|
We have a great pipeline. |
|
|
|
1:12:05.560 --> 1:12:07.120 |
|
But they're going to put me in the room |
|
|
|
1:12:07.120 --> 1:12:10.120 |
|
with some business development guy. |
|
|
|
1:12:10.120 --> 1:12:12.400 |
|
And I'm going to have to talk to this guy. |
|
|
|
1:12:12.400 --> 1:12:15.040 |
|
And he's not going to know most of the words I'm saying. |
|
|
|
1:12:15.040 --> 1:12:17.280 |
|
I'm not willing to tolerate that. |
|
|
|
1:12:17.280 --> 1:12:18.840 |
|
OK, Mick Jagger. |
|
|
|
1:12:18.840 --> 1:12:19.800 |
|
No, no, no, no. |
|
|
|
1:12:19.800 --> 1:12:21.040 |
|
But I think I agree with you. |
|
|
|
1:12:21.040 --> 1:12:21.720 |
|
I'm the same way. |
|
|
|
1:12:21.720 --> 1:12:22.960 |
|
But you just tell them the terms |
|
|
|
1:12:22.960 --> 1:12:24.640 |
|
and there's no discussion needed. |
|
|
|
1:12:24.640 --> 1:12:30.480 |
|
If I could just tell them the terms, then like, all right. |
|
|
|
1:12:30.480 --> 1:12:31.600 |
|
Who wants access to my data? |
|
|
|
1:12:31.600 --> 1:12:36.680 |
|
I will sell it to you for, let's say, |
|
|
|
1:12:36.680 --> 1:12:37.640 |
|
you want a subscription? |
|
|
|
1:12:37.640 --> 1:12:40.680 |
|
I'll sell you for 100k a month. |
|
|
|
1:12:40.680 --> 1:12:41.200 |
|
Anyone? |
|
|
|
1:12:41.200 --> 1:12:42.000 |
|
100k a month? |
|
|
|
1:12:42.000 --> 1:12:43.040 |
|
100k a month? |
|
|
|
1:12:43.040 --> 1:12:45.080 |
|
I'll give you access to the data subscription? |
|
|
|
1:12:45.080 --> 1:12:45.680 |
|
Yeah. |
|
|
|
1:12:45.680 --> 1:12:46.680 |
|
Yeah, I think that's kind of fair. |
|
|
|
1:12:46.680 --> 1:12:48.440 |
|
Came up with that number off the top of my head. |
|
|
|
1:12:48.440 --> 1:12:50.840 |
|
If somebody sends me like a three line email where it's like, |
|
|
|
1:12:50.840 --> 1:12:54.000 |
|
we would like to pay 100k a month to get access to your data. |
|
|
|
1:12:54.000 --> 1:12:56.160 |
|
We would agree to like reasonable privacy terms |
|
|
|
1:12:56.160 --> 1:12:58.360 |
|
of the people who are in the data set. |
|
|
|
1:12:58.360 --> 1:12:59.520 |
|
I would be happy to do it. |
|
|
|
1:12:59.520 --> 1:13:01.200 |
|
But that's not going to be the email. |
|
|
|
1:13:01.200 --> 1:13:03.120 |
|
The email is going to be, hey, do you |
|
|
|
1:13:03.120 --> 1:13:05.560 |
|
have some time in the next month where we can sit down |
|
|
|
1:13:05.560 --> 1:13:07.000 |
|
and we can, I don't have time for that. |
|
|
|
1:13:07.000 --> 1:13:08.360 |
|
We're moving too fast. |
|
|
|
1:13:08.360 --> 1:13:10.040 |
|
You could politely respond to that email, |
|
|
|
1:13:10.040 --> 1:13:13.240 |
|
but not saying I don't have any time for your bullshit. |
|
|
|
1:13:13.240 --> 1:13:15.440 |
|
You say, oh, well, unfortunately, these are the terms. |
|
|
|
1:13:15.440 --> 1:13:19.280 |
|
And so this is what we try to, we brought the cost down |
|
|
|
1:13:19.280 --> 1:13:22.320 |
|
for you in order to minimize the friction, the communication. |
|
|
|
1:13:22.320 --> 1:13:22.920 |
|
Yeah, absolutely. |
|
|
|
1:13:22.920 --> 1:13:26.720 |
|
Here's the whatever it is, $1, $2 million a year. |
|
|
|
1:13:26.720 --> 1:13:28.880 |
|
And you have access. |
|
|
|
1:13:28.880 --> 1:13:31.440 |
|
And it's not like I get that email from like, |
|
|
|
1:13:31.440 --> 1:13:32.720 |
|
but OK, am I going to reach out? |
|
|
|
1:13:32.720 --> 1:13:34.200 |
|
Am I going to hire a business development person |
|
|
|
1:13:34.200 --> 1:13:35.840 |
|
who's going to reach out to the automakers? |
|
|
|
1:13:35.840 --> 1:13:36.480 |
|
No way. |
|
|
|
1:13:36.480 --> 1:13:36.880 |
|
Yeah. |
|
|
|
1:13:36.880 --> 1:13:37.840 |
|
OK, I got you. |
|
|
|
1:13:37.840 --> 1:13:38.520 |
|
I admire. |
|
|
|
1:13:38.520 --> 1:13:39.680 |
|
If they reached into me, I'm not |
|
|
|
1:13:39.680 --> 1:13:40.600 |
|
going to ignore the email. |
|
|
|
1:13:40.600 --> 1:13:42.160 |
|
I'll come back with something like, yeah, |
|
|
|
1:13:42.160 --> 1:13:44.560 |
|
if you're willing to pay $100,000 for access to the data, |
|
|
|
1:13:44.560 --> 1:13:46.080 |
|
I'm happy to set that up. |
|
|
|
1:13:46.080 --> 1:13:48.200 |
|
That's worth my engineering time. |
|
|
|
1:13:48.200 --> 1:13:49.520 |
|
That's actually quite insightful of you. |
|
|
|
1:13:49.520 --> 1:13:50.440 |
|
You're right. |
|
|
|
1:13:50.440 --> 1:13:52.480 |
|
Probably because many of the automakers |
|
|
|
1:13:52.480 --> 1:13:54.480 |
|
are quite a bit old school, there |
|
|
|
1:13:54.480 --> 1:13:56.200 |
|
will be a need to reach out. |
|
|
|
1:13:56.200 --> 1:13:58.440 |
|
And they want it, but there will need |
|
|
|
1:13:58.440 --> 1:13:59.800 |
|
to be some communication. |
|
|
|
1:13:59.800 --> 1:14:00.160 |
|
You're right. |
|
|
|
1:14:00.160 --> 1:14:06.760 |
|
Mobileye circa 2015 had the lowest R&D spend of any chipmaker. |
|
|
|
1:14:06.760 --> 1:14:10.640 |
|
Like per, and you look at all the people who work for them, |
|
|
|
1:14:10.640 --> 1:14:12.120 |
|
and it's all business development people |
|
|
|
1:14:12.120 --> 1:14:15.320 |
|
because the car companies are impossible to work with. |
|
|
|
1:14:15.320 --> 1:14:17.880 |
|
Yeah, so you have no patience for that, |
|
|
|
1:14:17.880 --> 1:14:20.040 |
|
and you're a legit Android, huh? |
|
|
|
1:14:20.040 --> 1:14:21.440 |
|
I have something to do, right? |
|
|
|
1:14:21.440 --> 1:14:24.040 |
|
Like, it's not like I don't mean to be a dick and say, |
|
|
|
1:14:24.040 --> 1:14:25.920 |
|
I don't have patience for that, but it's like, |
|
|
|
1:14:25.920 --> 1:14:29.160 |
|
that stuff doesn't help us with our goal of winning |
|
|
|
1:14:29.160 --> 1:14:30.560 |
|
self driving cars. |
|
|
|
1:14:30.560 --> 1:14:33.800 |
|
If I want money in the short term, |
|
|
|
1:14:33.800 --> 1:14:38.040 |
|
if I showed off the actual learning tech that we have, |
|
|
|
1:14:38.040 --> 1:14:40.160 |
|
it's somewhat sad. |
|
|
|
1:14:40.160 --> 1:14:43.000 |
|
It's years and years ahead of everybody else's. |
|
|
|
1:14:43.000 --> 1:14:43.720 |
|
Maybe not Tesla's. |
|
|
|
1:14:43.720 --> 1:14:45.720 |
|
I think Tesla has similar stuff to us, actually. |
|
|
|
1:14:45.720 --> 1:14:47.640 |
|
I think Tesla has similar stuff, but when you compare it |
|
|
|
1:14:47.640 --> 1:14:50.920 |
|
to what the Toyota Research Institute has, |
|
|
|
1:14:50.920 --> 1:14:53.480 |
|
you're not even close to what we have. |
|
|
|
1:14:53.480 --> 1:14:55.840 |
|
No comments, but I also can't. |
|
|
|
1:14:55.840 --> 1:14:58.440 |
|
I have to take your comments. |
|
|
|
1:14:58.440 --> 1:15:01.960 |
|
I intuitively believe you, but I have |
|
|
|
1:15:01.960 --> 1:15:04.680 |
|
to take it with a grain of salt because, |
|
|
|
1:15:04.680 --> 1:15:07.440 |
|
I mean, you are an inspiration because you basically |
|
|
|
1:15:07.440 --> 1:15:10.000 |
|
don't care about a lot of things that other companies care |
|
|
|
1:15:10.000 --> 1:15:10.880 |
|
about. |
|
|
|
1:15:10.880 --> 1:15:16.600 |
|
You don't try to bullshit, in a sense, like make up stuff, |
|
|
|
1:15:16.600 --> 1:15:18.600 |
|
so to drive up valuation. |
|
|
|
1:15:18.600 --> 1:15:19.960 |
|
You're really very real, and you're |
|
|
|
1:15:19.960 --> 1:15:22.280 |
|
trying to solve the problem, and I admire that a lot. |
|
|
|
1:15:22.280 --> 1:15:26.520 |
|
What I don't necessarily fully can't trust you on about your |
|
|
|
1:15:26.520 --> 1:15:28.440 |
|
respect is how good it is, right? |
|
|
|
1:15:28.440 --> 1:15:33.320 |
|
I can only, but I also know how bad others are. |
|
|
|
1:15:33.320 --> 1:15:36.680 |
|
I'll say two things about, trust, but verify, right? |
|
|
|
1:15:36.680 --> 1:15:38.040 |
|
I'll say two things about that. |
|
|
|
1:15:38.040 --> 1:15:42.360 |
|
One is try, get in a 2020 Corolla, |
|
|
|
1:15:42.360 --> 1:15:46.680 |
|
and try OpenPilot 0.6 when it comes out next month. |
|
|
|
1:15:46.680 --> 1:15:48.400 |
|
I think already, you'll look at this, |
|
|
|
1:15:48.400 --> 1:15:51.400 |
|
and you'll be like, this is already really good. |
|
|
|
1:15:51.400 --> 1:15:54.240 |
|
And then, I could be doing that all with hand labelers |
|
|
|
1:15:54.240 --> 1:15:58.000 |
|
and all with the same approach that Mobileye uses. |
|
|
|
1:15:58.000 --> 1:16:00.040 |
|
When we release a model that no longer |
|
|
|
1:16:00.040 --> 1:16:05.000 |
|
has the lanes in it, that only outputs a path, |
|
|
|
1:16:05.000 --> 1:16:08.720 |
|
then think about how we did that machine learning, |
|
|
|
1:16:08.720 --> 1:16:10.080 |
|
and then right away, when you see, |
|
|
|
1:16:10.080 --> 1:16:11.240 |
|
and that's going to be an OpenPilot, |
|
|
|
1:16:11.240 --> 1:16:13.000 |
|
that's going to be an OpenPilot before 1.0, |
|
|
|
1:16:13.000 --> 1:16:14.400 |
|
when you see that model, you'll know |
|
|
|
1:16:14.400 --> 1:16:15.360 |
|
that everything I'm saying is true, |
|
|
|
1:16:15.360 --> 1:16:16.840 |
|
because how else did I get that model? |
|
|
|
1:16:16.840 --> 1:16:17.320 |
|
Good. |
|
|
|
1:16:17.320 --> 1:16:19.240 |
|
You know what I'm saying is true about the simulator. |
|
|
|
1:16:19.240 --> 1:16:20.600 |
|
Yeah, yeah, yeah, this is super exciting. |
|
|
|
1:16:20.600 --> 1:16:22.680 |
|
That's super exciting. |
|
|
|
1:16:22.680 --> 1:16:25.760 |
|
But I listened to your talk with Kyle, |
|
|
|
1:16:25.760 --> 1:16:30.480 |
|
and Kyle was originally building the aftermarket system, |
|
|
|
1:16:30.480 --> 1:16:34.920 |
|
and he gave up on it because of technical challenges, |
|
|
|
1:16:34.920 --> 1:16:37.360 |
|
because of the fact that he's going |
|
|
|
1:16:37.360 --> 1:16:39.160 |
|
to have to support 20 to 50 cars. |
|
|
|
1:16:39.160 --> 1:16:41.120 |
|
We support 45, because what is he |
|
|
|
1:16:41.120 --> 1:16:43.440 |
|
going to do when the manufacturer ABS system triggers? |
|
|
|
1:16:43.440 --> 1:16:45.480 |
|
We have alerts and warnings to deal with all of that |
|
|
|
1:16:45.480 --> 1:16:48.400 |
|
and all the cars, and how is he going to formally verify it? |
|
|
|
1:16:48.400 --> 1:16:49.800 |
|
Well, I got 10 million miles of data. |
|
|
|
1:16:49.800 --> 1:16:53.240 |
|
It's probably better verified than the spec. |
|
|
|
1:16:53.240 --> 1:16:57.720 |
|
Yeah, I'm glad you're here talking to me. |
|
|
|
1:16:57.720 --> 1:17:01.120 |
|
I'll remember this day, because it's interesting. |
|
|
|
1:17:01.120 --> 1:17:04.160 |
|
If you look at Kyle's from Cruise, |
|
|
|
1:17:04.160 --> 1:17:06.320 |
|
I'm sure they have a large number of business development |
|
|
|
1:17:06.320 --> 1:17:10.200 |
|
folks, and he's working with GM. |
|
|
|
1:17:10.200 --> 1:17:13.280 |
|
He could work with Argo AI, worked with Ford. |
|
|
|
1:17:13.280 --> 1:17:18.520 |
|
It's interesting, because chances that you fail businesswise, |
|
|
|
1:17:18.520 --> 1:17:21.120 |
|
like bankrupt, are pretty high. |
|
|
|
1:17:21.120 --> 1:17:23.880 |
|
And yet, it's the Android model, |
|
|
|
1:17:23.880 --> 1:17:26.440 |
|
is you're actually taking on the problem. |
|
|
|
1:17:26.440 --> 1:17:28.160 |
|
So that's really inspiring. |
|
|
|
1:17:28.160 --> 1:17:30.920 |
|
Well, I have a long term way for comedy to make money, too. |
|
|
|
1:17:30.920 --> 1:17:34.400 |
|
And one of the nice things when you really take on the problem, |
|
|
|
1:17:34.400 --> 1:17:36.760 |
|
which is my hope for autopilot, for example, |
|
|
|
1:17:36.760 --> 1:17:41.040 |
|
is things you don't expect, ways to make money, |
|
|
|
1:17:41.040 --> 1:17:44.160 |
|
or create value that you don't expect will pop up. |
|
|
|
1:17:44.160 --> 1:17:48.560 |
|
I've known how to do it since 2017 is the first time I said it. |
|
|
|
1:17:48.560 --> 1:17:50.440 |
|
Which part to know how to do which part? |
|
|
|
1:17:50.440 --> 1:17:52.520 |
|
Our long term plan is to be a car insurance company. |
|
|
|
1:17:52.520 --> 1:17:53.160 |
|
Insurance. |
|
|
|
1:17:53.160 --> 1:17:55.320 |
|
Yeah, I love it. |
|
|
|
1:17:55.320 --> 1:17:56.680 |
|
I make driving twice as safe. |
|
|
|
1:17:56.680 --> 1:17:57.680 |
|
Not only that, I have the best data |
|
|
|
1:17:57.680 --> 1:18:00.040 |
|
such to know who statistically is the safest drivers. |
|
|
|
1:18:00.040 --> 1:18:02.160 |
|
And oh, oh, we see you. |
|
|
|
1:18:02.160 --> 1:18:03.720 |
|
We see you driving unsafely. |
|
|
|
1:18:03.720 --> 1:18:05.360 |
|
We're not going to insure you. |
|
|
|
1:18:05.360 --> 1:18:08.960 |
|
And that causes a bifurcation in the market, |
|
|
|
1:18:08.960 --> 1:18:10.920 |
|
because the only people who can't get common insurance |
|
|
|
1:18:10.920 --> 1:18:12.760 |
|
or the bad drivers, Geico can insure them. |
|
|
|
1:18:12.760 --> 1:18:15.360 |
|
Their premiums are crazy high, our premiums are crazy low. |
|
|
|
1:18:15.360 --> 1:18:16.240 |
|
We win car insurance. |
|
|
|
1:18:16.240 --> 1:18:18.120 |
|
Take over that whole market. |
|
|
|
1:18:18.120 --> 1:18:21.560 |
|
OK, so if we win, if we win, but that's |
|
|
|
1:18:21.560 --> 1:18:23.800 |
|
I'm saying like how do you turn comma into a $10 billion |
|
|
|
1:18:23.800 --> 1:18:24.640 |
|
company is that. |
|
|
|
1:18:24.640 --> 1:18:25.600 |
|
That's right. |
|
|
|
1:18:25.600 --> 1:18:30.000 |
|
So you Elon Musk, who else? |
|
|
|
1:18:30.000 --> 1:18:32.720 |
|
Who else is thinking like this and working like this |
|
|
|
1:18:32.720 --> 1:18:33.160 |
|
in your view? |
|
|
|
1:18:33.160 --> 1:18:34.800 |
|
Who are the competitors? |
|
|
|
1:18:34.800 --> 1:18:36.160 |
|
Are there people seriously? |
|
|
|
1:18:36.160 --> 1:18:39.480 |
|
I don't think anyone that I'm aware of is seriously |
|
|
|
1:18:39.480 --> 1:18:45.280 |
|
taking on lane keeping, like to where it's a huge business that |
|
|
|
1:18:45.280 --> 1:18:51.400 |
|
turns eventually to full autonomy that then creates |
|
|
|
1:18:51.400 --> 1:18:53.440 |
|
other businesses on top of it and so on. |
|
|
|
1:18:53.440 --> 1:18:56.480 |
|
Thinks insurance, thinks all kinds of ideas like that. |
|
|
|
1:18:56.480 --> 1:19:00.480 |
|
Do you know anyone else thinking like this? |
|
|
|
1:19:00.480 --> 1:19:02.200 |
|
Not really. |
|
|
|
1:19:02.200 --> 1:19:02.960 |
|
That's interesting. |
|
|
|
1:19:02.960 --> 1:19:06.560 |
|
I mean, my sense is everybody turns to that in like four |
|
|
|
1:19:06.560 --> 1:19:07.800 |
|
or five years. |
|
|
|
1:19:07.800 --> 1:19:11.240 |
|
Like Ford, once the autonomy doesn't fall through. |
|
|
|
1:19:11.240 --> 1:19:12.600 |
|
But at this time. |
|
|
|
1:19:12.600 --> 1:19:14.120 |
|
Elon's the iOS. |
|
|
|
1:19:14.120 --> 1:19:16.720 |
|
By the way, he paved the way for all of us. |
|
|
|
1:19:16.720 --> 1:19:18.000 |
|
It's not iOS, true. |
|
|
|
1:19:18.000 --> 1:19:21.520 |
|
I would not be doing comma AI today if it was not |
|
|
|
1:19:21.520 --> 1:19:23.480 |
|
for those conversations with Elon. |
|
|
|
1:19:23.480 --> 1:19:26.840 |
|
And if it were not for him saying like, |
|
|
|
1:19:26.840 --> 1:19:28.600 |
|
I think he said like, well, obviously we're not |
|
|
|
1:19:28.600 --> 1:19:31.280 |
|
going to use LiDAR, we use cameras, humans use cameras. |
|
|
|
1:19:31.280 --> 1:19:32.600 |
|
So what do you think about that? |
|
|
|
1:19:32.600 --> 1:19:33.880 |
|
How important is LiDAR? |
|
|
|
1:19:33.880 --> 1:19:36.960 |
|
Everybody else's on L5 is using LiDAR. |
|
|
|
1:19:36.960 --> 1:19:39.160 |
|
What are your thoughts on his provocative statement |
|
|
|
1:19:39.160 --> 1:19:41.320 |
|
that LiDAR is a crutch? |
|
|
|
1:19:41.320 --> 1:19:43.520 |
|
See, sometimes they'll say dumb things like the driver |
|
|
|
1:19:43.520 --> 1:19:45.680 |
|
monitoring thing, but sometimes they'll say absolutely |
|
|
|
1:19:45.680 --> 1:19:48.400 |
|
completely 100% obviously true things. |
|
|
|
1:19:48.400 --> 1:19:50.840 |
|
Of course LiDAR is a crutch. |
|
|
|
1:19:50.840 --> 1:19:53.040 |
|
It's not even a good crutch. |
|
|
|
1:19:53.040 --> 1:19:54.200 |
|
You're not even using it. |
|
|
|
1:19:54.200 --> 1:19:56.920 |
|
They're using it for localization, |
|
|
|
1:19:56.920 --> 1:19:58.160 |
|
which isn't good in the first place. |
|
|
|
1:19:58.160 --> 1:20:00.480 |
|
If you have to localize your car to centimeters |
|
|
|
1:20:00.480 --> 1:20:04.280 |
|
in order to drive, that's not driving. |
|
|
|
1:20:04.280 --> 1:20:06.320 |
|
Currently not doing much machine learning. |
|
|
|
1:20:06.320 --> 1:20:09.280 |
|
I thought LiDAR data, meaning like to help you |
|
|
|
1:20:09.280 --> 1:20:12.840 |
|
in the task of general task of perception. |
|
|
|
1:20:12.840 --> 1:20:15.320 |
|
The main goal of those LiDARs on those cars |
|
|
|
1:20:15.320 --> 1:20:18.840 |
|
I think is actually localization more than perception, |
|
|
|
1:20:18.840 --> 1:20:20.080 |
|
or at least that's what they use them for. |
|
|
|
1:20:20.080 --> 1:20:20.920 |
|
Yeah, that's true. |
|
|
|
1:20:20.920 --> 1:20:22.480 |
|
If you want to localize to centimeters, |
|
|
|
1:20:22.480 --> 1:20:23.720 |
|
you can't use GPS. |
|
|
|
1:20:23.720 --> 1:20:25.120 |
|
The fancies GPS in the world can't do it, |
|
|
|
1:20:25.120 --> 1:20:26.960 |
|
especially if you're under tree cover and stuff. |
|
|
|
1:20:26.960 --> 1:20:28.480 |
|
LiDAR you can do this pretty easily. |
|
|
|
1:20:28.480 --> 1:20:30.240 |
|
So really they're not taking on, |
|
|
|
1:20:30.240 --> 1:20:33.200 |
|
I mean in some research they're using it for perception, |
|
|
|
1:20:33.200 --> 1:20:35.840 |
|
but and they're certainly not, which is sad, |
|
|
|
1:20:35.840 --> 1:20:38.680 |
|
they're not fusing it well with vision. |
|
|
|
1:20:38.680 --> 1:20:40.560 |
|
They do use it for perception. |
|
|
|
1:20:40.560 --> 1:20:42.400 |
|
I'm not saying they don't use it for perception, |
|
|
|
1:20:42.400 --> 1:20:45.480 |
|
but the thing that they have vision based |
|
|
|
1:20:45.480 --> 1:20:47.680 |
|
and radar based perception systems as well. |
|
|
|
1:20:47.680 --> 1:20:51.440 |
|
You could remove the LiDAR and keep around |
|
|
|
1:20:51.440 --> 1:20:54.040 |
|
a lot of the dynamic object perception. |
|
|
|
1:20:54.040 --> 1:20:56.320 |
|
You want to get centimeter accurate localization. |
|
|
|
1:20:56.320 --> 1:20:59.120 |
|
Good luck doing that with anything else. |
|
|
|
1:20:59.120 --> 1:21:02.880 |
|
So what should a cruise Waymo do? |
|
|
|
1:21:02.880 --> 1:21:05.360 |
|
Like what would be your advice to them now? |
|
|
|
1:21:06.400 --> 1:21:11.400 |
|
I mean Waymo is actually, they're serious. |
|
|
|
1:21:11.400 --> 1:21:13.120 |
|
Waymo out of the ball of them, |
|
|
|
1:21:13.120 --> 1:21:16.120 |
|
are quite serious about the long game. |
|
|
|
1:21:16.120 --> 1:21:20.680 |
|
If L5 is a lot, is requires 50 years, |
|
|
|
1:21:20.680 --> 1:21:24.000 |
|
I think Waymo will be the only one left standing at the end |
|
|
|
1:21:24.000 --> 1:21:26.560 |
|
with a given the financial backing that they have. |
|
|
|
1:21:26.560 --> 1:21:28.640 |
|
They're boo Google box. |
|
|
|
1:21:28.640 --> 1:21:31.040 |
|
I'll say nice things about both Waymo and cruise. |
|
|
|
1:21:32.320 --> 1:21:33.480 |
|
Let's do it. |
|
|
|
1:21:33.480 --> 1:21:34.320 |
|
Nice is good. |
|
|
|
1:21:35.720 --> 1:21:39.200 |
|
Waymo is by far the furthest along with technology. |
|
|
|
1:21:39.200 --> 1:21:41.160 |
|
Waymo has a three to five year lead |
|
|
|
1:21:41.160 --> 1:21:42.880 |
|
on all the competitors. |
|
|
|
1:21:43.960 --> 1:21:48.640 |
|
If the Waymo looking stack works, |
|
|
|
1:21:48.640 --> 1:21:49.720 |
|
maybe three year lead. |
|
|
|
1:21:49.720 --> 1:21:51.280 |
|
If the Waymo looking stack works, |
|
|
|
1:21:51.280 --> 1:21:52.800 |
|
they have a three year lead. |
|
|
|
1:21:52.800 --> 1:21:55.800 |
|
Now, I argue that Waymo has spent too much money |
|
|
|
1:21:55.800 --> 1:21:59.240 |
|
to recapitalize, to gain back their losses |
|
|
|
1:21:59.240 --> 1:22:00.160 |
|
in those three years. |
|
|
|
1:22:00.160 --> 1:22:03.600 |
|
Also self driving cars have no network effect like that. |
|
|
|
1:22:03.600 --> 1:22:04.800 |
|
Uber has a network effect. |
|
|
|
1:22:04.800 --> 1:22:07.120 |
|
You have a market, you have drivers and you have riders. |
|
|
|
1:22:07.120 --> 1:22:09.880 |
|
Self driving cars, you have capital and you have riders. |
|
|
|
1:22:09.880 --> 1:22:11.400 |
|
There's no network effect. |
|
|
|
1:22:11.400 --> 1:22:13.800 |
|
If I want to blanket a new city in self driving cars, |
|
|
|
1:22:13.800 --> 1:22:16.000 |
|
I buy the off the shelf Chinese knockoff self driving cars |
|
|
|
1:22:16.000 --> 1:22:17.160 |
|
and I buy enough of them in the city. |
|
|
|
1:22:17.160 --> 1:22:18.360 |
|
I can't do that with drivers. |
|
|
|
1:22:18.360 --> 1:22:20.840 |
|
And that's why Uber has a first mover advantage |
|
|
|
1:22:20.840 --> 1:22:22.640 |
|
that no self driving car company will. |
|
|
|
1:22:23.960 --> 1:22:26.520 |
|
Can you just a thing, let a little bit. |
|
|
|
1:22:26.520 --> 1:22:28.160 |
|
Uber, you're not talking about Uber, |
|
|
|
1:22:28.160 --> 1:22:29.240 |
|
the autonomous vehicle Uber. |
|
|
|
1:22:29.240 --> 1:22:30.960 |
|
You're talking about the Uber cars. |
|
|
|
1:22:30.960 --> 1:22:31.800 |
|
Yeah. |
|
|
|
1:22:31.800 --> 1:22:32.640 |
|
I'm Uber. |
|
|
|
1:22:32.640 --> 1:22:35.920 |
|
I open for business in Austin, Texas, let's say. |
|
|
|
1:22:35.920 --> 1:22:38.760 |
|
I need to attract both sides of the market. |
|
|
|
1:22:38.760 --> 1:22:41.200 |
|
I need to both get drivers on my platform |
|
|
|
1:22:41.200 --> 1:22:42.720 |
|
and riders on my platform. |
|
|
|
1:22:42.720 --> 1:22:45.320 |
|
And I need to keep them both sufficiently happy, right? |
|
|
|
1:22:45.320 --> 1:22:46.520 |
|
Riders aren't going to use it |
|
|
|
1:22:46.520 --> 1:22:48.960 |
|
if it takes more than five minutes for an Uber to show up. |
|
|
|
1:22:48.960 --> 1:22:50.120 |
|
Drivers aren't going to use it |
|
|
|
1:22:50.120 --> 1:22:52.120 |
|
if they have to sit around all day and there's no riders. |
|
|
|
1:22:52.120 --> 1:22:54.480 |
|
So you have to carefully balance a market. |
|
|
|
1:22:54.480 --> 1:22:56.240 |
|
And whenever you have to carefully balance a market, |
|
|
|
1:22:56.240 --> 1:22:58.280 |
|
there's a great first mover advantage |
|
|
|
1:22:58.280 --> 1:23:01.000 |
|
because there's a switching cost for everybody, right? |
|
|
|
1:23:01.000 --> 1:23:02.120 |
|
The drivers and the riders |
|
|
|
1:23:02.120 --> 1:23:04.080 |
|
would have to switch at the same time. |
|
|
|
1:23:04.080 --> 1:23:08.880 |
|
Let's even say that, let's say, Uber shows up. |
|
|
|
1:23:08.880 --> 1:23:13.880 |
|
And Uber somehow agrees to do things at a bigger, |
|
|
|
1:23:14.800 --> 1:23:17.440 |
|
we've done it more efficiently, right? |
|
|
|
1:23:17.440 --> 1:23:19.800 |
|
Uber only takes 5% of a car |
|
|
|
1:23:19.800 --> 1:23:21.600 |
|
instead of the 10% that Uber takes. |
|
|
|
1:23:21.600 --> 1:23:22.760 |
|
No one is going to switch |
|
|
|
1:23:22.760 --> 1:23:24.920 |
|
because the switching cost is higher than that 5%. |
|
|
|
1:23:24.920 --> 1:23:27.200 |
|
So you actually can, in markets like that, |
|
|
|
1:23:27.200 --> 1:23:28.520 |
|
you have a first mover advantage. |
|
|
|
1:23:28.520 --> 1:23:29.360 |
|
Yeah. |
|
|
|
1:23:30.160 --> 1:23:32.720 |
|
Autonomous vehicles of the level five variety |
|
|
|
1:23:32.720 --> 1:23:34.560 |
|
have no first mover advantage. |
|
|
|
1:23:34.560 --> 1:23:36.800 |
|
If the technology becomes commoditized, |
|
|
|
1:23:36.800 --> 1:23:39.520 |
|
say I want to go to a new city, look at the scooters. |
|
|
|
1:23:39.520 --> 1:23:41.480 |
|
It's going to look a lot more like scooters. |
|
|
|
1:23:41.480 --> 1:23:44.040 |
|
Every person with a checkbook |
|
|
|
1:23:44.040 --> 1:23:45.720 |
|
can blanket a city in scooters |
|
|
|
1:23:45.720 --> 1:23:47.920 |
|
and that's why you have 10 different scooter companies. |
|
|
|
1:23:47.920 --> 1:23:48.760 |
|
Which one's going to win? |
|
|
|
1:23:48.760 --> 1:23:49.600 |
|
It's a race to the bottom. |
|
|
|
1:23:49.600 --> 1:23:51.040 |
|
It's a terrible market to be in |
|
|
|
1:23:51.040 --> 1:23:53.160 |
|
because there's no market for scooters. |
|
|
|
1:23:54.960 --> 1:23:56.520 |
|
And the scooters don't get a say |
|
|
|
1:23:56.520 --> 1:23:57.480 |
|
in whether they want to be bought |
|
|
|
1:23:57.480 --> 1:23:58.440 |
|
and deployed to a city or not. |
|
|
|
1:23:58.440 --> 1:23:59.280 |
|
Right. |
|
|
|
1:23:59.280 --> 1:24:00.120 |
|
So yeah. |
|
|
|
1:24:00.120 --> 1:24:02.080 |
|
We're going to entice the scooters with subsidies |
|
|
|
1:24:02.080 --> 1:24:02.920 |
|
and deals. |
|
|
|
1:24:03.840 --> 1:24:05.480 |
|
So whenever you have to invest that capital, |
|
|
|
1:24:05.480 --> 1:24:06.720 |
|
it doesn't... |
|
|
|
1:24:06.720 --> 1:24:07.560 |
|
It doesn't come back. |
|
|
|
1:24:07.560 --> 1:24:08.600 |
|
Yeah. |
|
|
|
1:24:08.600 --> 1:24:12.320 |
|
They can't be your main criticism of the Waymo approach. |
|
|
|
1:24:12.320 --> 1:24:14.840 |
|
Oh, I'm saying even if it does technically work. |
|
|
|
1:24:14.840 --> 1:24:17.040 |
|
Even if it does technically work, that's a problem. |
|
|
|
1:24:17.040 --> 1:24:18.000 |
|
Yeah. |
|
|
|
1:24:18.000 --> 1:24:21.720 |
|
I don't know if I were to say, I would say, |
|
|
|
1:24:22.840 --> 1:24:23.520 |
|
you're already there. |
|
|
|
1:24:23.520 --> 1:24:24.560 |
|
I haven't even thought about that. |
|
|
|
1:24:24.560 --> 1:24:26.520 |
|
But I would say the bigger challenge |
|
|
|
1:24:26.520 --> 1:24:27.760 |
|
is the technical approach. |
|
|
|
1:24:29.760 --> 1:24:31.840 |
|
So Waymo's cruise is... |
|
|
|
1:24:31.840 --> 1:24:33.000 |
|
And not just the technical approach, |
|
|
|
1:24:33.000 --> 1:24:34.800 |
|
but of creating value. |
|
|
|
1:24:34.800 --> 1:24:39.800 |
|
I still don't understand how you beat Uber, |
|
|
|
1:24:40.760 --> 1:24:43.480 |
|
the human driven cars. |
|
|
|
1:24:43.480 --> 1:24:44.920 |
|
In terms of financially, |
|
|
|
1:24:44.920 --> 1:24:47.160 |
|
it doesn't make sense to me |
|
|
|
1:24:47.160 --> 1:24:50.080 |
|
that people want to get an autonomous vehicle. |
|
|
|
1:24:50.080 --> 1:24:52.800 |
|
I don't understand how you make money. |
|
|
|
1:24:52.800 --> 1:24:56.440 |
|
In the long term, yes, like real long term, |
|
|
|
1:24:56.440 --> 1:24:58.640 |
|
but it just feels like there's too much |
|
|
|
1:24:58.640 --> 1:24:59.960 |
|
capital investment needed. |
|
|
|
1:24:59.960 --> 1:25:01.200 |
|
Oh, and they're going to be worse than Ubers |
|
|
|
1:25:01.200 --> 1:25:02.440 |
|
because they're going to stop |
|
|
|
1:25:02.440 --> 1:25:04.760 |
|
for every little thing everywhere. |
|
|
|
1:25:06.320 --> 1:25:07.360 |
|
I'll say a nice thing about cruise. |
|
|
|
1:25:07.360 --> 1:25:08.440 |
|
That was my nice thing about Waymo. |
|
|
|
1:25:08.440 --> 1:25:09.280 |
|
They're three years ahead of me. |
|
|
|
1:25:09.280 --> 1:25:10.120 |
|
It was a nice... |
|
|
|
1:25:10.120 --> 1:25:10.960 |
|
Oh, because they're three years. |
|
|
|
1:25:10.960 --> 1:25:12.480 |
|
They're three years technically ahead of everybody. |
|
|
|
1:25:12.480 --> 1:25:13.960 |
|
Their tech stack is great. |
|
|
|
1:25:14.800 --> 1:25:17.920 |
|
My nice thing about cruise is GM buying them |
|
|
|
1:25:17.920 --> 1:25:19.160 |
|
was a great move for GM. |
|
|
|
1:25:20.600 --> 1:25:22.240 |
|
For $1 billion, |
|
|
|
1:25:22.240 --> 1:25:25.600 |
|
GM bought an insurance policy against Waymo. |
|
|
|
1:25:26.560 --> 1:25:30.000 |
|
They put cruise is three years behind Waymo. |
|
|
|
1:25:30.000 --> 1:25:32.600 |
|
That means Google will get a monopoly |
|
|
|
1:25:32.600 --> 1:25:35.160 |
|
on the technology for at most three years. |
|
|
|
1:25:36.840 --> 1:25:38.880 |
|
And if technology works, |
|
|
|
1:25:38.880 --> 1:25:40.840 |
|
you might not even be right about the three years. |
|
|
|
1:25:40.840 --> 1:25:41.840 |
|
It might be less. |
|
|
|
1:25:41.840 --> 1:25:42.680 |
|
Might be less. |
|
|
|
1:25:42.680 --> 1:25:44.320 |
|
Cruise actually might not be that far behind. |
|
|
|
1:25:44.320 --> 1:25:47.360 |
|
I don't know how much Waymo has waffled around |
|
|
|
1:25:47.360 --> 1:25:49.760 |
|
or how much of it actually is just that long tail. |
|
|
|
1:25:49.760 --> 1:25:50.600 |
|
Yeah, okay. |
|
|
|
1:25:50.600 --> 1:25:53.600 |
|
If that's the best you could say in terms of nice things, |
|
|
|
1:25:53.600 --> 1:25:55.200 |
|
that's more of a nice thing for GM |
|
|
|
1:25:55.200 --> 1:25:58.560 |
|
that that's a smart insurance policy. |
|
|
|
1:25:58.560 --> 1:25:59.680 |
|
It's a smart insurance policy. |
|
|
|
1:25:59.680 --> 1:26:01.880 |
|
I mean, I think that's how... |
|
|
|
1:26:01.880 --> 1:26:05.200 |
|
I can't see cruise working out any other. |
|
|
|
1:26:05.200 --> 1:26:07.840 |
|
For cruise to leapfrog Waymo would really surprise me. |
|
|
|
1:26:10.400 --> 1:26:13.000 |
|
Yeah, so let's talk about the underlying assumptions |
|
|
|
1:26:13.000 --> 1:26:13.840 |
|
of everything is... |
|
|
|
1:26:13.840 --> 1:26:15.440 |
|
We're not gonna leapfrog Tesla. |
|
|
|
1:26:17.560 --> 1:26:19.240 |
|
Tesla would have to seriously mess up |
|
|
|
1:26:19.240 --> 1:26:20.440 |
|
for us to leapfrog them. |
|
|
|
1:26:20.440 --> 1:26:23.240 |
|
Okay, so the way you leapfrog, right, |
|
|
|
1:26:23.240 --> 1:26:26.120 |
|
is you come up with an idea |
|
|
|
1:26:26.120 --> 1:26:28.560 |
|
or you take a direction, perhaps secretly, |
|
|
|
1:26:28.560 --> 1:26:30.640 |
|
that the other people aren't taking. |
|
|
|
1:26:31.640 --> 1:26:36.640 |
|
And so cruise, Waymo, even Aurora... |
|
|
|
1:26:38.080 --> 1:26:40.080 |
|
I don't know, Aurora, Zooks is the same stack as well. |
|
|
|
1:26:40.080 --> 1:26:41.720 |
|
They're all the same code base even. |
|
|
|
1:26:41.720 --> 1:26:44.120 |
|
They're all the same DARPA Urban Challenge code base. |
|
|
|
1:26:44.120 --> 1:26:45.360 |
|
It's... |
|
|
|
1:26:45.360 --> 1:26:47.760 |
|
So the question is, do you think there's a room |
|
|
|
1:26:47.760 --> 1:26:49.120 |
|
for brilliance and innovation there |
|
|
|
1:26:49.120 --> 1:26:50.560 |
|
that will change everything? |
|
|
|
1:26:51.560 --> 1:26:53.880 |
|
Like say, okay, so I'll give you examples. |
|
|
|
1:26:53.880 --> 1:26:58.880 |
|
It could be if revolution and mapping, for example, |
|
|
|
1:26:59.640 --> 1:27:03.040 |
|
that allow you to map things, |
|
|
|
1:27:03.040 --> 1:27:05.840 |
|
do HD maps of the whole world, |
|
|
|
1:27:05.840 --> 1:27:08.080 |
|
all weather conditions somehow really well, |
|
|
|
1:27:08.080 --> 1:27:13.080 |
|
or revolution and simulation, |
|
|
|
1:27:14.480 --> 1:27:18.840 |
|
to where all the way you said before becomes incorrect. |
|
|
|
1:27:20.480 --> 1:27:21.520 |
|
That kind of thing. |
|
|
|
1:27:21.520 --> 1:27:23.920 |
|
Any room for breakthrough innovation? |
|
|
|
1:27:24.920 --> 1:27:25.960 |
|
What I said before about, |
|
|
|
1:27:25.960 --> 1:27:28.280 |
|
oh, they actually get the whole thing, well, |
|
|
|
1:27:28.280 --> 1:27:32.600 |
|
I'll say this about we divide driving into three problems. |
|
|
|
1:27:32.600 --> 1:27:33.800 |
|
And I actually haven't solved the third yet, |
|
|
|
1:27:33.800 --> 1:27:34.800 |
|
but I haven't had any idea how to do it. |
|
|
|
1:27:34.800 --> 1:27:36.120 |
|
So there's the static. |
|
|
|
1:27:36.120 --> 1:27:38.000 |
|
The static driving problem is assuming |
|
|
|
1:27:38.000 --> 1:27:40.120 |
|
you are the only car on the road, right? |
|
|
|
1:27:40.120 --> 1:27:42.000 |
|
And this problem can be solved 100% |
|
|
|
1:27:42.000 --> 1:27:44.000 |
|
with mapping and localization. |
|
|
|
1:27:44.000 --> 1:27:45.760 |
|
This is why farms work the way they do. |
|
|
|
1:27:45.760 --> 1:27:48.440 |
|
If all you have to deal with is the static problem, |
|
|
|
1:27:48.440 --> 1:27:50.160 |
|
and you can statically schedule your machines, right? |
|
|
|
1:27:50.160 --> 1:27:52.680 |
|
It's the same as like statically scheduling processes. |
|
|
|
1:27:52.680 --> 1:27:54.040 |
|
You can statically schedule your tractors |
|
|
|
1:27:54.040 --> 1:27:56.160 |
|
to never hit each other on their paths, right? |
|
|
|
1:27:56.160 --> 1:27:57.520 |
|
Because then you know the speed they go at. |
|
|
|
1:27:57.520 --> 1:28:00.160 |
|
So that's the static driving problem. |
|
|
|
1:28:00.160 --> 1:28:03.160 |
|
Maps only helps you with the static driving problem. |
|
|
|
1:28:03.920 --> 1:28:06.960 |
|
Yeah, the question about static driving, |
|
|
|
1:28:06.960 --> 1:28:08.800 |
|
you've just made it sound like it's really easy. |
|
|
|
1:28:08.800 --> 1:28:10.160 |
|
Static driving is really easy. |
|
|
|
1:28:11.880 --> 1:28:13.040 |
|
How easy? |
|
|
|
1:28:13.040 --> 1:28:16.480 |
|
How, well, because the whole drifting out of lane, |
|
|
|
1:28:16.480 --> 1:28:18.760 |
|
when Tesla drifts out of lane, |
|
|
|
1:28:18.760 --> 1:28:21.960 |
|
it's failing on the fundamental static driving problem. |
|
|
|
1:28:21.960 --> 1:28:24.440 |
|
Tesla is drifting out of lane? |
|
|
|
1:28:24.440 --> 1:28:27.720 |
|
The static driving problem is not easy for the world. |
|
|
|
1:28:27.720 --> 1:28:30.320 |
|
The static driving problem is easy for one route. |
|
|
|
1:28:31.840 --> 1:28:33.920 |
|
One route in one weather condition |
|
|
|
1:28:33.920 --> 1:28:37.920 |
|
with one state of lane markings |
|
|
|
1:28:37.920 --> 1:28:40.920 |
|
and like no deterioration, no cracks in the road. |
|
|
|
1:28:40.920 --> 1:28:42.600 |
|
Well, I'm assuming you have a perfect localizer. |
|
|
|
1:28:42.600 --> 1:28:44.200 |
|
So that's all for the weather condition |
|
|
|
1:28:44.200 --> 1:28:45.560 |
|
and the lane marking condition. |
|
|
|
1:28:45.560 --> 1:28:46.640 |
|
But that's the problem. |
|
|
|
1:28:46.640 --> 1:28:47.680 |
|
How do you have a perfect localizer? |
|
|
|
1:28:47.680 --> 1:28:50.560 |
|
You can build, perfect localizers are not that hard to build. |
|
|
|
1:28:50.560 --> 1:28:53.360 |
|
Okay, come on now, with LIDAR. |
|
|
|
1:28:53.360 --> 1:28:54.200 |
|
LIDAR, yeah. |
|
|
|
1:28:54.200 --> 1:28:55.040 |
|
With LIDAR, okay. |
|
|
|
1:28:55.040 --> 1:28:56.440 |
|
LIDAR, yeah, but you use LIDAR, right? |
|
|
|
1:28:56.440 --> 1:28:58.640 |
|
Like you use LIDAR, build a perfect localizer. |
|
|
|
1:28:58.640 --> 1:29:00.960 |
|
Building a perfect localizer without LIDAR, |
|
|
|
1:29:03.000 --> 1:29:04.320 |
|
it's gonna be hard. |
|
|
|
1:29:04.320 --> 1:29:05.760 |
|
You can get 10 centimeters without LIDAR, |
|
|
|
1:29:05.760 --> 1:29:07.240 |
|
you can get one centimeter with LIDAR. |
|
|
|
1:29:07.240 --> 1:29:09.280 |
|
I'm not even concerned about the one or 10 centimeters. |
|
|
|
1:29:09.280 --> 1:29:12.680 |
|
I'm concerned if every once in a while you just weigh off. |
|
|
|
1:29:12.680 --> 1:29:17.480 |
|
Yeah, so this is why you have to carefully |
|
|
|
1:29:17.480 --> 1:29:20.040 |
|
make sure you're always tracking your position. |
|
|
|
1:29:20.040 --> 1:29:21.760 |
|
You wanna use LIDAR camera fusion, |
|
|
|
1:29:21.760 --> 1:29:24.480 |
|
but you can get the reliability of that system |
|
|
|
1:29:24.480 --> 1:29:28.000 |
|
up to 100,000 miles |
|
|
|
1:29:28.000 --> 1:29:29.720 |
|
and then you write some fallback condition |
|
|
|
1:29:29.720 --> 1:29:32.160 |
|
where it's not that bad if you're way off, right? |
|
|
|
1:29:32.160 --> 1:29:33.800 |
|
I think that you can get it to the point, |
|
|
|
1:29:33.800 --> 1:29:36.800 |
|
it's like ASL D that you're never in a case |
|
|
|
1:29:36.800 --> 1:29:38.480 |
|
where you're way off and you don't know it. |
|
|
|
1:29:38.480 --> 1:29:40.240 |
|
Yeah, okay, so this is brilliant. |
|
|
|
1:29:40.240 --> 1:29:41.160 |
|
So that's the static. |
|
|
|
1:29:41.160 --> 1:29:42.280 |
|
Static. |
|
|
|
1:29:42.280 --> 1:29:45.960 |
|
We can, especially with LIDAR and good HD maps, |
|
|
|
1:29:45.960 --> 1:29:47.080 |
|
you can solve that problem. |
|
|
|
1:29:47.080 --> 1:29:47.920 |
|
It's easy. |
|
|
|
1:29:47.920 --> 1:29:51.840 |
|
The static, the static problem is so easy. |
|
|
|
1:29:51.840 --> 1:29:54.000 |
|
It's very typical for you to say something's easy. |
|
|
|
1:29:54.000 --> 1:29:54.840 |
|
I got it. |
|
|
|
1:29:54.840 --> 1:29:56.920 |
|
It's not as challenging as the other ones, okay. |
|
|
|
1:29:56.920 --> 1:29:58.760 |
|
Well, okay, maybe it's obvious how to solve it. |
|
|
|
1:29:58.760 --> 1:29:59.760 |
|
The third one's the hardest. |
|
|
|
1:29:59.760 --> 1:30:01.920 |
|
And a lot of people don't even think about the third one |
|
|
|
1:30:01.920 --> 1:30:03.640 |
|
and even see it as different from the second one. |
|
|
|
1:30:03.640 --> 1:30:05.720 |
|
So the second one is dynamic. |
|
|
|
1:30:05.720 --> 1:30:08.560 |
|
The second one is like, say there's an obvious example, |
|
|
|
1:30:08.560 --> 1:30:10.360 |
|
it's like a car stopped at a red light, right? |
|
|
|
1:30:10.360 --> 1:30:12.520 |
|
You can't have that car in your map |
|
|
|
1:30:12.520 --> 1:30:13.720 |
|
because you don't know whether that car |
|
|
|
1:30:13.720 --> 1:30:14.880 |
|
is gonna be there or not. |
|
|
|
1:30:14.880 --> 1:30:17.960 |
|
So you have to detect that car in real time |
|
|
|
1:30:17.960 --> 1:30:21.600 |
|
and then you have to do the appropriate action, right? |
|
|
|
1:30:21.600 --> 1:30:24.800 |
|
Also, that car is not a fixed object. |
|
|
|
1:30:24.800 --> 1:30:26.600 |
|
That car may move and you have to predict |
|
|
|
1:30:26.600 --> 1:30:28.680 |
|
what that car will do, right? |
|
|
|
1:30:28.680 --> 1:30:30.840 |
|
So this is the dynamic problem. |
|
|
|
1:30:30.840 --> 1:30:31.680 |
|
Yeah. |
|
|
|
1:30:31.680 --> 1:30:32.800 |
|
So you have to deal with this. |
|
|
|
1:30:32.800 --> 1:30:36.640 |
|
This involves, again, like you're gonna need models |
|
|
|
1:30:36.640 --> 1:30:38.760 |
|
of other people's behavior. |
|
|
|
1:30:38.760 --> 1:30:40.160 |
|
Do you, are you including in that? |
|
|
|
1:30:40.160 --> 1:30:42.320 |
|
I don't wanna step on the third one. |
|
|
|
1:30:42.320 --> 1:30:46.600 |
|
Oh, but are you including in that your influence |
|
|
|
1:30:46.600 --> 1:30:47.440 |
|
on people? |
|
|
|
1:30:47.440 --> 1:30:48.280 |
|
Ah, that's the third one. |
|
|
|
1:30:48.280 --> 1:30:49.120 |
|
Okay. |
|
|
|
1:30:49.120 --> 1:30:49.960 |
|
That's the third one. |
|
|
|
1:30:49.960 --> 1:30:51.880 |
|
We call it the counterfactual. |
|
|
|
1:30:51.880 --> 1:30:52.720 |
|
Yeah, brilliant. |
|
|
|
1:30:52.720 --> 1:30:53.560 |
|
And that. |
|
|
|
1:30:53.560 --> 1:30:54.920 |
|
I just talked to Judea Pro who's obsessed |
|
|
|
1:30:54.920 --> 1:30:55.760 |
|
with counterfactuals. |
|
|
|
1:30:55.760 --> 1:30:58.640 |
|
Counterfactual, oh yeah, yeah, I read his books. |
|
|
|
1:30:58.640 --> 1:31:03.640 |
|
So the static and the dynamic are our approach right now |
|
|
|
1:31:03.960 --> 1:31:07.600 |
|
for lateral will scale completely to the static and dynamic. |
|
|
|
1:31:07.600 --> 1:31:10.760 |
|
The counterfactual, the only way I have to do it yet, |
|
|
|
1:31:10.760 --> 1:31:14.000 |
|
the thing that I wanna do once we have all of these cars |
|
|
|
1:31:14.000 --> 1:31:16.760 |
|
is I wanna do reinforcement learning on the world. |
|
|
|
1:31:16.760 --> 1:31:18.880 |
|
I'm always gonna turn the exploiter up to max. |
|
|
|
1:31:18.880 --> 1:31:20.440 |
|
I'm not gonna have them explore. |
|
|
|
1:31:20.440 --> 1:31:22.760 |
|
But the only real way to get at the counterfactual |
|
|
|
1:31:22.760 --> 1:31:24.080 |
|
is to do reinforcement learning |
|
|
|
1:31:24.080 --> 1:31:26.360 |
|
because the other agents are humans. |
|
|
|
1:31:27.760 --> 1:31:30.080 |
|
So that's fascinating that you break it down like that. |
|
|
|
1:31:30.080 --> 1:31:31.680 |
|
I agree completely. |
|
|
|
1:31:31.680 --> 1:31:33.600 |
|
I've spent my life thinking about this problem. |
|
|
|
1:31:33.600 --> 1:31:34.920 |
|
This is beautiful. |
|
|
|
1:31:34.920 --> 1:31:37.880 |
|
And part of it, cause you're slightly insane, |
|
|
|
1:31:37.880 --> 1:31:42.880 |
|
because not my life, just the last four years. |
|
|
|
1:31:43.120 --> 1:31:48.120 |
|
No, no, you have some non zero percent of your brain |
|
|
|
1:31:48.920 --> 1:31:52.360 |
|
has a madman in it, which is a really good feature. |
|
|
|
1:31:52.360 --> 1:31:55.920 |
|
But there's a safety component to it |
|
|
|
1:31:55.920 --> 1:31:57.320 |
|
that I think when there's sort of |
|
|
|
1:31:57.320 --> 1:31:59.040 |
|
with counterfactuals and so on, |
|
|
|
1:31:59.040 --> 1:32:00.280 |
|
that would just freak people out. |
|
|
|
1:32:00.280 --> 1:32:03.320 |
|
How do you even start to think about this in general? |
|
|
|
1:32:03.320 --> 1:32:07.600 |
|
I mean, you've had some friction with NHTSA and so on. |
|
|
|
1:32:07.600 --> 1:32:12.600 |
|
I am frankly exhausted by safety engineers. |
|
|
|
1:32:14.280 --> 1:32:19.280 |
|
The prioritization on safety over innovation |
|
|
|
1:32:21.360 --> 1:32:23.720 |
|
to a degree where it kills, in my view, |
|
|
|
1:32:23.720 --> 1:32:26.200 |
|
kills safety in the longterm. |
|
|
|
1:32:26.200 --> 1:32:28.080 |
|
So the counterfactual thing, |
|
|
|
1:32:28.080 --> 1:32:31.560 |
|
they just actually exploring this world |
|
|
|
1:32:31.560 --> 1:32:33.600 |
|
of how do you interact with dynamic objects and so on? |
|
|
|
1:32:33.600 --> 1:32:34.840 |
|
How do you think about safety? |
|
|
|
1:32:34.840 --> 1:32:38.120 |
|
You can do reinforcement learning without ever exploring. |
|
|
|
1:32:38.120 --> 1:32:39.200 |
|
And I said that, like, |
|
|
|
1:32:39.200 --> 1:32:41.560 |
|
so you can think about your, in like reinforcement learning, |
|
|
|
1:32:41.560 --> 1:32:44.320 |
|
it's usually called like a temperature parameter. |
|
|
|
1:32:44.320 --> 1:32:45.360 |
|
And your temperature parameter |
|
|
|
1:32:45.360 --> 1:32:48.080 |
|
is how often you deviate from the argmax. |
|
|
|
1:32:48.080 --> 1:32:50.720 |
|
I could always set that to zero and still learn. |
|
|
|
1:32:50.720 --> 1:32:52.840 |
|
And I feel that you'd always want that set to zero |
|
|
|
1:32:52.840 --> 1:32:54.080 |
|
on your actual system. |
|
|
|
1:32:54.080 --> 1:32:54.920 |
|
Gotcha. |
|
|
|
1:32:54.920 --> 1:32:58.160 |
|
But the problem is you first don't know very much |
|
|
|
1:32:58.160 --> 1:32:59.560 |
|
and so you're going to make mistakes. |
|
|
|
1:32:59.560 --> 1:33:01.680 |
|
So the learning, the exploration happens through mistakes. |
|
|
|
1:33:01.680 --> 1:33:03.240 |
|
We're all ready, yeah, but. |
|
|
|
1:33:03.240 --> 1:33:06.080 |
|
Okay, so the consequences of a mistake. |
|
|
|
1:33:06.080 --> 1:33:09.400 |
|
OpenPilot and Autopilot are making mistakes left and right. |
|
|
|
1:33:09.400 --> 1:33:12.560 |
|
We have 700 daily active users, |
|
|
|
1:33:12.560 --> 1:33:14.080 |
|
1,000 weekly active users. |
|
|
|
1:33:14.080 --> 1:33:18.920 |
|
OpenPilot makes tens of thousands of mistakes a week. |
|
|
|
1:33:18.920 --> 1:33:21.160 |
|
These mistakes have zero consequences. |
|
|
|
1:33:21.160 --> 1:33:22.520 |
|
These mistakes are, |
|
|
|
1:33:22.520 --> 1:33:26.800 |
|
oh, I wanted to take this exit and it went straight. |
|
|
|
1:33:26.800 --> 1:33:28.520 |
|
So I'm just going to carefully touch the wheel. |
|
|
|
1:33:28.520 --> 1:33:29.360 |
|
The humans catch them. |
|
|
|
1:33:29.360 --> 1:33:30.640 |
|
The humans catch them. |
|
|
|
1:33:30.640 --> 1:33:33.120 |
|
And the human disengagement is labeling |
|
|
|
1:33:33.120 --> 1:33:35.000 |
|
that reinforcement learning in a completely |
|
|
|
1:33:35.000 --> 1:33:36.200 |
|
consequence free way. |
|
|
|
1:33:37.240 --> 1:33:39.840 |
|
So driver monitoring is the way you ensure they keep. |
|
|
|
1:33:39.840 --> 1:33:40.680 |
|
Yes. |
|
|
|
1:33:40.680 --> 1:33:42.120 |
|
They keep paying attention. |
|
|
|
1:33:42.120 --> 1:33:43.240 |
|
How's your messaging? |
|
|
|
1:33:43.240 --> 1:33:45.200 |
|
Say I gave you a billion dollars, |
|
|
|
1:33:45.200 --> 1:33:46.960 |
|
so you would be scaling it now. |
|
|
|
1:33:47.800 --> 1:33:49.720 |
|
Oh, if I could scale, I couldn't scale with any amount of money. |
|
|
|
1:33:49.720 --> 1:33:51.640 |
|
I'd raise money if I could, if I had a way to scale it. |
|
|
|
1:33:51.640 --> 1:33:53.320 |
|
Yeah, you're not, no, I'm not focused on scale. |
|
|
|
1:33:53.320 --> 1:33:54.160 |
|
I don't know how to do. |
|
|
|
1:33:54.160 --> 1:33:55.760 |
|
Oh, like, I guess I could sell it to more people, |
|
|
|
1:33:55.760 --> 1:33:56.960 |
|
but I want to make the system better. |
|
|
|
1:33:56.960 --> 1:33:57.800 |
|
Better, better. |
|
|
|
1:33:57.800 --> 1:33:58.840 |
|
And I don't know how to. |
|
|
|
1:33:58.840 --> 1:34:01.080 |
|
But what's the messaging here? |
|
|
|
1:34:01.080 --> 1:34:02.560 |
|
I got a chance to talk to Elon. |
|
|
|
1:34:02.560 --> 1:34:07.560 |
|
And he basically said that the human factor doesn't matter. |
|
|
|
1:34:09.280 --> 1:34:10.360 |
|
You know, the human doesn't matter |
|
|
|
1:34:10.360 --> 1:34:12.280 |
|
because the system will perform. |
|
|
|
1:34:12.280 --> 1:34:14.760 |
|
There'll be sort of a, sorry to use the term, |
|
|
|
1:34:14.760 --> 1:34:16.120 |
|
but like a singular, like a point |
|
|
|
1:34:16.120 --> 1:34:17.920 |
|
where it gets just much better. |
|
|
|
1:34:17.920 --> 1:34:20.800 |
|
And so the human, it won't really matter. |
|
|
|
1:34:20.800 --> 1:34:25.000 |
|
But it seems like that human catching the system |
|
|
|
1:34:25.000 --> 1:34:29.360 |
|
when it gets into trouble is like the thing |
|
|
|
1:34:29.360 --> 1:34:32.720 |
|
which will make something like reinforcement learning work. |
|
|
|
1:34:32.720 --> 1:34:35.640 |
|
So how do you, how do you think messaging for Tesla, |
|
|
|
1:34:35.640 --> 1:34:39.080 |
|
for you, for the industry in general, should change? |
|
|
|
1:34:39.080 --> 1:34:40.840 |
|
I think my messaging is pretty clear, |
|
|
|
1:34:40.840 --> 1:34:43.080 |
|
at least like our messaging wasn't that clear |
|
|
|
1:34:43.080 --> 1:34:45.200 |
|
in the beginning and I do kind of fault myself for that. |
|
|
|
1:34:45.200 --> 1:34:48.480 |
|
We are proud right now to be a level two system. |
|
|
|
1:34:48.480 --> 1:34:50.360 |
|
We are proud to be level two. |
|
|
|
1:34:50.360 --> 1:34:51.640 |
|
If we talk about level four, |
|
|
|
1:34:51.640 --> 1:34:53.200 |
|
it's not with the current hardware. |
|
|
|
1:34:53.200 --> 1:34:55.920 |
|
It's not going to be just a magical OTA upgrade. |
|
|
|
1:34:55.920 --> 1:34:57.280 |
|
It's going to be new hardware. |
|
|
|
1:34:57.280 --> 1:35:00.000 |
|
It's going to be very carefully thought out right now. |
|
|
|
1:35:00.000 --> 1:35:01.560 |
|
We are proud to be level two. |
|
|
|
1:35:01.560 --> 1:35:03.320 |
|
And we have a rigorous safety model. |
|
|
|
1:35:03.320 --> 1:35:05.680 |
|
I mean, not like, like, okay, rigorous. |
|
|
|
1:35:05.680 --> 1:35:06.600 |
|
Who knows what that means? |
|
|
|
1:35:06.600 --> 1:35:08.600 |
|
But we at least have a safety model |
|
|
|
1:35:08.600 --> 1:35:09.560 |
|
and we make it explicit. |
|
|
|
1:35:09.560 --> 1:35:11.800 |
|
It's in safety.md and open pilot. |
|
|
|
1:35:11.800 --> 1:35:13.960 |
|
And it says, seriously though. |
|
|
|
1:35:13.960 --> 1:35:14.800 |
|
Safety.md. |
|
|
|
1:35:14.800 --> 1:35:15.840 |
|
Safety.md. |
|
|
|
1:35:16.840 --> 1:35:18.400 |
|
This is really, this is so Android. |
|
|
|
1:35:18.400 --> 1:35:21.800 |
|
So, well, this is, this is the safety model |
|
|
|
1:35:21.800 --> 1:35:25.520 |
|
and I like to have conversations like if, like, you know, |
|
|
|
1:35:25.520 --> 1:35:27.120 |
|
sometimes people will come to you and they're like, |
|
|
|
1:35:27.120 --> 1:35:29.240 |
|
your system's not safe. |
|
|
|
1:35:29.240 --> 1:35:30.080 |
|
Okay. |
|
|
|
1:35:30.080 --> 1:35:31.080 |
|
Have you read my safety docs? |
|
|
|
1:35:31.080 --> 1:35:32.720 |
|
Would you like to have an intelligent conversation |
|
|
|
1:35:32.720 --> 1:35:33.560 |
|
about this? |
|
|
|
1:35:33.560 --> 1:35:34.400 |
|
And the answer is always no. |
|
|
|
1:35:34.400 --> 1:35:36.880 |
|
They just like scream about, it runs Python. |
|
|
|
1:35:38.240 --> 1:35:39.080 |
|
Okay. What? |
|
|
|
1:35:39.080 --> 1:35:41.560 |
|
So you're saying that, that because Python's not real time, |
|
|
|
1:35:41.560 --> 1:35:44.240 |
|
Python not being real time never causes disengagement. |
|
|
|
1:35:44.240 --> 1:35:47.640 |
|
Disengagement's are caused by, you know, the model is QM. |
|
|
|
1:35:47.640 --> 1:35:49.760 |
|
But safety.md says the following. |
|
|
|
1:35:49.760 --> 1:35:50.600 |
|
First and foremost, |
|
|
|
1:35:50.600 --> 1:35:53.000 |
|
the driver must be paying attention at all times. |
|
|
|
1:35:54.240 --> 1:35:55.320 |
|
I don't consider, I do, |
|
|
|
1:35:55.320 --> 1:35:57.720 |
|
I still consider the software to be alpha software |
|
|
|
1:35:57.720 --> 1:36:00.080 |
|
until we can actually enforce that statement. |
|
|
|
1:36:00.080 --> 1:36:03.280 |
|
But I feel it's very well communicated to our users. |
|
|
|
1:36:03.280 --> 1:36:04.520 |
|
Two more things. |
|
|
|
1:36:04.520 --> 1:36:09.080 |
|
One is the user must be able to easily take control |
|
|
|
1:36:09.080 --> 1:36:10.880 |
|
of the vehicle at all times. |
|
|
|
1:36:10.880 --> 1:36:14.440 |
|
So if you step on the gas or brake with open pilot, |
|
|
|
1:36:14.440 --> 1:36:16.400 |
|
it gives full manual control back to the user |
|
|
|
1:36:16.400 --> 1:36:18.680 |
|
or press the cancel button. |
|
|
|
1:36:18.680 --> 1:36:23.240 |
|
Step two, the car will never react so quickly. |
|
|
|
1:36:23.240 --> 1:36:26.000 |
|
We define so quickly to be about one second |
|
|
|
1:36:26.000 --> 1:36:27.640 |
|
that you can't react in time. |
|
|
|
1:36:27.640 --> 1:36:29.480 |
|
And we do this by enforcing torque limits, |
|
|
|
1:36:29.480 --> 1:36:31.520 |
|
braking limits and acceleration limits. |
|
|
|
1:36:31.520 --> 1:36:36.520 |
|
So we have like our torque limits way lower than Tesla's. |
|
|
|
1:36:36.520 --> 1:36:39.080 |
|
This is another potential. |
|
|
|
1:36:39.080 --> 1:36:40.240 |
|
If I could tweak autopilot, |
|
|
|
1:36:40.240 --> 1:36:41.360 |
|
I would lower their torque limit |
|
|
|
1:36:41.360 --> 1:36:42.960 |
|
and I would add driver monitoring. |
|
|
|
1:36:42.960 --> 1:36:46.240 |
|
Because autopilot can jerk the wheel hard. |
|
|
|
1:36:46.240 --> 1:36:47.520 |
|
Open pilot can't. |
|
|
|
1:36:47.520 --> 1:36:52.080 |
|
It's, we limit and all this code is open source, readable. |
|
|
|
1:36:52.080 --> 1:36:54.880 |
|
And I believe now it's all MISRA C compliant. |
|
|
|
1:36:54.880 --> 1:36:55.800 |
|
What's that mean? |
|
|
|
1:36:57.080 --> 1:37:00.400 |
|
MISRA is like the automotive coding standard. |
|
|
|
1:37:00.400 --> 1:37:03.400 |
|
At first, I've come to respect, |
|
|
|
1:37:03.400 --> 1:37:04.960 |
|
I've been reading like the standards lately |
|
|
|
1:37:04.960 --> 1:37:05.920 |
|
and I've come to respect them. |
|
|
|
1:37:05.920 --> 1:37:07.800 |
|
They're actually written by very smart people. |
|
|
|
1:37:07.800 --> 1:37:09.920 |
|
Yeah, they're brilliant people actually. |
|
|
|
1:37:09.920 --> 1:37:11.320 |
|
They have a lot of experience. |
|
|
|
1:37:11.320 --> 1:37:13.360 |
|
They're sometimes a little too cautious, |
|
|
|
1:37:13.360 --> 1:37:16.800 |
|
but in this case, it pays off. |
|
|
|
1:37:16.800 --> 1:37:18.440 |
|
MISRA is written by like computer scientists |
|
|
|
1:37:18.440 --> 1:37:19.840 |
|
and you can tell by the language they use. |
|
|
|
1:37:19.840 --> 1:37:21.080 |
|
You can tell by the language they use. |
|
|
|
1:37:21.080 --> 1:37:24.440 |
|
They talk about like whether certain conditions in MISRA |
|
|
|
1:37:24.440 --> 1:37:26.520 |
|
are decidable or undecidable. |
|
|
|
1:37:26.520 --> 1:37:28.360 |
|
And you mean like the halting problem? |
|
|
|
1:37:28.360 --> 1:37:31.600 |
|
And yes, all right, you've earned my respect. |
|
|
|
1:37:31.600 --> 1:37:33.120 |
|
I will read carefully what you have to say |
|
|
|
1:37:33.120 --> 1:37:35.760 |
|
and we want to make our code compliant with that. |
|
|
|
1:37:35.760 --> 1:37:38.160 |
|
All right, so you're proud level two, beautiful. |
|
|
|
1:37:38.160 --> 1:37:42.320 |
|
So you were the founder and I think CEO of Kama AI, |
|
|
|
1:37:42.320 --> 1:37:44.320 |
|
then you were the head of research. |
|
|
|
1:37:44.320 --> 1:37:46.080 |
|
What the heck are you now? |
|
|
|
1:37:46.080 --> 1:37:47.480 |
|
What's your connection to Kama AI? |
|
|
|
1:37:47.480 --> 1:37:49.640 |
|
I'm the president, but I'm one of those like |
|
|
|
1:37:49.640 --> 1:37:53.440 |
|
unelected presidents of like a small dictatorship country, |
|
|
|
1:37:53.440 --> 1:37:55.200 |
|
not one of those like elected presidents. |
|
|
|
1:37:55.200 --> 1:37:57.640 |
|
Oh, so you're like Putin when he was like the, yeah, |
|
|
|
1:37:57.640 --> 1:37:58.980 |
|
I got you. |
|
|
|
1:37:58.980 --> 1:38:02.120 |
|
So there's, what's the governance structure? |
|
|
|
1:38:02.120 --> 1:38:04.800 |
|
What's the future of Kama AI finance? |
|
|
|
1:38:04.800 --> 1:38:08.120 |
|
I mean, yeah, as a business, do you want, |
|
|
|
1:38:08.120 --> 1:38:11.640 |
|
are you just focused on getting things right now, |
|
|
|
1:38:11.640 --> 1:38:14.920 |
|
making some small amount of money in the meantime |
|
|
|
1:38:14.920 --> 1:38:17.520 |
|
and then when it works, it works a new scale. |
|
|
|
1:38:17.520 --> 1:38:20.480 |
|
Our burn rate is about 200 K a month |
|
|
|
1:38:20.480 --> 1:38:23.040 |
|
and our revenue is about 100 K a month. |
|
|
|
1:38:23.040 --> 1:38:24.920 |
|
So we need to forex our revenue, |
|
|
|
1:38:24.920 --> 1:38:28.200 |
|
but we haven't like tried very hard at that yet. |
|
|
|
1:38:28.200 --> 1:38:30.160 |
|
And the revenue is basically selling stuff online. |
|
|
|
1:38:30.160 --> 1:38:32.360 |
|
Yeah, we sell stuff shop.com.ai. |
|
|
|
1:38:32.360 --> 1:38:33.920 |
|
Is there other, well, okay. |
|
|
|
1:38:33.920 --> 1:38:35.360 |
|
So you'll have to figure out. |
|
|
|
1:38:35.360 --> 1:38:37.880 |
|
That's our only, see, but to me, |
|
|
|
1:38:37.880 --> 1:38:40.400 |
|
that's like respectable revenues. |
|
|
|
1:38:40.400 --> 1:38:42.640 |
|
We make it by selling products to consumers |
|
|
|
1:38:42.640 --> 1:38:45.040 |
|
for honest and transparent about what they are. |
|
|
|
1:38:45.040 --> 1:38:49.000 |
|
Most actually level four companies, right? |
|
|
|
1:38:50.720 --> 1:38:54.320 |
|
Cause you could easily start blowing up like smoke, |
|
|
|
1:38:54.320 --> 1:38:57.080 |
|
like overselling the hype and feeding into, |
|
|
|
1:38:57.080 --> 1:38:59.080 |
|
getting some fundraisers. |
|
|
|
1:38:59.080 --> 1:39:00.520 |
|
Oh, you're the guy, you're a genius |
|
|
|
1:39:00.520 --> 1:39:01.800 |
|
because you hacked the iPhone. |
|
|
|
1:39:01.800 --> 1:39:02.920 |
|
Oh, I hate that. |
|
|
|
1:39:02.920 --> 1:39:03.760 |
|
I hate that. |
|
|
|
1:39:03.760 --> 1:39:06.360 |
|
Yeah, I can trade my social capital for more money. |
|
|
|
1:39:06.360 --> 1:39:07.320 |
|
I did it once. |
|
|
|
1:39:07.320 --> 1:39:10.320 |
|
I almost regret it doing it the first time. |
|
|
|
1:39:10.320 --> 1:39:11.640 |
|
Well, on a small tangent, |
|
|
|
1:39:11.640 --> 1:39:16.560 |
|
what's your, you seem to not like fame |
|
|
|
1:39:16.560 --> 1:39:18.840 |
|
and yet you're also drawn to fame. |
|
|
|
1:39:18.840 --> 1:39:23.840 |
|
What's, where have you on, where are you on that currently? |
|
|
|
1:39:24.560 --> 1:39:27.200 |
|
Have you had some introspection, some soul searching? |
|
|
|
1:39:27.200 --> 1:39:28.480 |
|
Yeah. |
|
|
|
1:39:28.480 --> 1:39:32.200 |
|
I actually, I've come to a pretty stable position on that. |
|
|
|
1:39:32.200 --> 1:39:33.880 |
|
Like after the first time, |
|
|
|
1:39:33.880 --> 1:39:36.840 |
|
I realized that I don't want attention from the masses. |
|
|
|
1:39:36.840 --> 1:39:39.160 |
|
I want attention from people who I respect. |
|
|
|
1:39:39.160 --> 1:39:41.960 |
|
Who do you respect? |
|
|
|
1:39:41.960 --> 1:39:43.960 |
|
I can give a list of people. |
|
|
|
1:39:43.960 --> 1:39:47.200 |
|
So are these like Elon Musk type characters? |
|
|
|
1:39:47.200 --> 1:39:49.040 |
|
Yeah. |
|
|
|
1:39:49.040 --> 1:39:50.000 |
|
Actually, you know what? |
|
|
|
1:39:50.000 --> 1:39:51.200 |
|
I'll make it more broad than that. |
|
|
|
1:39:51.200 --> 1:39:52.600 |
|
I won't make it about a person. |
|
|
|
1:39:52.600 --> 1:39:54.040 |
|
I respect skill. |
|
|
|
1:39:54.040 --> 1:39:56.880 |
|
I respect people who have skills, right? |
|
|
|
1:39:56.880 --> 1:40:00.280 |
|
And I would like to like be, |
|
|
|
1:40:00.280 --> 1:40:01.400 |
|
I'm not gonna say famous, |
|
|
|
1:40:01.400 --> 1:40:03.760 |
|
but be like known among more people |
|
|
|
1:40:03.760 --> 1:40:05.440 |
|
who have like real skills. |
|
|
|
1:40:05.440 --> 1:40:10.440 |
|
Who in cars, do you think have skill? |
|
|
|
1:40:12.560 --> 1:40:13.720 |
|
Not do you respect? |
|
|
|
1:40:15.000 --> 1:40:17.760 |
|
Oh, Kyle Voat has skill. |
|
|
|
1:40:17.760 --> 1:40:19.880 |
|
A lot of people at Waymo have skill. |
|
|
|
1:40:19.880 --> 1:40:20.840 |
|
And I respect them. |
|
|
|
1:40:20.840 --> 1:40:23.760 |
|
I respect them as engineers. |
|
|
|
1:40:23.760 --> 1:40:24.920 |
|
Like I can think, I mean, |
|
|
|
1:40:24.920 --> 1:40:26.280 |
|
I think about all the times in my life |
|
|
|
1:40:26.280 --> 1:40:27.960 |
|
where I've been like dead set on approaches |
|
|
|
1:40:27.960 --> 1:40:29.160 |
|
and they turn out to be wrong. |
|
|
|
1:40:29.160 --> 1:40:30.000 |
|
Yeah. |
|
|
|
1:40:30.000 --> 1:40:31.720 |
|
So I mean, this might, I might be wrong. |
|
|
|
1:40:31.720 --> 1:40:34.720 |
|
I accept that, I accept that there's a decent chance |
|
|
|
1:40:34.720 --> 1:40:36.600 |
|
that I'm wrong. |
|
|
|
1:40:36.600 --> 1:40:38.400 |
|
And actually, I mean, having talked to Chris Armson, |
|
|
|
1:40:38.400 --> 1:40:40.480 |
|
Sterling Anderson, those guys, |
|
|
|
1:40:40.480 --> 1:40:43.360 |
|
I mean, I deeply respect Chris. |
|
|
|
1:40:43.360 --> 1:40:44.640 |
|
I just admire the guy. |
|
|
|
1:40:46.040 --> 1:40:47.400 |
|
He's legit. |
|
|
|
1:40:47.400 --> 1:40:48.960 |
|
When you drive a car through the desert |
|
|
|
1:40:48.960 --> 1:40:52.400 |
|
when everybody thinks it's impossible, that's legit. |
|
|
|
1:40:52.400 --> 1:40:53.840 |
|
And then I also really respect the people |
|
|
|
1:40:53.840 --> 1:40:55.680 |
|
who are like writing the infrastructure of the world, |
|
|
|
1:40:55.680 --> 1:40:57.360 |
|
like the Linus Torvalds and the Chris Ladin. |
|
|
|
1:40:57.360 --> 1:40:59.080 |
|
Oh yeah, they were doing the real work. |
|
|
|
1:40:59.080 --> 1:41:00.800 |
|
I know they're doing the real work. |
|
|
|
1:41:02.000 --> 1:41:03.760 |
|
Having talked to Chris Ladin, |
|
|
|
1:41:03.760 --> 1:41:05.680 |
|
you realize, especially when they're humble, |
|
|
|
1:41:05.680 --> 1:41:07.680 |
|
it's like, you realize, oh, you guys, |
|
|
|
1:41:07.680 --> 1:41:09.640 |
|
we're just using your... |
|
|
|
1:41:09.640 --> 1:41:10.480 |
|
Oh yeah. |
|
|
|
1:41:10.480 --> 1:41:11.520 |
|
All the hard work that you did. |
|
|
|
1:41:11.520 --> 1:41:13.120 |
|
Yeah, that's incredible. |
|
|
|
1:41:13.120 --> 1:41:17.160 |
|
What do you think, Mr. Anthony Lewandowski? |
|
|
|
1:41:18.440 --> 1:41:21.640 |
|
What do you, he's a, he's another mad genius. |
|
|
|
1:41:21.640 --> 1:41:22.480 |
|
Sharp guy. |
|
|
|
1:41:22.480 --> 1:41:23.320 |
|
Oh yeah. |
|
|
|
1:41:23.320 --> 1:41:27.640 |
|
What, do you think he might long term become a competitor? |
|
|
|
1:41:27.640 --> 1:41:28.840 |
|
Oh, to comma? |
|
|
|
1:41:28.840 --> 1:41:32.400 |
|
Well, so I think that he has the other right approach. |
|
|
|
1:41:32.400 --> 1:41:35.280 |
|
I think that right now, there's two right approaches. |
|
|
|
1:41:35.280 --> 1:41:37.680 |
|
One is what we're doing and one is what he's doing. |
|
|
|
1:41:37.680 --> 1:41:39.800 |
|
Can you describe, I think it's called Pronto AI, |
|
|
|
1:41:39.800 --> 1:41:42.360 |
|
he's starting using, do you know what the approach is? |
|
|
|
1:41:42.360 --> 1:41:43.200 |
|
I actually don't know. |
|
|
|
1:41:43.200 --> 1:41:45.040 |
|
Embark is also doing the same sort of thing. |
|
|
|
1:41:45.040 --> 1:41:47.280 |
|
The idea is almost that you want to, |
|
|
|
1:41:47.280 --> 1:41:51.800 |
|
so if you're, I can't partner with Honda and Toyota. |
|
|
|
1:41:51.800 --> 1:41:56.800 |
|
Honda and Toyota are like 400,000 person companies. |
|
|
|
1:41:57.600 --> 1:41:59.400 |
|
It's not even a company at that point. |
|
|
|
1:41:59.400 --> 1:42:01.400 |
|
Like I don't think of it like, I don't personify it. |
|
|
|
1:42:01.400 --> 1:42:06.400 |
|
I think of it like an object, but a trucker drives for a fleet. |
|
|
|
1:42:07.120 --> 1:42:10.280 |
|
Maybe that has like, some truckers are independent. |
|
|
|
1:42:10.280 --> 1:42:12.080 |
|
Some truckers drive for fleets with a hundred trucks. |
|
|
|
1:42:12.080 --> 1:42:14.960 |
|
There are tons of independent trucking companies out there. |
|
|
|
1:42:14.960 --> 1:42:18.120 |
|
Start a trucking company and drive your costs down |
|
|
|
1:42:18.120 --> 1:42:23.120 |
|
or figure out how to drive down the cost of trucking. |
|
|
|
1:42:23.760 --> 1:42:26.560 |
|
Another company that I really respect is Nauto. |
|
|
|
1:42:26.560 --> 1:42:28.320 |
|
Actually, I respect their business model. |
|
|
|
1:42:28.320 --> 1:42:31.560 |
|
Nauto sells a driver monitoring camera |
|
|
|
1:42:31.560 --> 1:42:33.920 |
|
and they sell it to fleet owners. |
|
|
|
1:42:33.920 --> 1:42:38.920 |
|
If I owned a fleet of cars and I could pay 40 bucks a month |
|
|
|
1:42:39.120 --> 1:42:41.280 |
|
to monitor my employees, |
|
|
|
1:42:42.400 --> 1:42:45.520 |
|
this is gonna like reduces accidents 18%. |
|
|
|
1:42:45.520 --> 1:42:48.960 |
|
It's so like that in the space, |
|
|
|
1:42:48.960 --> 1:42:52.000 |
|
that is like the business model that I like most respect |
|
|
|
1:42:53.400 --> 1:42:55.360 |
|
because they're creating value today. |
|
|
|
1:42:55.360 --> 1:42:57.840 |
|
Yeah, which is, that's a huge one. |
|
|
|
1:42:57.840 --> 1:42:59.800 |
|
How do we create value today with some of this? |
|
|
|
1:42:59.800 --> 1:43:01.680 |
|
And the length keeping thing is huge. |
|
|
|
1:43:01.680 --> 1:43:03.800 |
|
And it sounds like you're creeping in |
|
|
|
1:43:03.800 --> 1:43:06.680 |
|
or full steam ahead on the drive of monitoring too. |
|
|
|
1:43:06.680 --> 1:43:09.240 |
|
Which I think actually where the short term value, |
|
|
|
1:43:09.240 --> 1:43:10.480 |
|
if you can get right. |
|
|
|
1:43:10.480 --> 1:43:12.800 |
|
I still, I'm not a huge fan of the statement |
|
|
|
1:43:12.800 --> 1:43:15.120 |
|
that everything is to have drive of monitoring. |
|
|
|
1:43:15.120 --> 1:43:16.120 |
|
I agree with that completely, |
|
|
|
1:43:16.120 --> 1:43:18.680 |
|
but that statement usually misses the point |
|
|
|
1:43:18.680 --> 1:43:21.920 |
|
that to get the experience of it right is not trivial. |
|
|
|
1:43:21.920 --> 1:43:22.840 |
|
Oh, no, not at all. |
|
|
|
1:43:22.840 --> 1:43:25.280 |
|
In fact, like, so right now we have, |
|
|
|
1:43:25.280 --> 1:43:28.480 |
|
I think the timeout depends on speed of the car, |
|
|
|
1:43:29.560 --> 1:43:32.520 |
|
but we want to depend on like the scene state. |
|
|
|
1:43:32.520 --> 1:43:35.440 |
|
If you're on like an empty highway, |
|
|
|
1:43:35.440 --> 1:43:37.680 |
|
it's very different if you don't pay attention |
|
|
|
1:43:37.680 --> 1:43:40.600 |
|
than if like you're like coming up to a traffic light. |
|
|
|
1:43:42.040 --> 1:43:45.720 |
|
And long term, it should probably learn from the driver |
|
|
|
1:43:45.720 --> 1:43:48.120 |
|
because that's to do, I watched a lot of video. |
|
|
|
1:43:48.120 --> 1:43:49.480 |
|
We've built a smartphone detector |
|
|
|
1:43:49.480 --> 1:43:51.520 |
|
just to analyze how people are using smartphones |
|
|
|
1:43:51.520 --> 1:43:53.400 |
|
and people are using it very differently. |
|
|
|
1:43:53.400 --> 1:43:57.760 |
|
And there's this, it's a texting styles. |
|
|
|
1:43:57.760 --> 1:44:00.320 |
|
We haven't watched nearly enough of the videos. |
|
|
|
1:44:00.320 --> 1:44:01.800 |
|
We haven't, I got millions of miles |
|
|
|
1:44:01.800 --> 1:44:02.960 |
|
of people driving cars. |
|
|
|
1:44:02.960 --> 1:44:05.960 |
|
In this moment, I spent a large fraction of my time |
|
|
|
1:44:05.960 --> 1:44:10.880 |
|
just watching videos because it's never fails to learn. |
|
|
|
1:44:10.880 --> 1:44:13.480 |
|
Like I've never failed from a video watching session |
|
|
|
1:44:13.480 --> 1:44:15.400 |
|
to learn something I didn't know before. |
|
|
|
1:44:15.400 --> 1:44:19.640 |
|
In fact, I usually, like when I eat lunch, I'll sit, |
|
|
|
1:44:19.640 --> 1:44:20.680 |
|
especially when the weather is good |
|
|
|
1:44:20.680 --> 1:44:22.080 |
|
and just watch pedestrians. |
|
|
|
1:44:22.080 --> 1:44:26.400 |
|
With an eye to understand like from a computer vision eye, |
|
|
|
1:44:26.400 --> 1:44:29.280 |
|
just to see, can this model, can you predict |
|
|
|
1:44:29.280 --> 1:44:30.480 |
|
what are the decisions made? |
|
|
|
1:44:30.480 --> 1:44:33.040 |
|
And there's so many things that we don't understand. |
|
|
|
1:44:33.040 --> 1:44:34.760 |
|
This is what I mean about state vector. |
|
|
|
1:44:34.760 --> 1:44:37.880 |
|
Yeah, it's, I'm trying to always think like, |
|
|
|
1:44:37.880 --> 1:44:40.280 |
|
because I'm understanding in my human brain, |
|
|
|
1:44:40.280 --> 1:44:42.000 |
|
how do we convert that into, |
|
|
|
1:44:43.000 --> 1:44:44.960 |
|
how hard is the learning problem here? |
|
|
|
1:44:44.960 --> 1:44:46.960 |
|
I guess is the fundamental question. |
|
|
|
1:44:46.960 --> 1:44:51.800 |
|
So something that's from a hacking perspective, |
|
|
|
1:44:51.800 --> 1:44:54.200 |
|
this is always comes up, especially with folks. |
|
|
|
1:44:54.200 --> 1:44:56.480 |
|
Well, first, the most popular question is |
|
|
|
1:44:56.480 --> 1:44:58.440 |
|
the trolley problem, right? |
|
|
|
1:44:58.440 --> 1:45:01.960 |
|
So that's not a sort of a serious problem. |
|
|
|
1:45:01.960 --> 1:45:05.000 |
|
There are some ethical questions, I think that arise. |
|
|
|
1:45:06.080 --> 1:45:09.600 |
|
Maybe you wanna, do you think there's any ethical, |
|
|
|
1:45:09.600 --> 1:45:11.280 |
|
serious ethical questions? |
|
|
|
1:45:11.280 --> 1:45:14.080 |
|
We have a solution to the trolley problem at com.ai. |
|
|
|
1:45:14.080 --> 1:45:15.920 |
|
Well, so there is actually an alert |
|
|
|
1:45:15.920 --> 1:45:18.000 |
|
in our code, ethical dilemma detected. |
|
|
|
1:45:18.000 --> 1:45:18.960 |
|
It's not triggered yet. |
|
|
|
1:45:18.960 --> 1:45:21.040 |
|
We don't know how yet to detect the ethical dilemmas, |
|
|
|
1:45:21.040 --> 1:45:22.360 |
|
but we're a level two system. |
|
|
|
1:45:22.360 --> 1:45:23.760 |
|
So we're going to disengage and leave |
|
|
|
1:45:23.760 --> 1:45:25.320 |
|
that decision to the human. |
|
|
|
1:45:25.320 --> 1:45:26.680 |
|
You're such a troll. |
|
|
|
1:45:26.680 --> 1:45:28.760 |
|
No, but the trolley problem deserves to be trolled. |
|
|
|
1:45:28.760 --> 1:45:32.040 |
|
Yeah, that's a beautiful answer actually. |
|
|
|
1:45:32.040 --> 1:45:34.440 |
|
I know, I gave it to someone who was like, |
|
|
|
1:45:34.440 --> 1:45:36.600 |
|
sometimes people ask like you asked about the trolley problem. |
|
|
|
1:45:36.600 --> 1:45:38.080 |
|
Like you can have a kind of discussion about it. |
|
|
|
1:45:38.080 --> 1:45:39.720 |
|
Like when you get someone who's like really like |
|
|
|
1:45:39.720 --> 1:45:43.600 |
|
earnest about it, because it's the kind of thing where |
|
|
|
1:45:43.600 --> 1:45:45.600 |
|
if you ask a bunch of people in an office, |
|
|
|
1:45:45.600 --> 1:45:48.360 |
|
whether we should use a SQL stack or no SQL stack, |
|
|
|
1:45:48.360 --> 1:45:50.600 |
|
if they're not that technical, they have no opinion. |
|
|
|
1:45:50.600 --> 1:45:52.360 |
|
But if you ask them what color they want to paint the office, |
|
|
|
1:45:52.360 --> 1:45:54.040 |
|
everyone has an opinion on that. |
|
|
|
1:45:54.040 --> 1:45:56.040 |
|
And that's why the trolley problem is. |
|
|
|
1:45:56.040 --> 1:45:57.280 |
|
I mean, that's a beautiful answer. |
|
|
|
1:45:57.280 --> 1:45:59.240 |
|
Yeah, we're able to detect the problem |
|
|
|
1:45:59.240 --> 1:46:01.960 |
|
and we're able to pass it on to the human. |
|
|
|
1:46:01.960 --> 1:46:03.760 |
|
Wow, I've never heard anyone say it. |
|
|
|
1:46:03.760 --> 1:46:06.160 |
|
This is your nice escape route. |
|
|
|
1:46:06.160 --> 1:46:07.320 |
|
Okay, but... |
|
|
|
1:46:07.320 --> 1:46:08.680 |
|
Proud level two. |
|
|
|
1:46:08.680 --> 1:46:09.760 |
|
I'm proud level two. |
|
|
|
1:46:09.760 --> 1:46:10.600 |
|
I love it. |
|
|
|
1:46:10.600 --> 1:46:14.400 |
|
So the other thing that people have some concern about |
|
|
|
1:46:14.400 --> 1:46:17.800 |
|
with AI in general is hacking. |
|
|
|
1:46:17.800 --> 1:46:21.400 |
|
So how hard is it, do you think, to hack an autonomous vehicle |
|
|
|
1:46:21.400 --> 1:46:25.000 |
|
either through physical access or through the more sort of |
|
|
|
1:46:25.000 --> 1:46:28.240 |
|
popular now, these adversarial examples on the sensors? |
|
|
|
1:46:28.240 --> 1:46:30.720 |
|
Okay, the adversarial examples one. |
|
|
|
1:46:30.720 --> 1:46:32.320 |
|
You want to see some adversarial examples |
|
|
|
1:46:32.320 --> 1:46:34.880 |
|
that affect humans, right? |
|
|
|
1:46:34.880 --> 1:46:38.040 |
|
Oh, well, there used to be a stop sign here, |
|
|
|
1:46:38.040 --> 1:46:40.000 |
|
but I put a black bag over the stop sign |
|
|
|
1:46:40.000 --> 1:46:43.520 |
|
and then people ran it, adversarial, right? |
|
|
|
1:46:43.520 --> 1:46:48.360 |
|
Like, there's tons of human adversarial examples too. |
|
|
|
1:46:48.360 --> 1:46:52.240 |
|
The question in general about security, if you saw, |
|
|
|
1:46:52.240 --> 1:46:54.040 |
|
something just came out today and there are always |
|
|
|
1:46:54.040 --> 1:46:57.560 |
|
such hypey headlines about how navigate on autopilot |
|
|
|
1:46:57.560 --> 1:47:00.960 |
|
was fooled by a GPS spoof to take an exit. |
|
|
|
1:47:00.960 --> 1:47:01.800 |
|
Right. |
|
|
|
1:47:01.800 --> 1:47:03.920 |
|
At least that's all they could do was take an exit. |
|
|
|
1:47:03.920 --> 1:47:06.720 |
|
If your car is relying on GPS in order |
|
|
|
1:47:06.720 --> 1:47:10.240 |
|
to have a safe driving policy, you're doing something wrong. |
|
|
|
1:47:10.240 --> 1:47:12.680 |
|
If you're relying, and this is why V2V |
|
|
|
1:47:12.680 --> 1:47:17.680 |
|
is such a terrible idea, V2V now relies on both parties |
|
|
|
1:47:18.160 --> 1:47:19.800 |
|
getting communication right. |
|
|
|
1:47:19.800 --> 1:47:24.800 |
|
This is not even, so I think of safety, |
|
|
|
1:47:26.080 --> 1:47:28.480 |
|
security is like a special case of safety, right? |
|
|
|
1:47:28.480 --> 1:47:31.880 |
|
Safety is like we put a little, you know, |
|
|
|
1:47:31.880 --> 1:47:33.360 |
|
piece of caution tape around the hole |
|
|
|
1:47:33.360 --> 1:47:35.560 |
|
so that people won't walk into it by accident. |
|
|
|
1:47:35.560 --> 1:47:38.240 |
|
Security is like put a 10 foot fence around the hole |
|
|
|
1:47:38.240 --> 1:47:40.120 |
|
so you actually physically cannot climb into it |
|
|
|
1:47:40.120 --> 1:47:42.360 |
|
with barbed wire on the top and stuff, right? |
|
|
|
1:47:42.360 --> 1:47:44.560 |
|
So like if you're designing systems |
|
|
|
1:47:44.560 --> 1:47:47.440 |
|
that are like unreliable, they're definitely not secure. |
|
|
|
1:47:48.440 --> 1:47:51.240 |
|
Your car should always do something safe |
|
|
|
1:47:51.240 --> 1:47:53.400 |
|
using its local sensors. |
|
|
|
1:47:53.400 --> 1:47:55.240 |
|
And then the local sensor should be hardwired. |
|
|
|
1:47:55.240 --> 1:47:57.400 |
|
And then could somebody hack into your can boss |
|
|
|
1:47:57.400 --> 1:47:58.640 |
|
and turn your steering wheel on your brakes? |
|
|
|
1:47:58.640 --> 1:48:01.240 |
|
Yes, but they could do it before comma AI too, so. |
|
|
|
1:48:02.800 --> 1:48:04.680 |
|
Let's think out of the box and some things. |
|
|
|
1:48:04.680 --> 1:48:09.400 |
|
So do you think teleoperation has a role in any of this? |
|
|
|
1:48:09.400 --> 1:48:13.880 |
|
So remotely stepping in and controlling the cars? |
|
|
|
1:48:13.880 --> 1:48:18.880 |
|
No, I think that if the safety operation |
|
|
|
1:48:21.320 --> 1:48:26.160 |
|
by design requires a constant link to the cars, |
|
|
|
1:48:26.160 --> 1:48:27.560 |
|
I think it doesn't work. |
|
|
|
1:48:27.560 --> 1:48:31.080 |
|
So that's the same argument used for V2I, V2V. |
|
|
|
1:48:31.080 --> 1:48:34.280 |
|
Well, there's a lot of non safety critical stuff |
|
|
|
1:48:34.280 --> 1:48:35.120 |
|
you can do with V2I. |
|
|
|
1:48:35.120 --> 1:48:37.440 |
|
I like V2I, I like V2I way more than V2V |
|
|
|
1:48:37.440 --> 1:48:39.280 |
|
because V2I is already like, |
|
|
|
1:48:39.280 --> 1:48:40.880 |
|
I already have internet in the car, right? |
|
|
|
1:48:40.880 --> 1:48:43.280 |
|
There's a lot of great stuff you can do with V2I. |
|
|
|
1:48:44.280 --> 1:48:46.320 |
|
Like for example, you can, |
|
|
|
1:48:46.320 --> 1:48:48.880 |
|
well, where I already have V2V, Waze is V2I, right? |
|
|
|
1:48:48.880 --> 1:48:50.520 |
|
Waze can route me around traffic jams. |
|
|
|
1:48:50.520 --> 1:48:52.760 |
|
That's a great example of V2I. |
|
|
|
1:48:52.760 --> 1:48:54.440 |
|
And then, okay, the car automatically talks |
|
|
|
1:48:54.440 --> 1:48:55.800 |
|
to that same service, like it works. |
|
|
|
1:48:55.800 --> 1:48:56.800 |
|
So it's improving the experience, |
|
|
|
1:48:56.800 --> 1:48:59.480 |
|
but it's not a fundamental fallback for safety. |
|
|
|
1:48:59.480 --> 1:49:04.160 |
|
No, if any of your things that require |
|
|
|
1:49:04.160 --> 1:49:07.480 |
|
wireless communication are more than QM, |
|
|
|
1:49:07.480 --> 1:49:10.640 |
|
like have an ASL rating, you shouldn't. |
|
|
|
1:49:10.640 --> 1:49:14.200 |
|
You previously said that life is work |
|
|
|
1:49:15.440 --> 1:49:17.480 |
|
and that you don't do anything to relax. |
|
|
|
1:49:17.480 --> 1:49:20.800 |
|
So how do you think about hard work? |
|
|
|
1:49:20.800 --> 1:49:22.200 |
|
Well, what is it? |
|
|
|
1:49:22.200 --> 1:49:24.720 |
|
What do you think it takes to accomplish great things? |
|
|
|
1:49:24.720 --> 1:49:25.840 |
|
And there's a lot of people saying |
|
|
|
1:49:25.840 --> 1:49:28.280 |
|
that there needs to be some balance. |
|
|
|
1:49:28.280 --> 1:49:29.600 |
|
You know, you need to, |
|
|
|
1:49:29.600 --> 1:49:31.120 |
|
in order to accomplish great things, |
|
|
|
1:49:31.120 --> 1:49:32.200 |
|
you need to take some time off, |
|
|
|
1:49:32.200 --> 1:49:34.640 |
|
you need to reflect and so on. |
|
|
|
1:49:34.640 --> 1:49:37.840 |
|
And then some people are just insanely working, |
|
|
|
1:49:37.840 --> 1:49:39.640 |
|
burning the candle on both ends. |
|
|
|
1:49:39.640 --> 1:49:41.360 |
|
How do you think about that? |
|
|
|
1:49:41.360 --> 1:49:43.400 |
|
I think I was trolling in the Suraj interview |
|
|
|
1:49:43.400 --> 1:49:45.600 |
|
when I said that off camera, |
|
|
|
1:49:45.600 --> 1:49:47.240 |
|
but right before I spoke to a little bit of weed, |
|
|
|
1:49:47.240 --> 1:49:49.800 |
|
like, you know, come on, this is a joke, right? |
|
|
|
1:49:49.800 --> 1:49:50.880 |
|
Like I do nothing to relax. |
|
|
|
1:49:50.880 --> 1:49:52.560 |
|
Look where I am, I'm at a party, right? |
|
|
|
1:49:52.560 --> 1:49:53.960 |
|
Yeah, yeah, yeah, sure. |
|
|
|
1:49:53.960 --> 1:49:55.200 |
|
That's true. |
|
|
|
1:49:55.200 --> 1:49:58.040 |
|
So no, no, of course I don't. |
|
|
|
1:49:58.040 --> 1:49:59.800 |
|
When I say that life is work though, |
|
|
|
1:49:59.800 --> 1:50:01.960 |
|
I mean that like, I think that |
|
|
|
1:50:01.960 --> 1:50:04.200 |
|
what gives my life meaning is work. |
|
|
|
1:50:04.200 --> 1:50:05.720 |
|
I don't mean that every minute of the day |
|
|
|
1:50:05.720 --> 1:50:06.560 |
|
you should be working. |
|
|
|
1:50:06.560 --> 1:50:08.000 |
|
I actually think this is not the best way |
|
|
|
1:50:08.000 --> 1:50:09.800 |
|
to maximize results. |
|
|
|
1:50:09.800 --> 1:50:12.040 |
|
I think that if you're working 12 hours a day, |
|
|
|
1:50:12.040 --> 1:50:14.920 |
|
you should be working smarter and not harder. |
|
|
|
1:50:14.920 --> 1:50:17.880 |
|
Well, so it gives work gives you meaning |
|
|
|
1:50:17.880 --> 1:50:20.520 |
|
for some people, other sorts of meaning |
|
|
|
1:50:20.520 --> 1:50:24.560 |
|
is personal relationships, like family and so on. |
|
|
|
1:50:24.560 --> 1:50:27.200 |
|
You've also in that interview with Suraj |
|
|
|
1:50:27.200 --> 1:50:30.720 |
|
or the trolling mentioned that one of the things |
|
|
|
1:50:30.720 --> 1:50:33.400 |
|
you look forward to in the future is AI girlfriends. |
|
|
|
1:50:33.400 --> 1:50:34.360 |
|
Yes. |
|
|
|
1:50:34.360 --> 1:50:38.800 |
|
So that's a topic that I'm very much fascinated by, |
|
|
|
1:50:38.800 --> 1:50:39.840 |
|
not necessarily girlfriends, |
|
|
|
1:50:39.840 --> 1:50:41.880 |
|
but just forming a deep connection with AI. |
|
|
|
1:50:41.880 --> 1:50:42.960 |
|
Yeah. |
|
|
|
1:50:42.960 --> 1:50:44.400 |
|
What kind of system do you imagine |
|
|
|
1:50:44.400 --> 1:50:46.240 |
|
when you say AI girlfriend, |
|
|
|
1:50:46.240 --> 1:50:47.800 |
|
whether you were trolling or not? |
|
|
|
1:50:47.800 --> 1:50:49.720 |
|
No, that one I'm very serious about. |
|
|
|
1:50:49.720 --> 1:50:52.360 |
|
And I'm serious about that on both a shallow level |
|
|
|
1:50:52.360 --> 1:50:53.680 |
|
and a deep level. |
|
|
|
1:50:53.680 --> 1:50:55.720 |
|
I think that VR brothels are coming soon |
|
|
|
1:50:55.720 --> 1:50:57.800 |
|
and are gonna be really cool. |
|
|
|
1:50:57.800 --> 1:50:59.760 |
|
It's not cheating if it's a robot. |
|
|
|
1:50:59.760 --> 1:51:01.080 |
|
I see the slogan already. |
|
|
|
1:51:01.080 --> 1:51:04.320 |
|
Um, but... |
|
|
|
1:51:04.320 --> 1:51:06.200 |
|
There's a, I don't know if you've watched |
|
|
|
1:51:06.200 --> 1:51:08.320 |
|
or just watched the Black Mirror episode. |
|
|
|
1:51:08.320 --> 1:51:09.320 |
|
I watched the latest one, yeah. |
|
|
|
1:51:09.320 --> 1:51:11.320 |
|
Yeah, yeah. |
|
|
|
1:51:11.320 --> 1:51:13.160 |
|
Oh, the Ashley 2 one? |
|
|
|
1:51:13.160 --> 1:51:15.120 |
|
Or the... |
|
|
|
1:51:15.120 --> 1:51:16.920 |
|
No, where there's two friends |
|
|
|
1:51:16.920 --> 1:51:20.160 |
|
who were having sex with each other in... |
|
|
|
1:51:20.160 --> 1:51:21.240 |
|
Oh, in the VR game. |
|
|
|
1:51:21.240 --> 1:51:23.560 |
|
In the VR game, it's the two guys, |
|
|
|
1:51:23.560 --> 1:51:26.720 |
|
but one of them was a female, yeah. |
|
|
|
1:51:26.720 --> 1:51:27.560 |
|
Yeah, the... |
|
|
|
1:51:27.560 --> 1:51:29.560 |
|
Which is another mind blowing concept. |
|
|
|
1:51:29.560 --> 1:51:33.320 |
|
That in VR, you don't have to be the form. |
|
|
|
1:51:33.320 --> 1:51:37.720 |
|
You can be two animals having sex, it's weird. |
|
|
|
1:51:37.720 --> 1:51:38.560 |
|
I mean, I'll see how nice |
|
|
|
1:51:38.560 --> 1:51:40.280 |
|
that the software maps the nerve endings, right? |
|
|
|
1:51:40.280 --> 1:51:41.600 |
|
Yeah, it's weird. |
|
|
|
1:51:41.600 --> 1:51:44.480 |
|
I mean, yeah, they sweep a lot of the fascinating, |
|
|
|
1:51:44.480 --> 1:51:46.440 |
|
really difficult technical challenges under the rug, |
|
|
|
1:51:46.440 --> 1:51:48.360 |
|
like assuming it's possible |
|
|
|
1:51:48.360 --> 1:51:51.160 |
|
to do the mapping of the nerve endings, then... |
|
|
|
1:51:51.160 --> 1:51:52.000 |
|
I wish, yeah, I saw that. |
|
|
|
1:51:52.000 --> 1:51:53.800 |
|
The way they did it with the little like stim unit |
|
|
|
1:51:53.800 --> 1:51:55.400 |
|
on the head, that'd be amazing. |
|
|
|
1:51:56.800 --> 1:51:58.760 |
|
So, well, no, no, on a shallow level, |
|
|
|
1:51:58.760 --> 1:52:01.640 |
|
like you could set up like almost a brothel |
|
|
|
1:52:01.640 --> 1:52:05.160 |
|
with like real dolls and Oculus quests, |
|
|
|
1:52:05.160 --> 1:52:06.200 |
|
write some good software. |
|
|
|
1:52:06.200 --> 1:52:08.280 |
|
I think it'd be a cool novelty experience. |
|
|
|
1:52:09.280 --> 1:52:11.400 |
|
But no, on a deeper, like emotional level. |
|
|
|
1:52:12.800 --> 1:52:16.960 |
|
I mean, yeah, I would really like to fall in love |
|
|
|
1:52:16.960 --> 1:52:18.120 |
|
with the machine. |
|
|
|
1:52:18.120 --> 1:52:23.120 |
|
Do you see yourself having a long term relationship |
|
|
|
1:52:23.120 --> 1:52:27.520 |
|
of the kind monogamous relationship that we have now |
|
|
|
1:52:27.520 --> 1:52:31.360 |
|
with the robot, with the AI system, even? |
|
|
|
1:52:31.360 --> 1:52:32.680 |
|
Not even just the robot. |
|
|
|
1:52:32.680 --> 1:52:37.680 |
|
So, I think about maybe my ideal future. |
|
|
|
1:52:38.200 --> 1:52:43.200 |
|
When I was 15, I read Eliezer Yudkowsky's early writings |
|
|
|
1:52:44.320 --> 1:52:49.120 |
|
on the singularity and like that AI |
|
|
|
1:52:49.120 --> 1:52:53.040 |
|
is going to surpass human intelligence massively. |
|
|
|
1:52:53.040 --> 1:52:55.480 |
|
He made some Moore's law based predictions |
|
|
|
1:52:55.480 --> 1:52:57.400 |
|
that I mostly agree with. |
|
|
|
1:52:57.400 --> 1:52:59.360 |
|
And then I really struggled |
|
|
|
1:52:59.360 --> 1:53:01.360 |
|
for the next couple of years of my life. |
|
|
|
1:53:01.360 --> 1:53:03.360 |
|
Like, why should I even bother to learn anything? |
|
|
|
1:53:03.360 --> 1:53:06.160 |
|
It's all gonna be meaningless when the machine show up. |
|
|
|
1:53:06.160 --> 1:53:07.000 |
|
Right. |
|
|
|
1:53:07.000 --> 1:53:10.520 |
|
Well, maybe when I was that young, |
|
|
|
1:53:10.520 --> 1:53:12.040 |
|
I was still a little bit more pure |
|
|
|
1:53:12.040 --> 1:53:13.160 |
|
and really like clung to that. |
|
|
|
1:53:13.160 --> 1:53:14.720 |
|
And then I'm like, well, the machine's ain't here yet. |
|
|
|
1:53:14.720 --> 1:53:16.800 |
|
You know, and I seem to be pretty good at this stuff. |
|
|
|
1:53:16.800 --> 1:53:18.520 |
|
Let's try my best, you know, |
|
|
|
1:53:18.520 --> 1:53:20.320 |
|
like what's the worst that happens? |
|
|
|
1:53:20.320 --> 1:53:23.440 |
|
But the best possible future I see |
|
|
|
1:53:23.440 --> 1:53:26.120 |
|
is me sort of merging with the machine. |
|
|
|
1:53:26.120 --> 1:53:28.120 |
|
And the way that I personify this |
|
|
|
1:53:28.120 --> 1:53:30.800 |
|
is in a longterm and augments relationship with the machine. |
|
|
|
1:53:32.160 --> 1:53:33.320 |
|
Oh, you don't think there's room |
|
|
|
1:53:33.320 --> 1:53:35.040 |
|
for another human in your life |
|
|
|
1:53:35.040 --> 1:53:37.440 |
|
if you really truly merge with another machine? |
|
|
|
1:53:38.440 --> 1:53:40.240 |
|
I mean, I see merging. |
|
|
|
1:53:40.240 --> 1:53:44.240 |
|
I see like the best interface to my brain |
|
|
|
1:53:45.520 --> 1:53:48.000 |
|
is like the same relationship interface |
|
|
|
1:53:48.000 --> 1:53:49.320 |
|
to merge with an AI, right? |
|
|
|
1:53:49.320 --> 1:53:51.440 |
|
What does that merging feel like? |
|
|
|
1:53:52.440 --> 1:53:55.320 |
|
I've seen couples who've been together for a long time |
|
|
|
1:53:55.320 --> 1:53:57.840 |
|
and like, I almost think of them as one person. |
|
|
|
1:53:57.840 --> 1:54:01.280 |
|
Like couples who spend all their time together and... |
|
|
|
1:54:01.280 --> 1:54:02.120 |
|
That's fascinating. |
|
|
|
1:54:02.120 --> 1:54:03.320 |
|
You're actually putting, |
|
|
|
1:54:03.320 --> 1:54:05.520 |
|
what does that merging actually looks like? |
|
|
|
1:54:05.520 --> 1:54:07.600 |
|
It's not just a nice channel. |
|
|
|
1:54:07.600 --> 1:54:11.640 |
|
Like a lot of people imagine it's just an efficient link, |
|
|
|
1:54:11.640 --> 1:54:13.800 |
|
search link to Wikipedia or something. |
|
|
|
1:54:13.800 --> 1:54:14.640 |
|
I don't believe in that. |
|
|
|
1:54:14.640 --> 1:54:17.120 |
|
But it's more, you're saying that there's the same kind of, |
|
|
|
1:54:17.120 --> 1:54:19.520 |
|
the same kind of relationship you have with another human |
|
|
|
1:54:19.520 --> 1:54:22.960 |
|
as a deep relationship is that's what merging looks like. |
|
|
|
1:54:22.960 --> 1:54:24.480 |
|
That's pretty... |
|
|
|
1:54:24.480 --> 1:54:26.680 |
|
I don't believe that link is possible. |
|
|
|
1:54:26.680 --> 1:54:28.120 |
|
I think that that link, so you're like, |
|
|
|
1:54:28.120 --> 1:54:30.160 |
|
oh, I'm gonna download Wikipedia right to my brain. |
|
|
|
1:54:30.160 --> 1:54:33.360 |
|
My reading speed is not limited by my eyes. |
|
|
|
1:54:33.360 --> 1:54:36.800 |
|
My reading speed is limited by my inner processing loop. |
|
|
|
1:54:36.800 --> 1:54:38.680 |
|
And to like bootstrap that |
|
|
|
1:54:38.680 --> 1:54:42.440 |
|
sounds kind of unclear how to do it and horrifying. |
|
|
|
1:54:42.440 --> 1:54:46.560 |
|
But if I am with somebody, and I'll use somebody |
|
|
|
1:54:46.560 --> 1:54:51.400 |
|
who is making a super sophisticated model of me |
|
|
|
1:54:51.400 --> 1:54:53.200 |
|
and then running simulations on that model, |
|
|
|
1:54:53.200 --> 1:54:54.120 |
|
I'm not gonna get into the question |
|
|
|
1:54:54.120 --> 1:54:55.880 |
|
whether the simulations are conscious or not. |
|
|
|
1:54:55.880 --> 1:54:58.240 |
|
I don't really wanna know what it's doing. |
|
|
|
1:54:58.240 --> 1:55:01.600 |
|
But using those simulations to play out hypothetical futures |
|
|
|
1:55:01.600 --> 1:55:04.880 |
|
for me, deciding what things to say to me |
|
|
|
1:55:04.880 --> 1:55:08.720 |
|
to guide me along a path and that's how I envision it. |
|
|
|
1:55:08.720 --> 1:55:13.720 |
|
So on that path to AI of super human level intelligence, |
|
|
|
1:55:13.720 --> 1:55:15.680 |
|
you've mentioned that you believe in the singularity, |
|
|
|
1:55:15.680 --> 1:55:17.280 |
|
that singularity is coming. |
|
|
|
1:55:17.280 --> 1:55:20.440 |
|
Again, could be trolling, could be not, could be part... |
|
|
|
1:55:20.440 --> 1:55:21.760 |
|
All trolling has truth in it. |
|
|
|
1:55:21.760 --> 1:55:22.840 |
|
I don't know what that means anymore. |
|
|
|
1:55:22.840 --> 1:55:24.520 |
|
What is the singularity? |
|
|
|
1:55:24.520 --> 1:55:26.720 |
|
So yeah, so that's really the question. |
|
|
|
1:55:26.720 --> 1:55:29.280 |
|
How many years do you think before the singularity |
|
|
|
1:55:29.280 --> 1:55:30.920 |
|
of what form do you think it will take? |
|
|
|
1:55:30.920 --> 1:55:34.200 |
|
Does that mean fundamental shifts in capabilities of AI? |
|
|
|
1:55:34.200 --> 1:55:36.960 |
|
Does it mean some other kind of ideas? |
|
|
|
1:55:36.960 --> 1:55:40.120 |
|
Maybe that's just my roots, but... |
|
|
|
1:55:40.120 --> 1:55:42.920 |
|
So I can buy a human being's worth of computers |
|
|
|
1:55:42.920 --> 1:55:46.000 |
|
for things worth of compute for like a million bucks a day. |
|
|
|
1:55:46.000 --> 1:55:47.800 |
|
It's about one TPU pod V3. |
|
|
|
1:55:47.800 --> 1:55:50.240 |
|
I want like, I think they claim a hundred pay to flops. |
|
|
|
1:55:50.240 --> 1:55:51.080 |
|
That's being generous. |
|
|
|
1:55:51.080 --> 1:55:52.320 |
|
I think humans are actually more like 20. |
|
|
|
1:55:52.320 --> 1:55:53.160 |
|
So that's like five humans. |
|
|
|
1:55:53.160 --> 1:55:54.040 |
|
That's pretty good. |
|
|
|
1:55:54.040 --> 1:55:55.560 |
|
Google needs to sell their TPUs. |
|
|
|
1:55:56.840 --> 1:55:58.640 |
|
But no, I could buy GPUs. |
|
|
|
1:55:58.640 --> 1:56:02.280 |
|
I could buy a stack of like, I buy 1080TIs, |
|
|
|
1:56:02.280 --> 1:56:03.880 |
|
build data center full of them. |
|
|
|
1:56:03.880 --> 1:56:07.280 |
|
And for a million bucks, I can get a human worth of compute. |
|
|
|
1:56:08.160 --> 1:56:12.280 |
|
But when you look at the total number of flops in the world, |
|
|
|
1:56:12.280 --> 1:56:14.400 |
|
when you look at human flops, |
|
|
|
1:56:14.400 --> 1:56:17.040 |
|
which goes up very, very slowly with the population, |
|
|
|
1:56:17.040 --> 1:56:19.760 |
|
and machine flops, which goes up exponentially, |
|
|
|
1:56:19.760 --> 1:56:22.360 |
|
but it's still nowhere near. |
|
|
|
1:56:22.360 --> 1:56:24.040 |
|
I think that's the key thing |
|
|
|
1:56:24.040 --> 1:56:25.880 |
|
to talk about when the singularity happened. |
|
|
|
1:56:25.880 --> 1:56:28.560 |
|
When most flops in the world are silicon |
|
|
|
1:56:28.560 --> 1:56:32.280 |
|
and not biological, that's kind of the crossing point. |
|
|
|
1:56:32.280 --> 1:56:35.480 |
|
Like they are now the dominant species on the planet. |
|
|
|
1:56:35.480 --> 1:56:38.720 |
|
And just looking at how technology is progressing, |
|
|
|
1:56:38.720 --> 1:56:40.360 |
|
when do you think that could possibly happen? |
|
|
|
1:56:40.360 --> 1:56:41.680 |
|
Do you think it would happen in your lifetime? |
|
|
|
1:56:41.680 --> 1:56:43.640 |
|
Oh yeah, definitely in my lifetime. |
|
|
|
1:56:43.640 --> 1:56:44.480 |
|
I've done the math. |
|
|
|
1:56:44.480 --> 1:56:47.560 |
|
I like 2038 because it's the UNIX timestamp roll over. |
|
|
|
1:56:49.920 --> 1:56:51.840 |
|
Yeah, beautifully put. |
|
|
|
1:56:52.680 --> 1:56:57.680 |
|
So you've said that the meaning of life is to win. |
|
|
|
1:56:58.000 --> 1:56:59.560 |
|
If you look five years into the future, |
|
|
|
1:56:59.560 --> 1:57:01.000 |
|
what does winning look like? |
|
|
|
1:57:02.640 --> 1:57:03.720 |
|
So... |
|
|
|
1:57:03.720 --> 1:57:08.720 |
|
I can go into technical depth to what I mean by that, to win. |
|
|
|
1:57:11.720 --> 1:57:12.720 |
|
It may not mean... |
|
|
|
1:57:12.720 --> 1:57:14.400 |
|
I was criticized for that in the comments. |
|
|
|
1:57:14.400 --> 1:57:17.720 |
|
Like, doesn't this guy want to save the penguins in Antarctica? |
|
|
|
1:57:17.720 --> 1:57:20.960 |
|
Or like, oh man, listen to what I'm saying. |
|
|
|
1:57:20.960 --> 1:57:23.720 |
|
I'm not talking about like I have a yacht or something. |
|
|
|
1:57:24.720 --> 1:57:26.720 |
|
I am an agent. |
|
|
|
1:57:26.720 --> 1:57:28.720 |
|
I am put into this world. |
|
|
|
1:57:28.720 --> 1:57:32.720 |
|
And I don't really know what my purpose is. |
|
|
|
1:57:33.720 --> 1:57:36.720 |
|
But if you're a reinforcement, if you're an intelligent agent |
|
|
|
1:57:36.720 --> 1:57:39.720 |
|
and you're put into a world, what is the ideal thing to do? |
|
|
|
1:57:39.720 --> 1:57:41.720 |
|
Well, the ideal thing, mathematically, |
|
|
|
1:57:41.720 --> 1:57:43.720 |
|
you can go back to like Schmidt Hoover theories about this, |
|
|
|
1:57:43.720 --> 1:57:46.720 |
|
is to build a compressive model of the world. |
|
|
|
1:57:46.720 --> 1:57:49.720 |
|
To build a maximally compressive to explore the world |
|
|
|
1:57:49.720 --> 1:57:52.720 |
|
such that your exploration function maximizes |
|
|
|
1:57:52.720 --> 1:57:55.720 |
|
the derivative of compression of the past. |
|
|
|
1:57:55.720 --> 1:57:58.720 |
|
Schmidt Hoover has a paper about this. |
|
|
|
1:57:58.720 --> 1:58:01.720 |
|
And like, I took that kind of as like a personal goal function. |
|
|
|
1:58:02.720 --> 1:58:04.720 |
|
So what I mean to win, I mean like, |
|
|
|
1:58:04.720 --> 1:58:08.720 |
|
maybe this is religious, but like I think that in the future |
|
|
|
1:58:08.720 --> 1:58:10.720 |
|
I might be given a real purpose. |
|
|
|
1:58:10.720 --> 1:58:12.720 |
|
Or I may decide this purpose myself. |
|
|
|
1:58:12.720 --> 1:58:14.720 |
|
And then at that point, now I know what the game is |
|
|
|
1:58:14.720 --> 1:58:15.720 |
|
and I know how to win. |
|
|
|
1:58:15.720 --> 1:58:18.720 |
|
I think right now I'm still just trying to figure out what the game is. |
|
|
|
1:58:18.720 --> 1:58:19.720 |
|
But once I know... |
|
|
|
1:58:20.720 --> 1:58:22.720 |
|
So you have... |
|
|
|
1:58:22.720 --> 1:58:25.720 |
|
You have imperfect information. |
|
|
|
1:58:25.720 --> 1:58:27.720 |
|
You have a lot of uncertainty about the reward function |
|
|
|
1:58:27.720 --> 1:58:28.720 |
|
and you're discovering it. |
|
|
|
1:58:28.720 --> 1:58:29.720 |
|
Exactly. |
|
|
|
1:58:29.720 --> 1:58:30.720 |
|
But the purpose is... |
|
|
|
1:58:30.720 --> 1:58:31.720 |
|
That's a better way to put it. |
|
|
|
1:58:31.720 --> 1:58:33.720 |
|
The purpose is to maximize it |
|
|
|
1:58:33.720 --> 1:58:36.720 |
|
while you have a lot of uncertainty around it. |
|
|
|
1:58:36.720 --> 1:58:38.720 |
|
And you're both reducing the uncertainty |
|
|
|
1:58:38.720 --> 1:58:40.720 |
|
and maximizing at the same time. |
|
|
|
1:58:40.720 --> 1:58:43.720 |
|
And so that's at the technical level. |
|
|
|
1:58:43.720 --> 1:58:44.720 |
|
What is the... |
|
|
|
1:58:44.720 --> 1:58:46.720 |
|
If you believe in the universal prior, |
|
|
|
1:58:46.720 --> 1:58:48.720 |
|
what is the universal reward function? |
|
|
|
1:58:48.720 --> 1:58:50.720 |
|
That's the better way to put it. |
|
|
|
1:58:50.720 --> 1:58:53.720 |
|
So that win is interesting. |
|
|
|
1:58:53.720 --> 1:58:56.720 |
|
I think I speak for everyone in saying that |
|
|
|
1:58:56.720 --> 1:59:01.720 |
|
I wonder what that reward function is for you. |
|
|
|
1:59:01.720 --> 1:59:06.720 |
|
And I look forward to seeing that in five years and ten years. |
|
|
|
1:59:06.720 --> 1:59:09.720 |
|
I think a lot of people including myself are cheering you on, man. |
|
|
|
1:59:09.720 --> 1:59:11.720 |
|
So I'm happy you exist. |
|
|
|
1:59:11.720 --> 1:59:13.720 |
|
And I wish you the best of luck. |
|
|
|
1:59:13.720 --> 1:59:14.720 |
|
Thanks for talking today, man. |
|
|
|
1:59:14.720 --> 1:59:15.720 |
|
Thank you. |
|
|
|
1:59:15.720 --> 1:59:20.720 |
|
This was a lot of fun. |
|
|
|
|