diff --git "a/vtt/episode_031_small.vtt" "b/vtt/episode_031_small.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_031_small.vtt" @@ -0,0 +1,9020 @@ +WEBVTT + +00:00.000 --> 00:02.520 + The following is a conversation with George Hotz. + +00:02.520 --> 00:04.480 + He's the founder of Comma AI, + +00:04.480 --> 00:07.400 + a machine learning based vehicle automation company. + +00:07.400 --> 00:10.200 + He is most certainly an outspoken personality + +00:10.200 --> 00:13.160 + in the field of AI and technology in general. + +00:13.160 --> 00:16.240 + He first gained recognition for being the first person + +00:16.240 --> 00:18.400 + to carry on lock and iPhone. + +00:18.400 --> 00:21.280 + And since then, he's done quite a few interesting things + +00:21.280 --> 00:24.400 + at the intersection of hardware and software. + +00:24.400 --> 00:27.440 + This is the artificial intelligence podcast. + +00:27.440 --> 00:29.560 + If you enjoy it, subscribe on YouTube, + +00:29.560 --> 00:32.920 + give it five stars on iTunes, support it on Patreon, + +00:32.920 --> 00:34.920 + or simply connect with me on Twitter. + +00:34.920 --> 00:39.120 + Alex Friedman, spelled F R I D M A N. + +00:39.120 --> 00:42.000 + And I'd like to give a special thank you to Jennifer + +00:42.000 --> 00:45.880 + from Canada for her support of the podcast on Patreon. + +00:45.880 --> 00:47.720 + Merci beaucoup, Jennifer. + +00:47.720 --> 00:50.600 + She's been a friend and an engineering colleague + +00:50.600 --> 00:52.800 + for many years since I was in grad school. + +00:52.800 --> 00:55.520 + Your support means a lot and inspires me + +00:55.520 --> 00:57.920 + to keep this series going. + +00:57.920 --> 01:01.600 + And now here's my conversation with George Hotz. + +01:02.720 --> 01:04.720 + Do you think we're living in a simulation? + +01:06.480 --> 01:10.080 + Yes, but it may be unfalsifiable. + +01:10.080 --> 01:12.440 + What do you mean by unfalsifiable? + +01:12.440 --> 01:16.840 + So if the simulation is designed in such a way + +01:16.840 --> 01:19.640 + that they did like a formal proof + +01:19.640 --> 01:22.320 + to show that no information can get in and out. + +01:22.320 --> 01:25.200 + And if their hardware is designed for the anything + +01:25.200 --> 01:27.880 + in the simulation to always keep the hardware in spec, + +01:27.880 --> 01:29.480 + it may be impossible to prove + +01:29.480 --> 01:31.280 + whether we're in a simulation or not. + +01:32.600 --> 01:35.680 + So they've designed it such that it's a closed system, + +01:35.680 --> 01:37.200 + you can't get outside the system. + +01:37.200 --> 01:38.760 + Well, maybe it's one of three worlds. + +01:38.760 --> 01:41.400 + We're either in a simulation which can be exploited, + +01:41.400 --> 01:44.200 + we're in a simulation which not only can't be exploited, + +01:44.200 --> 01:46.440 + but like the same thing's true about VMs. + +01:46.440 --> 01:48.160 + A really well designed VM, + +01:48.160 --> 01:50.520 + you can't even detect if you're in a VM or not. + +01:51.400 --> 01:52.520 + That's brilliant. + +01:52.520 --> 01:55.160 + So we're, yeah, so the simulation is running + +01:55.160 --> 01:56.800 + on a virtual machine. + +01:56.800 --> 01:59.440 + But now in reality, all VMs have ways to detect. + +01:59.440 --> 02:00.280 + That's the point. + +02:00.280 --> 02:04.840 + I mean, is it, you've done quite a bit of hacking yourself. + +02:04.840 --> 02:08.640 + So you should know that really any complicated system + +02:08.640 --> 02:11.000 + will have ways in and out. + +02:11.000 --> 02:14.240 + So this isn't necessarily true going forward. + +02:15.280 --> 02:18.080 + I spent my time away from comma, + +02:18.080 --> 02:21.240 + I learned a cock, it's a dependently typed, + +02:21.240 --> 02:24.360 + like it's a language for writing math proofs. + +02:24.360 --> 02:28.200 + And if you write code that compiles in a language like that, + +02:28.200 --> 02:30.840 + it is correct by definition. + +02:30.840 --> 02:33.560 + The types check it's correctance. + +02:33.560 --> 02:35.000 + So it's possible that the simulation + +02:35.000 --> 02:39.640 + is written in a language like this, in which case, yeah. + +02:39.640 --> 02:42.680 + Yeah, but that can't be sufficiently expressive + +02:42.680 --> 02:43.760 + of language like that. + +02:43.760 --> 02:44.600 + Oh, it can. + +02:44.600 --> 02:45.440 + It can be? + +02:45.440 --> 02:46.280 + Oh, yeah. + +02:46.280 --> 02:48.920 + Okay, well, so, all right, so. + +02:48.920 --> 02:50.640 + The simulation doesn't have to be tearing complete + +02:50.640 --> 02:52.320 + if it has a scheduled end date. + +02:52.320 --> 02:54.600 + Looks like it does actually with entropy. + +02:54.600 --> 02:58.520 + I mean, I don't think that a simulation + +02:58.520 --> 03:02.200 + that results in something as complicated as the universe + +03:03.080 --> 03:07.280 + would have a formal proof of correctness, right? + +03:08.240 --> 03:09.880 + It's possible, of course. + +03:09.880 --> 03:12.720 + We have no idea how good their tooling is. + +03:12.720 --> 03:14.640 + And we have no idea how complicated + +03:14.640 --> 03:16.280 + the universe computer really is. + +03:16.280 --> 03:17.920 + It may be quite simple. + +03:17.920 --> 03:19.680 + It's just very large, right? + +03:19.680 --> 03:22.160 + It's very, it's definitely very large. + +03:22.160 --> 03:24.480 + But the fundamental rules might be super simple. + +03:24.480 --> 03:26.240 + Yeah, Conway's gonna like kinda stop. + +03:26.240 --> 03:30.320 + Right, so if you could hack, + +03:30.320 --> 03:32.400 + so imagine the simulation that is hackable, + +03:32.400 --> 03:33.640 + if you could hack it, + +03:35.040 --> 03:37.960 + what would you change about the universe? + +03:37.960 --> 03:40.560 + Like how would you approach hacking a simulation? + +03:41.640 --> 03:44.360 + The reason I gave that talk? + +03:44.360 --> 03:46.680 + By the way, I'm not familiar with the talk you gave. + +03:46.680 --> 03:50.160 + I just read that you talked about escaping the simulation + +03:50.160 --> 03:51.280 + or something like that. + +03:51.280 --> 03:52.640 + So maybe you can tell me a little bit + +03:52.640 --> 03:55.360 + about the theme and the message there too. + +03:55.360 --> 03:57.680 + It wasn't a very practical talk + +03:57.680 --> 04:00.600 + about how to actually escape a simulation. + +04:00.600 --> 04:03.320 + It was more about a way of restructuring + +04:03.320 --> 04:05.120 + an us versus them narrative. + +04:05.120 --> 04:10.120 + If we continue on the path we're going with technology, + +04:12.360 --> 04:14.160 + I think we're in big trouble, + +04:14.160 --> 04:16.760 + like as a species and not just as a species, + +04:16.760 --> 04:19.480 + but even as me as an individual member of the species. + +04:19.480 --> 04:23.680 + So if we could change rhetoric to be more like, + +04:23.680 --> 04:24.920 + to think upwards, + +04:26.240 --> 04:29.080 + like to think about that we're in a simulation + +04:29.080 --> 04:30.360 + and how we could get out, + +04:30.360 --> 04:32.640 + already we'd be on the right path. + +04:32.640 --> 04:34.800 + What you actually do once you do that, + +04:34.800 --> 04:37.360 + well, I assume I would have acquired way more intelligence + +04:37.360 --> 04:39.760 + in the process of doing that, so I'll just ask that. + +04:39.760 --> 04:42.080 + So the thinking upwards, + +04:42.080 --> 04:43.760 + what kind of ideas, + +04:43.760 --> 04:45.640 + what kind of breakthrough ideas do you think thinking + +04:45.640 --> 04:47.280 + in that way could inspire? + +04:47.280 --> 04:49.800 + And why did you say upwards? + +04:49.800 --> 04:50.640 + Upwards. + +04:50.640 --> 04:51.480 + Into space? + +04:51.480 --> 04:54.120 + Are you thinking sort of exploration in all forms? + +04:54.120 --> 04:59.120 + The space narrative that held for the modernist generation + +04:59.880 --> 05:02.600 + doesn't hold as well for the postmodern generation. + +05:04.560 --> 05:05.480 + What's the space narrative? + +05:05.480 --> 05:06.520 + Are we talking about the same space? + +05:06.520 --> 05:07.360 + The three dimensional space? + +05:07.360 --> 05:08.840 + No, no, space, like going up space, + +05:08.840 --> 05:10.040 + like building like Elon Musk, + +05:10.040 --> 05:11.160 + like we're going to build rockets, + +05:11.160 --> 05:12.080 + we're going to go to Mars, + +05:12.080 --> 05:13.560 + we're going to colonize the universe. + +05:13.560 --> 05:14.720 + And the narrative you're referring, + +05:14.720 --> 05:16.040 + I was born in the Soviet Union, + +05:16.040 --> 05:18.000 + you're referring to the race to space? + +05:18.000 --> 05:18.840 + The race to space, yeah. + +05:18.840 --> 05:19.680 + Yes, explore, okay. + +05:19.680 --> 05:21.760 + That was a great modernist narrative. + +05:21.760 --> 05:23.360 + Yeah. + +05:23.360 --> 05:26.720 + It doesn't seem to hold the same weight in today's culture. + +05:27.640 --> 05:32.160 + I'm hoping for good postmodern narratives that replace it. + +05:32.160 --> 05:35.560 + So let's think, so you work a lot with AI. + +05:35.560 --> 05:39.080 + So AI is one formulation of that narrative. + +05:39.080 --> 05:40.080 + There could be also, + +05:40.080 --> 05:42.320 + I don't know how much you do in VR and AR. + +05:42.320 --> 05:43.160 + Yeah. + +05:43.160 --> 05:45.160 + That's another, I know less about it, + +05:45.160 --> 05:47.600 + but every time I play with it and our research, + +05:47.600 --> 05:49.640 + it's fascinating, that virtual world. + +05:49.640 --> 05:51.840 + Are you interested in the virtual world? + +05:51.840 --> 05:54.200 + I would like to move to virtual reality. + +05:55.360 --> 05:56.440 + In terms of your work? + +05:56.440 --> 05:58.760 + No, I would like to physically move there. + +05:58.760 --> 06:00.240 + The apartment I can rent in the cloud + +06:00.240 --> 06:03.240 + is way better than the apartment I can rent in the real world. + +06:03.240 --> 06:04.760 + Well, it's all relative, isn't it? + +06:04.760 --> 06:07.280 + Because others will have very nice apartments too, + +06:07.280 --> 06:09.200 + so you'll be inferior in the virtual world as well. + +06:09.200 --> 06:11.320 + But that's not how I view the world, right? + +06:11.320 --> 06:12.440 + I don't view the world. + +06:12.440 --> 06:15.640 + I mean, that's a very like, almost zero summish way + +06:15.640 --> 06:16.480 + to view the world. + +06:16.480 --> 06:18.800 + Say like, my great apartment isn't great + +06:18.800 --> 06:20.400 + because my neighbor has one too. + +06:20.400 --> 06:21.640 + No, my great apartment is great + +06:21.640 --> 06:24.320 + because like, look at this dishwasher, man. + +06:24.320 --> 06:26.640 + You just touch the dish and it's washed, right? + +06:26.640 --> 06:28.680 + And that is great in and of itself + +06:28.680 --> 06:30.120 + if I had the only apartment + +06:30.120 --> 06:31.520 + or if everybody had the apartment. + +06:31.520 --> 06:32.400 + I don't care. + +06:32.400 --> 06:34.760 + So you have fundamental gratitude. + +06:34.760 --> 06:39.080 + The world first learned of Geohot, George Hots + +06:39.080 --> 06:42.280 + in August 2007, maybe before then, + +06:42.280 --> 06:44.080 + but certainly in August 2007 + +06:44.080 --> 06:46.760 + when you were the first person to unlock, + +06:46.760 --> 06:48.880 + carry on lock an iPhone. + +06:48.880 --> 06:50.520 + How did you get into hacking? + +06:50.520 --> 06:53.080 + What was the first system you discovered + +06:53.080 --> 06:55.040 + vulnerabilities for and broke into? + +06:56.240 --> 07:01.240 + So that was really kind of the first thing. + +07:01.640 --> 07:06.640 + I had a book in 2006 called Gray Hat Hacking. + +07:07.480 --> 07:11.000 + And I guess I realized that + +07:11.000 --> 07:13.480 + if you acquired these sort of powers + +07:13.480 --> 07:15.280 + you could control the world. + +07:16.160 --> 07:18.920 + But I didn't really know that much + +07:18.920 --> 07:20.560 + about computers back then. + +07:20.560 --> 07:22.120 + I started with electronics. + +07:22.120 --> 07:24.200 + The first iPhone hack was physical. + +07:24.200 --> 07:25.040 + Cardware. + +07:25.040 --> 07:28.160 + You had to open it up and pull an address line high. + +07:28.160 --> 07:29.960 + And it was because I didn't really know + +07:29.960 --> 07:31.320 + about software exploitation. + +07:31.320 --> 07:32.960 + I learned that all in the next few years + +07:32.960 --> 07:33.920 + and I got very good at it. + +07:33.920 --> 07:36.560 + But back then I knew about like + +07:36.560 --> 07:38.920 + how memory chips are connected to processors and stuff. + +07:38.920 --> 07:41.040 + But you knew about software and programming. + +07:41.040 --> 07:43.200 + You didn't know. + +07:43.200 --> 07:46.160 + Oh really, so your view of the world + +07:46.160 --> 07:49.320 + and computers was physical, was hardware. + +07:49.320 --> 07:52.400 + Actually, if you read the code that I released with that + +07:52.400 --> 07:55.760 + in August 2007, it's atrocious. + +07:55.760 --> 07:56.760 + What language was it? + +07:56.760 --> 07:57.600 + C. + +07:57.600 --> 07:58.440 + C, nice. + +07:58.440 --> 08:01.480 + And in a broken sort of state machine, ask C. + +08:01.480 --> 08:02.960 + I didn't know how to program. + +08:02.960 --> 08:04.160 + Yeah. + +08:04.160 --> 08:06.600 + So how did you learn to program? + +08:07.520 --> 08:08.440 + What was your journey? + +08:08.440 --> 08:10.040 + I mean, we'll talk about it. + +08:10.040 --> 08:12.680 + You've live streamed some of your programming. + +08:12.680 --> 08:14.400 + This chaotic, beautiful mess. + +08:14.400 --> 08:16.480 + How did you arrive at that? + +08:16.480 --> 08:18.640 + Years and years of practice. + +08:18.640 --> 08:22.240 + I interned at Google after, + +08:22.240 --> 08:24.800 + the summer after the iPhone unlock. + +08:24.800 --> 08:26.720 + And I did a contract for them + +08:26.720 --> 08:29.040 + where I built a hardware for Street View + +08:29.040 --> 08:31.760 + and I wrote a software library to interact with it. + +08:32.680 --> 08:34.920 + And it was terrible code. + +08:34.920 --> 08:36.560 + And for the first time I got feedback + +08:36.560 --> 08:38.760 + from people who I respected saying, + +08:38.760 --> 08:41.160 + no, like, don't write code like this. + +08:42.680 --> 08:45.680 + Now, of course, just getting that feedback is not enough. + +08:45.680 --> 08:50.680 + The way that I really got good was, + +08:51.000 --> 08:54.800 + I wanted to write this thing that could emulate + +08:54.800 --> 08:58.440 + and then visualize like arm binaries + +08:58.440 --> 09:00.040 + because I wanted to hack the iPhone better. + +09:00.040 --> 09:01.960 + And I didn't like that I couldn't see what the, + +09:01.960 --> 09:03.800 + I couldn't single step through the processor + +09:03.800 --> 09:05.200 + because I had no debugger on there, + +09:05.200 --> 09:06.640 + especially for the low level things like the boot ROM + +09:06.640 --> 09:07.480 + and the boot loader. + +09:07.480 --> 09:09.440 + So I tried to build this tool to do it. + +09:10.920 --> 09:13.440 + And I built the tool once and it was terrible. + +09:13.440 --> 09:15.120 + I built the tool second times, it was terrible. + +09:15.120 --> 09:16.320 + I built the tool third time. + +09:16.320 --> 09:18.600 + This by the time I was at Facebook, it was kind of okay. + +09:18.600 --> 09:20.560 + And then I built the tool fourth time + +09:20.560 --> 09:22.560 + when I was a Google intern again in 2014. + +09:22.560 --> 09:24.320 + And that was the first time I was like, + +09:24.320 --> 09:25.880 + this is finally usable. + +09:25.880 --> 09:27.120 + How do you pronounce this, Kira? + +09:27.120 --> 09:28.360 + Kira, yeah. + +09:28.360 --> 09:31.840 + So it's essentially the most efficient way + +09:31.840 --> 09:35.720 + to visualize the change of state of the computer + +09:35.720 --> 09:37.200 + as the program is running. + +09:37.200 --> 09:38.920 + That's what you mean by debugger. + +09:38.920 --> 09:41.760 + Yeah, it's a timeless debugger. + +09:41.760 --> 09:45.080 + So you can rewind just as easily as going forward. + +09:45.080 --> 09:46.280 + Think about, if you're using GDB, + +09:46.280 --> 09:47.880 + you have to put a watch on a variable. + +09:47.880 --> 09:49.680 + If you want to see if that variable changes. + +09:49.680 --> 09:51.480 + In Kira, you can just click on that variable. + +09:51.480 --> 09:53.880 + And then it shows every single time + +09:53.880 --> 09:56.520 + when that variable was changed or accessed. + +09:56.520 --> 09:59.760 + Think about it like get for your computer's, the run lock. + +09:59.760 --> 10:04.760 + So there's like a deep log of the state of the computer + +10:05.640 --> 10:07.840 + as the program runs and you can rewind. + +10:07.840 --> 10:11.480 + Why isn't that, maybe it is, maybe you can educate me. + +10:11.480 --> 10:14.640 + Why isn't that kind of debugging used more often? + +10:14.640 --> 10:16.320 + Because the tooling's bad. + +10:16.320 --> 10:17.160 + Well, two things. + +10:17.160 --> 10:19.360 + One, if you're trying to debug Chrome, + +10:19.360 --> 10:22.920 + Chrome is a 200 megabyte binary + +10:22.920 --> 10:25.440 + that runs slowly on desktops. + +10:25.440 --> 10:27.760 + So that's gonna be really hard to use for that. + +10:27.760 --> 10:30.160 + But it's really good to use for like CTFs + +10:30.160 --> 10:33.200 + and for boot ROMs and for small parts of code. + +10:33.200 --> 10:36.360 + So it's hard if you're trying to debug like massive systems. + +10:36.360 --> 10:38.200 + What's a CTF and what's a boot ROM? + +10:38.200 --> 10:40.480 + A boot ROM is the first code that executes + +10:40.480 --> 10:42.280 + the minute you give power to your iPhone. + +10:42.280 --> 10:43.520 + Okay. + +10:43.520 --> 10:46.040 + And CTF were these competitions that I played. + +10:46.040 --> 10:46.880 + Capture the flag. + +10:46.880 --> 10:47.720 + Capture the flag. + +10:47.720 --> 10:48.560 + I was gonna ask you about that. + +10:48.560 --> 10:49.920 + What are those, those look at, + +10:49.920 --> 10:51.440 + I watched a couple of videos on YouTube. + +10:51.440 --> 10:52.920 + Those look fascinating. + +10:52.920 --> 10:55.560 + What have you learned about maybe at the high level + +10:55.560 --> 10:58.040 + in the vulnerability of systems from these competitions? + +11:00.840 --> 11:04.200 + I feel like in the heyday of CTFs, + +11:04.200 --> 11:08.160 + you had all of the best security people in the world + +11:08.160 --> 11:10.720 + challenging each other and coming up + +11:10.720 --> 11:13.640 + with new toy exploitable things over here. + +11:13.640 --> 11:15.400 + And then everybody, okay, who can break it? + +11:15.400 --> 11:17.160 + And when you break it, you get like, + +11:17.160 --> 11:19.360 + there's like a file in the server called flag. + +11:19.360 --> 11:20.960 + And then there's a program running, + +11:20.960 --> 11:22.680 + listening on a socket that's vulnerable. + +11:22.680 --> 11:25.000 + So you write an exploit, you get a shell, + +11:25.000 --> 11:27.160 + and then you cat flag, and then you type the flag + +11:27.160 --> 11:29.480 + into like a web based scoreboard and you get points. + +11:29.480 --> 11:33.000 + So the goal is essentially to find an exploit in the system + +11:33.000 --> 11:35.280 + that allows you to run shell, + +11:35.280 --> 11:38.040 + to run arbitrary code on that system. + +11:38.040 --> 11:40.200 + That's one of the categories. + +11:40.200 --> 11:41.960 + That's like the Poneable category. + +11:43.560 --> 11:44.400 + Poneable? + +11:44.400 --> 11:45.240 + Yeah, Poneable. + +11:45.240 --> 11:47.600 + It's like, you know, you Pone the program. + +11:47.600 --> 11:48.440 + It's a program. + +11:48.440 --> 11:51.760 + Oh, yeah. + +11:51.760 --> 11:55.360 + You know, first of all, I apologize, I'm gonna say, + +11:55.360 --> 11:56.280 + it's because I'm Russian, + +11:56.280 --> 11:59.120 + but maybe you can help educate me. + +12:00.120 --> 12:01.680 + Some video game like misspelled + +12:01.680 --> 12:02.840 + to own way back in the day. + +12:02.840 --> 12:04.880 + Yeah, and it's just, + +12:04.880 --> 12:06.280 + I wonder if there's a definition + +12:06.280 --> 12:08.000 + and I'll have to go to Urban Dictionary for it. + +12:08.000 --> 12:09.800 + Yeah, it'd be interesting to see what it says. + +12:09.800 --> 12:12.760 + Okay, so what was the heyday of CTL, by the way, + +12:12.760 --> 12:15.480 + but was it, what decade are we talking about? + +12:15.480 --> 12:18.400 + I think like, I mean, maybe I'm biased + +12:18.400 --> 12:21.120 + because it's the era that I played, + +12:21.120 --> 12:25.800 + but like 2011 to 2015, + +12:27.200 --> 12:30.320 + because the modern CTF scene + +12:30.320 --> 12:32.640 + is similar to the modern competitive programming scene. + +12:32.640 --> 12:34.280 + You have people who like do drills. + +12:34.280 --> 12:35.880 + You have people who practice. + +12:35.880 --> 12:37.040 + And then once you've done that, + +12:37.040 --> 12:40.040 + you've turned it less into a game of generic computer skill + +12:40.040 --> 12:42.440 + and more into a game of, okay, you memorize, + +12:42.440 --> 12:44.620 + you drill on these five categories. + +12:45.760 --> 12:48.920 + And then before that, it wasn't, + +12:48.920 --> 12:51.560 + it didn't have like as much attention as it had. + +12:52.800 --> 12:53.640 + I don't know, they were like, + +12:53.640 --> 12:55.200 + I won $30,000 once in Korea + +12:55.200 --> 12:56.120 + for one of these competitions. + +12:56.120 --> 12:56.960 + Holy crap. + +12:56.960 --> 12:57.920 + Yeah, they were, they were, that was... + +12:57.920 --> 12:59.520 + So that means, I mean, money is money, + +12:59.520 --> 13:02.320 + but that means there was probably good people there. + +13:02.320 --> 13:03.600 + Exactly, yeah. + +13:03.600 --> 13:06.800 + Are the challenges human constructed + +13:06.800 --> 13:10.760 + or are they grounded in some real flaws in real systems? + +13:10.760 --> 13:13.080 + Usually they're human constructed, + +13:13.080 --> 13:15.760 + but they're usually inspired by real flaws. + +13:15.760 --> 13:17.320 + What kind of systems are imagined + +13:17.320 --> 13:19.080 + is really focused on mobile? + +13:19.080 --> 13:20.920 + Like what has vulnerabilities these days? + +13:20.920 --> 13:25.120 + Is it primarily mobile systems like Android? + +13:25.120 --> 13:26.680 + Oh, everything does. + +13:26.680 --> 13:28.120 + Yeah, of course. + +13:28.120 --> 13:29.360 + The price has kind of gone up + +13:29.360 --> 13:31.280 + because less and less people can find them. + +13:31.280 --> 13:33.160 + And what's happened in security is now, + +13:33.160 --> 13:34.560 + if you want to like jailbreak an iPhone, + +13:34.560 --> 13:36.960 + you don't need one exploit anymore, you need nine. + +13:37.960 --> 13:39.160 + Nine change together? + +13:39.160 --> 13:40.000 + What would you mean? + +13:40.000 --> 13:40.840 + Yeah, wow. + +13:40.840 --> 13:44.800 + Okay, so it's really, what's the benefit? + +13:44.800 --> 13:48.240 + Speaking higher level philosophically about hacking. + +13:48.240 --> 13:50.400 + I mean, it sounds from everything I've seen about you, + +13:50.400 --> 13:55.040 + you just love the challenge and you don't want to do anything. + +13:55.040 --> 13:58.120 + You don't want to bring that exploit out into the world + +13:58.120 --> 14:01.680 + and do any actual, let it run wild. + +14:01.680 --> 14:02.760 + You just want to solve it + +14:02.760 --> 14:05.400 + and then you go on to the next thing. + +14:05.400 --> 14:08.440 + Oh yeah, I mean, doing criminal stuff's not really worth it. + +14:08.440 --> 14:10.520 + And I'll actually use the same argument + +14:10.520 --> 14:15.440 + for why I don't do defense for why I don't do crime. + +14:15.440 --> 14:16.840 + If you want to defend a system, + +14:16.840 --> 14:19.280 + say the system has 10 holes, right? + +14:19.280 --> 14:22.240 + If you find nine of those holes as a defender, + +14:22.240 --> 14:24.240 + you still lose because the attacker gets in + +14:24.240 --> 14:25.520 + through the last one. + +14:25.520 --> 14:26.360 + If you're an attacker, + +14:26.360 --> 14:28.720 + you only have to find one out of the 10. + +14:28.720 --> 14:30.760 + But if you're a criminal, + +14:30.760 --> 14:34.800 + if you log on with a VPN nine out of the 10 times, + +14:34.800 --> 14:37.760 + but one time you forget, you're done. + +14:37.760 --> 14:39.400 + Because you're caught, okay. + +14:39.400 --> 14:41.160 + Because you only have to mess up once + +14:41.160 --> 14:42.920 + to be caught as a criminal. + +14:42.920 --> 14:44.320 + That's why I'm not a criminal. + +14:45.920 --> 14:47.080 + But okay, let me, + +14:47.080 --> 14:49.520 + cause I was having a discussion with somebody + +14:49.520 --> 14:52.440 + just at a high level about nuclear weapons, + +14:52.440 --> 14:56.240 + actually why we're having blown ourselves up yet. + +14:56.240 --> 14:59.840 + And my feeling is all the smart people in the world, + +14:59.840 --> 15:04.120 + if you look at the distribution of smart people, + +15:04.120 --> 15:06.760 + smart people are generally good. + +15:06.760 --> 15:07.680 + And then the Southern person, + +15:07.680 --> 15:09.480 + I was talking to Sean Carroll, the physicist, + +15:09.480 --> 15:11.400 + and he was saying no good and bad people + +15:11.400 --> 15:14.080 + are evenly distributed amongst everybody. + +15:14.080 --> 15:18.080 + My sense was good hackers are in general good people + +15:18.080 --> 15:20.400 + and they don't want to mess with the world. + +15:20.400 --> 15:21.920 + What's your sense? + +15:21.920 --> 15:24.720 + I'm not even sure about that. + +15:25.920 --> 15:30.520 + Like, I have a nice life. + +15:30.520 --> 15:32.120 + Crime wouldn't get me anything. + +15:34.320 --> 15:36.520 + But if you're good and you have these skills, + +15:36.520 --> 15:38.720 + you probably have a nice life too, right? + +15:38.720 --> 15:40.160 + Right, you can use the father things. + +15:40.160 --> 15:41.120 + But is there an ethical, + +15:41.120 --> 15:44.200 + is there a little voice in your head that says, + +15:46.120 --> 15:49.040 + well, yeah, if you could hack something + +15:49.040 --> 15:50.720 + to where you could hurt people + +15:52.840 --> 15:54.960 + and you could earn a lot of money doing it though, + +15:54.960 --> 15:56.320 + not hurt physically perhaps, + +15:56.320 --> 15:59.000 + but disrupt their life in some kind of way. + +16:00.200 --> 16:02.360 + Isn't there a little voice that says, + +16:03.360 --> 16:04.560 + Well, two things. + +16:04.560 --> 16:06.800 + One, I don't really care about money. + +16:06.800 --> 16:08.680 + So like the money wouldn't be an incentive. + +16:08.680 --> 16:10.640 + The thrill might be an incentive. + +16:10.640 --> 16:14.440 + But when I was 19, I read crime and punishment. + +16:14.440 --> 16:16.120 + That was another great one + +16:16.120 --> 16:18.440 + that talked me out of ever really doing crime. + +16:19.400 --> 16:21.720 + Cause it's like, that's gonna be me. + +16:21.720 --> 16:25.040 + I'd get away with it, but it would just run through my head. + +16:25.040 --> 16:26.480 + Even if I got away with it, you know? + +16:26.480 --> 16:27.640 + And then you do crime for long enough, + +16:27.640 --> 16:28.960 + you'll never get away with it. + +16:28.960 --> 16:30.360 + That's right, in the end. + +16:30.360 --> 16:32.680 + That's a good reason to be good. + +16:32.680 --> 16:34.880 + I wouldn't say I'm good, I would just say I'm not bad. + +16:34.880 --> 16:38.080 + You're a talented programmer and a hacker + +16:38.080 --> 16:40.920 + in a good positive sense of the word. + +16:40.920 --> 16:43.360 + You've played around, found vulnerabilities + +16:43.360 --> 16:44.720 + in various systems. + +16:44.720 --> 16:46.120 + What have you learned broadly + +16:46.120 --> 16:49.480 + about the design of systems and so on + +16:49.480 --> 16:51.520 + from that whole process? + +16:53.280 --> 16:58.280 + You learn to not take things + +16:59.280 --> 17:02.160 + for what people say they are, + +17:02.160 --> 17:05.320 + but you look at things for what they actually are. + +17:07.040 --> 17:07.880 + Yeah. + +17:07.880 --> 17:10.080 + I understand that's what you tell me it is, + +17:10.080 --> 17:11.320 + but what does it do? + +17:12.960 --> 17:14.600 + And you have nice visualization tools + +17:14.600 --> 17:16.720 + to really know what it's really doing. + +17:16.720 --> 17:20.080 + Oh, I wish I'm a better programmer now than I was in 2014. + +17:20.080 --> 17:21.880 + I said, Kira, that was the first tool + +17:21.880 --> 17:23.440 + that I wrote that was usable. + +17:23.440 --> 17:25.360 + I wouldn't say the code was great. + +17:25.360 --> 17:27.360 + I still wouldn't say my code is great. + +17:28.840 --> 17:30.760 + So how was your evolution as a programmer? + +17:30.760 --> 17:32.280 + Except practice. + +17:32.280 --> 17:33.880 + You started with C, + +17:33.880 --> 17:35.560 + what point did you pick up Python? + +17:35.560 --> 17:37.080 + Because you're pretty big in Python now. + +17:37.080 --> 17:39.960 + Now, yeah, in college, + +17:39.960 --> 17:42.520 + I went to Carnegie Mellon when I was 22. + +17:42.520 --> 17:44.200 + I went back, I'm like, + +17:44.200 --> 17:46.640 + I'm gonna take all your hardest CS courses + +17:46.640 --> 17:47.640 + and we'll see how I do, right? + +17:47.640 --> 17:48.560 + Like, did I miss anything + +17:48.560 --> 17:51.520 + by not having a real undergraduate education? + +17:51.520 --> 17:54.240 + Took operating systems, compilers, AI, + +17:54.240 --> 17:56.880 + and they're like a freshman Weeder math course. + +17:56.880 --> 18:01.880 + And some of those classes you mentioned, + +18:03.320 --> 18:04.240 + pretty tough, actually. + +18:04.240 --> 18:05.640 + They're great. + +18:05.640 --> 18:07.640 + At least when the 2012, + +18:07.640 --> 18:10.240 + circa 2012 operating systems and compilers + +18:11.240 --> 18:14.440 + were two of the best classes I've ever taken in my life. + +18:14.440 --> 18:15.640 + Because you write an operating system + +18:15.640 --> 18:16.840 + and you write a compiler. + +18:18.080 --> 18:19.760 + I wrote my operating system in C + +18:19.760 --> 18:21.400 + and I wrote my compiler in Haskell, + +18:21.400 --> 18:26.400 + but somehow I picked up Python that semester as well. + +18:26.400 --> 18:28.080 + I started using it for the CTFs, actually. + +18:28.080 --> 18:30.320 + That's when I really started to get into CTFs + +18:30.320 --> 18:33.360 + and CTFs, you're all to race against the clock. + +18:33.360 --> 18:35.120 + So I can't write things and see. + +18:35.120 --> 18:36.240 + Oh, there's a clock component. + +18:36.240 --> 18:37.840 + So you really want to use the programming language + +18:37.840 --> 18:38.960 + just so you can be fastest. + +18:38.960 --> 18:40.080 + 48 hours. + +18:40.080 --> 18:41.440 + Pwn as many of these challenges as you can. + +18:41.440 --> 18:42.280 + Pwn. + +18:42.280 --> 18:43.120 + Yeah. + +18:43.120 --> 18:43.960 + You got like 100 points of challenge, + +18:43.960 --> 18:45.360 + whatever team gets the most. + +18:46.360 --> 18:50.240 + You were both at Facebook and Google for a brief stint. + +18:50.240 --> 18:51.080 + Yeah. + +18:51.080 --> 18:54.920 + With Project Zero, actually, at Google for five months + +18:54.920 --> 18:56.960 + where you develop Kira. + +18:56.960 --> 18:59.280 + What was Project Zero about in general? + +19:01.760 --> 19:05.160 + Just curious about the security efforts in these companies. + +19:05.160 --> 19:08.840 + Well, Project Zero started the same time I went there. + +19:08.840 --> 19:10.080 + What year is it there? + +19:11.080 --> 19:12.320 + 2015. + +19:12.320 --> 19:13.160 + 2015. + +19:13.160 --> 19:15.040 + So that was right at the beginning of Project Zero. + +19:15.040 --> 19:16.200 + It's small. + +19:16.200 --> 19:18.840 + It's Google's offensive security team. + +19:18.840 --> 19:23.840 + I'll try to give the best public facing explanation + +19:25.680 --> 19:26.520 + that I can. + +19:26.520 --> 19:30.960 + So the idea is basically, + +19:30.960 --> 19:33.240 + these vulnerabilities exist in the world. + +19:33.240 --> 19:35.240 + Nation states have them. + +19:35.240 --> 19:37.440 + Some high powered bad actors have them. + +19:39.840 --> 19:44.200 + Sometimes people will find these vulnerabilities + +19:44.200 --> 19:47.960 + and submit them in bug bounties to the companies. + +19:47.960 --> 19:49.440 + But a lot of the companies don't only care. + +19:49.440 --> 19:50.520 + They don't even fix the bug. + +19:50.520 --> 19:53.760 + It doesn't hurt for there to be a vulnerability. + +19:53.760 --> 19:55.880 + So Project Zero is like, we're going to do it different. + +19:55.880 --> 19:57.840 + We're going to announce a vulnerability + +19:57.840 --> 19:59.640 + and we're going to give them 90 days to fix it. + +19:59.640 --> 20:00.800 + And then whether they fix it or not, + +20:00.800 --> 20:03.200 + we're going to drop the Zero Day. + +20:03.200 --> 20:04.080 + Oh, wow. + +20:04.080 --> 20:05.240 + We're going to drop the weapon on the textbook. + +20:05.240 --> 20:06.080 + That's so cool. + +20:06.080 --> 20:07.480 + That is so cool. + +20:07.480 --> 20:09.200 + I love that deadlines. + +20:09.200 --> 20:10.040 + Oh, that's so cool. + +20:10.040 --> 20:10.880 + Give them real deadlines. + +20:10.880 --> 20:12.320 + Yeah. + +20:12.320 --> 20:15.800 + And I think it's done a lot for moving the industry forward. + +20:15.800 --> 20:20.360 + I watched your coding sessions on the streamed online. + +20:20.360 --> 20:25.280 + You code things up, the basic projects, usually from scratch. + +20:25.280 --> 20:28.200 + I would say, sort of as a programmer myself, + +20:28.200 --> 20:30.360 + just watching you, that you type really fast + +20:30.360 --> 20:34.440 + and your brain works in both brilliant and chaotic ways. + +20:34.440 --> 20:35.800 + I don't know if that's always true, + +20:35.800 --> 20:37.600 + but certainly for the live streams. + +20:37.600 --> 20:41.320 + So it's interesting to me because I'm much slower + +20:41.320 --> 20:43.520 + and systematic and careful. + +20:43.520 --> 20:48.040 + And you just move probably in order of magnitude faster. + +20:48.040 --> 20:51.800 + So I'm curious, is there a method to your madness? + +20:51.800 --> 20:53.040 + Or is it just who you are? + +20:53.040 --> 20:54.720 + There's pros and cons. + +20:54.720 --> 20:58.080 + There's pros and cons to my programming style. + +20:58.080 --> 21:00.360 + And I'm aware of them. + +21:00.360 --> 21:04.480 + If you ask me to get something up and working quickly + +21:04.480 --> 21:06.800 + with an API that's kind of undocumented, + +21:06.800 --> 21:08.880 + I will do this super fast because I will throw things + +21:08.880 --> 21:10.200 + at it until it works. + +21:10.200 --> 21:14.720 + If you ask me to take a vector and rotate it 90 degrees + +21:14.720 --> 21:19.320 + and then flip it over the X, Y plane, + +21:19.320 --> 21:22.280 + I'll spam program for two hours and won't get it. + +21:22.280 --> 21:23.480 + Oh, because it's something that you + +21:23.480 --> 21:26.240 + could do with a sheet of paper or think through design + +21:26.240 --> 21:30.400 + and then just you really just throw stuff at the wall + +21:30.400 --> 21:34.600 + and you get so good at it that it usually works. + +21:34.600 --> 21:36.920 + I should become better at the other kind as well. + +21:36.920 --> 21:39.440 + Sometimes I will do things methodically. + +21:39.440 --> 21:41.200 + It's nowhere near as entertaining on the Twitch streams. + +21:41.200 --> 21:43.520 + I do exaggerate it a bit on the Twitch streams as well. + +21:43.520 --> 21:45.480 + The Twitch streams, I mean, what do you want to see a game + +21:45.480 --> 21:46.840 + or you want to see actions permit, right? + +21:46.840 --> 21:48.200 + I'll show you APM for programming too. + +21:48.200 --> 21:50.280 + Yeah, I'd recommend people go to it. + +21:50.280 --> 21:53.800 + I think I watched probably several hours that you put, + +21:53.800 --> 21:57.480 + like I've actually left you programming in the background + +21:57.480 --> 22:00.400 + while I was programming because you made me, + +22:00.400 --> 22:03.120 + it was like watching a really good gamer. + +22:03.120 --> 22:06.240 + It's like energizes you because you're like moving so fast + +22:06.240 --> 22:08.840 + and so it's awesome, it's inspiring. + +22:08.840 --> 22:11.200 + It made me jealous that like, + +22:12.280 --> 22:14.280 + because my own programming is inadequate + +22:14.280 --> 22:16.960 + in terms of speed, so I was like. + +22:16.960 --> 22:20.520 + So I'm twice as frantic on the live streams + +22:20.520 --> 22:22.680 + as I am when I code without, oh. + +22:22.680 --> 22:23.720 + It's super entertaining. + +22:23.720 --> 22:26.400 + So I wasn't even paying attention to what you were coding, + +22:26.400 --> 22:29.760 + which is great, it's just watching you switch windows + +22:29.760 --> 22:31.400 + and Vim, I guess is the most way. + +22:31.400 --> 22:33.000 + Yeah, does Vim on screen? + +22:33.000 --> 22:35.640 + I've developed a workload Facebook and stuck with it. + +22:35.640 --> 22:37.320 + How do you learn new programming tools, + +22:37.320 --> 22:39.440 + ideas, techniques these days? + +22:39.440 --> 22:42.080 + What's your like methodology for learning new things? + +22:42.080 --> 22:45.920 + So I wrote for comma, + +22:47.200 --> 22:49.280 + the distributed file systems out in the world + +22:49.280 --> 22:50.720 + are extremely complex. + +22:50.720 --> 22:55.280 + Like if you want to install something like like like Ceph, + +22:55.280 --> 22:58.760 + Ceph is I think the like open infrastructure + +22:58.760 --> 23:03.040 + distributed file system or there's like newer ones + +23:03.040 --> 23:05.880 + like seaweed FS, but these are all like 10,000 + +23:05.880 --> 23:06.880 + plus line projects. + +23:06.880 --> 23:09.520 + I think some of them are even 100,000 line + +23:09.520 --> 23:11.120 + and just configuring them as a nightmare. + +23:11.120 --> 23:16.120 + So I wrote, I wrote one, it's 200 lines + +23:16.440 --> 23:18.880 + and it uses like engine X of the line servers + +23:18.880 --> 23:21.600 + and has this little master server that I wrote and go. + +23:21.600 --> 23:24.840 + And the way I go, this, if I would say + +23:24.840 --> 23:27.240 + that I'm proud per line of any code I wrote, + +23:27.240 --> 23:29.160 + maybe there's some exploits that I think are beautiful + +23:29.160 --> 23:31.320 + and then this, this is 200 lines + +23:31.320 --> 23:33.720 + and just the way that I thought about it, + +23:33.720 --> 23:35.560 + I think was very good and the reason it's very good + +23:35.560 --> 23:37.640 + is because that was the fourth version of it that I wrote + +23:37.640 --> 23:39.320 + and I had three versions that I threw away. + +23:39.320 --> 23:41.000 + You mentioned, did you say go? + +23:41.000 --> 23:41.840 + I wrote a go, yeah. + +23:41.840 --> 23:42.680 + And go. + +23:42.680 --> 23:43.880 + Is that a functional language? + +23:43.880 --> 23:45.280 + I forget what go is. + +23:45.280 --> 23:47.160 + Go is Google's language. + +23:47.160 --> 23:48.200 + Right. + +23:48.200 --> 23:49.480 + It's not functional. + +23:49.480 --> 23:54.480 + It's some, it's like, in a way it's C++, but easier. + +23:56.160 --> 23:58.200 + It's strongly typed. + +23:58.200 --> 23:59.760 + It has a nice ecosystem around it. + +23:59.760 --> 24:01.680 + When I first looked at it, I was like, + +24:01.680 --> 24:03.800 + this is like Python, but it takes twice as long + +24:03.800 --> 24:05.600 + to do anything. + +24:05.600 --> 24:09.600 + Now that I've open pilot is migrating to C, + +24:09.600 --> 24:11.000 + but it still has large Python components, + +24:11.000 --> 24:12.760 + I now understand why Python doesn't work + +24:12.760 --> 24:15.840 + for large code bases and why you want something like go. + +24:15.840 --> 24:16.680 + Interesting. + +24:16.680 --> 24:18.680 + So why, why doesn't Python work for, + +24:18.680 --> 24:21.720 + so even most, speaking for myself at least, + +24:21.720 --> 24:24.960 + like we do a lot of stuff, basically demo level work + +24:24.960 --> 24:29.240 + with autonomous vehicles and most of the work is Python. + +24:29.240 --> 24:32.440 + Why doesn't Python work for large code bases? + +24:32.440 --> 24:37.440 + Because, well, lack of type checking is a big one. + +24:37.920 --> 24:39.360 + So errors creep in. + +24:39.360 --> 24:41.920 + Yeah, and like you don't know, + +24:41.920 --> 24:45.320 + the compiler can tell you like nothing, right? + +24:45.320 --> 24:48.440 + So everything is either, you know, + +24:48.440 --> 24:49.880 + like syntax errors, fine, + +24:49.880 --> 24:51.800 + but if you misspell a variable in Python, + +24:51.800 --> 24:53.000 + the compiler won't catch that. + +24:53.000 --> 24:56.600 + There's like linters that can catch it some of the time. + +24:56.600 --> 24:57.560 + There's no types. + +24:57.560 --> 25:00.520 + This is really the biggest downside. + +25:00.520 --> 25:02.640 + And then we'll Python slow, but that's not related to it. + +25:02.640 --> 25:04.840 + Well, maybe it's kind of related to it, so it's lack of. + +25:04.840 --> 25:06.600 + So what's in your toolbox these days? + +25:06.600 --> 25:07.760 + Is it Python or what else? + +25:07.760 --> 25:08.600 + Go. + +25:08.600 --> 25:10.240 + I need to move to something else. + +25:10.240 --> 25:12.880 + My adventure into dependently typed languages, + +25:12.880 --> 25:14.240 + I love these languages. + +25:14.240 --> 25:17.520 + They just have like syntax from the 80s. + +25:18.520 --> 25:21.120 + What do you think about JavaScript? + +25:21.120 --> 25:24.000 + ES6, like the modern type script? + +25:24.000 --> 25:27.320 + JavaScript is, the whole ecosystem + +25:27.320 --> 25:29.320 + is unbelievably confusing. + +25:29.320 --> 25:32.840 + NPM updates a package from 022 to 025 + +25:32.840 --> 25:34.560 + and that breaks your Babel linter, + +25:34.560 --> 25:38.560 + which translates your ES5 into ES6, which doesn't run on. + +25:38.560 --> 25:42.480 + So why do I have to compile my JavaScript again, huh? + +25:42.480 --> 25:44.040 + It may be the future though. + +25:44.040 --> 25:45.800 + You think about, I mean, + +25:45.800 --> 25:47.400 + I've embraced JavaScript recently + +25:47.400 --> 25:52.280 + just because just like I've continually embraced PHP, + +25:52.280 --> 25:55.360 + it seems that these worst possible languages live on + +25:55.360 --> 25:57.480 + for the longest, like cockroaches never die. + +25:57.480 --> 26:00.760 + Yeah, well, it's in the browser and it's fast. + +26:00.760 --> 26:01.680 + It's fast. + +26:01.680 --> 26:02.520 + Yeah. + +26:02.520 --> 26:05.480 + It's in the browser and compute might stay become, + +26:05.480 --> 26:06.440 + you know, the browser, + +26:06.440 --> 26:09.040 + it's unclear what the role of the browser is + +26:09.040 --> 26:11.800 + in terms of distributed computation in the future. + +26:11.800 --> 26:12.640 + So. + +26:13.600 --> 26:15.240 + JavaScript is definitely here to stay. + +26:15.240 --> 26:16.080 + Yeah. + +26:16.080 --> 26:18.160 + It's interesting if autonomous vehicles + +26:18.160 --> 26:19.480 + will run on JavaScript one day. + +26:19.480 --> 26:21.760 + I mean, you have to consider these possibilities. + +26:21.760 --> 26:24.280 + Well, all our debug tools are JavaScript. + +26:24.280 --> 26:26.040 + We actually just open source them. + +26:26.040 --> 26:28.160 + We have a tool explorer, which you can annotate + +26:28.160 --> 26:30.080 + your disengagements and we have tool Kibana, + +26:30.080 --> 26:32.920 + which lets you analyze the can traffic from the car. + +26:32.920 --> 26:35.240 + So basically any time you're visualizing something + +26:35.240 --> 26:37.720 + about the log using JavaScript. + +26:37.720 --> 26:40.120 + Well, the web is the best UI toolkit by far. + +26:40.120 --> 26:40.960 + Yeah. + +26:40.960 --> 26:41.880 + So, and then, you know what? + +26:41.880 --> 26:42.760 + You're coding in JavaScript. + +26:42.760 --> 26:43.600 + We have a React guy. + +26:43.600 --> 26:44.440 + He's good. + +26:44.440 --> 26:46.080 + React, nice. + +26:46.080 --> 26:46.920 + Let's get into it. + +26:46.920 --> 26:49.120 + So let's talk autonomous vehicles. + +26:49.120 --> 26:50.640 + You found a comma AI. + +26:51.440 --> 26:54.920 + Let's, at a high level, + +26:54.920 --> 26:57.880 + how did you get into the world of vehicle automation? + +26:57.880 --> 26:59.920 + Can you also just, for people who don't know, + +26:59.920 --> 27:01.400 + tell the story of comma AI? + +27:01.400 --> 27:02.920 + Sure. + +27:02.920 --> 27:06.120 + So I was working at this AI startup + +27:06.120 --> 27:09.240 + and a friend approached me and he's like, + +27:09.240 --> 27:12.080 + dude, I don't know where this is going, + +27:12.080 --> 27:15.160 + but the coolest applied AI problem today + +27:15.160 --> 27:16.480 + is self driving cars. + +27:16.480 --> 27:17.720 + I'm like, well, absolutely. + +27:18.800 --> 27:20.520 + You wanna meet with Elon Musk + +27:20.520 --> 27:24.560 + and he's looking for somebody to build a vision system + +27:24.560 --> 27:27.600 + for autopilot. + +27:27.600 --> 27:29.320 + This is when they were still on AP one. + +27:29.320 --> 27:30.840 + They were still using Mobileye. + +27:30.840 --> 27:33.680 + Elon back then was looking for a replacement. + +27:33.680 --> 27:37.320 + And he brought me in and we talked about a contract + +27:37.320 --> 27:39.040 + where I would deliver something + +27:39.040 --> 27:41.640 + that meets Mobileye level performance. + +27:41.640 --> 27:43.920 + I would get paid $12 million if I could deliver it tomorrow + +27:43.920 --> 27:46.720 + and I would lose $1 million for every month I didn't deliver. + +27:47.720 --> 27:49.080 + So I was like, okay, this is a great deal. + +27:49.080 --> 27:50.800 + This is a super exciting challenge. + +27:52.360 --> 27:53.200 + You know what? + +27:53.200 --> 27:55.840 + It takes me 10 months, I get $2 million, it's good. + +27:55.840 --> 27:57.160 + Maybe I can finish up in five. + +27:57.160 --> 27:58.880 + Maybe I don't finish it at all and I get paid nothing + +27:58.880 --> 28:00.880 + and I'll work for 12 months for free. + +28:00.880 --> 28:02.960 + So maybe just take a pause on that. + +28:02.960 --> 28:04.280 + I'm also curious about this + +28:04.280 --> 28:06.360 + because I've been working in robotics for a long time. + +28:06.360 --> 28:08.320 + And I'm curious to see a person like you just step in + +28:08.320 --> 28:12.000 + and sort of somewhat naive, but brilliant, right? + +28:12.000 --> 28:14.000 + So that's the best place to be + +28:14.000 --> 28:17.240 + because you basically full steam take on a problem. + +28:17.240 --> 28:19.720 + How confident, from that time, + +28:19.720 --> 28:21.320 + because you know a lot more now, + +28:21.320 --> 28:23.440 + at that time, how hard do you think it is + +28:23.440 --> 28:25.880 + to solve all of autonomous driving? + +28:25.880 --> 28:30.440 + I remember I suggested to Elon in the meeting + +28:30.440 --> 28:33.120 + on putting a GPU behind each camera + +28:33.120 --> 28:35.120 + to keep the compute local. + +28:35.120 --> 28:38.000 + This is an incredibly stupid idea. + +28:38.000 --> 28:40.080 + I leave the meeting 10 minutes later and I'm like, + +28:40.080 --> 28:41.560 + I could have spent a little bit of time + +28:41.560 --> 28:42.880 + thinking about this problem before I went in. + +28:42.880 --> 28:44.200 + Why is this a stupid idea? + +28:44.200 --> 28:46.280 + Oh, just send all your cameras to one big GPU. + +28:46.280 --> 28:48.240 + You're much better off doing that. + +28:48.240 --> 28:50.160 + Oh, sorry, you said behind every camera. + +28:50.160 --> 28:51.000 + Every camera. + +28:51.000 --> 28:51.840 + Every small GPU. + +28:51.840 --> 28:52.720 + I was like, oh, I'll put the first few layers + +28:52.720 --> 28:54.520 + of my comms there. + +28:54.520 --> 28:56.080 + Like why did I say that? + +28:56.080 --> 28:56.920 + That's possible. + +28:56.920 --> 28:59.000 + It's possible, but it's a bad idea. + +28:59.000 --> 29:00.480 + It's not obviously a bad idea. + +29:00.480 --> 29:01.320 + Pretty obviously bad. + +29:01.320 --> 29:02.960 + But whether it's actually a bad idea or not, + +29:02.960 --> 29:05.240 + I left that meeting with Elon, like beating myself up. + +29:05.240 --> 29:07.080 + I'm like, why did I say something stupid? + +29:07.080 --> 29:09.360 + Yeah, you haven't, like you haven't at least + +29:09.360 --> 29:12.240 + like thought through every aspect fully. + +29:12.240 --> 29:13.200 + He's very sharp too. + +29:13.200 --> 29:15.760 + Like usually in life, I get away with saying stupid things + +29:15.760 --> 29:16.960 + and then kind of course, + +29:16.960 --> 29:18.560 + right away he called me out about it. + +29:18.560 --> 29:19.800 + And like, usually in life, + +29:19.800 --> 29:21.120 + I get away with saying stupid things. + +29:21.120 --> 29:24.640 + And then like people will, you know, + +29:24.640 --> 29:26.080 + a lot of times people don't even notice. + +29:26.080 --> 29:28.200 + And I'll like correct it and bring the conversation back. + +29:28.200 --> 29:30.600 + But with Elon, it was like, nope, like, okay. + +29:30.600 --> 29:33.520 + Well, that's not at all why the contract fell through. + +29:33.520 --> 29:35.520 + I was much more prepared the second time I met him. + +29:35.520 --> 29:36.360 + Yeah. + +29:36.360 --> 29:39.640 + But in general, how hard did you think it, + +29:39.640 --> 29:43.680 + like 12 months is a tough timeline? + +29:43.680 --> 29:45.720 + Oh, I just thought I'd clone Mobileye IQ three. + +29:45.720 --> 29:47.560 + I didn't think I'd solve level five self driving + +29:47.560 --> 29:48.400 + or anything. + +29:48.400 --> 29:51.000 + So the goal there was to do lane keeping, + +29:51.000 --> 29:52.840 + good lane keeping. + +29:52.840 --> 29:55.560 + I saw my friend showed me the outputs from Mobileye. + +29:55.560 --> 29:57.680 + And the outputs from Mobileye was just basically two lanes + +29:57.680 --> 29:59.440 + and a position of a lead car. + +29:59.440 --> 30:01.560 + I'm like, I can gather a data set + +30:01.560 --> 30:03.440 + and train this net in weeks. + +30:03.440 --> 30:04.840 + And I did. + +30:04.840 --> 30:07.600 + Well, first time I tried the implementation of Mobileye + +30:07.600 --> 30:11.240 + in a Tesla, I was really surprised how good it is. + +30:11.240 --> 30:12.320 + It's quite incredibly good. + +30:12.320 --> 30:14.080 + Cause I thought it's just cause I've done + +30:14.080 --> 30:14.920 + a lot of computer vision. + +30:14.920 --> 30:18.880 + I thought it'd be a lot harder to create a system + +30:18.880 --> 30:20.040 + that that's stable. + +30:21.000 --> 30:22.440 + So I was personally surprised. + +30:22.440 --> 30:25.000 + Just, you know, have to admit it. + +30:25.000 --> 30:27.840 + Cause I was kind of skeptical before trying it. + +30:27.840 --> 30:31.200 + Cause I thought it would go in and out a lot more. + +30:31.200 --> 30:33.160 + It would get disengaged a lot more. + +30:33.160 --> 30:35.000 + And it's pretty robust. + +30:36.200 --> 30:39.720 + So what, how, how, how hard is the problem + +30:39.720 --> 30:42.080 + when you, when you tackled it? + +30:42.080 --> 30:45.760 + So I think AP one was great. Like Elon talked + +30:45.760 --> 30:49.040 + about disengagements on the 405 down in LA + +30:49.040 --> 30:51.040 + with like the lane marks were kind of faded + +30:51.040 --> 30:52.960 + and the Mobileye system would drop out. + +30:53.960 --> 30:57.240 + Like I had something up and working + +30:57.240 --> 31:01.440 + that I would say was like the same quality in three months. + +31:02.480 --> 31:04.560 + Same quality, but how do you know? + +31:04.560 --> 31:07.400 + You say stuff like that confidently, but you can't, + +31:07.400 --> 31:12.120 + and I love it, but the question is you can't, + +31:12.120 --> 31:13.880 + you're kind of going by feel cause you just, + +31:13.880 --> 31:15.560 + You're going by feel, absolutely, absolutely. + +31:15.560 --> 31:17.280 + Like, like I would take, I hadn't, + +31:17.280 --> 31:18.480 + I borrowed my friend's Tesla. + +31:18.480 --> 31:20.760 + I would take AP one out for a drive. + +31:20.760 --> 31:22.320 + And then I would take my system out for a drive. + +31:22.320 --> 31:24.440 + And seems reasonably like the same. + +31:26.080 --> 31:30.480 + So the 405, how hard is it to create something + +31:30.480 --> 31:34.200 + that could actually be a product that's deployed? + +31:34.200 --> 31:39.200 + I mean, I've read an article where Elon, this respond, + +31:39.520 --> 31:41.880 + it said something about you saying that + +31:41.880 --> 31:46.880 + to build autopilot is more complicated + +31:47.080 --> 31:51.880 + than a single George Hodds level job. + +31:51.880 --> 31:55.520 + How hard is that job to create something + +31:55.520 --> 31:57.480 + that would work across the globally? + +31:58.960 --> 32:00.640 + Why don't the global is the challenge, + +32:00.640 --> 32:02.240 + but Elon followed that up by saying + +32:02.240 --> 32:04.920 + it's going to take two years and a company of 10 people. + +32:04.920 --> 32:07.920 + And here I am four years later with a company of 12 people. + +32:07.920 --> 32:09.960 + And I think we still have another two to go. + +32:09.960 --> 32:10.800 + Two years. + +32:10.800 --> 32:13.120 + So yeah, so what do you think, + +32:13.120 --> 32:15.960 + what do you think about how Tesla's progressing + +32:15.960 --> 32:19.200 + with autopilot of V2, V3? + +32:19.200 --> 32:23.120 + I think we've kept pace with them pretty well. + +32:24.080 --> 32:26.880 + I think navigating autopilot is terrible. + +32:26.880 --> 32:31.120 + We had some demo features internally of the same stuff + +32:31.120 --> 32:32.720 + and we would test it and I'm like, + +32:32.720 --> 32:34.720 + I'm not shipping this even as like open source software + +32:34.720 --> 32:35.560 + to people. + +32:35.560 --> 32:37.400 + What do you think is terrible? + +32:37.400 --> 32:39.600 + Consumer Reports does a great job of describing it. + +32:39.600 --> 32:41.240 + Like when it makes a lane change, + +32:41.240 --> 32:43.600 + it does it worse than a human. + +32:43.600 --> 32:46.960 + You shouldn't ship things like autopilot, open pilot, + +32:46.960 --> 32:49.760 + they lane keep better than a human. + +32:49.760 --> 32:53.440 + If you turn it on for a stretch of highway, + +32:53.440 --> 32:56.680 + like an hour long, it's never going to touch a lane line. + +32:56.680 --> 32:59.040 + Human will touch probably a lane line twice. + +32:59.040 --> 33:00.080 + You just inspired me. + +33:00.080 --> 33:02.200 + I don't know if you're grounded in data on that. + +33:02.200 --> 33:03.280 + I read your paper. + +33:03.280 --> 33:05.400 + Okay, but no, but that's interesting. + +33:06.720 --> 33:09.840 + I wonder actually how often we touch lane lines + +33:11.200 --> 33:13.400 + a little bit because it is. + +33:13.400 --> 33:14.960 + I could answer that question pretty easily + +33:14.960 --> 33:15.800 + with the common data side. + +33:15.800 --> 33:16.920 + Yeah, I'm curious. + +33:16.920 --> 33:17.760 + I've never answered it. + +33:17.760 --> 33:18.600 + I don't know. + +33:18.600 --> 33:20.000 + I just too was like my personal. + +33:20.000 --> 33:22.400 + It feels right, but that's interesting + +33:22.400 --> 33:23.800 + because every time you touch a lane, + +33:23.800 --> 33:26.760 + that's a source of a little bit of stress + +33:26.760 --> 33:29.320 + and kind of lane keeping is removing that stress. + +33:29.320 --> 33:31.840 + That's ultimately the biggest value add + +33:31.840 --> 33:34.240 + honestly is just removing the stress + +33:34.240 --> 33:35.480 + of having to stay in lane. + +33:35.480 --> 33:39.040 + And I think I don't think people fully realize + +33:39.040 --> 33:41.960 + first of all that that's a big value add, + +33:41.960 --> 33:45.000 + but also that that's all it is. + +33:45.000 --> 33:48.560 + And that not only I find it a huge value add. + +33:48.560 --> 33:50.440 + I drove down when we moved to San Diego, + +33:50.440 --> 33:52.640 + I drove down in an enterprise rental car + +33:52.640 --> 33:53.480 + and I missed it. + +33:53.480 --> 33:55.480 + So I missed having the system so much. + +33:55.480 --> 33:59.200 + It's so much more tiring to drive + +33:59.200 --> 34:00.320 + without it. + +34:00.320 --> 34:02.960 + It's, it is that lane centering. + +34:02.960 --> 34:04.840 + That's the key feature. + +34:04.840 --> 34:05.680 + Yeah. + +34:06.600 --> 34:08.960 + And in a way it's the only feature + +34:08.960 --> 34:11.040 + that actually adds value to people's lives + +34:11.040 --> 34:12.200 + in autonomous vehicles today. + +34:12.200 --> 34:13.840 + Waymo does not add value to people's lives. + +34:13.840 --> 34:15.880 + It's a more expensive, slower Uber. + +34:15.880 --> 34:18.640 + Maybe someday it'll be this big cliff where it adds value, + +34:18.640 --> 34:19.480 + but I don't usually. + +34:19.480 --> 34:20.320 + It's fascinating. + +34:20.320 --> 34:22.560 + I haven't talked to, this is good. + +34:22.560 --> 34:25.840 + Cause I haven't, I have intuitively, + +34:25.840 --> 34:28.320 + but I think we're making it explicit now. + +34:28.320 --> 34:33.320 + I actually believe that really good lane keeping + +34:35.480 --> 34:37.240 + is a reason to buy a car. + +34:37.240 --> 34:38.440 + Will be a reason to buy a car. + +34:38.440 --> 34:39.720 + It is a huge value add. + +34:39.720 --> 34:41.760 + I've never, until we just started talking about it, + +34:41.760 --> 34:43.880 + haven't really quite realized it, + +34:43.880 --> 34:48.880 + that I've felt with Elon's chase of level four + +34:49.440 --> 34:52.360 + is not the correct chase. + +34:52.360 --> 34:56.000 + It was on, cause you should just say Tesla has the best + +34:56.000 --> 34:58.320 + as if from a Tesla perspective say, + +34:58.320 --> 35:00.600 + Tesla has the best lane keeping. + +35:00.600 --> 35:04.160 + Kama AI should say Kama AI is the best lane keeping. + +35:04.160 --> 35:05.640 + And that is it. + +35:05.640 --> 35:06.480 + Yeah. + +35:06.480 --> 35:07.320 + Yeah. + +35:07.320 --> 35:08.160 + Do you think? + +35:08.160 --> 35:09.920 + You have to do the longitudinal as well. + +35:09.920 --> 35:10.960 + You can't just lane keep. + +35:10.960 --> 35:12.920 + You have to do ACC, + +35:12.920 --> 35:15.840 + but ACC is much more forgiving than lane keep, + +35:15.840 --> 35:17.400 + especially on the highway. + +35:17.400 --> 35:22.000 + By the way, are you Kama AI's camera only, correct? + +35:22.000 --> 35:23.440 + No, we use the radar. + +35:23.440 --> 35:26.960 + We, from the car, you're able to get to, okay. + +35:26.960 --> 35:28.800 + We can do it camera only now. + +35:28.800 --> 35:29.640 + It's gotten to the point, + +35:29.640 --> 35:31.600 + but we leave the radar there as like a, + +35:31.600 --> 35:33.440 + it's fusion now. + +35:33.440 --> 35:35.440 + Okay, so let's maybe talk through + +35:35.440 --> 35:37.920 + some of the system specs on the hardware. + +35:37.920 --> 35:42.880 + What's the hardware side of what you're providing? + +35:42.880 --> 35:44.720 + What's the capabilities on the software side + +35:44.720 --> 35:46.800 + with OpenPilot and so on? + +35:46.800 --> 35:51.800 + So OpenPilot as the box that we sell that it runs on, + +35:51.800 --> 35:53.920 + it's a phone in a plastic case. + +35:53.920 --> 35:54.840 + It's nothing special. + +35:54.840 --> 35:56.200 + We sell it without the software. + +35:56.200 --> 35:57.840 + So you're like, you know, you buy the phone, + +35:57.840 --> 35:58.920 + it's just easy. + +35:58.920 --> 36:00.240 + It'll be easy set up, + +36:00.240 --> 36:01.720 + but it's sold with no software. + +36:03.480 --> 36:06.600 + OpenPilot right now is about to be 0.6. + +36:06.600 --> 36:07.880 + When it gets to 1.0, + +36:07.880 --> 36:09.680 + I think we'll be ready for a consumer product. + +36:09.680 --> 36:11.120 + We're not gonna add any new features. + +36:11.120 --> 36:13.800 + We're just gonna make the lane keeping really, really good. + +36:13.800 --> 36:15.120 + Okay, I got it. + +36:15.120 --> 36:16.120 + So what do we have right now? + +36:16.120 --> 36:18.200 + It's a Snapdragon 820. + +36:18.200 --> 36:23.200 + It's a Sony IMX 298 forward facing camera, + +36:23.680 --> 36:24.720 + driver monitoring camera. + +36:24.720 --> 36:26.400 + It's just a selfie cam on the phone. + +36:26.400 --> 36:30.000 + And a can transceiver, + +36:30.000 --> 36:32.320 + maybe it's a little thing called pandas. + +36:32.320 --> 36:35.040 + And they talk over USB to the phone + +36:35.040 --> 36:36.400 + and then they have three can buses + +36:36.400 --> 36:37.560 + that they talk to the car. + +36:38.560 --> 36:40.920 + One of those can buses is the radar can bus. + +36:40.920 --> 36:42.920 + One of them is the main car can bus. + +36:42.920 --> 36:44.920 + And the other one is the proxy camera can bus. + +36:44.920 --> 36:47.320 + We leave the existing camera in place. + +36:47.320 --> 36:49.560 + So we don't turn AEB off. + +36:49.560 --> 36:51.040 + Right now we still turn AEB off + +36:51.040 --> 36:52.280 + if you're using our longitudinal, + +36:52.280 --> 36:54.320 + but we're gonna fix that before 1.0. + +36:54.320 --> 36:55.160 + Got it. + +36:55.160 --> 36:56.000 + Wow, that's cool. + +36:56.000 --> 36:57.960 + So in its can both ways. + +36:57.960 --> 37:02.120 + So how are you able to control vehicles? + +37:02.120 --> 37:05.520 + So we proxy the vehicles that we work with + +37:05.520 --> 37:08.960 + already have a lane keeping assist system. + +37:08.960 --> 37:12.520 + So lane keeping assist can mean a huge variety of things. + +37:12.520 --> 37:16.120 + It can mean it will apply a small torque + +37:16.120 --> 37:18.920 + to the wheel after you've already crossed a lane line + +37:18.920 --> 37:22.720 + by a foot, which is the system in the older Toyotas. + +37:22.720 --> 37:26.360 + Versus like, I think Tesla still calls it lane keeping assist + +37:26.360 --> 37:28.920 + where it'll keep you perfectly in the center of the lane + +37:28.920 --> 37:29.960 + on the highway. + +37:31.240 --> 37:34.000 + You can control like you with the joystick, the cars. + +37:34.000 --> 37:36.600 + So these cars already have the capability of drive by wire. + +37:36.600 --> 37:41.600 + So is it, is it trivial to convert a car + +37:41.600 --> 37:43.320 + that it operates with? + +37:43.320 --> 37:47.480 + It open pilot is able to control the steering. + +37:48.480 --> 37:49.720 + Oh, a new car or a car that we, + +37:49.720 --> 37:52.800 + so we have support now for 45 different makes of cars. + +37:52.800 --> 37:54.880 + What are the cars in general? + +37:54.880 --> 37:56.360 + Mostly Honda's and Toyotas. + +37:56.360 --> 38:00.640 + We support almost every Honda and Toyota made this year. + +38:01.680 --> 38:04.480 + And then bunch of GM's, bunch of Subaru's. + +38:04.480 --> 38:05.960 + But it doesn't have to be like a Prius. + +38:05.960 --> 38:07.320 + It could be Corolla as well. + +38:07.320 --> 38:10.760 + Oh, the 2020 Corolla is the best car with open pilot. + +38:10.760 --> 38:11.720 + It just came out there. + +38:11.720 --> 38:14.200 + The actuator has less lag than the older Corolla. + +38:15.840 --> 38:18.240 + I think I started watching a video with you. + +38:18.240 --> 38:21.480 + I mean, the way you make videos is awesome. + +38:21.480 --> 38:24.320 + It's just literally at the dealership streaming. + +38:25.320 --> 38:26.160 + I had my friend to follow him. + +38:26.160 --> 38:27.560 + I probably want to stream for an hour. + +38:27.560 --> 38:31.120 + Yeah, and basically like if stuff goes a little wrong, + +38:31.120 --> 38:33.160 + you just like, you just go with it. + +38:33.160 --> 38:34.000 + Yeah, I love it. + +38:34.000 --> 38:34.840 + It's real. + +38:34.840 --> 38:35.680 + Yeah, it's real. + +38:35.680 --> 38:42.000 + That's so beautiful and it's so in contrast to the way + +38:42.000 --> 38:44.600 + other companies would put together a video like that. + +38:44.600 --> 38:46.000 + Kind of why I like to do it like that. + +38:46.000 --> 38:46.840 + Good. + +38:46.840 --> 38:49.720 + I mean, if you become super rich one day and successful, + +38:49.720 --> 38:52.280 + I hope you keep it that way because I think that's actually + +38:52.280 --> 38:54.600 + what people love, that kind of genuine. + +38:54.600 --> 38:56.520 + Oh, it's all that has value to me. + +38:56.520 --> 38:59.840 + Money has no, if I sell out to like make money, + +38:59.840 --> 39:00.680 + I sold out. + +39:00.680 --> 39:01.520 + It doesn't matter. + +39:01.520 --> 39:02.360 + What do I get? + +39:02.360 --> 39:04.440 + Yacht, I don't want a yacht. + +39:04.440 --> 39:09.440 + And I think Tesla actually has a small inkling of that + +39:09.440 --> 39:11.240 + as well with autonomy day. + +39:11.240 --> 39:14.000 + They did reveal more than, I mean, of course, + +39:14.000 --> 39:15.680 + there's marketing communications, you could tell, + +39:15.680 --> 39:17.640 + but it's more than most companies would reveal, + +39:17.640 --> 39:20.960 + which is I hope they go towards that direction + +39:20.960 --> 39:23.000 + more other companies, GM, Ford. + +39:23.000 --> 39:25.400 + Oh, Tesla's going to win level five. + +39:25.400 --> 39:26.560 + They really are. + +39:26.560 --> 39:27.800 + So let's talk about it. + +39:27.800 --> 39:33.000 + You think, you're focused on level two currently, currently. + +39:33.000 --> 39:36.160 + We're going to be one to two years behind Tesla + +39:36.160 --> 39:37.160 + getting to level five. + +39:37.160 --> 39:38.520 + OK. + +39:38.520 --> 39:39.320 + We're Android, right? + +39:39.320 --> 39:39.880 + We're Android. + +39:39.880 --> 39:40.680 + You're Android. + +39:40.680 --> 39:42.240 + I'm just saying once Tesla gets it, + +39:42.240 --> 39:43.440 + we're one to two years behind. + +39:43.440 --> 39:45.680 + I'm not making any timeline on when Tesla's going to get it. + +39:45.680 --> 39:46.120 + That's right. + +39:46.120 --> 39:46.360 + You did. + +39:46.360 --> 39:46.960 + That's brilliant. + +39:46.960 --> 39:48.560 + I'm sorry, Tesla investors, if you + +39:48.560 --> 39:50.520 + think you're going to have an autonomous robot taxi + +39:50.520 --> 39:54.920 + fleet by the end of the year, I'll bet against that. + +39:54.920 --> 39:57.720 + So what do you think about this? + +39:57.720 --> 40:03.280 + The most level four companies are kind of just + +40:03.280 --> 40:08.360 + doing their usual safety driver, doing full autonomy kind + +40:08.360 --> 40:08.800 + of testing. + +40:08.800 --> 40:10.880 + And then Tesla does basically trying + +40:10.880 --> 40:15.280 + to go from lane keeping to full autonomy. + +40:15.280 --> 40:16.840 + What do you think about that approach? + +40:16.840 --> 40:18.360 + How successful would it be? + +40:18.360 --> 40:20.680 + It's a ton better approach. + +40:20.680 --> 40:23.960 + Because Tesla is gathering data on a scale + +40:23.960 --> 40:25.200 + that none of them are. + +40:25.200 --> 40:29.560 + They're putting real users behind the wheel of the cars. + +40:29.560 --> 40:34.440 + It's, I think, the only strategy that works, the incremental. + +40:34.440 --> 40:37.000 + Well, so there's a few components to Tesla approach + +40:37.000 --> 40:38.800 + that's more than just the incremental. + +40:38.800 --> 40:41.400 + What you spoke with is the software, + +40:41.400 --> 40:43.720 + so over the air software updates. + +40:43.720 --> 40:44.800 + Necessity. + +40:44.800 --> 40:46.440 + I mean, Waymo crews have those too. + +40:46.440 --> 40:47.560 + Those aren't. + +40:47.560 --> 40:48.080 + But no. + +40:48.080 --> 40:49.800 + Those differentiate from the automakers. + +40:49.800 --> 40:50.080 + Right. + +40:50.080 --> 40:53.440 + No lane keeping systems have no cars with lane keeping system + +40:53.440 --> 40:54.760 + have that except Tesla. + +40:54.760 --> 40:55.720 + Yeah. + +40:55.720 --> 40:59.760 + And the other one is the data, the other direction, + +40:59.760 --> 41:01.840 + which is the ability to query the data. + +41:01.840 --> 41:03.480 + I don't think they're actually collecting + +41:03.480 --> 41:05.240 + as much data as people think, but the ability + +41:05.240 --> 41:09.440 + to turn on collection and turn it off. + +41:09.440 --> 41:13.400 + So I'm both in the robotics world, in the psychology, + +41:13.400 --> 41:15.000 + human factors world. + +41:15.000 --> 41:17.320 + Many people believe that level two autonomy + +41:17.320 --> 41:20.040 + is problematic because of the human factor. + +41:20.040 --> 41:23.280 + Like the more the task is automated, + +41:23.280 --> 41:25.960 + the more there's a vigilance decrement. + +41:25.960 --> 41:27.200 + You start to fall asleep. + +41:27.200 --> 41:30.480 + You start to become complacent, start texting more and so on. + +41:30.480 --> 41:32.200 + Do you worry about that? + +41:32.200 --> 41:35.000 + Because if you're talking about transition from lane keeping + +41:35.000 --> 41:40.960 + to full autonomy, if you're spending 80% of the time + +41:40.960 --> 41:43.080 + not supervising the machine, do you + +41:43.080 --> 41:47.080 + worry about what that means for the safety of the drivers? + +41:47.080 --> 41:49.640 + One, we don't consider OpenPilot to be 1.0 + +41:49.640 --> 41:52.880 + until we have 100% driver monitoring. + +41:52.880 --> 41:55.000 + You can cheat right now, our driver monitoring system. + +41:55.000 --> 41:56.080 + There's a few ways to cheat it. + +41:56.080 --> 41:58.160 + They're pretty obvious. + +41:58.160 --> 41:59.680 + We're working on making that better. + +41:59.680 --> 42:02.520 + Before we ship a consumer product that can drive cars, + +42:02.520 --> 42:04.240 + I want to make sure that I have driver monitoring + +42:04.240 --> 42:05.440 + that you can't cheat. + +42:05.440 --> 42:09.000 + What's a successful driver monitoring system look like? + +42:09.000 --> 42:11.680 + Is it all about just keeping your eyes on the road? + +42:11.680 --> 42:12.760 + Well, a few things. + +42:12.760 --> 42:16.600 + So that's what we went with at first for driver monitoring. + +42:16.600 --> 42:17.160 + I'm checking. + +42:17.160 --> 42:19.000 + I'm actually looking at where your head is looking. + +42:19.000 --> 42:19.880 + The camera's not that high. + +42:19.880 --> 42:21.840 + Resolution eyes are a little bit hard to get. + +42:21.840 --> 42:22.880 + Well, head is big. + +42:22.880 --> 42:23.560 + I mean, that's just. + +42:23.560 --> 42:24.640 + Head is good. + +42:24.640 --> 42:28.720 + And actually, a lot of it, just psychology wise, + +42:28.720 --> 42:30.720 + to have that monitor constantly there, + +42:30.720 --> 42:33.400 + it reminds you that you have to be paying attention. + +42:33.400 --> 42:35.080 + But we want to go further. + +42:35.080 --> 42:36.760 + We just hired someone full time to come on + +42:36.760 --> 42:37.960 + to do the driver monitoring. + +42:37.960 --> 42:40.600 + I want to detect phone in frame, and I + +42:40.600 --> 42:42.600 + want to make sure you're not sleeping. + +42:42.600 --> 42:44.880 + How much does the camera see of the body? + +42:44.880 --> 42:47.480 + This one, not enough. + +42:47.480 --> 42:48.400 + Not enough. + +42:48.400 --> 42:50.720 + The next one, everything. + +42:50.720 --> 42:52.920 + What's interesting, FishEye, is we're + +42:52.920 --> 42:55.200 + doing just data collection, not real time. + +42:55.200 --> 42:59.200 + But FishEye is a beautiful being able to capture the body. + +42:59.200 --> 43:03.280 + And the smartphone is really the biggest problem. + +43:03.280 --> 43:03.880 + I'll show you. + +43:03.880 --> 43:07.800 + I can show you one of the pictures from our new system. + +43:07.800 --> 43:08.160 + Awesome. + +43:08.160 --> 43:10.520 + So you're basically saying the driver monitoring + +43:10.520 --> 43:13.080 + will be the answer to that. + +43:13.080 --> 43:15.320 + I think the other point that you raised in your paper + +43:15.320 --> 43:16.920 + is good as well. + +43:16.920 --> 43:20.400 + You're not asking a human to supervise a machine + +43:20.400 --> 43:23.920 + without giving them the they can take over at any time. + +43:23.920 --> 43:25.760 + Our safety model, you can take over. + +43:25.760 --> 43:27.720 + We disengage on both the gas or the brake. + +43:27.720 --> 43:28.880 + We don't disengage on steering. + +43:28.880 --> 43:29.920 + I don't feel you have to. + +43:29.920 --> 43:31.720 + But we disengage on gas or brake. + +43:31.720 --> 43:34.240 + So it's very easy for you to take over. + +43:34.240 --> 43:36.400 + And it's very easy for you to reengage. + +43:36.400 --> 43:39.320 + That switching should be super cheap. + +43:39.320 --> 43:40.800 + The cars that require, even autopilot, + +43:40.800 --> 43:42.400 + requires a double press. + +43:42.400 --> 43:44.360 + That's almost, I see, I don't like that. + +43:44.360 --> 43:46.440 + And then the cancel. + +43:46.440 --> 43:48.320 + To cancel in autopilot, you either + +43:48.320 --> 43:49.920 + have to press cancel, which no one knows where that is. + +43:49.920 --> 43:51.000 + So they press the brake. + +43:51.000 --> 43:53.360 + But a lot of times you don't want to press the brake. + +43:53.360 --> 43:54.560 + You want to press the gas. + +43:54.560 --> 43:56.880 + So you should cancel on gas or wiggle the steering wheel, + +43:56.880 --> 43:57.960 + which is bad as well. + +43:57.960 --> 43:58.920 + Wow, that's brilliant. + +43:58.920 --> 44:01.440 + I haven't heard anyone articulate that point. + +44:01.440 --> 44:04.960 + Oh, there's a lot I think about. + +44:04.960 --> 44:09.800 + Because I think actually Tesla has done a better job + +44:09.800 --> 44:12.920 + than most automakers at making that frictionless. + +44:12.920 --> 44:16.600 + But you just described that it could be even better. + +44:16.600 --> 44:19.320 + I love Super Cruise as an experience. + +44:19.320 --> 44:21.120 + Once it's engaged. + +44:21.120 --> 44:22.800 + I don't know if you've used it, but getting the thing + +44:22.800 --> 44:25.040 + to try to engage. + +44:25.040 --> 44:27.480 + Yeah, I've used the driven Super Cruise a lot. + +44:27.480 --> 44:29.680 + So what's your thoughts on the Super Cruise system in general? + +44:29.680 --> 44:32.640 + You disengage Super Cruise, and it falls back to ACC. + +44:32.640 --> 44:34.600 + So my car is still accelerating. + +44:34.600 --> 44:36.280 + It feels weird. + +44:36.280 --> 44:39.000 + Otherwise, when you actually have Super Cruise engaged + +44:39.000 --> 44:41.200 + on the highway, it is phenomenal. + +44:41.200 --> 44:42.320 + We bought that Cadillac. + +44:42.320 --> 44:43.240 + We just sold it. + +44:43.240 --> 44:45.600 + But we bought it just to experience this. + +44:45.600 --> 44:47.440 + And I wanted everyone in the office to be like, + +44:47.440 --> 44:49.360 + this is what we're striving to build. + +44:49.360 --> 44:52.800 + GM pioneering with the driver monitoring. + +44:52.800 --> 44:55.040 + You like their driver monitoring system? + +44:55.040 --> 44:56.440 + It has some bugs. + +44:56.440 --> 45:01.960 + If there's a sun shining back here, it'll be blind to you. + +45:01.960 --> 45:03.360 + But overall, mostly, yeah. + +45:03.360 --> 45:05.960 + That's so cool that you know all this stuff. + +45:05.960 --> 45:09.960 + I don't often talk to people that because it's such a rare car, + +45:09.960 --> 45:10.960 + unfortunately, currently. + +45:10.960 --> 45:12.760 + We bought one explicitly for that. + +45:12.760 --> 45:15.040 + We lost like $25K in the deprecation, + +45:15.040 --> 45:16.720 + but it feels worth it. + +45:16.720 --> 45:21.280 + I was very pleasantly surprised that our GM system + +45:21.280 --> 45:26.320 + was so innovative and really wasn't advertised much, + +45:26.320 --> 45:28.480 + wasn't talked about much. + +45:28.480 --> 45:31.840 + And I was nervous that it would die, that it would disappear. + +45:31.840 --> 45:33.520 + Well, they put it on the wrong car. + +45:33.520 --> 45:35.680 + They should have put it on the bolt and not some weird Cadillac + +45:35.680 --> 45:36.640 + that nobody bought. + +45:36.640 --> 45:39.520 + I think that's going to be into, they're saying at least + +45:39.520 --> 45:41.840 + it's going to be into their entire fleet. + +45:41.840 --> 45:44.320 + So what do you think about, as long as we're + +45:44.320 --> 45:46.920 + on the driver monitoring, what do you think + +45:46.920 --> 45:51.920 + about Elon Musk's claim that driver monitoring is not needed? + +45:51.920 --> 45:53.680 + Normally, I love his claims. + +45:53.680 --> 45:55.560 + That one is stupid. + +45:55.560 --> 45:56.560 + That one is stupid. + +45:56.560 --> 46:00.320 + And he's not going to have his level five fleet + +46:00.320 --> 46:01.320 + by the end of the year. + +46:01.320 --> 46:04.880 + Hopefully, he's like, OK, I was wrong. + +46:04.880 --> 46:06.280 + I'm going to add driver monitoring. + +46:06.280 --> 46:08.240 + Because when these systems get to the point + +46:08.240 --> 46:10.320 + that they're only messing up once every 1,000 miles, + +46:10.320 --> 46:14.080 + you absolutely need driver monitoring. + +46:14.080 --> 46:15.880 + So let me play, because I agree with you, + +46:15.880 --> 46:17.320 + but let me play devil's advocate. + +46:17.320 --> 46:22.440 + One possibility is that without driver monitoring, + +46:22.440 --> 46:29.400 + people are able to self regulate, monitor themselves. + +46:29.400 --> 46:30.680 + So your idea is, I'm just. + +46:30.680 --> 46:34.160 + You're seeing all the people sleeping in Teslas? + +46:34.160 --> 46:35.280 + Yeah. + +46:35.280 --> 46:38.320 + Well, I'm a little skeptical of all the people sleeping + +46:38.320 --> 46:43.960 + in Teslas because I've stopped paying attention to that kind + +46:43.960 --> 46:45.680 + of stuff because I want to see real data. + +46:45.680 --> 46:47.240 + It's too much glorified. + +46:47.240 --> 46:48.720 + It doesn't feel scientific to me. + +46:48.720 --> 46:52.560 + So I want to know how many people are really sleeping + +46:52.560 --> 46:55.080 + in Teslas versus sleeping. + +46:55.080 --> 46:57.640 + I was driving here, sleep deprived, + +46:57.640 --> 46:59.520 + in a car with no automation. + +46:59.520 --> 47:01.040 + I was falling asleep. + +47:01.040 --> 47:02.120 + I agree that it's hypey. + +47:02.120 --> 47:04.840 + It's just like, you know what? + +47:04.840 --> 47:08.480 + If Elon put driver monitoring, my last autopilot experience + +47:08.480 --> 47:12.200 + was I rented a Model 3 in March and drove it around. + +47:12.200 --> 47:13.640 + The wheel thing is annoying. + +47:13.640 --> 47:15.440 + And the reason the wheel thing is annoying. + +47:15.440 --> 47:17.080 + We use the wheel thing as well, but we + +47:17.080 --> 47:18.720 + don't disengage on wheel. + +47:18.720 --> 47:21.720 + For Tesla, you have to touch the wheel just enough + +47:21.720 --> 47:25.320 + to trigger the torque sensor to tell it that you're there, + +47:25.320 --> 47:29.720 + but not enough as to disengage it, which don't use it + +47:29.720 --> 47:30.440 + for two things. + +47:30.440 --> 47:31.360 + Don't disengage on wheel. + +47:31.360 --> 47:32.400 + You don't have to. + +47:32.400 --> 47:35.360 + That whole experience, wow, beautifully put. + +47:35.360 --> 47:38.360 + All those elements, even if you don't have driver monitoring, + +47:38.360 --> 47:41.080 + that whole experience needs to be better. + +47:41.080 --> 47:43.760 + Driver monitoring, I think would make, + +47:43.760 --> 47:46.200 + I mean, I think supercruise is a better experience + +47:46.200 --> 47:48.440 + once it's engaged over autopilot. + +47:48.440 --> 47:51.600 + I think supercruise is a transition to engagement + +47:51.600 --> 47:55.200 + and disengagement are significantly worse. + +47:55.200 --> 47:57.880 + There's a tricky thing, because if I were to criticize + +47:57.880 --> 48:00.800 + supercruise, it's a little too crude. + +48:00.800 --> 48:03.640 + And I think it's like six seconds or something. + +48:03.640 --> 48:06.080 + If you look off road, it'll start warning you. + +48:06.080 --> 48:09.120 + It's some ridiculously long period of time. + +48:09.120 --> 48:14.120 + And just the way, I think it's basically, it's a binary. + +48:15.840 --> 48:17.440 + It should be adapted. + +48:17.440 --> 48:19.880 + Yeah, it needs to learn more about you. + +48:19.880 --> 48:23.160 + It needs to communicate what it sees about you more. + +48:23.160 --> 48:25.800 + I'm not, you know, Tesla shows what it sees + +48:25.800 --> 48:27.160 + about the external world. + +48:27.160 --> 48:29.120 + It would be nice if supercruise would tell us + +48:29.120 --> 48:30.840 + what it sees about the internal world. + +48:30.840 --> 48:31.960 + It's even worse than that. + +48:31.960 --> 48:33.320 + You press the button to engage + +48:33.320 --> 48:35.480 + and it just says supercruise unavailable. + +48:35.480 --> 48:36.320 + Yeah, why? + +48:36.320 --> 48:37.800 + Why? + +48:37.800 --> 48:41.480 + Yeah, that transparency is good. + +48:41.480 --> 48:43.520 + We've renamed the driver monitoring packet + +48:43.520 --> 48:45.360 + to driver state. + +48:45.360 --> 48:46.280 + Driver state. + +48:46.280 --> 48:48.360 + We have car state packet, which has the state of the car + +48:48.360 --> 48:51.040 + and driver state packet, which has state of the driver. + +48:51.040 --> 48:52.240 + So what is it? + +48:52.240 --> 48:54.080 + Estimate their BAC. + +48:54.080 --> 48:54.920 + What's BAC? + +48:54.920 --> 48:55.920 + Blood alcohol, kind of. + +48:57.360 --> 48:59.240 + You think that's possible with computer vision? + +48:59.240 --> 49:00.080 + Absolutely. + +49:02.560 --> 49:04.520 + It's a, to me, it's an open question. + +49:04.520 --> 49:06.600 + I haven't looked into too much. + +49:06.600 --> 49:08.440 + Actually, I quite seriously looked at the literature. + +49:08.440 --> 49:10.840 + It's not obvious to me that from the eyes and so on, + +49:10.840 --> 49:11.680 + you can tell. + +49:11.680 --> 49:13.440 + You might need stuff from the car as well. + +49:13.440 --> 49:15.760 + You might need how they're controlling the car, right? + +49:15.760 --> 49:17.360 + And that's fundamentally at the end of the day + +49:17.360 --> 49:18.640 + what you care about. + +49:18.640 --> 49:21.640 + But I think, especially when people are really drunk, + +49:21.640 --> 49:23.640 + they're not controlling the car nearly as smoothly + +49:23.640 --> 49:25.160 + as they would look at them walking, right? + +49:25.160 --> 49:27.240 + They're, the car is like an extension of the body. + +49:27.240 --> 49:29.360 + So I think you could totally detect. + +49:29.360 --> 49:30.880 + And if you could fix people who are drunk, + +49:30.880 --> 49:32.840 + distracted, asleep, if you fix those three. + +49:32.840 --> 49:35.480 + Yeah, that's a huge, that's huge. + +49:35.480 --> 49:38.240 + So what are the current limitations of OpenPilot? + +49:38.240 --> 49:41.720 + What are the main problems that still need to be solved? + +49:41.720 --> 49:45.440 + We're hopefully fixing a few of them in zero six. + +49:45.440 --> 49:48.400 + We're not as good as autopilot at stop cars. + +49:49.440 --> 49:54.240 + So if you're coming up to a red light at like 55, + +49:55.200 --> 49:56.880 + so it's the radar stopped car problem, + +49:56.880 --> 49:59.200 + which is responsible for two autopilot accidents, + +49:59.200 --> 50:01.480 + it's hard to differentiate a stopped car + +50:01.480 --> 50:03.640 + from a like signpost. + +50:03.640 --> 50:05.320 + Yeah, static object. + +50:05.320 --> 50:07.520 + So you have to fuse, you have to do this visually. + +50:07.520 --> 50:09.600 + There's no way from the radar data to tell the difference. + +50:09.600 --> 50:10.680 + Maybe you can make a map, + +50:10.680 --> 50:13.840 + but I don't really believe in mapping at all anymore. + +50:13.840 --> 50:14.920 + Wait, wait, wait, what? + +50:14.920 --> 50:16.040 + You don't believe in mapping? + +50:16.040 --> 50:16.880 + No. + +50:16.880 --> 50:21.120 + So you're basically, the OpenPilot solution is saying, + +50:21.120 --> 50:22.480 + react to the environment as you see it, + +50:22.480 --> 50:24.480 + just like human doing beings do. + +50:24.480 --> 50:26.200 + And then eventually when you want to do navigate + +50:26.200 --> 50:30.400 + on OpenPilot, I'll train the net to look at ways. + +50:30.400 --> 50:31.360 + I'll run ways in the background, + +50:31.360 --> 50:32.200 + I'll train and come down a way. + +50:32.200 --> 50:33.560 + Are you using GPS at all? + +50:33.560 --> 50:34.840 + We use it to ground truth. + +50:34.840 --> 50:37.440 + We use it to very carefully ground truth the paths. + +50:37.440 --> 50:39.560 + We have a stack which can recover relative + +50:39.560 --> 50:41.800 + to 10 centimeters over one minute. + +50:41.800 --> 50:43.440 + And then we use that to ground truth + +50:43.440 --> 50:45.880 + exactly where the car went in that local part + +50:45.880 --> 50:47.800 + of the environment, but it's all local. + +50:47.800 --> 50:49.160 + How are you testing in general? + +50:49.160 --> 50:51.400 + Just for yourself, like experiments and stuff. + +50:51.400 --> 50:54.000 + Where are you located? + +50:54.000 --> 50:54.840 + San Diego. + +50:54.840 --> 50:55.680 + San Diego. + +50:55.680 --> 50:56.520 + Yeah. + +50:56.520 --> 50:57.360 + Okay. + +50:57.360 --> 50:59.760 + So you basically drive around there, + +50:59.760 --> 51:02.200 + collect some data and watch the performance? + +51:02.200 --> 51:04.800 + We have a simulator now and we have, + +51:04.800 --> 51:06.440 + our simulator is really cool. + +51:06.440 --> 51:08.120 + Our simulator is not, + +51:08.120 --> 51:09.720 + it's not like a Unity based simulator. + +51:09.720 --> 51:11.840 + Our simulator lets us load in real estate. + +51:12.880 --> 51:13.720 + What do you mean? + +51:13.720 --> 51:16.760 + We can load in a drive and simulate + +51:16.760 --> 51:20.280 + what the system would have done on the historical data. + +51:20.280 --> 51:21.480 + Ooh, nice. + +51:22.520 --> 51:24.360 + Interesting. + +51:24.360 --> 51:26.080 + Right now we're only using it for testing, + +51:26.080 --> 51:28.640 + but as soon as we start using it for training. + +51:28.640 --> 51:29.480 + That's it. + +51:29.480 --> 51:30.840 + That's all set up for us. + +51:30.840 --> 51:33.040 + What's your feeling about the real world versus simulation? + +51:33.040 --> 51:34.320 + Do you like simulation for training? + +51:34.320 --> 51:35.720 + If this moves to training? + +51:35.720 --> 51:40.040 + So we have to distinguish two types of simulators, right? + +51:40.040 --> 51:44.720 + There's a simulator that like is completely fake. + +51:44.720 --> 51:46.720 + I could get my car to drive around in GTA. + +51:47.800 --> 51:51.080 + I feel that this kind of simulator is useless. + +51:51.880 --> 51:53.640 + You're never, there's so many. + +51:54.640 --> 51:57.000 + My analogy here is like, okay, fine. + +51:57.000 --> 51:59.920 + You're not solving the computer vision problem, + +51:59.920 --> 52:02.440 + but you're solving the computer graphics problem. + +52:02.440 --> 52:03.280 + Right. + +52:03.280 --> 52:04.600 + And you don't think you can get very far + +52:04.600 --> 52:08.040 + by creating ultra realistic graphics? + +52:08.040 --> 52:10.360 + No, because you can create ultra realistic graphics + +52:10.360 --> 52:13.160 + or the road, now create ultra realistic behavioral models + +52:13.160 --> 52:14.600 + of the other cars. + +52:14.600 --> 52:16.920 + Oh, well, I'll just use myself driving. + +52:16.920 --> 52:18.280 + No, you won't. + +52:18.280 --> 52:21.640 + You need real, you need actual human behavior + +52:21.640 --> 52:23.320 + because that's what you're trying to learn. + +52:23.320 --> 52:25.840 + The driving does not have a spec. + +52:25.840 --> 52:29.920 + The definition of driving is what humans do when they drive. + +52:29.920 --> 52:32.800 + Whatever Waymo does, I don't think it's driving. + +52:32.800 --> 52:33.640 + Right. + +52:33.640 --> 52:36.400 + Well, I think actually Waymo and others, + +52:36.400 --> 52:38.920 + if there's any use for reinforcement learning, + +52:38.920 --> 52:40.360 + I've seen it used quite well. + +52:40.360 --> 52:41.640 + I studied pedestrians a lot too, + +52:41.640 --> 52:44.360 + is try to train models from real data + +52:44.360 --> 52:46.920 + of how pedestrians move and try to use reinforcement learning + +52:46.920 --> 52:50.040 + models to make pedestrians move in human like ways. + +52:50.040 --> 52:53.520 + By that point, you've already gone so many layers, + +52:53.520 --> 52:55.680 + you detected a pedestrian. + +52:55.680 --> 52:59.640 + Did you hand code the feature vector of their state? + +52:59.640 --> 53:00.480 + Right. + +53:00.480 --> 53:02.880 + Did you guys learn anything from computer vision + +53:02.880 --> 53:04.600 + before deep learning? + +53:04.600 --> 53:07.160 + Well, okay, I feel like this is... + +53:07.160 --> 53:10.840 + So perception to you is the sticking point. + +53:10.840 --> 53:13.760 + I mean, what's the hardest part of the stack here? + +53:13.760 --> 53:18.760 + There is no human understandable feature vector + +53:19.680 --> 53:22.000 + separating perception and planning. + +53:23.040 --> 53:25.120 + That's the best way I can put that. + +53:25.120 --> 53:25.960 + There is no... + +53:25.960 --> 53:29.600 + So it's all together and it's a joint problem. + +53:29.600 --> 53:31.480 + So you can take localization. + +53:31.480 --> 53:32.960 + Localization and planning, + +53:32.960 --> 53:34.760 + there is a human understandable feature vector + +53:34.760 --> 53:36.000 + between these two things. + +53:36.000 --> 53:38.720 + I mean, okay, so I have like three degrees position, + +53:38.720 --> 53:40.560 + three degrees orientation and those derivatives, + +53:40.560 --> 53:42.000 + maybe those second derivatives, right? + +53:42.000 --> 53:44.520 + That's human understandable, that's physical. + +53:44.520 --> 53:48.560 + The between perception and planning. + +53:49.520 --> 53:53.600 + So like Waymo has a perception stack and then a planner. + +53:53.600 --> 53:55.560 + And one of the things Waymo does right + +53:55.560 --> 54:00.000 + is they have a simulator that can separate those two. + +54:00.000 --> 54:02.920 + They can like replay their perception data + +54:02.920 --> 54:03.920 + and test their system, + +54:03.920 --> 54:04.880 + which is what I'm talking about + +54:04.880 --> 54:06.520 + about like the two different kinds of simulators. + +54:06.520 --> 54:08.240 + There's the kind that can work on real data + +54:08.240 --> 54:10.920 + and there's the kind that can't work on real data. + +54:10.920 --> 54:13.880 + Now, the problem is that I don't think + +54:13.880 --> 54:16.160 + you can hand code a feature vector, right? + +54:16.160 --> 54:17.360 + Like you have some list of like, + +54:17.360 --> 54:19.040 + well, here's my list of cars in the scenes. + +54:19.040 --> 54:21.280 + Here's my list of pedestrians in the scene. + +54:21.280 --> 54:23.240 + This isn't what humans are doing. + +54:23.240 --> 54:24.920 + What are humans doing? + +54:24.920 --> 54:25.760 + Global. + +54:27.200 --> 54:28.040 + Some, some. + +54:28.040 --> 54:31.960 + You're saying that's too difficult to hand engineer. + +54:31.960 --> 54:34.120 + I'm saying that there is no state vector. + +54:34.120 --> 54:36.560 + Given a perfect, I could give you the best team + +54:36.560 --> 54:38.520 + of engineers in the world to build a perception system + +54:38.520 --> 54:40.640 + and the best team to build a planner. + +54:40.640 --> 54:42.640 + All you have to do is define the state vector + +54:42.640 --> 54:43.960 + that separates those two. + +54:43.960 --> 54:48.560 + I'm missing the state vector that separates those two. + +54:48.560 --> 54:49.400 + What do you mean? + +54:49.400 --> 54:54.000 + So what is the output of your perception system? + +54:54.000 --> 54:56.880 + Output of the perception system. + +54:56.880 --> 55:01.560 + It's, there's, okay, well, there's several ways to do it. + +55:01.560 --> 55:03.840 + One is the slam component is localization. + +55:03.840 --> 55:05.920 + The other is drivable area, drivable space. + +55:05.920 --> 55:06.760 + Drivable space, yep. + +55:06.760 --> 55:09.000 + And then there's the different objects in the scene. + +55:09.000 --> 55:09.840 + Yep. + +55:11.000 --> 55:16.000 + And different objects in the scene over time maybe + +55:16.000 --> 55:18.720 + to give you input to then try to start + +55:18.720 --> 55:21.560 + modeling the trajectories of those objects. + +55:21.560 --> 55:22.400 + Sure. + +55:22.400 --> 55:23.240 + That's it. + +55:23.240 --> 55:25.160 + I can give you a concrete example of something you missed. + +55:25.160 --> 55:26.000 + What's that? + +55:26.000 --> 55:28.640 + So say there's a bush in the scene. + +55:28.640 --> 55:30.920 + Humans understand that when they see this bush + +55:30.920 --> 55:34.680 + that there may or may not be a car behind that bush. + +55:34.680 --> 55:37.280 + Drivable area and a list of objects does not include that. + +55:37.280 --> 55:38.920 + Humans are doing this constantly + +55:38.920 --> 55:40.920 + at the simplest intersections. + +55:40.920 --> 55:43.880 + So now you have to talk about occluded area. + +55:43.880 --> 55:44.720 + Right. + +55:44.720 --> 55:47.800 + Right, but even that, what do you mean by occluded? + +55:47.800 --> 55:49.640 + Okay, so I can't see it. + +55:49.640 --> 55:51.840 + Well, if it's the other side of a house, I don't care. + +55:51.840 --> 55:53.560 + What's the likelihood that there's a car + +55:53.560 --> 55:55.280 + in that occluded area, right? + +55:55.280 --> 55:58.080 + And if you say, okay, we'll add that, + +55:58.080 --> 56:00.680 + I can come up with 10 more examples that you can't add. + +56:01.680 --> 56:03.960 + Certainly occluded area would be something + +56:03.960 --> 56:06.760 + that simulator would have because it's simulating + +56:06.760 --> 56:11.320 + the entire, you know, occlusion is part of it. + +56:11.320 --> 56:12.680 + Occlusion is part of a vision stack. + +56:12.680 --> 56:13.520 + Vision stack. + +56:13.520 --> 56:16.600 + But what I'm saying is if you have a hand engineered, + +56:16.600 --> 56:20.040 + if your perception system output can be written + +56:20.040 --> 56:22.240 + in a spec document, it is incomplete. + +56:23.120 --> 56:27.800 + Yeah, I mean, I certainly, it's hard to argue with that + +56:27.800 --> 56:30.120 + because in the end, that's going to be true. + +56:30.120 --> 56:31.760 + Yeah, and I'll tell you what the output + +56:31.760 --> 56:32.720 + of our perception system is. + +56:32.720 --> 56:33.560 + What's that? + +56:33.560 --> 56:37.120 + It's a 1024 dimensional vector. + +56:37.120 --> 56:38.000 + Transparent neural net. + +56:38.000 --> 56:39.000 + Oh, you know that. + +56:39.000 --> 56:42.000 + No, that's the 1024 dimensions of who knows what. + +56:43.520 --> 56:45.160 + Because it's operating on real data. + +56:45.160 --> 56:47.000 + Yeah. + +56:47.000 --> 56:48.320 + And that's the perception. + +56:48.320 --> 56:50.360 + That's the perception state, right? + +56:50.360 --> 56:53.520 + Think about an autoencoder for faces, right? + +56:53.520 --> 56:54.720 + If you have an autoencoder for faces + +56:54.720 --> 56:59.720 + and you say it has 256 dimensions in the middle, + +56:59.720 --> 57:00.680 + and I'm taking a face over here + +57:00.680 --> 57:02.800 + and projecting it to a face over here. + +57:02.800 --> 57:05.360 + Can you hand label all 256 of those dimensions? + +57:06.280 --> 57:09.240 + Well, no, but those are generated automatically. + +57:09.240 --> 57:11.360 + But even if you tried to do it by hand, + +57:11.360 --> 57:15.520 + could you come up with a spec between your encoder + +57:15.520 --> 57:16.360 + and your decoder? + +57:17.400 --> 57:20.720 + No, no, because it wasn't designed, but they're... + +57:20.720 --> 57:22.600 + No, no, no, but if you could design it, + +57:23.600 --> 57:26.480 + if you could design a face reconstructor system, + +57:26.480 --> 57:28.080 + could you come up with a spec? + +57:29.240 --> 57:32.320 + No, but I think we're missing here a little bit. + +57:32.320 --> 57:35.080 + I think you're just being very poetic + +57:35.080 --> 57:37.880 + about expressing a fundamental problem of simulators, + +57:37.880 --> 57:41.640 + that they are going to be missing so much + +57:42.480 --> 57:44.680 + that the feature of actually + +57:44.680 --> 57:47.080 + would just look fundamentally different + +57:47.080 --> 57:50.440 + from in the simulated world than the real world. + +57:51.280 --> 57:53.800 + I'm not making a claim about simulators. + +57:53.800 --> 57:57.120 + I'm making a claim about the spec division + +57:57.120 --> 57:58.800 + between perception and planning. + +57:58.800 --> 57:59.640 + And planning. + +57:59.640 --> 58:00.840 + Even in your system. + +58:00.840 --> 58:01.800 + Just in general. + +58:01.800 --> 58:03.360 + Right, just in general. + +58:03.360 --> 58:05.680 + If you're trying to build a car that drives, + +58:05.680 --> 58:07.280 + if you're trying to hand code + +58:07.280 --> 58:08.760 + the output of your perception system, + +58:08.760 --> 58:10.960 + like saying, here's a list of all the cars in the scene. + +58:10.960 --> 58:11.920 + Here's a list of all the people. + +58:11.920 --> 58:13.120 + Here's a list of the occluded areas. + +58:13.120 --> 58:14.920 + Here's a vector of drivable areas. + +58:14.920 --> 58:16.600 + It's insufficient. + +58:16.600 --> 58:18.000 + And if you start to believe that, + +58:18.000 --> 58:20.840 + you realize that what Waymo and Cruz are doing is impossible. + +58:20.840 --> 58:24.320 + Currently, what we're doing is the perception problem + +58:24.320 --> 58:28.200 + is converting the scene into a chessboard. + +58:29.200 --> 58:31.720 + And then you reason some basic reasoning + +58:31.720 --> 58:33.400 + around that chessboard. + +58:33.400 --> 58:38.080 + And you're saying that really there's a lot missing there. + +58:38.080 --> 58:40.240 + First of all, why are we talking about this? + +58:40.240 --> 58:42.840 + Because isn't this a full autonomy? + +58:42.840 --> 58:44.720 + Is this something you think about? + +58:44.720 --> 58:47.680 + Oh, I want to win self driving cars. + +58:47.680 --> 58:52.680 + So your definition of win includes the full five. + +58:53.680 --> 58:55.800 + I don't think level four is a real thing. + +58:55.800 --> 58:59.720 + I want to build the AlphaGo of driving. + +58:59.720 --> 59:04.720 + So AlphaGo is really end to end. + +59:06.160 --> 59:07.000 + Yeah. + +59:07.000 --> 59:09.840 + Is, yeah, it's end to end. + +59:09.840 --> 59:12.480 + And do you think this whole problem, + +59:12.480 --> 59:14.680 + is that also kind of what you're getting at + +59:14.680 --> 59:16.640 + with the perception and the planning? + +59:16.640 --> 59:19.440 + Is that this whole problem, the right way to do it, + +59:19.440 --> 59:21.600 + is really to learn the entire thing? + +59:21.600 --> 59:23.680 + I'll argue that not only is it the right way, + +59:23.680 --> 59:27.640 + it's the only way that's going to exceed human performance. + +59:27.640 --> 59:29.960 + Well, it's certainly true for Go. + +59:29.960 --> 59:31.520 + Everyone who tried to hand code Go things + +59:31.520 --> 59:33.440 + built human inferior things. + +59:33.440 --> 59:36.200 + And then someone came along and wrote some 10,000 line thing + +59:36.200 --> 59:39.800 + that doesn't know anything about Go that beat everybody. + +59:39.800 --> 59:41.080 + It's 10,000 lines. + +59:41.080 --> 59:43.360 + True, in that sense. + +59:43.360 --> 59:47.520 + The open question then that maybe I can ask you + +59:47.520 --> 59:52.520 + is driving is much harder than Go. + +59:53.440 --> 59:56.240 + The open question is how much harder? + +59:56.240 --> 59:59.480 + So how, because I think the Elon Musk approach here + +59:59.480 --> 1:00:01.600 + with planning and perception is similar + +1:00:01.600 --> 1:00:02.960 + to what you're describing, + +1:00:02.960 --> 1:00:07.960 + which is really turning into not some kind of modular thing, + +1:00:08.280 --> 1:00:11.120 + but really do formulate as a learning problem + +1:00:11.120 --> 1:00:13.360 + and solve the learning problem with scale. + +1:00:13.360 --> 1:00:17.120 + So how many years, put one, + +1:00:17.120 --> 1:00:18.880 + how many years would it take to solve this problem + +1:00:18.880 --> 1:00:21.680 + or just how hard is this freaking problem? + +1:00:21.680 --> 1:00:24.560 + Well, the cool thing is, + +1:00:24.560 --> 1:00:27.800 + I think there's a lot of value + +1:00:27.800 --> 1:00:29.840 + that we can deliver along the way. + +1:00:30.840 --> 1:00:35.840 + I think that you can build lame keeping assist + +1:00:36.600 --> 1:00:41.440 + actually plus adaptive cruise control plus, okay, + +1:00:41.440 --> 1:00:46.000 + looking at ways extends to like all of driving. + +1:00:46.000 --> 1:00:47.920 + Yeah, most of driving, right? + +1:00:47.920 --> 1:00:49.760 + Oh, your adaptive cruise control treats red lights + +1:00:49.760 --> 1:00:51.200 + like cars, okay. + +1:00:51.200 --> 1:00:53.480 + So let's jump around with you mentioned + +1:00:53.480 --> 1:00:55.760 + that you didn't like navigate an autopilot. + +1:00:55.760 --> 1:00:57.760 + What advice, how would you make it better? + +1:00:57.760 --> 1:01:00.560 + Do you think as a feature that if it's done really well, + +1:01:00.560 --> 1:01:02.360 + it's a good feature? + +1:01:02.360 --> 1:01:07.360 + I think that it's too reliant on like hand coded hacks + +1:01:07.520 --> 1:01:10.400 + for like, how does navigate an autopilot do a lane change? + +1:01:10.400 --> 1:01:13.400 + It actually does the same lane change every time + +1:01:13.400 --> 1:01:14.320 + and it feels mechanical. + +1:01:14.320 --> 1:01:15.920 + Humans do different lane changes. + +1:01:15.920 --> 1:01:17.360 + Humans, sometimes we'll do a slow one, + +1:01:17.360 --> 1:01:18.920 + sometimes do a fast one. + +1:01:18.920 --> 1:01:20.880 + Navigate an autopilot at least every time I use it + +1:01:20.880 --> 1:01:23.040 + is it the identical lane change? + +1:01:23.040 --> 1:01:24.280 + How do you learn? + +1:01:24.280 --> 1:01:26.800 + I mean, this is a fundamental thing actually + +1:01:26.800 --> 1:01:30.400 + is the breaking and accelerating, + +1:01:30.400 --> 1:01:33.960 + something that still, Tesla probably does it better + +1:01:33.960 --> 1:01:36.800 + than most cars, but it still doesn't do a great job + +1:01:36.800 --> 1:01:39.960 + of creating a comfortable natural experience + +1:01:39.960 --> 1:01:42.680 + and navigate an autopilot is just lane changes + +1:01:42.680 --> 1:01:44.120 + and extension of that. + +1:01:44.120 --> 1:01:49.120 + So how do you learn to do natural lane change? + +1:01:49.120 --> 1:01:52.920 + So we have it and I can talk about how it works. + +1:01:52.920 --> 1:01:57.920 + So I feel that we have the solution for lateral + +1:01:58.720 --> 1:02:00.640 + but we don't yet have the solution for longitudinal. + +1:02:00.640 --> 1:02:03.360 + There's a few reasons longitudinal is harder than lateral. + +1:02:03.360 --> 1:02:06.920 + The lane change component, the way that we train on it + +1:02:06.920 --> 1:02:10.840 + very simply is like our model has an input + +1:02:10.840 --> 1:02:14.040 + for whether it's doing a lane change or not. + +1:02:14.040 --> 1:02:16.360 + And then when we train the end to end model, + +1:02:16.360 --> 1:02:19.560 + we hand label all the lane changes because you have to. + +1:02:19.560 --> 1:02:22.440 + I've struggled a long time about not wanting to do that + +1:02:22.440 --> 1:02:24.280 + but I think you have to. + +1:02:24.280 --> 1:02:25.320 + Or the training data. + +1:02:25.320 --> 1:02:26.520 + For the training data, right? + +1:02:26.520 --> 1:02:28.280 + We actually have an automatic ground truth + +1:02:28.280 --> 1:02:30.600 + or which automatically labels all the lane changes. + +1:02:30.600 --> 1:02:31.680 + Was that possible? + +1:02:31.680 --> 1:02:32.720 + To automatically label lane changes? + +1:02:32.720 --> 1:02:33.560 + Yeah. + +1:02:33.560 --> 1:02:34.800 + And detect the lane I see when it crosses it, right? + +1:02:34.800 --> 1:02:36.680 + And I don't have to get that high percent accuracy + +1:02:36.680 --> 1:02:38.080 + but it's like 95 good enough. + +1:02:38.080 --> 1:02:38.960 + Okay. + +1:02:38.960 --> 1:02:43.200 + Now I set the bit when it's doing the lane change + +1:02:43.200 --> 1:02:44.840 + in the end to end learning. + +1:02:44.840 --> 1:02:47.920 + And then I set it to zero when it's not doing a lane change. + +1:02:47.920 --> 1:02:49.720 + So now if I want us to do a lane change a test time, + +1:02:49.720 --> 1:02:52.360 + I just put the bit to a one and it'll do a lane change. + +1:02:52.360 --> 1:02:54.640 + Yeah, but so if you look at the space of lane change, + +1:02:54.640 --> 1:02:57.320 + you know some percentage, not a hundred percent, + +1:02:57.320 --> 1:03:01.120 + that we make as humans is not a pleasant experience + +1:03:01.120 --> 1:03:02.800 + because we messed some part of it up. + +1:03:02.800 --> 1:03:04.320 + It's nerve wracking to change. + +1:03:04.320 --> 1:03:05.760 + If you look, you have to see, + +1:03:05.760 --> 1:03:06.920 + it has to accelerate. + +1:03:06.920 --> 1:03:09.920 + How do we label the ones that are natural and feel good? + +1:03:09.920 --> 1:03:11.560 + You know, that's the, + +1:03:11.560 --> 1:03:13.360 + because that's your ultimate criticism, + +1:03:13.360 --> 1:03:17.000 + the current navigate and autopilot just doesn't feel good. + +1:03:17.000 --> 1:03:18.520 + Well, the current navigate and autopilot + +1:03:18.520 --> 1:03:21.720 + is a hand coded policy written by an engineer in a room + +1:03:21.720 --> 1:03:25.080 + who probably went out and tested it a few times on the 280. + +1:03:25.080 --> 1:03:28.560 + Probably a more, a better version of that. + +1:03:28.560 --> 1:03:29.400 + But yes. + +1:03:29.400 --> 1:03:30.560 + That's how we would have written it. + +1:03:30.560 --> 1:03:31.400 + Yeah. + +1:03:31.400 --> 1:03:33.480 + Maybe Tesla did a Tesla, they tested it in. + +1:03:33.480 --> 1:03:34.920 + That might have been two engineers. + +1:03:34.920 --> 1:03:35.760 + Two engineers. + +1:03:35.760 --> 1:03:37.400 + Yeah. + +1:03:37.400 --> 1:03:40.120 + No, but so if you learn the lane change, + +1:03:40.120 --> 1:03:42.480 + if you learn how to do a lane change from data, + +1:03:42.480 --> 1:03:44.680 + just like you have a label that says lane change + +1:03:44.680 --> 1:03:48.040 + and then you put it in when you want it to do the lane change, + +1:03:48.040 --> 1:03:49.640 + it'll automatically do the lane change + +1:03:49.640 --> 1:03:51.600 + that's appropriate for the situation. + +1:03:51.600 --> 1:03:54.720 + Now, to get at the problem of some humans + +1:03:54.720 --> 1:03:55.960 + do bad lane changes, + +1:03:57.400 --> 1:03:59.920 + we haven't worked too much on this problem yet. + +1:03:59.920 --> 1:04:03.120 + It's not that much of a problem in practice. + +1:04:03.120 --> 1:04:06.160 + My theory is that all good drivers are good in the same way + +1:04:06.160 --> 1:04:08.440 + and all bad drivers are bad in different ways. + +1:04:09.360 --> 1:04:11.320 + And we've seen some data to back this up. + +1:04:11.320 --> 1:04:12.400 + Well, beautifully put. + +1:04:12.400 --> 1:04:16.560 + So you just basically, if that's true hypothesis, + +1:04:16.560 --> 1:04:19.920 + then your task is to discover the good drivers. + +1:04:19.920 --> 1:04:21.800 + The good drivers stand out + +1:04:21.800 --> 1:04:23.360 + because they're in one cluster + +1:04:23.360 --> 1:04:25.200 + and the bad drivers are scattered all over the place + +1:04:25.200 --> 1:04:27.240 + and your net learns the cluster. + +1:04:27.240 --> 1:04:28.080 + Yeah. + +1:04:28.080 --> 1:04:30.800 + So you just learn from the good drivers + +1:04:30.800 --> 1:04:32.200 + and they're easy to cluster. + +1:04:33.200 --> 1:04:34.240 + In fact, we learned from all of them + +1:04:34.240 --> 1:04:35.840 + and the net automatically learns the policy + +1:04:35.840 --> 1:04:36.920 + that's like the majority. + +1:04:36.920 --> 1:04:38.440 + But we'll eventually probably have to build some out. + +1:04:38.440 --> 1:04:41.560 + So if that theory is true, I hope it's true + +1:04:41.560 --> 1:04:46.440 + because the counter theory is there is many clusters, + +1:04:49.480 --> 1:04:53.680 + maybe arbitrarily many clusters of good drivers. + +1:04:53.680 --> 1:04:55.840 + Because if there's one cluster of good drivers, + +1:04:55.840 --> 1:04:57.600 + you can at least discover a set of policies. + +1:04:57.600 --> 1:04:59.000 + You can learn a set of policies + +1:04:59.000 --> 1:05:00.640 + which would be good universally. + +1:05:00.640 --> 1:05:01.640 + Yeah. + +1:05:01.640 --> 1:05:04.560 + That would be nice if it's true. + +1:05:04.560 --> 1:05:06.560 + And you're saying that there is some evidence that... + +1:05:06.560 --> 1:05:09.720 + Let's say lane changes can be clustered into four clusters. + +1:05:09.720 --> 1:05:10.560 + Right. + +1:05:10.560 --> 1:05:12.040 + There's a finite level of... + +1:05:12.040 --> 1:05:15.280 + I would argue that all four of those are good clusters. + +1:05:15.280 --> 1:05:18.360 + All the things that are random are noise and probably bad. + +1:05:18.360 --> 1:05:20.360 + And which one of the four you pick? + +1:05:20.360 --> 1:05:21.920 + Or maybe it's 10 or maybe it's 20. + +1:05:21.920 --> 1:05:22.760 + You can learn that. + +1:05:22.760 --> 1:05:23.800 + It's context dependent. + +1:05:23.800 --> 1:05:25.040 + It depends on the scene. + +1:05:26.760 --> 1:05:30.440 + And the hope is it's not too dependent on the driver. + +1:05:31.400 --> 1:05:34.240 + Yeah, the hope is that it all washes out. + +1:05:34.240 --> 1:05:36.960 + The hope is that the distribution is not bimodal. + +1:05:36.960 --> 1:05:39.080 + The hope is that it's a nice Gaussian. + +1:05:39.080 --> 1:05:41.640 + So what advice would you give to Tesla? + +1:05:41.640 --> 1:05:45.000 + How to fix, how to improve, navigate an autopilot? + +1:05:45.000 --> 1:05:48.240 + That's the lessons that you've learned from Kamii. + +1:05:48.240 --> 1:05:50.560 + The only real advice I would give to Tesla + +1:05:50.560 --> 1:05:52.920 + is please put driver monitoring in your cars. + +1:05:53.920 --> 1:05:55.160 + With respect to improving it. + +1:05:55.160 --> 1:05:56.000 + You can't do that anymore. + +1:05:56.000 --> 1:05:57.280 + I started to interrupt. + +1:05:57.280 --> 1:06:01.760 + But there's a practical nature of many of hundreds of thousands + +1:06:01.760 --> 1:06:05.760 + of cars being produced that don't have a good driver facing camera. + +1:06:05.760 --> 1:06:07.520 + The Model 3 has a selfie cam. + +1:06:07.520 --> 1:06:08.680 + Is it not good enough? + +1:06:08.680 --> 1:06:10.800 + Did they not have put IR LEDs for night? + +1:06:10.800 --> 1:06:11.640 + That's a good question. + +1:06:11.640 --> 1:06:13.360 + But I do know that it's fish eye + +1:06:13.360 --> 1:06:15.800 + and it's relatively low resolution. + +1:06:15.800 --> 1:06:16.760 + So it's really not designed. + +1:06:16.760 --> 1:06:18.760 + It wasn't designed for driver monitoring. + +1:06:18.760 --> 1:06:21.760 + You can hope that you can kind of scrape up + +1:06:21.760 --> 1:06:24.400 + and have something from it. + +1:06:24.400 --> 1:06:27.520 + But why didn't they put it in today? + +1:06:27.520 --> 1:06:28.280 + Put it in today. + +1:06:28.280 --> 1:06:29.520 + Put it in today. + +1:06:29.520 --> 1:06:31.520 + Every time I've heard Carpathian talk about the problem + +1:06:31.520 --> 1:06:33.240 + and talking about like software 2.0 + +1:06:33.240 --> 1:06:35.240 + and how the machine learning is gobbling up everything, + +1:06:35.240 --> 1:06:37.440 + I think this is absolutely the right strategy. + +1:06:37.440 --> 1:06:40.160 + I think that he didn't write and navigate on autopilot. + +1:06:40.160 --> 1:06:43.240 + I think somebody else did and kind of hacked it on top of that stuff. + +1:06:43.240 --> 1:06:45.680 + I think when Carpathian says, wait a second, + +1:06:45.680 --> 1:06:47.440 + why did we hand code this lane change policy + +1:06:47.440 --> 1:06:48.360 + with all these magic numbers? + +1:06:48.360 --> 1:06:49.360 + We're going to learn it from data. + +1:06:49.360 --> 1:06:49.840 + They'll fix it. + +1:06:49.840 --> 1:06:51.040 + They already know what to do there. + +1:06:51.040 --> 1:06:54.360 + Well, that's Andre's job is to turn everything + +1:06:54.360 --> 1:06:57.480 + into a learning problem and collect a huge amount of data. + +1:06:57.480 --> 1:07:01.120 + The reality is, though, not every problem + +1:07:01.120 --> 1:07:04.080 + can be turned into a learning problem in the short term. + +1:07:04.080 --> 1:07:07.280 + In the end, everything will be a learning problem. + +1:07:07.280 --> 1:07:12.880 + The reality is, like if you want to build L5 vehicles today, + +1:07:12.880 --> 1:07:15.600 + it will likely involve no learning. + +1:07:15.600 --> 1:07:20.320 + And that's the reality is, so at which point does learning start? + +1:07:20.320 --> 1:07:23.480 + It's the crutch statement that LiDAR is a crutch. + +1:07:23.480 --> 1:07:27.240 + Which point will learning get up to part of human performance? + +1:07:27.240 --> 1:07:31.960 + It's over human performance on ImageNet, classification, + +1:07:31.960 --> 1:07:34.000 + on driving, it's a question still. + +1:07:34.000 --> 1:07:35.760 + It is a question. + +1:07:35.760 --> 1:07:39.160 + I'll say this, I'm here to play for 10 years. + +1:07:39.160 --> 1:07:40.280 + I'm not here to try to. + +1:07:40.280 --> 1:07:42.960 + I'm here to play for 10 years and make money along the way. + +1:07:42.960 --> 1:07:45.040 + I'm not here to try to promise people + +1:07:45.040 --> 1:07:47.600 + that I'm going to have my L5 taxi network up and working + +1:07:47.600 --> 1:07:48.200 + in two years. + +1:07:48.200 --> 1:07:49.400 + Do you think that was a mistake? + +1:07:49.400 --> 1:07:50.520 + Yes. + +1:07:50.520 --> 1:07:53.160 + What do you think was the motivation behind saying + +1:07:53.160 --> 1:07:56.640 + that other companies are also promising L5 vehicles + +1:07:56.640 --> 1:08:01.880 + with their different approaches in 2020, 2021, 2022? + +1:08:01.880 --> 1:08:05.720 + If anybody would like to bet me that those things do not pan out, + +1:08:05.720 --> 1:08:07.000 + I will bet you. + +1:08:07.000 --> 1:08:10.800 + Even money, even money, I'll bet you as much as you want. + +1:08:10.800 --> 1:08:13.600 + So are you worried about what's going to happen? + +1:08:13.600 --> 1:08:16.040 + Because you're not in full agreement on that. + +1:08:16.040 --> 1:08:19.160 + What's going to happen when 2022, 2021 come around + +1:08:19.160 --> 1:08:22.800 + and nobody has fleets of autonomous vehicles? + +1:08:22.800 --> 1:08:25.000 + Well, you can look at the history. + +1:08:25.000 --> 1:08:26.880 + If you go back five years ago, they + +1:08:26.880 --> 1:08:29.880 + were all promised by 2018 and 2017. + +1:08:29.880 --> 1:08:32.200 + But they weren't that strong of promises. + +1:08:32.200 --> 1:08:36.240 + I mean, Ford really declared. + +1:08:36.240 --> 1:08:40.560 + I think not many have declared as definitively + +1:08:40.560 --> 1:08:42.600 + as they have now these dates. + +1:08:42.600 --> 1:08:43.320 + Well, OK. + +1:08:43.320 --> 1:08:45.040 + So let's separate L4 and L5. + +1:08:45.040 --> 1:08:46.800 + Do I think that it's possible for Waymo + +1:08:46.800 --> 1:08:50.960 + to continue to hack on their system + +1:08:50.960 --> 1:08:53.400 + until it gets to level four in Chandler, Arizona? + +1:08:53.400 --> 1:08:55.040 + Yes. + +1:08:55.040 --> 1:08:56.800 + No safety driver? + +1:08:56.800 --> 1:08:57.600 + Chandler, Arizona? + +1:08:57.600 --> 1:08:59.600 + Yeah. + +1:08:59.600 --> 1:09:02.440 + By which year are we talking about? + +1:09:02.440 --> 1:09:06.120 + Oh, I even think that's possible by like 2020, 2021. + +1:09:06.120 --> 1:09:09.480 + But level four, Chandler, Arizona, not level five, + +1:09:09.480 --> 1:09:11.480 + New York City. + +1:09:11.480 --> 1:09:15.920 + Level four, meaning some very defined streets. + +1:09:15.920 --> 1:09:17.400 + It works out really well. + +1:09:17.400 --> 1:09:18.280 + Very defined streets. + +1:09:18.280 --> 1:09:20.680 + And then practically, these streets are pretty empty. + +1:09:20.680 --> 1:09:24.680 + If most of the streets are covered in Waymos, + +1:09:24.680 --> 1:09:28.360 + Waymo can kind of change the definition of what driving is. + +1:09:28.360 --> 1:09:28.920 + Right? + +1:09:28.920 --> 1:09:31.720 + If your self driving network is the majority + +1:09:31.720 --> 1:09:34.120 + of cars in an area, they only need + +1:09:34.120 --> 1:09:35.720 + to be safe with respect to each other, + +1:09:35.720 --> 1:09:38.640 + and all the humans will need to learn to adapt to them. + +1:09:38.640 --> 1:09:41.120 + Now go drive in downtown New York. + +1:09:41.120 --> 1:09:42.200 + Oh, yeah, that's. + +1:09:42.200 --> 1:09:43.440 + I mean, already. + +1:09:43.440 --> 1:09:46.040 + You can talk about autonomy and like on farms, + +1:09:46.040 --> 1:09:48.520 + it already works great, because you can really just + +1:09:48.520 --> 1:09:51.320 + follow the GPS line. + +1:09:51.320 --> 1:09:56.800 + So what does success look like for Kama AI? + +1:09:56.800 --> 1:09:58.200 + What are the milestones like where + +1:09:58.200 --> 1:09:59.800 + you can sit back with some champagne + +1:09:59.800 --> 1:10:04.120 + and say, we did it, boys and girls? + +1:10:04.120 --> 1:10:06.320 + Well, it's never over. + +1:10:06.320 --> 1:10:07.800 + Yeah, but don't be so. + +1:10:07.800 --> 1:10:10.400 + You must drink champagne every time you celebrate. + +1:10:10.400 --> 1:10:11.440 + So what is good? + +1:10:11.440 --> 1:10:13.160 + What are some wins? + +1:10:13.160 --> 1:10:19.480 + A big milestone that we're hoping for by mid next year + +1:10:19.480 --> 1:10:20.680 + is profitability of the company. + +1:10:20.680 --> 1:10:28.560 + And we're going to have to revisit the idea of selling + +1:10:28.560 --> 1:10:30.280 + a consumer product. + +1:10:30.280 --> 1:10:32.720 + But it's not going to be like the Kama One. + +1:10:32.720 --> 1:10:36.240 + When we do it, it's going to be perfect. + +1:10:36.240 --> 1:10:39.600 + OpenPilot has gotten so much better in the last two years. + +1:10:39.600 --> 1:10:41.680 + We're going to have a few features. + +1:10:41.680 --> 1:10:43.760 + We're going to have 100% driver monitoring. + +1:10:43.760 --> 1:10:46.720 + We're going to disable no safety features in the car. + +1:10:46.720 --> 1:10:48.760 + Actually, I think it'd be really cool what we're doing right + +1:10:48.760 --> 1:10:51.600 + now, our project this week is we're analyzing the data set + +1:10:51.600 --> 1:10:53.240 + and looking for all the AEB triggers + +1:10:53.240 --> 1:10:55.640 + from the manufacturer systems. + +1:10:55.640 --> 1:10:59.440 + We have better data set on that than the manufacturers. + +1:10:59.440 --> 1:11:02.960 + How much does Toyota have 10 million miles of real world + +1:11:02.960 --> 1:11:05.360 + driving to know how many times they're AEB triggered? + +1:11:05.360 --> 1:11:10.880 + So let me give you, because you asked, financial advice. + +1:11:10.880 --> 1:11:12.440 + Because I work with a lot of automakers + +1:11:12.440 --> 1:11:15.840 + and one possible source of money for you, + +1:11:15.840 --> 1:11:21.400 + which I'll be excited to see you take on, is basically + +1:11:21.400 --> 1:11:29.120 + selling the data, which is something that most people, + +1:11:29.120 --> 1:11:31.800 + and not selling in a way where here, here at Automaker, + +1:11:31.800 --> 1:11:33.000 + but creating. + +1:11:33.000 --> 1:11:35.480 + We've done this actually at MIT, not for money purposes, + +1:11:35.480 --> 1:11:37.760 + but you could do it for significant money purposes + +1:11:37.760 --> 1:11:39.440 + and make the world a better place + +1:11:39.440 --> 1:11:44.240 + by creating a consortia where automakers would pay in + +1:11:44.240 --> 1:11:46.960 + and then they get to have free access to the data. + +1:11:46.960 --> 1:11:52.400 + And I think a lot of people are really hungry for that + +1:11:52.400 --> 1:11:54.200 + and would pay significant amount of money for it. + +1:11:54.200 --> 1:11:55.400 + Here's the problem with that. + +1:11:55.400 --> 1:11:56.840 + I like this idea all in theory. + +1:11:56.840 --> 1:11:59.640 + It'd be very easy for me to give them access to my servers. + +1:11:59.640 --> 1:12:02.280 + And we already have all open source tools to access this data. + +1:12:02.280 --> 1:12:03.400 + It's in a great format. + +1:12:03.400 --> 1:12:05.560 + We have a great pipeline. + +1:12:05.560 --> 1:12:07.120 + But they're going to put me in the room + +1:12:07.120 --> 1:12:10.120 + with some business development guy. + +1:12:10.120 --> 1:12:12.400 + And I'm going to have to talk to this guy. + +1:12:12.400 --> 1:12:15.040 + And he's not going to know most of the words I'm saying. + +1:12:15.040 --> 1:12:17.280 + I'm not willing to tolerate that. + +1:12:17.280 --> 1:12:18.840 + OK, Mick Jagger. + +1:12:18.840 --> 1:12:19.800 + No, no, no, no. + +1:12:19.800 --> 1:12:21.040 + But I think I agree with you. + +1:12:21.040 --> 1:12:21.720 + I'm the same way. + +1:12:21.720 --> 1:12:22.960 + But you just tell them the terms + +1:12:22.960 --> 1:12:24.640 + and there's no discussion needed. + +1:12:24.640 --> 1:12:30.480 + If I could just tell them the terms, then like, all right. + +1:12:30.480 --> 1:12:31.600 + Who wants access to my data? + +1:12:31.600 --> 1:12:36.680 + I will sell it to you for, let's say, + +1:12:36.680 --> 1:12:37.640 + you want a subscription? + +1:12:37.640 --> 1:12:40.680 + I'll sell you for 100k a month. + +1:12:40.680 --> 1:12:41.200 + Anyone? + +1:12:41.200 --> 1:12:42.000 + 100k a month? + +1:12:42.000 --> 1:12:43.040 + 100k a month? + +1:12:43.040 --> 1:12:45.080 + I'll give you access to the data subscription? + +1:12:45.080 --> 1:12:45.680 + Yeah. + +1:12:45.680 --> 1:12:46.680 + Yeah, I think that's kind of fair. + +1:12:46.680 --> 1:12:48.440 + Came up with that number off the top of my head. + +1:12:48.440 --> 1:12:50.840 + If somebody sends me like a three line email where it's like, + +1:12:50.840 --> 1:12:54.000 + we would like to pay 100k a month to get access to your data. + +1:12:54.000 --> 1:12:56.160 + We would agree to like reasonable privacy terms + +1:12:56.160 --> 1:12:58.360 + of the people who are in the data set. + +1:12:58.360 --> 1:12:59.520 + I would be happy to do it. + +1:12:59.520 --> 1:13:01.200 + But that's not going to be the email. + +1:13:01.200 --> 1:13:03.120 + The email is going to be, hey, do you + +1:13:03.120 --> 1:13:05.560 + have some time in the next month where we can sit down + +1:13:05.560 --> 1:13:07.000 + and we can, I don't have time for that. + +1:13:07.000 --> 1:13:08.360 + We're moving too fast. + +1:13:08.360 --> 1:13:10.040 + You could politely respond to that email, + +1:13:10.040 --> 1:13:13.240 + but not saying I don't have any time for your bullshit. + +1:13:13.240 --> 1:13:15.440 + You say, oh, well, unfortunately, these are the terms. + +1:13:15.440 --> 1:13:19.280 + And so this is what we try to, we brought the cost down + +1:13:19.280 --> 1:13:22.320 + for you in order to minimize the friction, the communication. + +1:13:22.320 --> 1:13:22.920 + Yeah, absolutely. + +1:13:22.920 --> 1:13:26.720 + Here's the whatever it is, $1, $2 million a year. + +1:13:26.720 --> 1:13:28.880 + And you have access. + +1:13:28.880 --> 1:13:31.440 + And it's not like I get that email from like, + +1:13:31.440 --> 1:13:32.720 + but OK, am I going to reach out? + +1:13:32.720 --> 1:13:34.200 + Am I going to hire a business development person + +1:13:34.200 --> 1:13:35.840 + who's going to reach out to the automakers? + +1:13:35.840 --> 1:13:36.480 + No way. + +1:13:36.480 --> 1:13:36.880 + Yeah. + +1:13:36.880 --> 1:13:37.840 + OK, I got you. + +1:13:37.840 --> 1:13:38.520 + I admire. + +1:13:38.520 --> 1:13:39.680 + If they reached into me, I'm not + +1:13:39.680 --> 1:13:40.600 + going to ignore the email. + +1:13:40.600 --> 1:13:42.160 + I'll come back with something like, yeah, + +1:13:42.160 --> 1:13:44.560 + if you're willing to pay $100,000 for access to the data, + +1:13:44.560 --> 1:13:46.080 + I'm happy to set that up. + +1:13:46.080 --> 1:13:48.200 + That's worth my engineering time. + +1:13:48.200 --> 1:13:49.520 + That's actually quite insightful of you. + +1:13:49.520 --> 1:13:50.440 + You're right. + +1:13:50.440 --> 1:13:52.480 + Probably because many of the automakers + +1:13:52.480 --> 1:13:54.480 + are quite a bit old school, there + +1:13:54.480 --> 1:13:56.200 + will be a need to reach out. + +1:13:56.200 --> 1:13:58.440 + And they want it, but there will need + +1:13:58.440 --> 1:13:59.800 + to be some communication. + +1:13:59.800 --> 1:14:00.160 + You're right. + +1:14:00.160 --> 1:14:06.760 + Mobileye circa 2015 had the lowest R&D spend of any chipmaker. + +1:14:06.760 --> 1:14:10.640 + Like per, and you look at all the people who work for them, + +1:14:10.640 --> 1:14:12.120 + and it's all business development people + +1:14:12.120 --> 1:14:15.320 + because the car companies are impossible to work with. + +1:14:15.320 --> 1:14:17.880 + Yeah, so you have no patience for that, + +1:14:17.880 --> 1:14:20.040 + and you're a legit Android, huh? + +1:14:20.040 --> 1:14:21.440 + I have something to do, right? + +1:14:21.440 --> 1:14:24.040 + Like, it's not like I don't mean to be a dick and say, + +1:14:24.040 --> 1:14:25.920 + I don't have patience for that, but it's like, + +1:14:25.920 --> 1:14:29.160 + that stuff doesn't help us with our goal of winning + +1:14:29.160 --> 1:14:30.560 + self driving cars. + +1:14:30.560 --> 1:14:33.800 + If I want money in the short term, + +1:14:33.800 --> 1:14:38.040 + if I showed off the actual learning tech that we have, + +1:14:38.040 --> 1:14:40.160 + it's somewhat sad. + +1:14:40.160 --> 1:14:43.000 + It's years and years ahead of everybody else's. + +1:14:43.000 --> 1:14:43.720 + Maybe not Tesla's. + +1:14:43.720 --> 1:14:45.720 + I think Tesla has similar stuff to us, actually. + +1:14:45.720 --> 1:14:47.640 + I think Tesla has similar stuff, but when you compare it + +1:14:47.640 --> 1:14:50.920 + to what the Toyota Research Institute has, + +1:14:50.920 --> 1:14:53.480 + you're not even close to what we have. + +1:14:53.480 --> 1:14:55.840 + No comments, but I also can't. + +1:14:55.840 --> 1:14:58.440 + I have to take your comments. + +1:14:58.440 --> 1:15:01.960 + I intuitively believe you, but I have + +1:15:01.960 --> 1:15:04.680 + to take it with a grain of salt because, + +1:15:04.680 --> 1:15:07.440 + I mean, you are an inspiration because you basically + +1:15:07.440 --> 1:15:10.000 + don't care about a lot of things that other companies care + +1:15:10.000 --> 1:15:10.880 + about. + +1:15:10.880 --> 1:15:16.600 + You don't try to bullshit, in a sense, like make up stuff, + +1:15:16.600 --> 1:15:18.600 + so to drive up valuation. + +1:15:18.600 --> 1:15:19.960 + You're really very real, and you're + +1:15:19.960 --> 1:15:22.280 + trying to solve the problem, and I admire that a lot. + +1:15:22.280 --> 1:15:26.520 + What I don't necessarily fully can't trust you on about your + +1:15:26.520 --> 1:15:28.440 + respect is how good it is, right? + +1:15:28.440 --> 1:15:33.320 + I can only, but I also know how bad others are. + +1:15:33.320 --> 1:15:36.680 + I'll say two things about, trust, but verify, right? + +1:15:36.680 --> 1:15:38.040 + I'll say two things about that. + +1:15:38.040 --> 1:15:42.360 + One is try, get in a 2020 Corolla, + +1:15:42.360 --> 1:15:46.680 + and try OpenPilot 0.6 when it comes out next month. + +1:15:46.680 --> 1:15:48.400 + I think already, you'll look at this, + +1:15:48.400 --> 1:15:51.400 + and you'll be like, this is already really good. + +1:15:51.400 --> 1:15:54.240 + And then, I could be doing that all with hand labelers + +1:15:54.240 --> 1:15:58.000 + and all with the same approach that Mobileye uses. + +1:15:58.000 --> 1:16:00.040 + When we release a model that no longer + +1:16:00.040 --> 1:16:05.000 + has the lanes in it, that only outputs a path, + +1:16:05.000 --> 1:16:08.720 + then think about how we did that machine learning, + +1:16:08.720 --> 1:16:10.080 + and then right away, when you see, + +1:16:10.080 --> 1:16:11.240 + and that's going to be an OpenPilot, + +1:16:11.240 --> 1:16:13.000 + that's going to be an OpenPilot before 1.0, + +1:16:13.000 --> 1:16:14.400 + when you see that model, you'll know + +1:16:14.400 --> 1:16:15.360 + that everything I'm saying is true, + +1:16:15.360 --> 1:16:16.840 + because how else did I get that model? + +1:16:16.840 --> 1:16:17.320 + Good. + +1:16:17.320 --> 1:16:19.240 + You know what I'm saying is true about the simulator. + +1:16:19.240 --> 1:16:20.600 + Yeah, yeah, yeah, this is super exciting. + +1:16:20.600 --> 1:16:22.680 + That's super exciting. + +1:16:22.680 --> 1:16:25.760 + But I listened to your talk with Kyle, + +1:16:25.760 --> 1:16:30.480 + and Kyle was originally building the aftermarket system, + +1:16:30.480 --> 1:16:34.920 + and he gave up on it because of technical challenges, + +1:16:34.920 --> 1:16:37.360 + because of the fact that he's going + +1:16:37.360 --> 1:16:39.160 + to have to support 20 to 50 cars. + +1:16:39.160 --> 1:16:41.120 + We support 45, because what is he + +1:16:41.120 --> 1:16:43.440 + going to do when the manufacturer ABS system triggers? + +1:16:43.440 --> 1:16:45.480 + We have alerts and warnings to deal with all of that + +1:16:45.480 --> 1:16:48.400 + and all the cars, and how is he going to formally verify it? + +1:16:48.400 --> 1:16:49.800 + Well, I got 10 million miles of data. + +1:16:49.800 --> 1:16:53.240 + It's probably better verified than the spec. + +1:16:53.240 --> 1:16:57.720 + Yeah, I'm glad you're here talking to me. + +1:16:57.720 --> 1:17:01.120 + I'll remember this day, because it's interesting. + +1:17:01.120 --> 1:17:04.160 + If you look at Kyle's from Cruise, + +1:17:04.160 --> 1:17:06.320 + I'm sure they have a large number of business development + +1:17:06.320 --> 1:17:10.200 + folks, and he's working with GM. + +1:17:10.200 --> 1:17:13.280 + He could work with Argo AI, worked with Ford. + +1:17:13.280 --> 1:17:18.520 + It's interesting, because chances that you fail businesswise, + +1:17:18.520 --> 1:17:21.120 + like bankrupt, are pretty high. + +1:17:21.120 --> 1:17:23.880 + And yet, it's the Android model, + +1:17:23.880 --> 1:17:26.440 + is you're actually taking on the problem. + +1:17:26.440 --> 1:17:28.160 + So that's really inspiring. + +1:17:28.160 --> 1:17:30.920 + Well, I have a long term way for comedy to make money, too. + +1:17:30.920 --> 1:17:34.400 + And one of the nice things when you really take on the problem, + +1:17:34.400 --> 1:17:36.760 + which is my hope for autopilot, for example, + +1:17:36.760 --> 1:17:41.040 + is things you don't expect, ways to make money, + +1:17:41.040 --> 1:17:44.160 + or create value that you don't expect will pop up. + +1:17:44.160 --> 1:17:48.560 + I've known how to do it since 2017 is the first time I said it. + +1:17:48.560 --> 1:17:50.440 + Which part to know how to do which part? + +1:17:50.440 --> 1:17:52.520 + Our long term plan is to be a car insurance company. + +1:17:52.520 --> 1:17:53.160 + Insurance. + +1:17:53.160 --> 1:17:55.320 + Yeah, I love it. + +1:17:55.320 --> 1:17:56.680 + I make driving twice as safe. + +1:17:56.680 --> 1:17:57.680 + Not only that, I have the best data + +1:17:57.680 --> 1:18:00.040 + such to know who statistically is the safest drivers. + +1:18:00.040 --> 1:18:02.160 + And oh, oh, we see you. + +1:18:02.160 --> 1:18:03.720 + We see you driving unsafely. + +1:18:03.720 --> 1:18:05.360 + We're not going to insure you. + +1:18:05.360 --> 1:18:08.960 + And that causes a bifurcation in the market, + +1:18:08.960 --> 1:18:10.920 + because the only people who can't get common insurance + +1:18:10.920 --> 1:18:12.760 + or the bad drivers, Geico can insure them. + +1:18:12.760 --> 1:18:15.360 + Their premiums are crazy high, our premiums are crazy low. + +1:18:15.360 --> 1:18:16.240 + We win car insurance. + +1:18:16.240 --> 1:18:18.120 + Take over that whole market. + +1:18:18.120 --> 1:18:21.560 + OK, so if we win, if we win, but that's + +1:18:21.560 --> 1:18:23.800 + I'm saying like how do you turn comma into a $10 billion + +1:18:23.800 --> 1:18:24.640 + company is that. + +1:18:24.640 --> 1:18:25.600 + That's right. + +1:18:25.600 --> 1:18:30.000 + So you Elon Musk, who else? + +1:18:30.000 --> 1:18:32.720 + Who else is thinking like this and working like this + +1:18:32.720 --> 1:18:33.160 + in your view? + +1:18:33.160 --> 1:18:34.800 + Who are the competitors? + +1:18:34.800 --> 1:18:36.160 + Are there people seriously? + +1:18:36.160 --> 1:18:39.480 + I don't think anyone that I'm aware of is seriously + +1:18:39.480 --> 1:18:45.280 + taking on lane keeping, like to where it's a huge business that + +1:18:45.280 --> 1:18:51.400 + turns eventually to full autonomy that then creates + +1:18:51.400 --> 1:18:53.440 + other businesses on top of it and so on. + +1:18:53.440 --> 1:18:56.480 + Thinks insurance, thinks all kinds of ideas like that. + +1:18:56.480 --> 1:19:00.480 + Do you know anyone else thinking like this? + +1:19:00.480 --> 1:19:02.200 + Not really. + +1:19:02.200 --> 1:19:02.960 + That's interesting. + +1:19:02.960 --> 1:19:06.560 + I mean, my sense is everybody turns to that in like four + +1:19:06.560 --> 1:19:07.800 + or five years. + +1:19:07.800 --> 1:19:11.240 + Like Ford, once the autonomy doesn't fall through. + +1:19:11.240 --> 1:19:12.600 + But at this time. + +1:19:12.600 --> 1:19:14.120 + Elon's the iOS. + +1:19:14.120 --> 1:19:16.720 + By the way, he paved the way for all of us. + +1:19:16.720 --> 1:19:18.000 + It's not iOS, true. + +1:19:18.000 --> 1:19:21.520 + I would not be doing comma AI today if it was not + +1:19:21.520 --> 1:19:23.480 + for those conversations with Elon. + +1:19:23.480 --> 1:19:26.840 + And if it were not for him saying like, + +1:19:26.840 --> 1:19:28.600 + I think he said like, well, obviously we're not + +1:19:28.600 --> 1:19:31.280 + going to use LiDAR, we use cameras, humans use cameras. + +1:19:31.280 --> 1:19:32.600 + So what do you think about that? + +1:19:32.600 --> 1:19:33.880 + How important is LiDAR? + +1:19:33.880 --> 1:19:36.960 + Everybody else's on L5 is using LiDAR. + +1:19:36.960 --> 1:19:39.160 + What are your thoughts on his provocative statement + +1:19:39.160 --> 1:19:41.320 + that LiDAR is a crutch? + +1:19:41.320 --> 1:19:43.520 + See, sometimes they'll say dumb things like the driver + +1:19:43.520 --> 1:19:45.680 + monitoring thing, but sometimes they'll say absolutely + +1:19:45.680 --> 1:19:48.400 + completely 100% obviously true things. + +1:19:48.400 --> 1:19:50.840 + Of course LiDAR is a crutch. + +1:19:50.840 --> 1:19:53.040 + It's not even a good crutch. + +1:19:53.040 --> 1:19:54.200 + You're not even using it. + +1:19:54.200 --> 1:19:56.920 + They're using it for localization, + +1:19:56.920 --> 1:19:58.160 + which isn't good in the first place. + +1:19:58.160 --> 1:20:00.480 + If you have to localize your car to centimeters + +1:20:00.480 --> 1:20:04.280 + in order to drive, that's not driving. + +1:20:04.280 --> 1:20:06.320 + Currently not doing much machine learning. + +1:20:06.320 --> 1:20:09.280 + I thought LiDAR data, meaning like to help you + +1:20:09.280 --> 1:20:12.840 + in the task of general task of perception. + +1:20:12.840 --> 1:20:15.320 + The main goal of those LiDARs on those cars + +1:20:15.320 --> 1:20:18.840 + I think is actually localization more than perception, + +1:20:18.840 --> 1:20:20.080 + or at least that's what they use them for. + +1:20:20.080 --> 1:20:20.920 + Yeah, that's true. + +1:20:20.920 --> 1:20:22.480 + If you want to localize to centimeters, + +1:20:22.480 --> 1:20:23.720 + you can't use GPS. + +1:20:23.720 --> 1:20:25.120 + The fancies GPS in the world can't do it, + +1:20:25.120 --> 1:20:26.960 + especially if you're under tree cover and stuff. + +1:20:26.960 --> 1:20:28.480 + LiDAR you can do this pretty easily. + +1:20:28.480 --> 1:20:30.240 + So really they're not taking on, + +1:20:30.240 --> 1:20:33.200 + I mean in some research they're using it for perception, + +1:20:33.200 --> 1:20:35.840 + but and they're certainly not, which is sad, + +1:20:35.840 --> 1:20:38.680 + they're not fusing it well with vision. + +1:20:38.680 --> 1:20:40.560 + They do use it for perception. + +1:20:40.560 --> 1:20:42.400 + I'm not saying they don't use it for perception, + +1:20:42.400 --> 1:20:45.480 + but the thing that they have vision based + +1:20:45.480 --> 1:20:47.680 + and radar based perception systems as well. + +1:20:47.680 --> 1:20:51.440 + You could remove the LiDAR and keep around + +1:20:51.440 --> 1:20:54.040 + a lot of the dynamic object perception. + +1:20:54.040 --> 1:20:56.320 + You want to get centimeter accurate localization. + +1:20:56.320 --> 1:20:59.120 + Good luck doing that with anything else. + +1:20:59.120 --> 1:21:02.880 + So what should a cruise Waymo do? + +1:21:02.880 --> 1:21:05.360 + Like what would be your advice to them now? + +1:21:06.400 --> 1:21:11.400 + I mean Waymo is actually, they're serious. + +1:21:11.400 --> 1:21:13.120 + Waymo out of the ball of them, + +1:21:13.120 --> 1:21:16.120 + are quite serious about the long game. + +1:21:16.120 --> 1:21:20.680 + If L5 is a lot, is requires 50 years, + +1:21:20.680 --> 1:21:24.000 + I think Waymo will be the only one left standing at the end + +1:21:24.000 --> 1:21:26.560 + with a given the financial backing that they have. + +1:21:26.560 --> 1:21:28.640 + They're boo Google box. + +1:21:28.640 --> 1:21:31.040 + I'll say nice things about both Waymo and cruise. + +1:21:32.320 --> 1:21:33.480 + Let's do it. + +1:21:33.480 --> 1:21:34.320 + Nice is good. + +1:21:35.720 --> 1:21:39.200 + Waymo is by far the furthest along with technology. + +1:21:39.200 --> 1:21:41.160 + Waymo has a three to five year lead + +1:21:41.160 --> 1:21:42.880 + on all the competitors. + +1:21:43.960 --> 1:21:48.640 + If the Waymo looking stack works, + +1:21:48.640 --> 1:21:49.720 + maybe three year lead. + +1:21:49.720 --> 1:21:51.280 + If the Waymo looking stack works, + +1:21:51.280 --> 1:21:52.800 + they have a three year lead. + +1:21:52.800 --> 1:21:55.800 + Now, I argue that Waymo has spent too much money + +1:21:55.800 --> 1:21:59.240 + to recapitalize, to gain back their losses + +1:21:59.240 --> 1:22:00.160 + in those three years. + +1:22:00.160 --> 1:22:03.600 + Also self driving cars have no network effect like that. + +1:22:03.600 --> 1:22:04.800 + Uber has a network effect. + +1:22:04.800 --> 1:22:07.120 + You have a market, you have drivers and you have riders. + +1:22:07.120 --> 1:22:09.880 + Self driving cars, you have capital and you have riders. + +1:22:09.880 --> 1:22:11.400 + There's no network effect. + +1:22:11.400 --> 1:22:13.800 + If I want to blanket a new city in self driving cars, + +1:22:13.800 --> 1:22:16.000 + I buy the off the shelf Chinese knockoff self driving cars + +1:22:16.000 --> 1:22:17.160 + and I buy enough of them in the city. + +1:22:17.160 --> 1:22:18.360 + I can't do that with drivers. + +1:22:18.360 --> 1:22:20.840 + And that's why Uber has a first mover advantage + +1:22:20.840 --> 1:22:22.640 + that no self driving car company will. + +1:22:23.960 --> 1:22:26.520 + Can you just a thing, let a little bit. + +1:22:26.520 --> 1:22:28.160 + Uber, you're not talking about Uber, + +1:22:28.160 --> 1:22:29.240 + the autonomous vehicle Uber. + +1:22:29.240 --> 1:22:30.960 + You're talking about the Uber cars. + +1:22:30.960 --> 1:22:31.800 + Yeah. + +1:22:31.800 --> 1:22:32.640 + I'm Uber. + +1:22:32.640 --> 1:22:35.920 + I open for business in Austin, Texas, let's say. + +1:22:35.920 --> 1:22:38.760 + I need to attract both sides of the market. + +1:22:38.760 --> 1:22:41.200 + I need to both get drivers on my platform + +1:22:41.200 --> 1:22:42.720 + and riders on my platform. + +1:22:42.720 --> 1:22:45.320 + And I need to keep them both sufficiently happy, right? + +1:22:45.320 --> 1:22:46.520 + Riders aren't going to use it + +1:22:46.520 --> 1:22:48.960 + if it takes more than five minutes for an Uber to show up. + +1:22:48.960 --> 1:22:50.120 + Drivers aren't going to use it + +1:22:50.120 --> 1:22:52.120 + if they have to sit around all day and there's no riders. + +1:22:52.120 --> 1:22:54.480 + So you have to carefully balance a market. + +1:22:54.480 --> 1:22:56.240 + And whenever you have to carefully balance a market, + +1:22:56.240 --> 1:22:58.280 + there's a great first mover advantage + +1:22:58.280 --> 1:23:01.000 + because there's a switching cost for everybody, right? + +1:23:01.000 --> 1:23:02.120 + The drivers and the riders + +1:23:02.120 --> 1:23:04.080 + would have to switch at the same time. + +1:23:04.080 --> 1:23:08.880 + Let's even say that, let's say, Uber shows up. + +1:23:08.880 --> 1:23:13.880 + And Uber somehow agrees to do things at a bigger, + +1:23:14.800 --> 1:23:17.440 + we've done it more efficiently, right? + +1:23:17.440 --> 1:23:19.800 + Uber only takes 5% of a car + +1:23:19.800 --> 1:23:21.600 + instead of the 10% that Uber takes. + +1:23:21.600 --> 1:23:22.760 + No one is going to switch + +1:23:22.760 --> 1:23:24.920 + because the switching cost is higher than that 5%. + +1:23:24.920 --> 1:23:27.200 + So you actually can, in markets like that, + +1:23:27.200 --> 1:23:28.520 + you have a first mover advantage. + +1:23:28.520 --> 1:23:29.360 + Yeah. + +1:23:30.160 --> 1:23:32.720 + Autonomous vehicles of the level five variety + +1:23:32.720 --> 1:23:34.560 + have no first mover advantage. + +1:23:34.560 --> 1:23:36.800 + If the technology becomes commoditized, + +1:23:36.800 --> 1:23:39.520 + say I want to go to a new city, look at the scooters. + +1:23:39.520 --> 1:23:41.480 + It's going to look a lot more like scooters. + +1:23:41.480 --> 1:23:44.040 + Every person with a checkbook + +1:23:44.040 --> 1:23:45.720 + can blanket a city in scooters + +1:23:45.720 --> 1:23:47.920 + and that's why you have 10 different scooter companies. + +1:23:47.920 --> 1:23:48.760 + Which one's going to win? + +1:23:48.760 --> 1:23:49.600 + It's a race to the bottom. + +1:23:49.600 --> 1:23:51.040 + It's a terrible market to be in + +1:23:51.040 --> 1:23:53.160 + because there's no market for scooters. + +1:23:54.960 --> 1:23:56.520 + And the scooters don't get a say + +1:23:56.520 --> 1:23:57.480 + in whether they want to be bought + +1:23:57.480 --> 1:23:58.440 + and deployed to a city or not. + +1:23:58.440 --> 1:23:59.280 + Right. + +1:23:59.280 --> 1:24:00.120 + So yeah. + +1:24:00.120 --> 1:24:02.080 + We're going to entice the scooters with subsidies + +1:24:02.080 --> 1:24:02.920 + and deals. + +1:24:03.840 --> 1:24:05.480 + So whenever you have to invest that capital, + +1:24:05.480 --> 1:24:06.720 + it doesn't... + +1:24:06.720 --> 1:24:07.560 + It doesn't come back. + +1:24:07.560 --> 1:24:08.600 + Yeah. + +1:24:08.600 --> 1:24:12.320 + They can't be your main criticism of the Waymo approach. + +1:24:12.320 --> 1:24:14.840 + Oh, I'm saying even if it does technically work. + +1:24:14.840 --> 1:24:17.040 + Even if it does technically work, that's a problem. + +1:24:17.040 --> 1:24:18.000 + Yeah. + +1:24:18.000 --> 1:24:21.720 + I don't know if I were to say, I would say, + +1:24:22.840 --> 1:24:23.520 + you're already there. + +1:24:23.520 --> 1:24:24.560 + I haven't even thought about that. + +1:24:24.560 --> 1:24:26.520 + But I would say the bigger challenge + +1:24:26.520 --> 1:24:27.760 + is the technical approach. + +1:24:29.760 --> 1:24:31.840 + So Waymo's cruise is... + +1:24:31.840 --> 1:24:33.000 + And not just the technical approach, + +1:24:33.000 --> 1:24:34.800 + but of creating value. + +1:24:34.800 --> 1:24:39.800 + I still don't understand how you beat Uber, + +1:24:40.760 --> 1:24:43.480 + the human driven cars. + +1:24:43.480 --> 1:24:44.920 + In terms of financially, + +1:24:44.920 --> 1:24:47.160 + it doesn't make sense to me + +1:24:47.160 --> 1:24:50.080 + that people want to get an autonomous vehicle. + +1:24:50.080 --> 1:24:52.800 + I don't understand how you make money. + +1:24:52.800 --> 1:24:56.440 + In the long term, yes, like real long term, + +1:24:56.440 --> 1:24:58.640 + but it just feels like there's too much + +1:24:58.640 --> 1:24:59.960 + capital investment needed. + +1:24:59.960 --> 1:25:01.200 + Oh, and they're going to be worse than Ubers + +1:25:01.200 --> 1:25:02.440 + because they're going to stop + +1:25:02.440 --> 1:25:04.760 + for every little thing everywhere. + +1:25:06.320 --> 1:25:07.360 + I'll say a nice thing about cruise. + +1:25:07.360 --> 1:25:08.440 + That was my nice thing about Waymo. + +1:25:08.440 --> 1:25:09.280 + They're three years ahead of me. + +1:25:09.280 --> 1:25:10.120 + It was a nice... + +1:25:10.120 --> 1:25:10.960 + Oh, because they're three years. + +1:25:10.960 --> 1:25:12.480 + They're three years technically ahead of everybody. + +1:25:12.480 --> 1:25:13.960 + Their tech stack is great. + +1:25:14.800 --> 1:25:17.920 + My nice thing about cruise is GM buying them + +1:25:17.920 --> 1:25:19.160 + was a great move for GM. + +1:25:20.600 --> 1:25:22.240 + For $1 billion, + +1:25:22.240 --> 1:25:25.600 + GM bought an insurance policy against Waymo. + +1:25:26.560 --> 1:25:30.000 + They put cruise is three years behind Waymo. + +1:25:30.000 --> 1:25:32.600 + That means Google will get a monopoly + +1:25:32.600 --> 1:25:35.160 + on the technology for at most three years. + +1:25:36.840 --> 1:25:38.880 + And if technology works, + +1:25:38.880 --> 1:25:40.840 + you might not even be right about the three years. + +1:25:40.840 --> 1:25:41.840 + It might be less. + +1:25:41.840 --> 1:25:42.680 + Might be less. + +1:25:42.680 --> 1:25:44.320 + Cruise actually might not be that far behind. + +1:25:44.320 --> 1:25:47.360 + I don't know how much Waymo has waffled around + +1:25:47.360 --> 1:25:49.760 + or how much of it actually is just that long tail. + +1:25:49.760 --> 1:25:50.600 + Yeah, okay. + +1:25:50.600 --> 1:25:53.600 + If that's the best you could say in terms of nice things, + +1:25:53.600 --> 1:25:55.200 + that's more of a nice thing for GM + +1:25:55.200 --> 1:25:58.560 + that that's a smart insurance policy. + +1:25:58.560 --> 1:25:59.680 + It's a smart insurance policy. + +1:25:59.680 --> 1:26:01.880 + I mean, I think that's how... + +1:26:01.880 --> 1:26:05.200 + I can't see cruise working out any other. + +1:26:05.200 --> 1:26:07.840 + For cruise to leapfrog Waymo would really surprise me. + +1:26:10.400 --> 1:26:13.000 + Yeah, so let's talk about the underlying assumptions + +1:26:13.000 --> 1:26:13.840 + of everything is... + +1:26:13.840 --> 1:26:15.440 + We're not gonna leapfrog Tesla. + +1:26:17.560 --> 1:26:19.240 + Tesla would have to seriously mess up + +1:26:19.240 --> 1:26:20.440 + for us to leapfrog them. + +1:26:20.440 --> 1:26:23.240 + Okay, so the way you leapfrog, right, + +1:26:23.240 --> 1:26:26.120 + is you come up with an idea + +1:26:26.120 --> 1:26:28.560 + or you take a direction, perhaps secretly, + +1:26:28.560 --> 1:26:30.640 + that the other people aren't taking. + +1:26:31.640 --> 1:26:36.640 + And so cruise, Waymo, even Aurora... + +1:26:38.080 --> 1:26:40.080 + I don't know, Aurora, Zooks is the same stack as well. + +1:26:40.080 --> 1:26:41.720 + They're all the same code base even. + +1:26:41.720 --> 1:26:44.120 + They're all the same DARPA Urban Challenge code base. + +1:26:44.120 --> 1:26:45.360 + It's... + +1:26:45.360 --> 1:26:47.760 + So the question is, do you think there's a room + +1:26:47.760 --> 1:26:49.120 + for brilliance and innovation there + +1:26:49.120 --> 1:26:50.560 + that will change everything? + +1:26:51.560 --> 1:26:53.880 + Like say, okay, so I'll give you examples. + +1:26:53.880 --> 1:26:58.880 + It could be if revolution and mapping, for example, + +1:26:59.640 --> 1:27:03.040 + that allow you to map things, + +1:27:03.040 --> 1:27:05.840 + do HD maps of the whole world, + +1:27:05.840 --> 1:27:08.080 + all weather conditions somehow really well, + +1:27:08.080 --> 1:27:13.080 + or revolution and simulation, + +1:27:14.480 --> 1:27:18.840 + to where all the way you said before becomes incorrect. + +1:27:20.480 --> 1:27:21.520 + That kind of thing. + +1:27:21.520 --> 1:27:23.920 + Any room for breakthrough innovation? + +1:27:24.920 --> 1:27:25.960 + What I said before about, + +1:27:25.960 --> 1:27:28.280 + oh, they actually get the whole thing, well, + +1:27:28.280 --> 1:27:32.600 + I'll say this about we divide driving into three problems. + +1:27:32.600 --> 1:27:33.800 + And I actually haven't solved the third yet, + +1:27:33.800 --> 1:27:34.800 + but I haven't had any idea how to do it. + +1:27:34.800 --> 1:27:36.120 + So there's the static. + +1:27:36.120 --> 1:27:38.000 + The static driving problem is assuming + +1:27:38.000 --> 1:27:40.120 + you are the only car on the road, right? + +1:27:40.120 --> 1:27:42.000 + And this problem can be solved 100% + +1:27:42.000 --> 1:27:44.000 + with mapping and localization. + +1:27:44.000 --> 1:27:45.760 + This is why farms work the way they do. + +1:27:45.760 --> 1:27:48.440 + If all you have to deal with is the static problem, + +1:27:48.440 --> 1:27:50.160 + and you can statically schedule your machines, right? + +1:27:50.160 --> 1:27:52.680 + It's the same as like statically scheduling processes. + +1:27:52.680 --> 1:27:54.040 + You can statically schedule your tractors + +1:27:54.040 --> 1:27:56.160 + to never hit each other on their paths, right? + +1:27:56.160 --> 1:27:57.520 + Because then you know the speed they go at. + +1:27:57.520 --> 1:28:00.160 + So that's the static driving problem. + +1:28:00.160 --> 1:28:03.160 + Maps only helps you with the static driving problem. + +1:28:03.920 --> 1:28:06.960 + Yeah, the question about static driving, + +1:28:06.960 --> 1:28:08.800 + you've just made it sound like it's really easy. + +1:28:08.800 --> 1:28:10.160 + Static driving is really easy. + +1:28:11.880 --> 1:28:13.040 + How easy? + +1:28:13.040 --> 1:28:16.480 + How, well, because the whole drifting out of lane, + +1:28:16.480 --> 1:28:18.760 + when Tesla drifts out of lane, + +1:28:18.760 --> 1:28:21.960 + it's failing on the fundamental static driving problem. + +1:28:21.960 --> 1:28:24.440 + Tesla is drifting out of lane? + +1:28:24.440 --> 1:28:27.720 + The static driving problem is not easy for the world. + +1:28:27.720 --> 1:28:30.320 + The static driving problem is easy for one route. + +1:28:31.840 --> 1:28:33.920 + One route in one weather condition + +1:28:33.920 --> 1:28:37.920 + with one state of lane markings + +1:28:37.920 --> 1:28:40.920 + and like no deterioration, no cracks in the road. + +1:28:40.920 --> 1:28:42.600 + Well, I'm assuming you have a perfect localizer. + +1:28:42.600 --> 1:28:44.200 + So that's all for the weather condition + +1:28:44.200 --> 1:28:45.560 + and the lane marking condition. + +1:28:45.560 --> 1:28:46.640 + But that's the problem. + +1:28:46.640 --> 1:28:47.680 + How do you have a perfect localizer? + +1:28:47.680 --> 1:28:50.560 + You can build, perfect localizers are not that hard to build. + +1:28:50.560 --> 1:28:53.360 + Okay, come on now, with LIDAR. + +1:28:53.360 --> 1:28:54.200 + LIDAR, yeah. + +1:28:54.200 --> 1:28:55.040 + With LIDAR, okay. + +1:28:55.040 --> 1:28:56.440 + LIDAR, yeah, but you use LIDAR, right? + +1:28:56.440 --> 1:28:58.640 + Like you use LIDAR, build a perfect localizer. + +1:28:58.640 --> 1:29:00.960 + Building a perfect localizer without LIDAR, + +1:29:03.000 --> 1:29:04.320 + it's gonna be hard. + +1:29:04.320 --> 1:29:05.760 + You can get 10 centimeters without LIDAR, + +1:29:05.760 --> 1:29:07.240 + you can get one centimeter with LIDAR. + +1:29:07.240 --> 1:29:09.280 + I'm not even concerned about the one or 10 centimeters. + +1:29:09.280 --> 1:29:12.680 + I'm concerned if every once in a while you just weigh off. + +1:29:12.680 --> 1:29:17.480 + Yeah, so this is why you have to carefully + +1:29:17.480 --> 1:29:20.040 + make sure you're always tracking your position. + +1:29:20.040 --> 1:29:21.760 + You wanna use LIDAR camera fusion, + +1:29:21.760 --> 1:29:24.480 + but you can get the reliability of that system + +1:29:24.480 --> 1:29:28.000 + up to 100,000 miles + +1:29:28.000 --> 1:29:29.720 + and then you write some fallback condition + +1:29:29.720 --> 1:29:32.160 + where it's not that bad if you're way off, right? + +1:29:32.160 --> 1:29:33.800 + I think that you can get it to the point, + +1:29:33.800 --> 1:29:36.800 + it's like ASL D that you're never in a case + +1:29:36.800 --> 1:29:38.480 + where you're way off and you don't know it. + +1:29:38.480 --> 1:29:40.240 + Yeah, okay, so this is brilliant. + +1:29:40.240 --> 1:29:41.160 + So that's the static. + +1:29:41.160 --> 1:29:42.280 + Static. + +1:29:42.280 --> 1:29:45.960 + We can, especially with LIDAR and good HD maps, + +1:29:45.960 --> 1:29:47.080 + you can solve that problem. + +1:29:47.080 --> 1:29:47.920 + It's easy. + +1:29:47.920 --> 1:29:51.840 + The static, the static problem is so easy. + +1:29:51.840 --> 1:29:54.000 + It's very typical for you to say something's easy. + +1:29:54.000 --> 1:29:54.840 + I got it. + +1:29:54.840 --> 1:29:56.920 + It's not as challenging as the other ones, okay. + +1:29:56.920 --> 1:29:58.760 + Well, okay, maybe it's obvious how to solve it. + +1:29:58.760 --> 1:29:59.760 + The third one's the hardest. + +1:29:59.760 --> 1:30:01.920 + And a lot of people don't even think about the third one + +1:30:01.920 --> 1:30:03.640 + and even see it as different from the second one. + +1:30:03.640 --> 1:30:05.720 + So the second one is dynamic. + +1:30:05.720 --> 1:30:08.560 + The second one is like, say there's an obvious example, + +1:30:08.560 --> 1:30:10.360 + it's like a car stopped at a red light, right? + +1:30:10.360 --> 1:30:12.520 + You can't have that car in your map + +1:30:12.520 --> 1:30:13.720 + because you don't know whether that car + +1:30:13.720 --> 1:30:14.880 + is gonna be there or not. + +1:30:14.880 --> 1:30:17.960 + So you have to detect that car in real time + +1:30:17.960 --> 1:30:21.600 + and then you have to do the appropriate action, right? + +1:30:21.600 --> 1:30:24.800 + Also, that car is not a fixed object. + +1:30:24.800 --> 1:30:26.600 + That car may move and you have to predict + +1:30:26.600 --> 1:30:28.680 + what that car will do, right? + +1:30:28.680 --> 1:30:30.840 + So this is the dynamic problem. + +1:30:30.840 --> 1:30:31.680 + Yeah. + +1:30:31.680 --> 1:30:32.800 + So you have to deal with this. + +1:30:32.800 --> 1:30:36.640 + This involves, again, like you're gonna need models + +1:30:36.640 --> 1:30:38.760 + of other people's behavior. + +1:30:38.760 --> 1:30:40.160 + Do you, are you including in that? + +1:30:40.160 --> 1:30:42.320 + I don't wanna step on the third one. + +1:30:42.320 --> 1:30:46.600 + Oh, but are you including in that your influence + +1:30:46.600 --> 1:30:47.440 + on people? + +1:30:47.440 --> 1:30:48.280 + Ah, that's the third one. + +1:30:48.280 --> 1:30:49.120 + Okay. + +1:30:49.120 --> 1:30:49.960 + That's the third one. + +1:30:49.960 --> 1:30:51.880 + We call it the counterfactual. + +1:30:51.880 --> 1:30:52.720 + Yeah, brilliant. + +1:30:52.720 --> 1:30:53.560 + And that. + +1:30:53.560 --> 1:30:54.920 + I just talked to Judea Pro who's obsessed + +1:30:54.920 --> 1:30:55.760 + with counterfactuals. + +1:30:55.760 --> 1:30:58.640 + Counterfactual, oh yeah, yeah, I read his books. + +1:30:58.640 --> 1:31:03.640 + So the static and the dynamic are our approach right now + +1:31:03.960 --> 1:31:07.600 + for lateral will scale completely to the static and dynamic. + +1:31:07.600 --> 1:31:10.760 + The counterfactual, the only way I have to do it yet, + +1:31:10.760 --> 1:31:14.000 + the thing that I wanna do once we have all of these cars + +1:31:14.000 --> 1:31:16.760 + is I wanna do reinforcement learning on the world. + +1:31:16.760 --> 1:31:18.880 + I'm always gonna turn the exploiter up to max. + +1:31:18.880 --> 1:31:20.440 + I'm not gonna have them explore. + +1:31:20.440 --> 1:31:22.760 + But the only real way to get at the counterfactual + +1:31:22.760 --> 1:31:24.080 + is to do reinforcement learning + +1:31:24.080 --> 1:31:26.360 + because the other agents are humans. + +1:31:27.760 --> 1:31:30.080 + So that's fascinating that you break it down like that. + +1:31:30.080 --> 1:31:31.680 + I agree completely. + +1:31:31.680 --> 1:31:33.600 + I've spent my life thinking about this problem. + +1:31:33.600 --> 1:31:34.920 + This is beautiful. + +1:31:34.920 --> 1:31:37.880 + And part of it, cause you're slightly insane, + +1:31:37.880 --> 1:31:42.880 + because not my life, just the last four years. + +1:31:43.120 --> 1:31:48.120 + No, no, you have some non zero percent of your brain + +1:31:48.920 --> 1:31:52.360 + has a madman in it, which is a really good feature. + +1:31:52.360 --> 1:31:55.920 + But there's a safety component to it + +1:31:55.920 --> 1:31:57.320 + that I think when there's sort of + +1:31:57.320 --> 1:31:59.040 + with counterfactuals and so on, + +1:31:59.040 --> 1:32:00.280 + that would just freak people out. + +1:32:00.280 --> 1:32:03.320 + How do you even start to think about this in general? + +1:32:03.320 --> 1:32:07.600 + I mean, you've had some friction with NHTSA and so on. + +1:32:07.600 --> 1:32:12.600 + I am frankly exhausted by safety engineers. + +1:32:14.280 --> 1:32:19.280 + The prioritization on safety over innovation + +1:32:21.360 --> 1:32:23.720 + to a degree where it kills, in my view, + +1:32:23.720 --> 1:32:26.200 + kills safety in the longterm. + +1:32:26.200 --> 1:32:28.080 + So the counterfactual thing, + +1:32:28.080 --> 1:32:31.560 + they just actually exploring this world + +1:32:31.560 --> 1:32:33.600 + of how do you interact with dynamic objects and so on? + +1:32:33.600 --> 1:32:34.840 + How do you think about safety? + +1:32:34.840 --> 1:32:38.120 + You can do reinforcement learning without ever exploring. + +1:32:38.120 --> 1:32:39.200 + And I said that, like, + +1:32:39.200 --> 1:32:41.560 + so you can think about your, in like reinforcement learning, + +1:32:41.560 --> 1:32:44.320 + it's usually called like a temperature parameter. + +1:32:44.320 --> 1:32:45.360 + And your temperature parameter + +1:32:45.360 --> 1:32:48.080 + is how often you deviate from the argmax. + +1:32:48.080 --> 1:32:50.720 + I could always set that to zero and still learn. + +1:32:50.720 --> 1:32:52.840 + And I feel that you'd always want that set to zero + +1:32:52.840 --> 1:32:54.080 + on your actual system. + +1:32:54.080 --> 1:32:54.920 + Gotcha. + +1:32:54.920 --> 1:32:58.160 + But the problem is you first don't know very much + +1:32:58.160 --> 1:32:59.560 + and so you're going to make mistakes. + +1:32:59.560 --> 1:33:01.680 + So the learning, the exploration happens through mistakes. + +1:33:01.680 --> 1:33:03.240 + We're all ready, yeah, but. + +1:33:03.240 --> 1:33:06.080 + Okay, so the consequences of a mistake. + +1:33:06.080 --> 1:33:09.400 + OpenPilot and Autopilot are making mistakes left and right. + +1:33:09.400 --> 1:33:12.560 + We have 700 daily active users, + +1:33:12.560 --> 1:33:14.080 + 1,000 weekly active users. + +1:33:14.080 --> 1:33:18.920 + OpenPilot makes tens of thousands of mistakes a week. + +1:33:18.920 --> 1:33:21.160 + These mistakes have zero consequences. + +1:33:21.160 --> 1:33:22.520 + These mistakes are, + +1:33:22.520 --> 1:33:26.800 + oh, I wanted to take this exit and it went straight. + +1:33:26.800 --> 1:33:28.520 + So I'm just going to carefully touch the wheel. + +1:33:28.520 --> 1:33:29.360 + The humans catch them. + +1:33:29.360 --> 1:33:30.640 + The humans catch them. + +1:33:30.640 --> 1:33:33.120 + And the human disengagement is labeling + +1:33:33.120 --> 1:33:35.000 + that reinforcement learning in a completely + +1:33:35.000 --> 1:33:36.200 + consequence free way. + +1:33:37.240 --> 1:33:39.840 + So driver monitoring is the way you ensure they keep. + +1:33:39.840 --> 1:33:40.680 + Yes. + +1:33:40.680 --> 1:33:42.120 + They keep paying attention. + +1:33:42.120 --> 1:33:43.240 + How's your messaging? + +1:33:43.240 --> 1:33:45.200 + Say I gave you a billion dollars, + +1:33:45.200 --> 1:33:46.960 + so you would be scaling it now. + +1:33:47.800 --> 1:33:49.720 + Oh, if I could scale, I couldn't scale with any amount of money. + +1:33:49.720 --> 1:33:51.640 + I'd raise money if I could, if I had a way to scale it. + +1:33:51.640 --> 1:33:53.320 + Yeah, you're not, no, I'm not focused on scale. + +1:33:53.320 --> 1:33:54.160 + I don't know how to do. + +1:33:54.160 --> 1:33:55.760 + Oh, like, I guess I could sell it to more people, + +1:33:55.760 --> 1:33:56.960 + but I want to make the system better. + +1:33:56.960 --> 1:33:57.800 + Better, better. + +1:33:57.800 --> 1:33:58.840 + And I don't know how to. + +1:33:58.840 --> 1:34:01.080 + But what's the messaging here? + +1:34:01.080 --> 1:34:02.560 + I got a chance to talk to Elon. + +1:34:02.560 --> 1:34:07.560 + And he basically said that the human factor doesn't matter. + +1:34:09.280 --> 1:34:10.360 + You know, the human doesn't matter + +1:34:10.360 --> 1:34:12.280 + because the system will perform. + +1:34:12.280 --> 1:34:14.760 + There'll be sort of a, sorry to use the term, + +1:34:14.760 --> 1:34:16.120 + but like a singular, like a point + +1:34:16.120 --> 1:34:17.920 + where it gets just much better. + +1:34:17.920 --> 1:34:20.800 + And so the human, it won't really matter. + +1:34:20.800 --> 1:34:25.000 + But it seems like that human catching the system + +1:34:25.000 --> 1:34:29.360 + when it gets into trouble is like the thing + +1:34:29.360 --> 1:34:32.720 + which will make something like reinforcement learning work. + +1:34:32.720 --> 1:34:35.640 + So how do you, how do you think messaging for Tesla, + +1:34:35.640 --> 1:34:39.080 + for you, for the industry in general, should change? + +1:34:39.080 --> 1:34:40.840 + I think my messaging is pretty clear, + +1:34:40.840 --> 1:34:43.080 + at least like our messaging wasn't that clear + +1:34:43.080 --> 1:34:45.200 + in the beginning and I do kind of fault myself for that. + +1:34:45.200 --> 1:34:48.480 + We are proud right now to be a level two system. + +1:34:48.480 --> 1:34:50.360 + We are proud to be level two. + +1:34:50.360 --> 1:34:51.640 + If we talk about level four, + +1:34:51.640 --> 1:34:53.200 + it's not with the current hardware. + +1:34:53.200 --> 1:34:55.920 + It's not going to be just a magical OTA upgrade. + +1:34:55.920 --> 1:34:57.280 + It's going to be new hardware. + +1:34:57.280 --> 1:35:00.000 + It's going to be very carefully thought out right now. + +1:35:00.000 --> 1:35:01.560 + We are proud to be level two. + +1:35:01.560 --> 1:35:03.320 + And we have a rigorous safety model. + +1:35:03.320 --> 1:35:05.680 + I mean, not like, like, okay, rigorous. + +1:35:05.680 --> 1:35:06.600 + Who knows what that means? + +1:35:06.600 --> 1:35:08.600 + But we at least have a safety model + +1:35:08.600 --> 1:35:09.560 + and we make it explicit. + +1:35:09.560 --> 1:35:11.800 + It's in safety.md and open pilot. + +1:35:11.800 --> 1:35:13.960 + And it says, seriously though. + +1:35:13.960 --> 1:35:14.800 + Safety.md. + +1:35:14.800 --> 1:35:15.840 + Safety.md. + +1:35:16.840 --> 1:35:18.400 + This is really, this is so Android. + +1:35:18.400 --> 1:35:21.800 + So, well, this is, this is the safety model + +1:35:21.800 --> 1:35:25.520 + and I like to have conversations like if, like, you know, + +1:35:25.520 --> 1:35:27.120 + sometimes people will come to you and they're like, + +1:35:27.120 --> 1:35:29.240 + your system's not safe. + +1:35:29.240 --> 1:35:30.080 + Okay. + +1:35:30.080 --> 1:35:31.080 + Have you read my safety docs? + +1:35:31.080 --> 1:35:32.720 + Would you like to have an intelligent conversation + +1:35:32.720 --> 1:35:33.560 + about this? + +1:35:33.560 --> 1:35:34.400 + And the answer is always no. + +1:35:34.400 --> 1:35:36.880 + They just like scream about, it runs Python. + +1:35:38.240 --> 1:35:39.080 + Okay. What? + +1:35:39.080 --> 1:35:41.560 + So you're saying that, that because Python's not real time, + +1:35:41.560 --> 1:35:44.240 + Python not being real time never causes disengagement. + +1:35:44.240 --> 1:35:47.640 + Disengagement's are caused by, you know, the model is QM. + +1:35:47.640 --> 1:35:49.760 + But safety.md says the following. + +1:35:49.760 --> 1:35:50.600 + First and foremost, + +1:35:50.600 --> 1:35:53.000 + the driver must be paying attention at all times. + +1:35:54.240 --> 1:35:55.320 + I don't consider, I do, + +1:35:55.320 --> 1:35:57.720 + I still consider the software to be alpha software + +1:35:57.720 --> 1:36:00.080 + until we can actually enforce that statement. + +1:36:00.080 --> 1:36:03.280 + But I feel it's very well communicated to our users. + +1:36:03.280 --> 1:36:04.520 + Two more things. + +1:36:04.520 --> 1:36:09.080 + One is the user must be able to easily take control + +1:36:09.080 --> 1:36:10.880 + of the vehicle at all times. + +1:36:10.880 --> 1:36:14.440 + So if you step on the gas or brake with open pilot, + +1:36:14.440 --> 1:36:16.400 + it gives full manual control back to the user + +1:36:16.400 --> 1:36:18.680 + or press the cancel button. + +1:36:18.680 --> 1:36:23.240 + Step two, the car will never react so quickly. + +1:36:23.240 --> 1:36:26.000 + We define so quickly to be about one second + +1:36:26.000 --> 1:36:27.640 + that you can't react in time. + +1:36:27.640 --> 1:36:29.480 + And we do this by enforcing torque limits, + +1:36:29.480 --> 1:36:31.520 + braking limits and acceleration limits. + +1:36:31.520 --> 1:36:36.520 + So we have like our torque limits way lower than Tesla's. + +1:36:36.520 --> 1:36:39.080 + This is another potential. + +1:36:39.080 --> 1:36:40.240 + If I could tweak autopilot, + +1:36:40.240 --> 1:36:41.360 + I would lower their torque limit + +1:36:41.360 --> 1:36:42.960 + and I would add driver monitoring. + +1:36:42.960 --> 1:36:46.240 + Because autopilot can jerk the wheel hard. + +1:36:46.240 --> 1:36:47.520 + Open pilot can't. + +1:36:47.520 --> 1:36:52.080 + It's, we limit and all this code is open source, readable. + +1:36:52.080 --> 1:36:54.880 + And I believe now it's all MISRA C compliant. + +1:36:54.880 --> 1:36:55.800 + What's that mean? + +1:36:57.080 --> 1:37:00.400 + MISRA is like the automotive coding standard. + +1:37:00.400 --> 1:37:03.400 + At first, I've come to respect, + +1:37:03.400 --> 1:37:04.960 + I've been reading like the standards lately + +1:37:04.960 --> 1:37:05.920 + and I've come to respect them. + +1:37:05.920 --> 1:37:07.800 + They're actually written by very smart people. + +1:37:07.800 --> 1:37:09.920 + Yeah, they're brilliant people actually. + +1:37:09.920 --> 1:37:11.320 + They have a lot of experience. + +1:37:11.320 --> 1:37:13.360 + They're sometimes a little too cautious, + +1:37:13.360 --> 1:37:16.800 + but in this case, it pays off. + +1:37:16.800 --> 1:37:18.440 + MISRA is written by like computer scientists + +1:37:18.440 --> 1:37:19.840 + and you can tell by the language they use. + +1:37:19.840 --> 1:37:21.080 + You can tell by the language they use. + +1:37:21.080 --> 1:37:24.440 + They talk about like whether certain conditions in MISRA + +1:37:24.440 --> 1:37:26.520 + are decidable or undecidable. + +1:37:26.520 --> 1:37:28.360 + And you mean like the halting problem? + +1:37:28.360 --> 1:37:31.600 + And yes, all right, you've earned my respect. + +1:37:31.600 --> 1:37:33.120 + I will read carefully what you have to say + +1:37:33.120 --> 1:37:35.760 + and we want to make our code compliant with that. + +1:37:35.760 --> 1:37:38.160 + All right, so you're proud level two, beautiful. + +1:37:38.160 --> 1:37:42.320 + So you were the founder and I think CEO of Kama AI, + +1:37:42.320 --> 1:37:44.320 + then you were the head of research. + +1:37:44.320 --> 1:37:46.080 + What the heck are you now? + +1:37:46.080 --> 1:37:47.480 + What's your connection to Kama AI? + +1:37:47.480 --> 1:37:49.640 + I'm the president, but I'm one of those like + +1:37:49.640 --> 1:37:53.440 + unelected presidents of like a small dictatorship country, + +1:37:53.440 --> 1:37:55.200 + not one of those like elected presidents. + +1:37:55.200 --> 1:37:57.640 + Oh, so you're like Putin when he was like the, yeah, + +1:37:57.640 --> 1:37:58.980 + I got you. + +1:37:58.980 --> 1:38:02.120 + So there's, what's the governance structure? + +1:38:02.120 --> 1:38:04.800 + What's the future of Kama AI finance? + +1:38:04.800 --> 1:38:08.120 + I mean, yeah, as a business, do you want, + +1:38:08.120 --> 1:38:11.640 + are you just focused on getting things right now, + +1:38:11.640 --> 1:38:14.920 + making some small amount of money in the meantime + +1:38:14.920 --> 1:38:17.520 + and then when it works, it works a new scale. + +1:38:17.520 --> 1:38:20.480 + Our burn rate is about 200 K a month + +1:38:20.480 --> 1:38:23.040 + and our revenue is about 100 K a month. + +1:38:23.040 --> 1:38:24.920 + So we need to forex our revenue, + +1:38:24.920 --> 1:38:28.200 + but we haven't like tried very hard at that yet. + +1:38:28.200 --> 1:38:30.160 + And the revenue is basically selling stuff online. + +1:38:30.160 --> 1:38:32.360 + Yeah, we sell stuff shop.com.ai. + +1:38:32.360 --> 1:38:33.920 + Is there other, well, okay. + +1:38:33.920 --> 1:38:35.360 + So you'll have to figure out. + +1:38:35.360 --> 1:38:37.880 + That's our only, see, but to me, + +1:38:37.880 --> 1:38:40.400 + that's like respectable revenues. + +1:38:40.400 --> 1:38:42.640 + We make it by selling products to consumers + +1:38:42.640 --> 1:38:45.040 + for honest and transparent about what they are. + +1:38:45.040 --> 1:38:49.000 + Most actually level four companies, right? + +1:38:50.720 --> 1:38:54.320 + Cause you could easily start blowing up like smoke, + +1:38:54.320 --> 1:38:57.080 + like overselling the hype and feeding into, + +1:38:57.080 --> 1:38:59.080 + getting some fundraisers. + +1:38:59.080 --> 1:39:00.520 + Oh, you're the guy, you're a genius + +1:39:00.520 --> 1:39:01.800 + because you hacked the iPhone. + +1:39:01.800 --> 1:39:02.920 + Oh, I hate that. + +1:39:02.920 --> 1:39:03.760 + I hate that. + +1:39:03.760 --> 1:39:06.360 + Yeah, I can trade my social capital for more money. + +1:39:06.360 --> 1:39:07.320 + I did it once. + +1:39:07.320 --> 1:39:10.320 + I almost regret it doing it the first time. + +1:39:10.320 --> 1:39:11.640 + Well, on a small tangent, + +1:39:11.640 --> 1:39:16.560 + what's your, you seem to not like fame + +1:39:16.560 --> 1:39:18.840 + and yet you're also drawn to fame. + +1:39:18.840 --> 1:39:23.840 + What's, where have you on, where are you on that currently? + +1:39:24.560 --> 1:39:27.200 + Have you had some introspection, some soul searching? + +1:39:27.200 --> 1:39:28.480 + Yeah. + +1:39:28.480 --> 1:39:32.200 + I actually, I've come to a pretty stable position on that. + +1:39:32.200 --> 1:39:33.880 + Like after the first time, + +1:39:33.880 --> 1:39:36.840 + I realized that I don't want attention from the masses. + +1:39:36.840 --> 1:39:39.160 + I want attention from people who I respect. + +1:39:39.160 --> 1:39:41.960 + Who do you respect? + +1:39:41.960 --> 1:39:43.960 + I can give a list of people. + +1:39:43.960 --> 1:39:47.200 + So are these like Elon Musk type characters? + +1:39:47.200 --> 1:39:49.040 + Yeah. + +1:39:49.040 --> 1:39:50.000 + Actually, you know what? + +1:39:50.000 --> 1:39:51.200 + I'll make it more broad than that. + +1:39:51.200 --> 1:39:52.600 + I won't make it about a person. + +1:39:52.600 --> 1:39:54.040 + I respect skill. + +1:39:54.040 --> 1:39:56.880 + I respect people who have skills, right? + +1:39:56.880 --> 1:40:00.280 + And I would like to like be, + +1:40:00.280 --> 1:40:01.400 + I'm not gonna say famous, + +1:40:01.400 --> 1:40:03.760 + but be like known among more people + +1:40:03.760 --> 1:40:05.440 + who have like real skills. + +1:40:05.440 --> 1:40:10.440 + Who in cars, do you think have skill? + +1:40:12.560 --> 1:40:13.720 + Not do you respect? + +1:40:15.000 --> 1:40:17.760 + Oh, Kyle Voat has skill. + +1:40:17.760 --> 1:40:19.880 + A lot of people at Waymo have skill. + +1:40:19.880 --> 1:40:20.840 + And I respect them. + +1:40:20.840 --> 1:40:23.760 + I respect them as engineers. + +1:40:23.760 --> 1:40:24.920 + Like I can think, I mean, + +1:40:24.920 --> 1:40:26.280 + I think about all the times in my life + +1:40:26.280 --> 1:40:27.960 + where I've been like dead set on approaches + +1:40:27.960 --> 1:40:29.160 + and they turn out to be wrong. + +1:40:29.160 --> 1:40:30.000 + Yeah. + +1:40:30.000 --> 1:40:31.720 + So I mean, this might, I might be wrong. + +1:40:31.720 --> 1:40:34.720 + I accept that, I accept that there's a decent chance + +1:40:34.720 --> 1:40:36.600 + that I'm wrong. + +1:40:36.600 --> 1:40:38.400 + And actually, I mean, having talked to Chris Armson, + +1:40:38.400 --> 1:40:40.480 + Sterling Anderson, those guys, + +1:40:40.480 --> 1:40:43.360 + I mean, I deeply respect Chris. + +1:40:43.360 --> 1:40:44.640 + I just admire the guy. + +1:40:46.040 --> 1:40:47.400 + He's legit. + +1:40:47.400 --> 1:40:48.960 + When you drive a car through the desert + +1:40:48.960 --> 1:40:52.400 + when everybody thinks it's impossible, that's legit. + +1:40:52.400 --> 1:40:53.840 + And then I also really respect the people + +1:40:53.840 --> 1:40:55.680 + who are like writing the infrastructure of the world, + +1:40:55.680 --> 1:40:57.360 + like the Linus Torvalds and the Chris Ladin. + +1:40:57.360 --> 1:40:59.080 + Oh yeah, they were doing the real work. + +1:40:59.080 --> 1:41:00.800 + I know they're doing the real work. + +1:41:02.000 --> 1:41:03.760 + Having talked to Chris Ladin, + +1:41:03.760 --> 1:41:05.680 + you realize, especially when they're humble, + +1:41:05.680 --> 1:41:07.680 + it's like, you realize, oh, you guys, + +1:41:07.680 --> 1:41:09.640 + we're just using your... + +1:41:09.640 --> 1:41:10.480 + Oh yeah. + +1:41:10.480 --> 1:41:11.520 + All the hard work that you did. + +1:41:11.520 --> 1:41:13.120 + Yeah, that's incredible. + +1:41:13.120 --> 1:41:17.160 + What do you think, Mr. Anthony Lewandowski? + +1:41:18.440 --> 1:41:21.640 + What do you, he's a, he's another mad genius. + +1:41:21.640 --> 1:41:22.480 + Sharp guy. + +1:41:22.480 --> 1:41:23.320 + Oh yeah. + +1:41:23.320 --> 1:41:27.640 + What, do you think he might long term become a competitor? + +1:41:27.640 --> 1:41:28.840 + Oh, to comma? + +1:41:28.840 --> 1:41:32.400 + Well, so I think that he has the other right approach. + +1:41:32.400 --> 1:41:35.280 + I think that right now, there's two right approaches. + +1:41:35.280 --> 1:41:37.680 + One is what we're doing and one is what he's doing. + +1:41:37.680 --> 1:41:39.800 + Can you describe, I think it's called Pronto AI, + +1:41:39.800 --> 1:41:42.360 + he's starting using, do you know what the approach is? + +1:41:42.360 --> 1:41:43.200 + I actually don't know. + +1:41:43.200 --> 1:41:45.040 + Embark is also doing the same sort of thing. + +1:41:45.040 --> 1:41:47.280 + The idea is almost that you want to, + +1:41:47.280 --> 1:41:51.800 + so if you're, I can't partner with Honda and Toyota. + +1:41:51.800 --> 1:41:56.800 + Honda and Toyota are like 400,000 person companies. + +1:41:57.600 --> 1:41:59.400 + It's not even a company at that point. + +1:41:59.400 --> 1:42:01.400 + Like I don't think of it like, I don't personify it. + +1:42:01.400 --> 1:42:06.400 + I think of it like an object, but a trucker drives for a fleet. + +1:42:07.120 --> 1:42:10.280 + Maybe that has like, some truckers are independent. + +1:42:10.280 --> 1:42:12.080 + Some truckers drive for fleets with a hundred trucks. + +1:42:12.080 --> 1:42:14.960 + There are tons of independent trucking companies out there. + +1:42:14.960 --> 1:42:18.120 + Start a trucking company and drive your costs down + +1:42:18.120 --> 1:42:23.120 + or figure out how to drive down the cost of trucking. + +1:42:23.760 --> 1:42:26.560 + Another company that I really respect is Nauto. + +1:42:26.560 --> 1:42:28.320 + Actually, I respect their business model. + +1:42:28.320 --> 1:42:31.560 + Nauto sells a driver monitoring camera + +1:42:31.560 --> 1:42:33.920 + and they sell it to fleet owners. + +1:42:33.920 --> 1:42:38.920 + If I owned a fleet of cars and I could pay 40 bucks a month + +1:42:39.120 --> 1:42:41.280 + to monitor my employees, + +1:42:42.400 --> 1:42:45.520 + this is gonna like reduces accidents 18%. + +1:42:45.520 --> 1:42:48.960 + It's so like that in the space, + +1:42:48.960 --> 1:42:52.000 + that is like the business model that I like most respect + +1:42:53.400 --> 1:42:55.360 + because they're creating value today. + +1:42:55.360 --> 1:42:57.840 + Yeah, which is, that's a huge one. + +1:42:57.840 --> 1:42:59.800 + How do we create value today with some of this? + +1:42:59.800 --> 1:43:01.680 + And the length keeping thing is huge. + +1:43:01.680 --> 1:43:03.800 + And it sounds like you're creeping in + +1:43:03.800 --> 1:43:06.680 + or full steam ahead on the drive of monitoring too. + +1:43:06.680 --> 1:43:09.240 + Which I think actually where the short term value, + +1:43:09.240 --> 1:43:10.480 + if you can get right. + +1:43:10.480 --> 1:43:12.800 + I still, I'm not a huge fan of the statement + +1:43:12.800 --> 1:43:15.120 + that everything is to have drive of monitoring. + +1:43:15.120 --> 1:43:16.120 + I agree with that completely, + +1:43:16.120 --> 1:43:18.680 + but that statement usually misses the point + +1:43:18.680 --> 1:43:21.920 + that to get the experience of it right is not trivial. + +1:43:21.920 --> 1:43:22.840 + Oh, no, not at all. + +1:43:22.840 --> 1:43:25.280 + In fact, like, so right now we have, + +1:43:25.280 --> 1:43:28.480 + I think the timeout depends on speed of the car, + +1:43:29.560 --> 1:43:32.520 + but we want to depend on like the scene state. + +1:43:32.520 --> 1:43:35.440 + If you're on like an empty highway, + +1:43:35.440 --> 1:43:37.680 + it's very different if you don't pay attention + +1:43:37.680 --> 1:43:40.600 + than if like you're like coming up to a traffic light. + +1:43:42.040 --> 1:43:45.720 + And long term, it should probably learn from the driver + +1:43:45.720 --> 1:43:48.120 + because that's to do, I watched a lot of video. + +1:43:48.120 --> 1:43:49.480 + We've built a smartphone detector + +1:43:49.480 --> 1:43:51.520 + just to analyze how people are using smartphones + +1:43:51.520 --> 1:43:53.400 + and people are using it very differently. + +1:43:53.400 --> 1:43:57.760 + And there's this, it's a texting styles. + +1:43:57.760 --> 1:44:00.320 + We haven't watched nearly enough of the videos. + +1:44:00.320 --> 1:44:01.800 + We haven't, I got millions of miles + +1:44:01.800 --> 1:44:02.960 + of people driving cars. + +1:44:02.960 --> 1:44:05.960 + In this moment, I spent a large fraction of my time + +1:44:05.960 --> 1:44:10.880 + just watching videos because it's never fails to learn. + +1:44:10.880 --> 1:44:13.480 + Like I've never failed from a video watching session + +1:44:13.480 --> 1:44:15.400 + to learn something I didn't know before. + +1:44:15.400 --> 1:44:19.640 + In fact, I usually, like when I eat lunch, I'll sit, + +1:44:19.640 --> 1:44:20.680 + especially when the weather is good + +1:44:20.680 --> 1:44:22.080 + and just watch pedestrians. + +1:44:22.080 --> 1:44:26.400 + With an eye to understand like from a computer vision eye, + +1:44:26.400 --> 1:44:29.280 + just to see, can this model, can you predict + +1:44:29.280 --> 1:44:30.480 + what are the decisions made? + +1:44:30.480 --> 1:44:33.040 + And there's so many things that we don't understand. + +1:44:33.040 --> 1:44:34.760 + This is what I mean about state vector. + +1:44:34.760 --> 1:44:37.880 + Yeah, it's, I'm trying to always think like, + +1:44:37.880 --> 1:44:40.280 + because I'm understanding in my human brain, + +1:44:40.280 --> 1:44:42.000 + how do we convert that into, + +1:44:43.000 --> 1:44:44.960 + how hard is the learning problem here? + +1:44:44.960 --> 1:44:46.960 + I guess is the fundamental question. + +1:44:46.960 --> 1:44:51.800 + So something that's from a hacking perspective, + +1:44:51.800 --> 1:44:54.200 + this is always comes up, especially with folks. + +1:44:54.200 --> 1:44:56.480 + Well, first, the most popular question is + +1:44:56.480 --> 1:44:58.440 + the trolley problem, right? + +1:44:58.440 --> 1:45:01.960 + So that's not a sort of a serious problem. + +1:45:01.960 --> 1:45:05.000 + There are some ethical questions, I think that arise. + +1:45:06.080 --> 1:45:09.600 + Maybe you wanna, do you think there's any ethical, + +1:45:09.600 --> 1:45:11.280 + serious ethical questions? + +1:45:11.280 --> 1:45:14.080 + We have a solution to the trolley problem at com.ai. + +1:45:14.080 --> 1:45:15.920 + Well, so there is actually an alert + +1:45:15.920 --> 1:45:18.000 + in our code, ethical dilemma detected. + +1:45:18.000 --> 1:45:18.960 + It's not triggered yet. + +1:45:18.960 --> 1:45:21.040 + We don't know how yet to detect the ethical dilemmas, + +1:45:21.040 --> 1:45:22.360 + but we're a level two system. + +1:45:22.360 --> 1:45:23.760 + So we're going to disengage and leave + +1:45:23.760 --> 1:45:25.320 + that decision to the human. + +1:45:25.320 --> 1:45:26.680 + You're such a troll. + +1:45:26.680 --> 1:45:28.760 + No, but the trolley problem deserves to be trolled. + +1:45:28.760 --> 1:45:32.040 + Yeah, that's a beautiful answer actually. + +1:45:32.040 --> 1:45:34.440 + I know, I gave it to someone who was like, + +1:45:34.440 --> 1:45:36.600 + sometimes people ask like you asked about the trolley problem. + +1:45:36.600 --> 1:45:38.080 + Like you can have a kind of discussion about it. + +1:45:38.080 --> 1:45:39.720 + Like when you get someone who's like really like + +1:45:39.720 --> 1:45:43.600 + earnest about it, because it's the kind of thing where + +1:45:43.600 --> 1:45:45.600 + if you ask a bunch of people in an office, + +1:45:45.600 --> 1:45:48.360 + whether we should use a SQL stack or no SQL stack, + +1:45:48.360 --> 1:45:50.600 + if they're not that technical, they have no opinion. + +1:45:50.600 --> 1:45:52.360 + But if you ask them what color they want to paint the office, + +1:45:52.360 --> 1:45:54.040 + everyone has an opinion on that. + +1:45:54.040 --> 1:45:56.040 + And that's why the trolley problem is. + +1:45:56.040 --> 1:45:57.280 + I mean, that's a beautiful answer. + +1:45:57.280 --> 1:45:59.240 + Yeah, we're able to detect the problem + +1:45:59.240 --> 1:46:01.960 + and we're able to pass it on to the human. + +1:46:01.960 --> 1:46:03.760 + Wow, I've never heard anyone say it. + +1:46:03.760 --> 1:46:06.160 + This is your nice escape route. + +1:46:06.160 --> 1:46:07.320 + Okay, but... + +1:46:07.320 --> 1:46:08.680 + Proud level two. + +1:46:08.680 --> 1:46:09.760 + I'm proud level two. + +1:46:09.760 --> 1:46:10.600 + I love it. + +1:46:10.600 --> 1:46:14.400 + So the other thing that people have some concern about + +1:46:14.400 --> 1:46:17.800 + with AI in general is hacking. + +1:46:17.800 --> 1:46:21.400 + So how hard is it, do you think, to hack an autonomous vehicle + +1:46:21.400 --> 1:46:25.000 + either through physical access or through the more sort of + +1:46:25.000 --> 1:46:28.240 + popular now, these adversarial examples on the sensors? + +1:46:28.240 --> 1:46:30.720 + Okay, the adversarial examples one. + +1:46:30.720 --> 1:46:32.320 + You want to see some adversarial examples + +1:46:32.320 --> 1:46:34.880 + that affect humans, right? + +1:46:34.880 --> 1:46:38.040 + Oh, well, there used to be a stop sign here, + +1:46:38.040 --> 1:46:40.000 + but I put a black bag over the stop sign + +1:46:40.000 --> 1:46:43.520 + and then people ran it, adversarial, right? + +1:46:43.520 --> 1:46:48.360 + Like, there's tons of human adversarial examples too. + +1:46:48.360 --> 1:46:52.240 + The question in general about security, if you saw, + +1:46:52.240 --> 1:46:54.040 + something just came out today and there are always + +1:46:54.040 --> 1:46:57.560 + such hypey headlines about how navigate on autopilot + +1:46:57.560 --> 1:47:00.960 + was fooled by a GPS spoof to take an exit. + +1:47:00.960 --> 1:47:01.800 + Right. + +1:47:01.800 --> 1:47:03.920 + At least that's all they could do was take an exit. + +1:47:03.920 --> 1:47:06.720 + If your car is relying on GPS in order + +1:47:06.720 --> 1:47:10.240 + to have a safe driving policy, you're doing something wrong. + +1:47:10.240 --> 1:47:12.680 + If you're relying, and this is why V2V + +1:47:12.680 --> 1:47:17.680 + is such a terrible idea, V2V now relies on both parties + +1:47:18.160 --> 1:47:19.800 + getting communication right. + +1:47:19.800 --> 1:47:24.800 + This is not even, so I think of safety, + +1:47:26.080 --> 1:47:28.480 + security is like a special case of safety, right? + +1:47:28.480 --> 1:47:31.880 + Safety is like we put a little, you know, + +1:47:31.880 --> 1:47:33.360 + piece of caution tape around the hole + +1:47:33.360 --> 1:47:35.560 + so that people won't walk into it by accident. + +1:47:35.560 --> 1:47:38.240 + Security is like put a 10 foot fence around the hole + +1:47:38.240 --> 1:47:40.120 + so you actually physically cannot climb into it + +1:47:40.120 --> 1:47:42.360 + with barbed wire on the top and stuff, right? + +1:47:42.360 --> 1:47:44.560 + So like if you're designing systems + +1:47:44.560 --> 1:47:47.440 + that are like unreliable, they're definitely not secure. + +1:47:48.440 --> 1:47:51.240 + Your car should always do something safe + +1:47:51.240 --> 1:47:53.400 + using its local sensors. + +1:47:53.400 --> 1:47:55.240 + And then the local sensor should be hardwired. + +1:47:55.240 --> 1:47:57.400 + And then could somebody hack into your can boss + +1:47:57.400 --> 1:47:58.640 + and turn your steering wheel on your brakes? + +1:47:58.640 --> 1:48:01.240 + Yes, but they could do it before comma AI too, so. + +1:48:02.800 --> 1:48:04.680 + Let's think out of the box and some things. + +1:48:04.680 --> 1:48:09.400 + So do you think teleoperation has a role in any of this? + +1:48:09.400 --> 1:48:13.880 + So remotely stepping in and controlling the cars? + +1:48:13.880 --> 1:48:18.880 + No, I think that if the safety operation + +1:48:21.320 --> 1:48:26.160 + by design requires a constant link to the cars, + +1:48:26.160 --> 1:48:27.560 + I think it doesn't work. + +1:48:27.560 --> 1:48:31.080 + So that's the same argument used for V2I, V2V. + +1:48:31.080 --> 1:48:34.280 + Well, there's a lot of non safety critical stuff + +1:48:34.280 --> 1:48:35.120 + you can do with V2I. + +1:48:35.120 --> 1:48:37.440 + I like V2I, I like V2I way more than V2V + +1:48:37.440 --> 1:48:39.280 + because V2I is already like, + +1:48:39.280 --> 1:48:40.880 + I already have internet in the car, right? + +1:48:40.880 --> 1:48:43.280 + There's a lot of great stuff you can do with V2I. + +1:48:44.280 --> 1:48:46.320 + Like for example, you can, + +1:48:46.320 --> 1:48:48.880 + well, where I already have V2V, Waze is V2I, right? + +1:48:48.880 --> 1:48:50.520 + Waze can route me around traffic jams. + +1:48:50.520 --> 1:48:52.760 + That's a great example of V2I. + +1:48:52.760 --> 1:48:54.440 + And then, okay, the car automatically talks + +1:48:54.440 --> 1:48:55.800 + to that same service, like it works. + +1:48:55.800 --> 1:48:56.800 + So it's improving the experience, + +1:48:56.800 --> 1:48:59.480 + but it's not a fundamental fallback for safety. + +1:48:59.480 --> 1:49:04.160 + No, if any of your things that require + +1:49:04.160 --> 1:49:07.480 + wireless communication are more than QM, + +1:49:07.480 --> 1:49:10.640 + like have an ASL rating, you shouldn't. + +1:49:10.640 --> 1:49:14.200 + You previously said that life is work + +1:49:15.440 --> 1:49:17.480 + and that you don't do anything to relax. + +1:49:17.480 --> 1:49:20.800 + So how do you think about hard work? + +1:49:20.800 --> 1:49:22.200 + Well, what is it? + +1:49:22.200 --> 1:49:24.720 + What do you think it takes to accomplish great things? + +1:49:24.720 --> 1:49:25.840 + And there's a lot of people saying + +1:49:25.840 --> 1:49:28.280 + that there needs to be some balance. + +1:49:28.280 --> 1:49:29.600 + You know, you need to, + +1:49:29.600 --> 1:49:31.120 + in order to accomplish great things, + +1:49:31.120 --> 1:49:32.200 + you need to take some time off, + +1:49:32.200 --> 1:49:34.640 + you need to reflect and so on. + +1:49:34.640 --> 1:49:37.840 + And then some people are just insanely working, + +1:49:37.840 --> 1:49:39.640 + burning the candle on both ends. + +1:49:39.640 --> 1:49:41.360 + How do you think about that? + +1:49:41.360 --> 1:49:43.400 + I think I was trolling in the Suraj interview + +1:49:43.400 --> 1:49:45.600 + when I said that off camera, + +1:49:45.600 --> 1:49:47.240 + but right before I spoke to a little bit of weed, + +1:49:47.240 --> 1:49:49.800 + like, you know, come on, this is a joke, right? + +1:49:49.800 --> 1:49:50.880 + Like I do nothing to relax. + +1:49:50.880 --> 1:49:52.560 + Look where I am, I'm at a party, right? + +1:49:52.560 --> 1:49:53.960 + Yeah, yeah, yeah, sure. + +1:49:53.960 --> 1:49:55.200 + That's true. + +1:49:55.200 --> 1:49:58.040 + So no, no, of course I don't. + +1:49:58.040 --> 1:49:59.800 + When I say that life is work though, + +1:49:59.800 --> 1:50:01.960 + I mean that like, I think that + +1:50:01.960 --> 1:50:04.200 + what gives my life meaning is work. + +1:50:04.200 --> 1:50:05.720 + I don't mean that every minute of the day + +1:50:05.720 --> 1:50:06.560 + you should be working. + +1:50:06.560 --> 1:50:08.000 + I actually think this is not the best way + +1:50:08.000 --> 1:50:09.800 + to maximize results. + +1:50:09.800 --> 1:50:12.040 + I think that if you're working 12 hours a day, + +1:50:12.040 --> 1:50:14.920 + you should be working smarter and not harder. + +1:50:14.920 --> 1:50:17.880 + Well, so it gives work gives you meaning + +1:50:17.880 --> 1:50:20.520 + for some people, other sorts of meaning + +1:50:20.520 --> 1:50:24.560 + is personal relationships, like family and so on. + +1:50:24.560 --> 1:50:27.200 + You've also in that interview with Suraj + +1:50:27.200 --> 1:50:30.720 + or the trolling mentioned that one of the things + +1:50:30.720 --> 1:50:33.400 + you look forward to in the future is AI girlfriends. + +1:50:33.400 --> 1:50:34.360 + Yes. + +1:50:34.360 --> 1:50:38.800 + So that's a topic that I'm very much fascinated by, + +1:50:38.800 --> 1:50:39.840 + not necessarily girlfriends, + +1:50:39.840 --> 1:50:41.880 + but just forming a deep connection with AI. + +1:50:41.880 --> 1:50:42.960 + Yeah. + +1:50:42.960 --> 1:50:44.400 + What kind of system do you imagine + +1:50:44.400 --> 1:50:46.240 + when you say AI girlfriend, + +1:50:46.240 --> 1:50:47.800 + whether you were trolling or not? + +1:50:47.800 --> 1:50:49.720 + No, that one I'm very serious about. + +1:50:49.720 --> 1:50:52.360 + And I'm serious about that on both a shallow level + +1:50:52.360 --> 1:50:53.680 + and a deep level. + +1:50:53.680 --> 1:50:55.720 + I think that VR brothels are coming soon + +1:50:55.720 --> 1:50:57.800 + and are gonna be really cool. + +1:50:57.800 --> 1:50:59.760 + It's not cheating if it's a robot. + +1:50:59.760 --> 1:51:01.080 + I see the slogan already. + +1:51:01.080 --> 1:51:04.320 + Um, but... + +1:51:04.320 --> 1:51:06.200 + There's a, I don't know if you've watched + +1:51:06.200 --> 1:51:08.320 + or just watched the Black Mirror episode. + +1:51:08.320 --> 1:51:09.320 + I watched the latest one, yeah. + +1:51:09.320 --> 1:51:11.320 + Yeah, yeah. + +1:51:11.320 --> 1:51:13.160 + Oh, the Ashley 2 one? + +1:51:13.160 --> 1:51:15.120 + Or the... + +1:51:15.120 --> 1:51:16.920 + No, where there's two friends + +1:51:16.920 --> 1:51:20.160 + who were having sex with each other in... + +1:51:20.160 --> 1:51:21.240 + Oh, in the VR game. + +1:51:21.240 --> 1:51:23.560 + In the VR game, it's the two guys, + +1:51:23.560 --> 1:51:26.720 + but one of them was a female, yeah. + +1:51:26.720 --> 1:51:27.560 + Yeah, the... + +1:51:27.560 --> 1:51:29.560 + Which is another mind blowing concept. + +1:51:29.560 --> 1:51:33.320 + That in VR, you don't have to be the form. + +1:51:33.320 --> 1:51:37.720 + You can be two animals having sex, it's weird. + +1:51:37.720 --> 1:51:38.560 + I mean, I'll see how nice + +1:51:38.560 --> 1:51:40.280 + that the software maps the nerve endings, right? + +1:51:40.280 --> 1:51:41.600 + Yeah, it's weird. + +1:51:41.600 --> 1:51:44.480 + I mean, yeah, they sweep a lot of the fascinating, + +1:51:44.480 --> 1:51:46.440 + really difficult technical challenges under the rug, + +1:51:46.440 --> 1:51:48.360 + like assuming it's possible + +1:51:48.360 --> 1:51:51.160 + to do the mapping of the nerve endings, then... + +1:51:51.160 --> 1:51:52.000 + I wish, yeah, I saw that. + +1:51:52.000 --> 1:51:53.800 + The way they did it with the little like stim unit + +1:51:53.800 --> 1:51:55.400 + on the head, that'd be amazing. + +1:51:56.800 --> 1:51:58.760 + So, well, no, no, on a shallow level, + +1:51:58.760 --> 1:52:01.640 + like you could set up like almost a brothel + +1:52:01.640 --> 1:52:05.160 + with like real dolls and Oculus quests, + +1:52:05.160 --> 1:52:06.200 + write some good software. + +1:52:06.200 --> 1:52:08.280 + I think it'd be a cool novelty experience. + +1:52:09.280 --> 1:52:11.400 + But no, on a deeper, like emotional level. + +1:52:12.800 --> 1:52:16.960 + I mean, yeah, I would really like to fall in love + +1:52:16.960 --> 1:52:18.120 + with the machine. + +1:52:18.120 --> 1:52:23.120 + Do you see yourself having a long term relationship + +1:52:23.120 --> 1:52:27.520 + of the kind monogamous relationship that we have now + +1:52:27.520 --> 1:52:31.360 + with the robot, with the AI system, even? + +1:52:31.360 --> 1:52:32.680 + Not even just the robot. + +1:52:32.680 --> 1:52:37.680 + So, I think about maybe my ideal future. + +1:52:38.200 --> 1:52:43.200 + When I was 15, I read Eliezer Yudkowsky's early writings + +1:52:44.320 --> 1:52:49.120 + on the singularity and like that AI + +1:52:49.120 --> 1:52:53.040 + is going to surpass human intelligence massively. + +1:52:53.040 --> 1:52:55.480 + He made some Moore's law based predictions + +1:52:55.480 --> 1:52:57.400 + that I mostly agree with. + +1:52:57.400 --> 1:52:59.360 + And then I really struggled + +1:52:59.360 --> 1:53:01.360 + for the next couple of years of my life. + +1:53:01.360 --> 1:53:03.360 + Like, why should I even bother to learn anything? + +1:53:03.360 --> 1:53:06.160 + It's all gonna be meaningless when the machine show up. + +1:53:06.160 --> 1:53:07.000 + Right. + +1:53:07.000 --> 1:53:10.520 + Well, maybe when I was that young, + +1:53:10.520 --> 1:53:12.040 + I was still a little bit more pure + +1:53:12.040 --> 1:53:13.160 + and really like clung to that. + +1:53:13.160 --> 1:53:14.720 + And then I'm like, well, the machine's ain't here yet. + +1:53:14.720 --> 1:53:16.800 + You know, and I seem to be pretty good at this stuff. + +1:53:16.800 --> 1:53:18.520 + Let's try my best, you know, + +1:53:18.520 --> 1:53:20.320 + like what's the worst that happens? + +1:53:20.320 --> 1:53:23.440 + But the best possible future I see + +1:53:23.440 --> 1:53:26.120 + is me sort of merging with the machine. + +1:53:26.120 --> 1:53:28.120 + And the way that I personify this + +1:53:28.120 --> 1:53:30.800 + is in a longterm and augments relationship with the machine. + +1:53:32.160 --> 1:53:33.320 + Oh, you don't think there's room + +1:53:33.320 --> 1:53:35.040 + for another human in your life + +1:53:35.040 --> 1:53:37.440 + if you really truly merge with another machine? + +1:53:38.440 --> 1:53:40.240 + I mean, I see merging. + +1:53:40.240 --> 1:53:44.240 + I see like the best interface to my brain + +1:53:45.520 --> 1:53:48.000 + is like the same relationship interface + +1:53:48.000 --> 1:53:49.320 + to merge with an AI, right? + +1:53:49.320 --> 1:53:51.440 + What does that merging feel like? + +1:53:52.440 --> 1:53:55.320 + I've seen couples who've been together for a long time + +1:53:55.320 --> 1:53:57.840 + and like, I almost think of them as one person. + +1:53:57.840 --> 1:54:01.280 + Like couples who spend all their time together and... + +1:54:01.280 --> 1:54:02.120 + That's fascinating. + +1:54:02.120 --> 1:54:03.320 + You're actually putting, + +1:54:03.320 --> 1:54:05.520 + what does that merging actually looks like? + +1:54:05.520 --> 1:54:07.600 + It's not just a nice channel. + +1:54:07.600 --> 1:54:11.640 + Like a lot of people imagine it's just an efficient link, + +1:54:11.640 --> 1:54:13.800 + search link to Wikipedia or something. + +1:54:13.800 --> 1:54:14.640 + I don't believe in that. + +1:54:14.640 --> 1:54:17.120 + But it's more, you're saying that there's the same kind of, + +1:54:17.120 --> 1:54:19.520 + the same kind of relationship you have with another human + +1:54:19.520 --> 1:54:22.960 + as a deep relationship is that's what merging looks like. + +1:54:22.960 --> 1:54:24.480 + That's pretty... + +1:54:24.480 --> 1:54:26.680 + I don't believe that link is possible. + +1:54:26.680 --> 1:54:28.120 + I think that that link, so you're like, + +1:54:28.120 --> 1:54:30.160 + oh, I'm gonna download Wikipedia right to my brain. + +1:54:30.160 --> 1:54:33.360 + My reading speed is not limited by my eyes. + +1:54:33.360 --> 1:54:36.800 + My reading speed is limited by my inner processing loop. + +1:54:36.800 --> 1:54:38.680 + And to like bootstrap that + +1:54:38.680 --> 1:54:42.440 + sounds kind of unclear how to do it and horrifying. + +1:54:42.440 --> 1:54:46.560 + But if I am with somebody, and I'll use somebody + +1:54:46.560 --> 1:54:51.400 + who is making a super sophisticated model of me + +1:54:51.400 --> 1:54:53.200 + and then running simulations on that model, + +1:54:53.200 --> 1:54:54.120 + I'm not gonna get into the question + +1:54:54.120 --> 1:54:55.880 + whether the simulations are conscious or not. + +1:54:55.880 --> 1:54:58.240 + I don't really wanna know what it's doing. + +1:54:58.240 --> 1:55:01.600 + But using those simulations to play out hypothetical futures + +1:55:01.600 --> 1:55:04.880 + for me, deciding what things to say to me + +1:55:04.880 --> 1:55:08.720 + to guide me along a path and that's how I envision it. + +1:55:08.720 --> 1:55:13.720 + So on that path to AI of super human level intelligence, + +1:55:13.720 --> 1:55:15.680 + you've mentioned that you believe in the singularity, + +1:55:15.680 --> 1:55:17.280 + that singularity is coming. + +1:55:17.280 --> 1:55:20.440 + Again, could be trolling, could be not, could be part... + +1:55:20.440 --> 1:55:21.760 + All trolling has truth in it. + +1:55:21.760 --> 1:55:22.840 + I don't know what that means anymore. + +1:55:22.840 --> 1:55:24.520 + What is the singularity? + +1:55:24.520 --> 1:55:26.720 + So yeah, so that's really the question. + +1:55:26.720 --> 1:55:29.280 + How many years do you think before the singularity + +1:55:29.280 --> 1:55:30.920 + of what form do you think it will take? + +1:55:30.920 --> 1:55:34.200 + Does that mean fundamental shifts in capabilities of AI? + +1:55:34.200 --> 1:55:36.960 + Does it mean some other kind of ideas? + +1:55:36.960 --> 1:55:40.120 + Maybe that's just my roots, but... + +1:55:40.120 --> 1:55:42.920 + So I can buy a human being's worth of computers + +1:55:42.920 --> 1:55:46.000 + for things worth of compute for like a million bucks a day. + +1:55:46.000 --> 1:55:47.800 + It's about one TPU pod V3. + +1:55:47.800 --> 1:55:50.240 + I want like, I think they claim a hundred pay to flops. + +1:55:50.240 --> 1:55:51.080 + That's being generous. + +1:55:51.080 --> 1:55:52.320 + I think humans are actually more like 20. + +1:55:52.320 --> 1:55:53.160 + So that's like five humans. + +1:55:53.160 --> 1:55:54.040 + That's pretty good. + +1:55:54.040 --> 1:55:55.560 + Google needs to sell their TPUs. + +1:55:56.840 --> 1:55:58.640 + But no, I could buy GPUs. + +1:55:58.640 --> 1:56:02.280 + I could buy a stack of like, I buy 1080TIs, + +1:56:02.280 --> 1:56:03.880 + build data center full of them. + +1:56:03.880 --> 1:56:07.280 + And for a million bucks, I can get a human worth of compute. + +1:56:08.160 --> 1:56:12.280 + But when you look at the total number of flops in the world, + +1:56:12.280 --> 1:56:14.400 + when you look at human flops, + +1:56:14.400 --> 1:56:17.040 + which goes up very, very slowly with the population, + +1:56:17.040 --> 1:56:19.760 + and machine flops, which goes up exponentially, + +1:56:19.760 --> 1:56:22.360 + but it's still nowhere near. + +1:56:22.360 --> 1:56:24.040 + I think that's the key thing + +1:56:24.040 --> 1:56:25.880 + to talk about when the singularity happened. + +1:56:25.880 --> 1:56:28.560 + When most flops in the world are silicon + +1:56:28.560 --> 1:56:32.280 + and not biological, that's kind of the crossing point. + +1:56:32.280 --> 1:56:35.480 + Like they are now the dominant species on the planet. + +1:56:35.480 --> 1:56:38.720 + And just looking at how technology is progressing, + +1:56:38.720 --> 1:56:40.360 + when do you think that could possibly happen? + +1:56:40.360 --> 1:56:41.680 + Do you think it would happen in your lifetime? + +1:56:41.680 --> 1:56:43.640 + Oh yeah, definitely in my lifetime. + +1:56:43.640 --> 1:56:44.480 + I've done the math. + +1:56:44.480 --> 1:56:47.560 + I like 2038 because it's the UNIX timestamp roll over. + +1:56:49.920 --> 1:56:51.840 + Yeah, beautifully put. + +1:56:52.680 --> 1:56:57.680 + So you've said that the meaning of life is to win. + +1:56:58.000 --> 1:56:59.560 + If you look five years into the future, + +1:56:59.560 --> 1:57:01.000 + what does winning look like? + +1:57:02.640 --> 1:57:03.720 + So... + +1:57:03.720 --> 1:57:08.720 + I can go into technical depth to what I mean by that, to win. + +1:57:11.720 --> 1:57:12.720 + It may not mean... + +1:57:12.720 --> 1:57:14.400 + I was criticized for that in the comments. + +1:57:14.400 --> 1:57:17.720 + Like, doesn't this guy want to save the penguins in Antarctica? + +1:57:17.720 --> 1:57:20.960 + Or like, oh man, listen to what I'm saying. + +1:57:20.960 --> 1:57:23.720 + I'm not talking about like I have a yacht or something. + +1:57:24.720 --> 1:57:26.720 + I am an agent. + +1:57:26.720 --> 1:57:28.720 + I am put into this world. + +1:57:28.720 --> 1:57:32.720 + And I don't really know what my purpose is. + +1:57:33.720 --> 1:57:36.720 + But if you're a reinforcement, if you're an intelligent agent + +1:57:36.720 --> 1:57:39.720 + and you're put into a world, what is the ideal thing to do? + +1:57:39.720 --> 1:57:41.720 + Well, the ideal thing, mathematically, + +1:57:41.720 --> 1:57:43.720 + you can go back to like Schmidt Hoover theories about this, + +1:57:43.720 --> 1:57:46.720 + is to build a compressive model of the world. + +1:57:46.720 --> 1:57:49.720 + To build a maximally compressive to explore the world + +1:57:49.720 --> 1:57:52.720 + such that your exploration function maximizes + +1:57:52.720 --> 1:57:55.720 + the derivative of compression of the past. + +1:57:55.720 --> 1:57:58.720 + Schmidt Hoover has a paper about this. + +1:57:58.720 --> 1:58:01.720 + And like, I took that kind of as like a personal goal function. + +1:58:02.720 --> 1:58:04.720 + So what I mean to win, I mean like, + +1:58:04.720 --> 1:58:08.720 + maybe this is religious, but like I think that in the future + +1:58:08.720 --> 1:58:10.720 + I might be given a real purpose. + +1:58:10.720 --> 1:58:12.720 + Or I may decide this purpose myself. + +1:58:12.720 --> 1:58:14.720 + And then at that point, now I know what the game is + +1:58:14.720 --> 1:58:15.720 + and I know how to win. + +1:58:15.720 --> 1:58:18.720 + I think right now I'm still just trying to figure out what the game is. + +1:58:18.720 --> 1:58:19.720 + But once I know... + +1:58:20.720 --> 1:58:22.720 + So you have... + +1:58:22.720 --> 1:58:25.720 + You have imperfect information. + +1:58:25.720 --> 1:58:27.720 + You have a lot of uncertainty about the reward function + +1:58:27.720 --> 1:58:28.720 + and you're discovering it. + +1:58:28.720 --> 1:58:29.720 + Exactly. + +1:58:29.720 --> 1:58:30.720 + But the purpose is... + +1:58:30.720 --> 1:58:31.720 + That's a better way to put it. + +1:58:31.720 --> 1:58:33.720 + The purpose is to maximize it + +1:58:33.720 --> 1:58:36.720 + while you have a lot of uncertainty around it. + +1:58:36.720 --> 1:58:38.720 + And you're both reducing the uncertainty + +1:58:38.720 --> 1:58:40.720 + and maximizing at the same time. + +1:58:40.720 --> 1:58:43.720 + And so that's at the technical level. + +1:58:43.720 --> 1:58:44.720 + What is the... + +1:58:44.720 --> 1:58:46.720 + If you believe in the universal prior, + +1:58:46.720 --> 1:58:48.720 + what is the universal reward function? + +1:58:48.720 --> 1:58:50.720 + That's the better way to put it. + +1:58:50.720 --> 1:58:53.720 + So that win is interesting. + +1:58:53.720 --> 1:58:56.720 + I think I speak for everyone in saying that + +1:58:56.720 --> 1:59:01.720 + I wonder what that reward function is for you. + +1:59:01.720 --> 1:59:06.720 + And I look forward to seeing that in five years and ten years. + +1:59:06.720 --> 1:59:09.720 + I think a lot of people including myself are cheering you on, man. + +1:59:09.720 --> 1:59:11.720 + So I'm happy you exist. + +1:59:11.720 --> 1:59:13.720 + And I wish you the best of luck. + +1:59:13.720 --> 1:59:14.720 + Thanks for talking today, man. + +1:59:14.720 --> 1:59:15.720 + Thank you. + +1:59:15.720 --> 1:59:20.720 + This was a lot of fun. +