diff --git "a/vtt/episode_021_small.vtt" "b/vtt/episode_021_small.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_021_small.vtt" @@ -0,0 +1,5168 @@ +WEBVTT + +00:00.000 --> 00:02.680 + The following is a conversation with Chris Latner. + +00:02.680 --> 00:05.680 + Currently, he's a senior director of Google working + +00:05.680 --> 00:09.560 + on several projects, including CPU, GPU, TPU accelerators + +00:09.560 --> 00:12.080 + for TensorFlow, Swift for TensorFlow, + +00:12.080 --> 00:14.400 + and all kinds of machine learning compiler magic + +00:14.400 --> 00:16.360 + going on behind the scenes. + +00:16.360 --> 00:18.440 + He's one of the top experts in the world + +00:18.440 --> 00:20.640 + on compiler technologies, which means + +00:20.640 --> 00:23.960 + he deeply understands the intricacies of how + +00:23.960 --> 00:26.280 + hardware and software come together + +00:26.280 --> 00:27.960 + to create efficient code. + +00:27.960 --> 00:31.480 + He created the LLVM compiler infrastructure project + +00:31.480 --> 00:33.400 + and the Clang compiler. + +00:33.400 --> 00:36.040 + He led major engineering efforts at Apple, + +00:36.040 --> 00:39.040 + including the creation of the Swift programming language. + +00:39.040 --> 00:42.600 + He also briefly spent time at Tesla as vice president + +00:42.600 --> 00:45.920 + of autopilot software during the transition from autopilot + +00:45.920 --> 00:49.640 + hardware one to hardware two, when Tesla essentially + +00:49.640 --> 00:52.640 + started from scratch to build an in house software + +00:52.640 --> 00:54.880 + infrastructure for autopilot. + +00:54.880 --> 00:58.040 + I could have easily talked to Chris for many more hours. + +00:58.040 --> 01:01.240 + Compiling code down across the level's abstraction + +01:01.240 --> 01:04.120 + is one of the most fundamental and fascinating aspects + +01:04.120 --> 01:07.160 + of what computers do, and he is one of the world experts + +01:07.160 --> 01:08.640 + in this process. + +01:08.640 --> 01:12.920 + It's rigorous science, and it's messy, beautiful art. + +01:12.920 --> 01:15.920 + This conversation is part of the Artificial Intelligence + +01:15.920 --> 01:16.760 + podcast. + +01:16.760 --> 01:18.920 + If you enjoy it, subscribe on YouTube, + +01:18.920 --> 01:21.560 + iTunes, or simply connect with me on Twitter + +01:21.560 --> 01:25.760 + at Lex Friedman, spelled F R I D. And now, here's + +01:25.760 --> 01:29.400 + my conversation with Chris Ladner. + +01:29.400 --> 01:33.240 + What was the first program you've ever written? + +01:33.240 --> 01:34.680 + My first program back. + +01:34.680 --> 01:35.400 + And when was it? + +01:35.400 --> 01:39.160 + I think I started as a kid, and my parents + +01:39.160 --> 01:41.640 + got a basic programming book. + +01:41.640 --> 01:45.400 + And so when I started, it was typing out programs from a book + +01:45.400 --> 01:49.320 + and seeing how they worked, and then typing them in wrong + +01:49.320 --> 01:51.600 + and trying to figure out why they were not working right, + +01:51.600 --> 01:52.960 + and that kind of stuff. + +01:52.960 --> 01:54.880 + So basic, what was the first language + +01:54.880 --> 01:58.360 + that you remember yourself maybe falling in love with, + +01:58.360 --> 01:59.960 + like really connecting with? + +01:59.960 --> 02:00.480 + I don't know. + +02:00.480 --> 02:02.480 + I mean, I feel like I've learned a lot along the way, + +02:02.480 --> 02:06.720 + and each of them have a different, special thing about them. + +02:06.720 --> 02:09.880 + So I started in basic, and then went like GW basic, which + +02:09.880 --> 02:11.440 + was the thing back in the DOS days, + +02:11.440 --> 02:15.320 + and then upgraded to Q basic, and eventually Quick basic, + +02:15.320 --> 02:17.760 + which are all slightly more fancy versions + +02:17.760 --> 02:20.920 + of Microsoft basic, made the jump to Pascal, + +02:20.920 --> 02:23.960 + and started doing machine language programming and assembly + +02:23.960 --> 02:25.280 + in Pascal, which was really cool. + +02:25.280 --> 02:28.120 + Turbo Pascal was amazing for its day. + +02:28.120 --> 02:31.600 + Eventually, gone to C, C++, and then kind of did + +02:31.600 --> 02:33.440 + lots of other weird things. + +02:33.440 --> 02:36.640 + I feel like you took the dark path, which is the, + +02:36.640 --> 02:41.520 + you could have gone Lisp, you could have gone a higher level + +02:41.520 --> 02:44.600 + sort of functional, philosophical, hippie route. + +02:44.600 --> 02:48.040 + Instead, you went into like the dark arts of the C. + +02:48.040 --> 02:49.720 + It was straight into the machine. + +02:49.720 --> 02:50.680 + Straight into the machine. + +02:50.680 --> 02:53.880 + So started with basic Pascal and then assembly, + +02:53.880 --> 02:58.120 + and then wrote a lot of assembly, and eventually did + +02:58.120 --> 03:00.120 + small talk and other things like that, + +03:00.120 --> 03:01.920 + but that was not the starting point. + +03:01.920 --> 03:05.080 + But so what is this journey to see? + +03:05.080 --> 03:06.360 + Is that in high school? + +03:06.360 --> 03:07.560 + Is that in college? + +03:07.560 --> 03:08.800 + That was in high school, yeah. + +03:08.800 --> 03:13.760 + So, and then that was really about trying + +03:13.760 --> 03:16.120 + to be able to do more powerful things than what Pascal + +03:16.120 --> 03:18.920 + could do and also to learn a different world. + +03:18.920 --> 03:20.720 + So he was really confusing to me with pointers + +03:20.720 --> 03:22.800 + and the syntax and everything, and it took a while, + +03:22.800 --> 03:28.720 + but Pascal's much more principled in various ways, + +03:28.720 --> 03:33.360 + sees more, I mean, it has its historical roots, + +03:33.360 --> 03:35.440 + but it's not as easy to learn. + +03:35.440 --> 03:39.840 + With pointers, there's this memory management thing + +03:39.840 --> 03:41.640 + that you have to become conscious of. + +03:41.640 --> 03:43.880 + Is that the first time you start to understand + +03:43.880 --> 03:46.480 + that there's resources that you're supposed to manage? + +03:46.480 --> 03:48.480 + Well, so you have that in Pascal as well, + +03:48.480 --> 03:51.920 + but in Pascal, the carrot instead of the star, + +03:51.920 --> 03:53.160 + there's some small differences like that, + +03:53.160 --> 03:55.640 + but it's not about pointer arithmetic. + +03:55.640 --> 03:58.920 + And see, you end up thinking about how things get laid + +03:58.920 --> 04:00.800 + out in memory a lot more. + +04:00.800 --> 04:04.160 + And so in Pascal, you have allocating and deallocating + +04:04.160 --> 04:07.520 + and owning the memory, but just the programs are simpler + +04:07.520 --> 04:10.880 + and you don't have to, well, for example, + +04:10.880 --> 04:12.600 + Pascal has a string type. + +04:12.600 --> 04:14.040 + And so you can think about a string + +04:14.040 --> 04:15.880 + instead of an array of characters + +04:15.880 --> 04:17.680 + which are consecutive in memory. + +04:17.680 --> 04:20.360 + So it's a little bit of a higher level abstraction. + +04:20.360 --> 04:22.760 + So let's get into it. + +04:22.760 --> 04:25.600 + Let's talk about LLVM, Selang, and compilers. + +04:25.600 --> 04:26.520 + Sure. + +04:26.520 --> 04:31.520 + So can you tell me first what LLVM and Selang are, + +04:32.120 --> 04:34.560 + and how is it that you find yourself the creator + +04:34.560 --> 04:37.720 + and lead developer, one of the most powerful + +04:37.720 --> 04:40.080 + compiler optimization systems in use today? + +04:40.080 --> 04:43.240 + Sure, so I guess they're different things. + +04:43.240 --> 04:47.040 + So let's start with what is a compiler? + +04:47.040 --> 04:48.840 + Is that a good place to start? + +04:48.840 --> 04:50.200 + What are the phases of a compiler? + +04:50.200 --> 04:51.040 + Where are the parts? + +04:51.040 --> 04:51.880 + Yeah, what is it? + +04:51.880 --> 04:53.400 + So what is even a compiler used for? + +04:53.400 --> 04:57.560 + So the way I look at this is you have a two sided problem + +04:57.560 --> 05:00.440 + of you have humans that need to write code. + +05:00.440 --> 05:02.360 + And then you have machines that need to run the program + +05:02.360 --> 05:03.360 + that the human wrote. + +05:03.360 --> 05:05.720 + And for lots of reasons, the humans don't want to be + +05:05.720 --> 05:08.320 + writing in binary and want to think about every piece + +05:08.320 --> 05:09.160 + of hardware. + +05:09.160 --> 05:12.080 + So at the same time that you have lots of humans, + +05:12.080 --> 05:14.760 + you also have lots of kinds of hardware. + +05:14.760 --> 05:17.760 + And so compilers are the art of allowing humans + +05:17.760 --> 05:19.200 + to think at a level of abstraction + +05:19.200 --> 05:20.880 + that they want to think about. + +05:20.880 --> 05:23.600 + And then get that program, get the thing that they wrote + +05:23.600 --> 05:26.040 + to run on a specific piece of hardware. + +05:26.040 --> 05:29.480 + And the interesting and exciting part of all this + +05:29.480 --> 05:31.960 + is that there's now lots of different kinds of hardware, + +05:31.960 --> 05:35.880 + chips like x86 and PowerPC and ARM and things like that, + +05:35.880 --> 05:37.720 + but also high performance accelerators for machine + +05:37.720 --> 05:39.920 + learning and other things like that are also just different + +05:39.920 --> 05:42.880 + kinds of hardware, GPUs, these are new kinds of hardware. + +05:42.880 --> 05:45.600 + And at the same time on the programming side of it, + +05:45.600 --> 05:48.640 + you have basic, you have C, you have JavaScript, + +05:48.640 --> 05:51.480 + you have Python, you have Swift, you have like lots + +05:51.480 --> 05:54.440 + of other languages that are all trying to talk to the human + +05:54.440 --> 05:57.040 + in a different way to make them more expressive + +05:57.040 --> 05:58.320 + and capable and powerful. + +05:58.320 --> 06:02.080 + And so compilers are the thing that goes from one + +06:02.080 --> 06:03.440 + to the other now. + +06:03.440 --> 06:05.200 + And to end from the very beginning to the very end. + +06:05.200 --> 06:08.120 + And to end, and so you go from what the human wrote + +06:08.120 --> 06:12.600 + and programming languages end up being about expressing intent, + +06:12.600 --> 06:15.960 + not just for the compiler and the hardware, + +06:15.960 --> 06:20.320 + but the programming language's job is really to capture + +06:20.320 --> 06:22.640 + an expression of what the programmer wanted + +06:22.640 --> 06:25.080 + that then can be maintained and adapted + +06:25.080 --> 06:28.240 + and evolved by other humans, as well as by the, + +06:28.240 --> 06:29.680 + interpreted by the compiler. + +06:29.680 --> 06:31.520 + So when you look at this problem, + +06:31.520 --> 06:34.160 + you have on the one hand humans, which are complicated, + +06:34.160 --> 06:36.720 + you have hardware, which is complicated. + +06:36.720 --> 06:39.880 + And so compilers typically work in multiple phases. + +06:39.880 --> 06:42.720 + And so the software engineering challenge + +06:42.720 --> 06:44.960 + that you have here is try to get maximum reuse + +06:44.960 --> 06:47.080 + out of the amount of code that you write + +06:47.080 --> 06:49.720 + because these compilers are very complicated. + +06:49.720 --> 06:51.840 + And so the way it typically works out is that + +06:51.840 --> 06:54.440 + you have something called a front end or a parser + +06:54.440 --> 06:56.600 + that is language specific. + +06:56.600 --> 06:59.440 + And so you'll have a C parser, that's what Clang is, + +07:00.360 --> 07:03.440 + or C++ or JavaScript or Python or whatever, + +07:03.440 --> 07:04.960 + that's the front end. + +07:04.960 --> 07:07.080 + Then you'll have a middle part, + +07:07.080 --> 07:09.000 + which is often the optimizer. + +07:09.000 --> 07:11.120 + And then you'll have a late part, + +07:11.120 --> 07:13.320 + which is hardware specific. + +07:13.320 --> 07:16.680 + And so compilers end up, there's many different layers often, + +07:16.680 --> 07:20.880 + but these three big groups are very common in compilers. + +07:20.880 --> 07:23.440 + And what LLVM is trying to do is trying to standardize + +07:23.440 --> 07:25.360 + that middle and last part. + +07:25.360 --> 07:27.880 + And so one of the cool things about LLVM + +07:27.880 --> 07:29.760 + is that there are a lot of different languages + +07:29.760 --> 07:31.080 + that compile through to it. + +07:31.080 --> 07:35.640 + And so things like Swift, but also Julia, Rust, + +07:36.520 --> 07:39.120 + Clang for C, C++, Subjective C, + +07:39.120 --> 07:40.920 + like these are all very different languages + +07:40.920 --> 07:43.800 + and they can all use the same optimization infrastructure, + +07:43.800 --> 07:45.400 + which gets better performance, + +07:45.400 --> 07:47.240 + and the same code generation infrastructure + +07:47.240 --> 07:48.800 + for hardware support. + +07:48.800 --> 07:52.240 + And so LLVM is really that layer that is common, + +07:52.240 --> 07:55.560 + that all these different specific compilers can use. + +07:55.560 --> 07:59.280 + And is it a standard, like a specification, + +07:59.280 --> 08:01.160 + or is it literally an implementation? + +08:01.160 --> 08:02.120 + It's an implementation. + +08:02.120 --> 08:05.880 + And so it's, I think there's a couple of different ways + +08:05.880 --> 08:06.720 + of looking at it, right? + +08:06.720 --> 08:09.680 + Because it depends on which angle you're looking at it from. + +08:09.680 --> 08:12.600 + LLVM ends up being a bunch of code, okay? + +08:12.600 --> 08:14.440 + So it's a bunch of code that people reuse + +08:14.440 --> 08:16.520 + and they build compilers with. + +08:16.520 --> 08:18.040 + We call it a compiler infrastructure + +08:18.040 --> 08:20.000 + because it's kind of the underlying platform + +08:20.000 --> 08:22.520 + that you build a concrete compiler on top of. + +08:22.520 --> 08:23.680 + But it's also a community. + +08:23.680 --> 08:26.800 + And the LLVM community is hundreds of people + +08:26.800 --> 08:27.920 + that all collaborate. + +08:27.920 --> 08:30.560 + And one of the most fascinating things about LLVM + +08:30.560 --> 08:34.280 + over the course of time is that we've managed somehow + +08:34.280 --> 08:37.080 + to successfully get harsh competitors + +08:37.080 --> 08:39.080 + in the commercial space to collaborate + +08:39.080 --> 08:41.120 + on shared infrastructure. + +08:41.120 --> 08:43.880 + And so you have Google and Apple. + +08:43.880 --> 08:45.880 + You have AMD and Intel. + +08:45.880 --> 08:48.880 + You have NVIDIA and AMD on the graphics side. + +08:48.880 --> 08:52.640 + You have Cray and everybody else doing these things. + +08:52.640 --> 08:55.400 + And like all these companies are collaborating together + +08:55.400 --> 08:57.480 + to make that shared infrastructure + +08:57.480 --> 08:58.520 + really, really great. + +08:58.520 --> 09:01.400 + And they do this not out of the goodness of their heart + +09:01.400 --> 09:03.440 + but they do it because it's in their commercial interest + +09:03.440 --> 09:05.160 + of having really great infrastructure + +09:05.160 --> 09:06.800 + that they can build on top of. + +09:06.800 --> 09:09.120 + And facing the reality that it's so expensive + +09:09.120 --> 09:11.200 + that no one company, even the big companies, + +09:11.200 --> 09:14.600 + no one company really wants to implement it all themselves. + +09:14.600 --> 09:16.120 + Expensive or difficult? + +09:16.120 --> 09:16.960 + Both. + +09:16.960 --> 09:20.600 + That's a great point because it's also about the skill sets. + +09:20.600 --> 09:25.600 + And these, the skill sets are very hard to find. + +09:25.600 --> 09:27.960 + How big is the LLVM? + +09:27.960 --> 09:30.400 + It always seems like with open source projects, + +09:30.400 --> 09:33.480 + the kind, and LLVM is open source? + +09:33.480 --> 09:34.440 + Yes, it's open source. + +09:34.440 --> 09:36.320 + It's about, it's 19 years old now. + +09:36.320 --> 09:38.640 + So it's fairly old. + +09:38.640 --> 09:40.960 + It seems like the magic often happens + +09:40.960 --> 09:43.040 + within a very small circle of people. + +09:43.040 --> 09:43.880 + Yes. + +09:43.880 --> 09:46.080 + At least at early birth and whatever. + +09:46.080 --> 09:46.920 + Yes. + +09:46.920 --> 09:49.640 + So the LLVM came from a university project. + +09:49.640 --> 09:51.640 + And so I was at the University of Illinois. + +09:51.640 --> 09:53.880 + And there it was myself, my advisor, + +09:53.880 --> 09:57.480 + and then a team of two or three research students + +09:57.480 --> 09:58.360 + in the research group. + +09:58.360 --> 10:02.080 + And we built many of the core pieces initially. + +10:02.080 --> 10:03.720 + I then graduated and went to Apple. + +10:03.720 --> 10:06.480 + And then Apple brought it to the products, + +10:06.480 --> 10:09.320 + first in the OpenGL graphics stack, + +10:09.320 --> 10:11.600 + but eventually to the C compiler realm + +10:11.600 --> 10:12.760 + and eventually built Clang + +10:12.760 --> 10:14.640 + and eventually built Swift and these things. + +10:14.640 --> 10:16.360 + Along the way, building a team of people + +10:16.360 --> 10:18.600 + that are really amazing compiler engineers + +10:18.600 --> 10:20.120 + that helped build a lot of that. + +10:20.120 --> 10:21.840 + And so as it was gaining momentum + +10:21.840 --> 10:24.800 + and as Apple was using it, being open source and public + +10:24.800 --> 10:27.040 + and encouraging contribution, many others, + +10:27.040 --> 10:30.400 + for example, at Google, came in and started contributing. + +10:30.400 --> 10:33.680 + And in some cases, Google effectively owns Clang now + +10:33.680 --> 10:35.520 + because it cares so much about C++ + +10:35.520 --> 10:37.280 + and the evolution of that ecosystem. + +10:37.280 --> 10:41.400 + And so it's investing a lot in the C++ world + +10:41.400 --> 10:42.960 + and the tooling and things like that. + +10:42.960 --> 10:47.840 + And so likewise, NVIDIA cares a lot about CUDA. + +10:47.840 --> 10:52.840 + And so CUDA uses Clang and uses LVM for graphics and GPGPU. + +10:54.960 --> 10:59.880 + And so when you first started as a master's project, I guess, + +10:59.880 --> 11:02.920 + did you think it was gonna go as far as it went? + +11:02.920 --> 11:06.280 + Were you crazy ambitious about it? + +11:06.280 --> 11:07.120 + No. + +11:07.120 --> 11:09.760 + It seems like a really difficult undertaking, a brave one. + +11:09.760 --> 11:11.280 + Yeah, no, no, it was nothing like that. + +11:11.280 --> 11:13.640 + So I mean, my goal when I went to University of Illinois + +11:13.640 --> 11:16.120 + was to get in and out with the non thesis masters + +11:16.120 --> 11:18.680 + in a year and get back to work. + +11:18.680 --> 11:22.160 + So I was not planning to stay for five years + +11:22.160 --> 11:24.440 + and build this massive infrastructure. + +11:24.440 --> 11:27.400 + I got nerd sniped into staying. + +11:27.400 --> 11:29.480 + And a lot of it was because LVM was fun + +11:29.480 --> 11:30.920 + and I was building cool stuff + +11:30.920 --> 11:33.400 + and learning really interesting things + +11:33.400 --> 11:36.880 + and facing both software engineering challenges + +11:36.880 --> 11:38.520 + but also learning how to work in a team + +11:38.520 --> 11:40.120 + and things like that. + +11:40.120 --> 11:43.600 + I had worked at many companies as interns before that, + +11:43.600 --> 11:45.840 + but it was really a different thing + +11:45.840 --> 11:48.120 + to have a team of people that were working together + +11:48.120 --> 11:50.480 + and trying to collaborate in version control + +11:50.480 --> 11:52.400 + and it was just a little bit different. + +11:52.400 --> 11:54.080 + Like I said, I just talked to Don Knuth + +11:54.080 --> 11:56.840 + and he believes that 2% of the world population + +11:56.840 --> 11:59.600 + have something weird with their brain, that they're geeks, + +11:59.600 --> 12:02.560 + they understand computers, they're connected with computers. + +12:02.560 --> 12:04.360 + He put it at exactly 2%. + +12:04.360 --> 12:05.560 + Okay, so... + +12:05.560 --> 12:06.560 + Is this a specific act? + +12:06.560 --> 12:08.760 + It's very specific. + +12:08.760 --> 12:10.200 + Well, he says, I can't prove it, + +12:10.200 --> 12:11.800 + but it's very empirically there. + +12:13.040 --> 12:14.480 + Is there something that attracts you + +12:14.480 --> 12:16.920 + to the idea of optimizing code? + +12:16.920 --> 12:19.120 + And he seems like that's one of the biggest, + +12:19.120 --> 12:20.920 + coolest things about LVM. + +12:20.920 --> 12:22.480 + Yeah, that's one of the major things it does. + +12:22.480 --> 12:26.440 + So I got into that because of a person, actually. + +12:26.440 --> 12:28.200 + So when I was in my undergraduate, + +12:28.200 --> 12:32.040 + I had an advisor or a professor named Steve Vegdahl + +12:32.040 --> 12:35.760 + and I went to this little tiny private school. + +12:35.760 --> 12:38.280 + There were like seven or nine people + +12:38.280 --> 12:40.320 + in my computer science department, + +12:40.320 --> 12:43.080 + students in my class. + +12:43.080 --> 12:47.440 + So it was a very tiny, very small school. + +12:47.440 --> 12:49.960 + It was kind of a work on the side of the math department + +12:49.960 --> 12:51.240 + kind of a thing at the time. + +12:51.240 --> 12:53.800 + I think it's evolved a lot in the many years since then, + +12:53.800 --> 12:58.280 + but Steve Vegdahl was a compiler guy + +12:58.280 --> 12:59.600 + and he was super passionate + +12:59.600 --> 13:02.720 + and his passion rubbed off on me + +13:02.720 --> 13:04.440 + and one of the things I like about compilers + +13:04.440 --> 13:09.120 + is that they're large, complicated software pieces. + +13:09.120 --> 13:12.920 + And so one of the culminating classes + +13:12.920 --> 13:14.520 + that many computer science departments + +13:14.520 --> 13:16.680 + at least at the time did was to say + +13:16.680 --> 13:18.400 + that you would take algorithms and data structures + +13:18.400 --> 13:19.480 + and all these core classes, + +13:19.480 --> 13:20.720 + but then the compilers class + +13:20.720 --> 13:22.160 + was one of the last classes you take + +13:22.160 --> 13:24.360 + because it pulls everything together + +13:24.360 --> 13:27.000 + and then you work on one piece of code + +13:27.000 --> 13:28.680 + over the entire semester. + +13:28.680 --> 13:32.080 + And so you keep building on your own work, + +13:32.080 --> 13:34.800 + which is really interesting and it's also very challenging + +13:34.800 --> 13:37.520 + because in many classes, if you don't get a project done, + +13:37.520 --> 13:39.320 + you just forget about it and move on to the next one + +13:39.320 --> 13:41.320 + and get your B or whatever it is, + +13:41.320 --> 13:43.880 + but here you have to live with the decisions you make + +13:43.880 --> 13:45.280 + and continue to reinvest in it. + +13:45.280 --> 13:46.880 + And I really like that. + +13:46.880 --> 13:51.080 + And so I did a extra study project with him + +13:51.080 --> 13:53.960 + the following semester and he was just really great + +13:53.960 --> 13:56.920 + and he was also a great mentor in a lot of ways. + +13:56.920 --> 13:59.560 + And so from him and from his advice, + +13:59.560 --> 14:01.520 + he encouraged me to go to graduate school. + +14:01.520 --> 14:03.200 + I wasn't super excited about going to grad school. + +14:03.200 --> 14:05.240 + I wanted the master's degree, + +14:05.240 --> 14:07.440 + but I didn't want to be an academic. + +14:09.000 --> 14:11.160 + But like I said, I kind of got tricked into saying + +14:11.160 --> 14:14.560 + I was having a lot of fun and I definitely do not regret it. + +14:14.560 --> 14:15.840 + Well, the aspects of compilers + +14:15.840 --> 14:17.960 + were the things you connected with. + +14:17.960 --> 14:22.120 + So LVM, there's also the other part + +14:22.120 --> 14:23.440 + that's just really interesting + +14:23.440 --> 14:27.640 + if you're interested in languages is parsing and just analyzing + +14:27.640 --> 14:29.640 + like, yeah, analyzing the language, + +14:29.640 --> 14:31.240 + breaking it down, parsing and so on. + +14:31.240 --> 14:32.280 + Was that interesting to you + +14:32.280 --> 14:34.080 + or were you more interested in optimization? + +14:34.080 --> 14:37.400 + For me, it was more, so I'm not really a math person. + +14:37.400 --> 14:39.600 + I can do math, I understand some bits of it + +14:39.600 --> 14:41.600 + when I get into it, + +14:41.600 --> 14:43.960 + but math is never the thing that attracted me. + +14:43.960 --> 14:46.160 + And so a lot of the parser part of the compiler + +14:46.160 --> 14:48.960 + has a lot of good formal theories that Dawn, for example, + +14:48.960 --> 14:50.440 + knows quite well. + +14:50.440 --> 14:51.920 + Still waiting for his book on that. + +14:51.920 --> 14:56.080 + But I just like building a thing + +14:56.080 --> 14:59.200 + and seeing what it could do and exploring + +14:59.200 --> 15:00.800 + and getting it to do more things + +15:00.800 --> 15:02.880 + and then setting new goals and reaching for them. + +15:02.880 --> 15:08.880 + And in the case of LVM, when I started working on that, + +15:08.880 --> 15:13.360 + my research advisor that I was working for was a compiler guy. + +15:13.360 --> 15:15.600 + And so he and I specifically found each other + +15:15.600 --> 15:16.920 + because we both interested in compilers + +15:16.920 --> 15:19.480 + and so I started working with them and taking his class. + +15:19.480 --> 15:21.800 + And a lot of LVM initially was it's fun + +15:21.800 --> 15:23.560 + implementing all the standard algorithms + +15:23.560 --> 15:26.360 + and all the things that people had been talking about + +15:26.360 --> 15:28.920 + and were well known and they were in the curricula + +15:28.920 --> 15:31.320 + for advanced studies in compilers. + +15:31.320 --> 15:34.560 + And so just being able to build that was really fun + +15:34.560 --> 15:36.160 + and I was learning a lot + +15:36.160 --> 15:38.640 + by instead of reading about it, just building. + +15:38.640 --> 15:40.200 + And so I enjoyed that. + +15:40.200 --> 15:42.800 + So you said compilers are these complicated systems. + +15:42.800 --> 15:47.240 + Can you even just with language try to describe + +15:48.240 --> 15:52.240 + how you turn a C++ program into code? + +15:52.240 --> 15:53.480 + Like what are the hard parts? + +15:53.480 --> 15:54.640 + Why is it so hard? + +15:54.640 --> 15:56.840 + So I'll give you examples of the hard parts along the way. + +15:56.840 --> 16:01.040 + So C++ is a very complicated programming language. + +16:01.040 --> 16:03.480 + It's something like 1,400 pages in the spec. + +16:03.480 --> 16:06.120 + So C++ by itself is crazy complicated. + +16:06.120 --> 16:07.160 + Can we just, sorry, pause. + +16:07.160 --> 16:08.720 + What makes the language complicated + +16:08.720 --> 16:11.520 + in terms of what's syntactically? + +16:11.520 --> 16:14.320 + Like, so it's what they call syntax. + +16:14.320 --> 16:16.280 + So the actual how the characters are arranged. + +16:16.280 --> 16:20.080 + Yes, it's also semantics, how it behaves. + +16:20.080 --> 16:21.720 + It's also in the case of C++. + +16:21.720 --> 16:23.400 + There's a huge amount of history. + +16:23.400 --> 16:25.560 + C++ build on top of C. + +16:25.560 --> 16:28.720 + You play that forward and then a bunch of suboptimal + +16:28.720 --> 16:30.360 + in some cases decisions were made + +16:30.360 --> 16:33.400 + and they compound and then more and more and more things + +16:33.400 --> 16:35.080 + keep getting added to C++ + +16:35.080 --> 16:37.040 + and it will probably never stop. + +16:37.040 --> 16:39.440 + But the language is very complicated from that perspective. + +16:39.440 --> 16:41.200 + And so the interactions between subsystems + +16:41.200 --> 16:42.360 + is very complicated. + +16:42.360 --> 16:43.560 + There's just a lot there. + +16:43.560 --> 16:45.640 + And when you talk about the front end, + +16:45.640 --> 16:48.560 + one of the major challenges which playing as a project, + +16:48.560 --> 16:52.280 + the C++ compiler that I built, I and many people built. + +16:53.320 --> 16:57.560 + One of the challenges we took on was we looked at GCC. + +16:57.560 --> 17:01.120 + I think GCC at the time was like a really good + +17:01.120 --> 17:05.320 + industry standardized compiler that had really consolidated + +17:05.320 --> 17:06.760 + a lot of the other compilers in the world + +17:06.760 --> 17:08.360 + and was a standard. + +17:08.360 --> 17:10.640 + But it wasn't really great for research. + +17:10.640 --> 17:12.600 + The design was very difficult to work with + +17:12.600 --> 17:16.640 + and it was full of global variables and other things + +17:16.640 --> 17:18.120 + that made it very difficult to reuse + +17:18.120 --> 17:20.400 + in ways that it wasn't originally designed for. + +17:20.400 --> 17:22.560 + And so with Clang, one of the things that we wanted to do + +17:22.560 --> 17:25.520 + is push forward on better user interface. + +17:25.520 --> 17:28.160 + So make error messages that are just better than GCCs. + +17:28.160 --> 17:29.920 + And that's actually hard because you have to do + +17:29.920 --> 17:31.880 + a lot of bookkeeping in an efficient way + +17:32.800 --> 17:33.640 + to be able to do that. + +17:33.640 --> 17:35.160 + We want to make compile time better. + +17:35.160 --> 17:37.520 + And so compile time is about making it efficient, + +17:37.520 --> 17:38.920 + which is also really hard when you're keeping + +17:38.920 --> 17:40.520 + track of extra information. + +17:40.520 --> 17:43.400 + We wanted to make new tools available. + +17:43.400 --> 17:46.400 + So refactoring tools and other analysis tools + +17:46.400 --> 17:48.400 + that the GCC never supported, + +17:48.400 --> 17:51.160 + also leveraging the extra information we kept, + +17:52.200 --> 17:54.080 + but enabling those new classes of tools + +17:54.080 --> 17:55.960 + that then get built into IDEs. + +17:55.960 --> 17:58.560 + And so that's been one of the areas + +17:58.560 --> 18:01.320 + that Clang has really helped push the world forward in + +18:01.320 --> 18:05.080 + is in the tooling for C and C++ and things like that. + +18:05.080 --> 18:07.760 + But C++ in the front end piece is complicated + +18:07.760 --> 18:09.040 + and you have to build syntax trees + +18:09.040 --> 18:11.360 + and you have to check every rule in the spec + +18:11.360 --> 18:14.000 + and you have to turn that back into an error message + +18:14.000 --> 18:16.040 + to the human that the human can understand + +18:16.040 --> 18:17.840 + when they do something wrong. + +18:17.840 --> 18:20.760 + But then you start doing what's called lowering. + +18:20.760 --> 18:23.440 + So going from C++ in the way that it represents code + +18:23.440 --> 18:24.960 + down to the machine. + +18:24.960 --> 18:25.800 + And when you do that, + +18:25.800 --> 18:28.240 + there's many different phases you go through. + +18:29.640 --> 18:33.040 + Often there are, I think LVM has something like 150 + +18:33.040 --> 18:36.240 + different, what are called passes in the compiler + +18:36.240 --> 18:38.760 + that the code passes through + +18:38.760 --> 18:41.880 + and these get organized in very complicated ways, + +18:41.880 --> 18:44.360 + which affect the generated code and the performance + +18:44.360 --> 18:46.000 + and compile time and many of the things. + +18:46.000 --> 18:47.320 + What are they passing through? + +18:47.320 --> 18:51.840 + So after you do the Clang parsing, + +18:51.840 --> 18:53.960 + what's the graph? + +18:53.960 --> 18:54.800 + What does it look like? + +18:54.800 --> 18:55.960 + What's the data structure here? + +18:55.960 --> 18:59.040 + Yeah, so in the parser, it's usually a tree + +18:59.040 --> 19:01.040 + and it's called an abstract syntax tree. + +19:01.040 --> 19:04.560 + And so the idea is you have a node for the plus + +19:04.560 --> 19:06.800 + that the human wrote in their code + +19:06.800 --> 19:09.000 + or the function call, you'll have a node for call + +19:09.000 --> 19:11.840 + with the function that they call in the arguments they pass. + +19:11.840 --> 19:12.680 + Things like that. + +19:14.440 --> 19:16.840 + This then gets lowered into what's called + +19:16.840 --> 19:18.600 + an intermediate representation + +19:18.600 --> 19:22.080 + and intermediate representations are like LVM has one. + +19:22.080 --> 19:26.920 + And there it's a, it's what's called a control flow graph. + +19:26.920 --> 19:31.200 + And so you represent each operation in the program + +19:31.200 --> 19:34.480 + as a very simple, like this is gonna add two numbers. + +19:34.480 --> 19:35.880 + This is gonna multiply two things. + +19:35.880 --> 19:37.480 + This maybe we'll do a call, + +19:37.480 --> 19:40.280 + but then they get put in what are called blocks. + +19:40.280 --> 19:43.600 + And so you get blocks of these straight line operations + +19:43.600 --> 19:45.320 + where instead of being nested like in a tree, + +19:45.320 --> 19:46.920 + it's straight line operations. + +19:46.920 --> 19:47.920 + And so there's a sequence + +19:47.920 --> 19:49.760 + in ordering to these operations. + +19:49.760 --> 19:51.840 + So within the block or outside the block? + +19:51.840 --> 19:53.240 + That's within the block. + +19:53.240 --> 19:55.000 + And so it's a straight line sequence of operations + +19:55.000 --> 19:55.840 + within the block. + +19:55.840 --> 19:57.520 + And then you have branches, + +19:57.520 --> 20:00.160 + like conditional branches between blocks. + +20:00.160 --> 20:02.760 + And so when you write a loop, for example, + +20:04.120 --> 20:07.080 + in a syntax tree, you would have a four node + +20:07.080 --> 20:09.080 + like for a four statement in a C like language, + +20:09.080 --> 20:10.840 + you'd have a four node. + +20:10.840 --> 20:12.200 + And you have a pointer to the expression + +20:12.200 --> 20:14.120 + for the initializer, a pointer to the expression + +20:14.120 --> 20:15.840 + for the increment, a pointer to the expression + +20:15.840 --> 20:18.720 + for the comparison, a pointer to the body. + +20:18.720 --> 20:21.040 + Okay, and these are all nested underneath it. + +20:21.040 --> 20:22.880 + In a control flow graph, you get a block + +20:22.880 --> 20:25.960 + for the code that runs before the loop. + +20:25.960 --> 20:27.600 + So the initializer code. + +20:27.600 --> 20:30.280 + Then you have a block for the body of the loop. + +20:30.280 --> 20:33.760 + And so the body of the loop code goes in there, + +20:33.760 --> 20:35.520 + but also the increment and other things like that. + +20:35.520 --> 20:37.800 + And then you have a branch that goes back to the top + +20:37.800 --> 20:39.840 + and a comparison and a branch that goes out. + +20:39.840 --> 20:44.000 + And so it's more of a assembly level kind of representation. + +20:44.000 --> 20:46.040 + But the nice thing about this level of representation + +20:46.040 --> 20:48.680 + is it's much more language independent. + +20:48.680 --> 20:51.880 + And so there's lots of different kinds of languages + +20:51.880 --> 20:54.520 + with different kinds of, you know, + +20:54.520 --> 20:56.600 + JavaScript has a lot of different ideas + +20:56.600 --> 20:58.160 + of what is false, for example, + +20:58.160 --> 21:00.760 + and all that can stay in the front end, + +21:00.760 --> 21:04.200 + but then that middle part can be shared across all of those. + +21:04.200 --> 21:07.520 + How close is that intermediate representation + +21:07.520 --> 21:10.280 + to new networks, for example? + +21:10.280 --> 21:14.320 + Are they, because everything you describe as a kind of + +21:14.320 --> 21:16.080 + a close of a neural network graph, + +21:16.080 --> 21:18.920 + are they neighbors or what? + +21:18.920 --> 21:20.960 + They're quite different in details, + +21:20.960 --> 21:22.480 + but they're very similar in idea. + +21:22.480 --> 21:24.000 + So one of the things that normal networks do + +21:24.000 --> 21:26.880 + is they learn representations for data + +21:26.880 --> 21:29.120 + at different levels of abstraction, right? + +21:29.120 --> 21:32.360 + And then they transform those through layers, right? + +21:33.920 --> 21:35.680 + So the compiler does very similar things, + +21:35.680 --> 21:37.120 + but one of the things the compiler does + +21:37.120 --> 21:40.640 + is it has relatively few different representations. + +21:40.640 --> 21:42.480 + Where a neural network, often as you get deeper, + +21:42.480 --> 21:44.800 + for example, you get many different representations + +21:44.800 --> 21:47.400 + and each, you know, layer or set of ops + +21:47.400 --> 21:50.200 + is transforming between these different representations. + +21:50.200 --> 21:53.080 + In a compiler, often you get one representation + +21:53.080 --> 21:55.240 + and they do many transformations to it. + +21:55.240 --> 21:59.520 + And these transformations are often applied iteratively. + +21:59.520 --> 22:02.920 + And for programmers, they're familiar types of things. + +22:02.920 --> 22:06.160 + For example, trying to find expressions inside of a loop + +22:06.160 --> 22:07.320 + and pulling them out of a loop. + +22:07.320 --> 22:08.560 + So if they execute fairer times + +22:08.560 --> 22:10.760 + or find redundant computation + +22:10.760 --> 22:15.360 + or find constant folding or other simplifications + +22:15.360 --> 22:19.040 + turning, you know, two times X into X shift left by one + +22:19.040 --> 22:21.960 + and things like this are all the examples + +22:21.960 --> 22:23.360 + of the things that happen. + +22:23.360 --> 22:26.200 + But compilers end up getting a lot of theorem proving + +22:26.200 --> 22:27.640 + and other kinds of algorithms + +22:27.640 --> 22:29.960 + that try to find higher level properties of the program + +22:29.960 --> 22:32.320 + that then can be used by the optimizer. + +22:32.320 --> 22:35.920 + Cool, so what's like the biggest bang for the buck + +22:35.920 --> 22:37.680 + with optimization? + +22:37.680 --> 22:38.720 + What's a day? + +22:38.720 --> 22:39.560 + Yeah. + +22:39.560 --> 22:40.920 + Well, no, not even today. + +22:40.920 --> 22:42.800 + At the very beginning, the 80s, I don't know. + +22:42.800 --> 22:43.960 + Yeah, so for the 80s, + +22:43.960 --> 22:46.440 + a lot of it was things like register allocation. + +22:46.440 --> 22:51.000 + So the idea of in a modern, like a microprocessor, + +22:51.000 --> 22:52.760 + what you'll end up having is you'll end up having memory, + +22:52.760 --> 22:54.320 + which is relatively slow. + +22:54.320 --> 22:57.080 + And then you have registers relatively fast, + +22:57.080 --> 22:59.920 + but registers, you don't have very many of them. + +22:59.920 --> 23:02.600 + Okay, and so when you're writing a bunch of code, + +23:02.600 --> 23:04.200 + you're just saying like, compute this, + +23:04.200 --> 23:05.520 + put in temporary variable, compute this, + +23:05.520 --> 23:07.800 + compute this, put in temporary variable, + +23:07.800 --> 23:09.760 + I have a loop, I have some other stuff going on. + +23:09.760 --> 23:11.680 + Well, now you're running on an x86, + +23:11.680 --> 23:13.920 + like a desktop PC or something. + +23:13.920 --> 23:16.160 + Well, it only has, in some cases, + +23:16.160 --> 23:18.720 + some modes, eight registers, right? + +23:18.720 --> 23:20.800 + And so now the compiler has to choose + +23:20.800 --> 23:22.800 + what values get put in what registers, + +23:22.800 --> 23:24.840 + at what points in the program. + +23:24.840 --> 23:26.480 + And this is actually a really big deal. + +23:26.480 --> 23:28.560 + So if you think about, you have a loop, + +23:28.560 --> 23:31.640 + an inner loop that executes millions of times maybe. + +23:31.640 --> 23:33.600 + If you're doing loads and stores inside that loop, + +23:33.600 --> 23:34.920 + then it's gonna be really slow. + +23:34.920 --> 23:37.080 + But if you can somehow fit all the values + +23:37.080 --> 23:40.200 + inside that loop in registers, now it's really fast. + +23:40.200 --> 23:43.400 + And so getting that right requires a lot of work, + +23:43.400 --> 23:44.960 + because there's many different ways to do that. + +23:44.960 --> 23:47.000 + And often what the compiler ends up doing + +23:47.000 --> 23:48.880 + is it ends up thinking about things + +23:48.880 --> 23:51.920 + in a different representation than what the human wrote. + +23:51.920 --> 23:53.320 + All right, you wrote into x. + +23:53.320 --> 23:56.800 + Well, the compiler thinks about that as four different values, + +23:56.800 --> 23:58.360 + each which have different lifetimes + +23:58.360 --> 24:00.400 + across the function that it's in. + +24:00.400 --> 24:02.640 + And each of those could be put in a register + +24:02.640 --> 24:05.840 + or memory or different memory, or maybe in some parts + +24:05.840 --> 24:08.760 + of the code, recompute it instead of stored and reloaded. + +24:08.760 --> 24:10.000 + And there are many of these different kinds + +24:10.000 --> 24:11.440 + of techniques that can be used. + +24:11.440 --> 24:14.840 + So it's adding almost like a time dimension + +24:14.840 --> 24:18.320 + to it's trying to optimize across time. + +24:18.320 --> 24:20.360 + So it's considering when you're programming, + +24:20.360 --> 24:21.920 + you're not thinking in that way. + +24:21.920 --> 24:23.200 + Yeah, absolutely. + +24:23.200 --> 24:28.200 + And so the risk era made things, so risk chips, RISC, + +24:28.200 --> 24:33.200 + RISC, the risk chips as opposed to SISC chips, + +24:33.680 --> 24:36.000 + the risk chips made things more complicated + +24:36.000 --> 24:39.720 + for the compiler because what they ended up doing + +24:39.720 --> 24:42.360 + is ending up adding pipelines to the processor + +24:42.360 --> 24:45.000 + where the processor can do more than one thing at a time. + +24:45.000 --> 24:47.600 + But this means that the order of operations matters a lot. + +24:47.600 --> 24:49.720 + And so one of the classical compiler techniques + +24:49.720 --> 24:52.000 + that you use is called scheduling. + +24:52.000 --> 24:54.200 + And so moving the instructions around + +24:54.200 --> 24:57.400 + so that the processor can like keep its pipelines full + +24:57.400 --> 24:59.600 + instead of stalling and getting blocked. + +24:59.600 --> 25:00.960 + And so there's a lot of things like that + +25:00.960 --> 25:03.600 + that are kind of bread and butter or compiler techniques + +25:03.600 --> 25:06.240 + that have been studied a lot over the course of decades now. + +25:06.240 --> 25:08.520 + But the engineering side of making them real + +25:08.520 --> 25:10.680 + is also still quite hard. + +25:10.680 --> 25:12.400 + And you talk about machine learning, + +25:12.400 --> 25:14.400 + this is a huge opportunity for machine learning + +25:14.400 --> 25:16.520 + because many of these algorithms + +25:16.520 --> 25:19.120 + are full of these like hokey hand rolled heuristics + +25:19.120 --> 25:20.880 + which work well on specific benchmarks + +25:20.880 --> 25:23.920 + that don't generalize and full of magic numbers. + +25:23.920 --> 25:26.520 + And I hear there's some techniques + +25:26.520 --> 25:28.000 + that are good at handling that. + +25:28.000 --> 25:29.880 + So what would be the, + +25:29.880 --> 25:33.040 + if you were to apply machine learning to this, + +25:33.040 --> 25:34.720 + what's the thing you try to optimize? + +25:34.720 --> 25:38.080 + Is it ultimately the running time? + +25:38.080 --> 25:39.960 + Yeah, you can pick your metric + +25:39.960 --> 25:42.240 + and there's running time, there's memory use, + +25:42.240 --> 25:44.760 + there's lots of different things that you can optimize + +25:44.760 --> 25:47.200 + for code size is another one that some people care about + +25:47.200 --> 25:48.800 + in the embedded space. + +25:48.800 --> 25:51.680 + Is this like the thinking into the future + +25:51.680 --> 25:55.600 + or has somebody actually been crazy enough to try + +25:55.600 --> 25:59.080 + to have machine learning based parameter tuning + +25:59.080 --> 26:01.040 + for optimization of compilers? + +26:01.040 --> 26:04.840 + So this is something that is, I would say research right now. + +26:04.840 --> 26:06.800 + There are a lot of research systems + +26:06.800 --> 26:09.080 + that have been applying search in various forms + +26:09.080 --> 26:11.440 + and using reinforcement learning as one form, + +26:11.440 --> 26:14.400 + but also brute force search has been tried for quite a while. + +26:14.400 --> 26:18.160 + And usually these are in small problem spaces. + +26:18.160 --> 26:21.480 + So find the optimal way to code generate + +26:21.480 --> 26:23.680 + a matrix multiply for a GPU, right? + +26:23.680 --> 26:25.480 + Something like that where you say, + +26:25.480 --> 26:28.080 + there there's a lot of design space + +26:28.080 --> 26:29.920 + of do you unroll loops a lot? + +26:29.920 --> 26:32.600 + Do you execute multiple things in parallel? + +26:32.600 --> 26:35.320 + And there's many different confounding factors here + +26:35.320 --> 26:38.120 + because graphics cards have different numbers of threads + +26:38.120 --> 26:41.040 + and registers and execution ports and memory bandwidth + +26:41.040 --> 26:42.760 + and many different constraints to interact + +26:42.760 --> 26:44.280 + in nonlinear ways. + +26:44.280 --> 26:46.480 + And so search is very powerful for that + +26:46.480 --> 26:49.840 + and it gets used in certain ways, + +26:49.840 --> 26:51.240 + but it's not very structured. + +26:51.240 --> 26:52.640 + This is something that we need, + +26:52.640 --> 26:54.520 + we as an industry need to fix. + +26:54.520 --> 26:56.240 + So you said 80s, but like, + +26:56.240 --> 26:59.960 + so have there been like big jumps in improvement + +26:59.960 --> 27:01.280 + and optimization? + +27:01.280 --> 27:02.360 + Yeah. + +27:02.360 --> 27:05.320 + Yeah, since then, what's the coolest thing about it? + +27:05.320 --> 27:07.120 + It's largely been driven by hardware. + +27:07.120 --> 27:09.880 + So hardware and software. + +27:09.880 --> 27:13.880 + So in the mid 90s, Java totally changed the world, right? + +27:13.880 --> 27:17.520 + And I'm still amazed by how much change was introduced + +27:17.520 --> 27:19.320 + by Java in a good way or in a good way. + +27:19.320 --> 27:20.600 + So like reflecting back, + +27:20.600 --> 27:23.800 + Java introduced things like all at once introduced things + +27:23.800 --> 27:25.680 + like JIT compilation. + +27:25.680 --> 27:26.920 + None of these were novel, + +27:26.920 --> 27:28.640 + but it pulled it together and made it mainstream + +27:28.640 --> 27:30.600 + and made people invest in it. + +27:30.600 --> 27:32.680 + JIT compilation, garbage collection, + +27:32.680 --> 27:36.680 + portable code, safe code, like memory safe code, + +27:37.680 --> 27:41.480 + like a very dynamic dispatch execution model. + +27:41.480 --> 27:42.680 + Like many of these things, + +27:42.680 --> 27:44.120 + which had been done in research systems + +27:44.120 --> 27:46.960 + and had been done in small ways in various places, + +27:46.960 --> 27:48.040 + really came to the forefront + +27:48.040 --> 27:49.840 + and really changed how things worked. + +27:49.840 --> 27:52.040 + And therefore changed the way people thought + +27:52.040 --> 27:53.120 + about the problem. + +27:53.120 --> 27:56.360 + JavaScript was another major world change + +27:56.360 --> 27:57.780 + based on the way it works. + +27:59.320 --> 28:01.240 + But also on the hardware side of things, + +28:02.240 --> 28:05.200 + multi core and vector instructions + +28:05.200 --> 28:07.520 + really change the problem space + +28:07.520 --> 28:10.800 + and are very, they don't remove any of the problems + +28:10.800 --> 28:12.360 + that compilers faced in the past, + +28:12.360 --> 28:14.560 + but they add new kinds of problems + +28:14.560 --> 28:16.400 + of how do you find enough work + +28:16.400 --> 28:20.040 + to keep a four wide vector busy, right? + +28:20.040 --> 28:22.640 + Or if you're doing a matrix multiplication, + +28:22.640 --> 28:25.360 + how do you do different columns out of that matrix + +28:25.360 --> 28:26.680 + in at the same time? + +28:26.680 --> 28:30.160 + And how do you maximum utilize the arithmetic compute + +28:30.160 --> 28:31.440 + that one core has? + +28:31.440 --> 28:33.480 + And then how do you take it to multiple cores? + +28:33.480 --> 28:35.040 + How did the whole virtual machine thing + +28:35.040 --> 28:37.960 + change the compilation pipeline? + +28:37.960 --> 28:40.440 + Yeah, so what the Java virtual machine does + +28:40.440 --> 28:44.160 + is it splits, just like I was talking about before, + +28:44.160 --> 28:46.280 + where you have a front end that parses the code + +28:46.280 --> 28:47.960 + and then you have an intermediate representation + +28:47.960 --> 28:49.400 + that gets transformed. + +28:49.400 --> 28:50.960 + What Java did was they said, + +28:50.960 --> 28:52.720 + we will parse the code and then compile + +28:52.720 --> 28:55.480 + to what's known as Java bytecode. + +28:55.480 --> 28:58.560 + And that bytecode is now a portable code representation + +28:58.560 --> 29:02.400 + that is industry standard and locked down and can't change. + +29:02.400 --> 29:05.040 + And then the back part of the compiler + +29:05.040 --> 29:07.280 + that does optimization and code generation + +29:07.280 --> 29:09.440 + can now be built by different vendors. + +29:09.440 --> 29:12.080 + Okay, and Java bytecode can be shipped around + +29:12.080 --> 29:15.840 + across the wire, it's memory safe and relatively trusted. + +29:16.840 --> 29:18.680 + And because of that it can run in the browser. + +29:18.680 --> 29:20.480 + And that's why it runs in the browser, right? + +29:20.480 --> 29:22.960 + And so that way you can be in, you know, + +29:22.960 --> 29:25.000 + again, back in the day, you would write a Java applet + +29:25.000 --> 29:27.720 + and you'd use it as a web developer, + +29:27.720 --> 29:30.840 + you'd build this mini app that would run on a web page. + +29:30.840 --> 29:33.600 + Well, a user of that is running a web browser + +29:33.600 --> 29:36.160 + on their computer, you download that Java bytecode, + +29:36.160 --> 29:39.280 + which can be trusted, and then you do + +29:39.280 --> 29:41.040 + all the compiler stuff on your machine + +29:41.040 --> 29:42.400 + so that you know that you trust that. + +29:42.400 --> 29:44.080 + Is that a good idea or a bad idea? + +29:44.080 --> 29:44.920 + It's a great idea, I mean, + +29:44.920 --> 29:46.200 + it's a great idea for certain problems. + +29:46.200 --> 29:48.200 + And I'm very much a believer + +29:48.200 --> 29:50.480 + that the technology is itself neither good nor bad, + +29:50.480 --> 29:51.600 + it's how you apply it. + +29:52.920 --> 29:54.600 + You know, this would be a very, very bad thing + +29:54.600 --> 29:56.960 + for very low levels of the software stack, + +29:56.960 --> 30:00.280 + but in terms of solving some of these software portability + +30:00.280 --> 30:02.760 + and transparency or portability problems, + +30:02.760 --> 30:04.200 + I think it's been really good. + +30:04.200 --> 30:06.560 + Now Java ultimately didn't win out on the desktop + +30:06.560 --> 30:09.400 + and like there are good reasons for that, + +30:09.400 --> 30:13.200 + but it's been very successful on servers and in many places, + +30:13.200 --> 30:16.280 + it's been a very successful thing over decades. + +30:16.280 --> 30:21.280 + So what has been LLVM's and Selang's improvements + +30:24.480 --> 30:28.720 + and optimization that throughout its history, + +30:28.720 --> 30:31.080 + what are some moments we had set back + +30:31.080 --> 30:33.280 + and really proud of what's been accomplished? + +30:33.280 --> 30:36.200 + Yeah, I think that the interesting thing about LLVM + +30:36.200 --> 30:40.120 + is not the innovations in compiler research, + +30:40.120 --> 30:41.880 + it has very good implementations + +30:41.880 --> 30:43.880 + of very important algorithms, no doubt. + +30:43.880 --> 30:48.280 + And a lot of really smart people have worked on it, + +30:48.280 --> 30:50.560 + but I think that the thing that's most profound about LLVM + +30:50.560 --> 30:52.600 + is that through standardization, + +30:52.600 --> 30:55.720 + it made things possible that otherwise wouldn't have happened. + +30:55.720 --> 30:56.560 + Okay. + +30:56.560 --> 30:59.120 + And so interesting things that have happened with LLVM, + +30:59.120 --> 31:01.280 + for example, Sony has picked up LLVM + +31:01.280 --> 31:03.920 + and used it to do all the graphics compilation + +31:03.920 --> 31:06.080 + in their movie production pipeline. + +31:06.080 --> 31:07.920 + And so now they're able to have better special effects + +31:07.920 --> 31:09.680 + because of LLVM. + +31:09.680 --> 31:11.200 + That's kind of cool. + +31:11.200 --> 31:13.000 + That's not what it was designed for, right? + +31:13.000 --> 31:15.480 + But that's the sign of good infrastructure + +31:15.480 --> 31:18.800 + when it can be used in ways it was never designed for + +31:18.800 --> 31:20.960 + because it has good layering and software engineering + +31:20.960 --> 31:23.440 + and it's composable and things like that. + +31:23.440 --> 31:26.120 + Which is where, as you said, it differs from GCC. + +31:26.120 --> 31:28.240 + Yes, GCC is also great in various ways, + +31:28.240 --> 31:31.800 + but it's not as good as infrastructure technology. + +31:31.800 --> 31:36.120 + It's really a C compiler, or it's a 4 train compiler. + +31:36.120 --> 31:39.200 + It's not infrastructure in the same way. + +31:39.200 --> 31:40.400 + Is it, now you can tell, + +31:40.400 --> 31:41.560 + I don't know what I'm talking about + +31:41.560 --> 31:43.680 + because I keep saying C lang. + +31:44.520 --> 31:48.080 + You can always tell when a person is closed, + +31:48.080 --> 31:49.400 + by the way, pronounce something. + +31:49.400 --> 31:52.600 + I don't think, have I ever used Clang? + +31:52.600 --> 31:53.440 + Entirely possible. + +31:53.440 --> 31:55.680 + Have you, well, so you've used code, + +31:55.680 --> 31:58.200 + it's generated probably. + +31:58.200 --> 32:01.760 + So Clang is an LLVM or used to compile + +32:01.760 --> 32:05.240 + all the apps on the iPhone effectively and the OSes. + +32:05.240 --> 32:09.360 + It compiles Google's production server applications. + +32:09.360 --> 32:14.360 + It's used to build GameCube games and PlayStation 4 + +32:14.880 --> 32:16.720 + and things like that. + +32:16.720 --> 32:17.920 + Those are the user I have, + +32:17.920 --> 32:20.800 + but just everything I've done that I experienced + +32:20.800 --> 32:23.600 + with Linux has been, I believe, always GCC. + +32:23.600 --> 32:25.720 + Yeah, I think Linux still defaults to GCC. + +32:25.720 --> 32:27.840 + And is there a reason for that? + +32:27.840 --> 32:29.480 + Or is it, I mean, is there a reason? + +32:29.480 --> 32:32.080 + It's a combination of technical and social reasons. + +32:32.080 --> 32:36.000 + Many Linux developers do use Clang, + +32:36.000 --> 32:40.600 + but the distributions, for lots of reasons, + +32:40.600 --> 32:44.280 + use GCC historically and they've not switched, yeah. + +32:44.280 --> 32:46.680 + Because it's just anecdotally online, + +32:46.680 --> 32:50.680 + it seems that LLVM has either reached the level of GCC + +32:50.680 --> 32:53.560 + or superseded on different features or whatever. + +32:53.560 --> 32:55.240 + The way I would say it is that they're so close + +32:55.240 --> 32:56.080 + it doesn't matter. + +32:56.080 --> 32:56.920 + Yeah, exactly. + +32:56.920 --> 32:58.160 + Like they're slightly better in some ways, + +32:58.160 --> 32:59.200 + slightly worse than otherwise, + +32:59.200 --> 33:03.320 + but it doesn't actually really matter anymore at that level. + +33:03.320 --> 33:06.320 + So in terms of optimization, breakthroughs, + +33:06.320 --> 33:09.200 + it's just been solid incremental work. + +33:09.200 --> 33:12.200 + Yeah, yeah, which describes a lot of compilers. + +33:12.200 --> 33:14.360 + The hard thing about compilers, + +33:14.360 --> 33:16.000 + in my experience, is the engineering, + +33:16.000 --> 33:18.680 + the software engineering, making it + +33:18.680 --> 33:20.920 + so that you can have hundreds of people collaborating + +33:20.920 --> 33:25.400 + on really detailed low level work and scaling that. + +33:25.400 --> 33:27.880 + And that's really hard. + +33:27.880 --> 33:30.720 + And that's one of the things I think LLVM has done well. + +33:30.720 --> 33:34.160 + And that kind of goes back to the original design goals + +33:34.160 --> 33:37.160 + with it to be modular and things like that. + +33:37.160 --> 33:38.840 + And incidentally, I don't want to take all the credit + +33:38.840 --> 33:39.680 + for this, right? + +33:39.680 --> 33:41.760 + I mean, some of the best parts about LLVM + +33:41.760 --> 33:43.600 + is that it was designed to be modular. + +33:43.600 --> 33:44.960 + And when I started, I would write, + +33:44.960 --> 33:46.840 + for example, a register allocator, + +33:46.840 --> 33:49.040 + and then somebody much smarter than me would come in + +33:49.040 --> 33:51.320 + and pull it out and replace it with something else + +33:51.320 --> 33:52.640 + that they would come up with. + +33:52.640 --> 33:55.160 + And because it's modular, they were able to do that. + +33:55.160 --> 33:58.240 + And that's one of the challenges with GCC, for example, + +33:58.240 --> 34:01.240 + is replacing subsystems is incredibly difficult. + +34:01.240 --> 34:04.640 + It can be done, but it wasn't designed for that. + +34:04.640 --> 34:06.040 + And that's one of the reasons that LLVM has been + +34:06.040 --> 34:08.720 + very successful in the research world as well. + +34:08.720 --> 34:11.040 + But in the community sense, + +34:11.040 --> 34:12.960 + Guido van Rasen, right? + +34:12.960 --> 34:16.880 + From Python, just retired from, + +34:18.080 --> 34:20.480 + what is it, benevolent, dictated for life, right? + +34:20.480 --> 34:24.720 + So in managing this community of brilliant compiler folks, + +34:24.720 --> 34:28.640 + is there, did it, for a time at least, + +34:28.640 --> 34:31.480 + fall on you to approve things? + +34:31.480 --> 34:34.240 + Oh yeah, so I mean, I still have something like + +34:34.240 --> 34:38.000 + an order of magnitude more patches in LLVM + +34:38.000 --> 34:39.000 + than anybody else. + +34:40.000 --> 34:42.760 + And many of those I wrote myself. + +34:42.760 --> 34:43.840 + But you're still right. + +34:43.840 --> 34:48.360 + I mean, you're still close to the, + +34:48.360 --> 34:50.040 + I don't know what the expression is to the metal. + +34:50.040 --> 34:51.040 + You're still right, Ko. + +34:51.040 --> 34:52.200 + Yeah, I'm still right, Ko. + +34:52.200 --> 34:54.240 + Not as much as I was able to in grad school, + +34:54.240 --> 34:56.760 + but that's an important part of my identity. + +34:56.760 --> 34:58.880 + But the way that LLVM has worked over time + +34:58.880 --> 35:00.440 + is that when I was a grad student, + +35:00.440 --> 35:03.000 + I could do all the work and steer everything + +35:03.000 --> 35:05.800 + and review every patch and make sure everything was done + +35:05.800 --> 35:09.040 + exactly the way my opinionated sense + +35:09.040 --> 35:10.640 + felt like it should be done. + +35:10.640 --> 35:11.760 + And that was fine. + +35:11.760 --> 35:14.320 + But as things scale, you can't do that, right? + +35:14.320 --> 35:18.040 + And so what ends up happening is LLVM has a hierarchical + +35:18.040 --> 35:20.520 + system of what's called code owners. + +35:20.520 --> 35:22.880 + These code owners are given the responsibility + +35:22.880 --> 35:24.920 + not to do all the work, + +35:24.920 --> 35:26.680 + not necessarily to review all the patches, + +35:26.680 --> 35:28.840 + but to make sure that the patches do get reviewed + +35:28.840 --> 35:30.360 + and make sure that the right thing's happening + +35:30.360 --> 35:32.200 + architecturally in their area. + +35:32.200 --> 35:34.200 + And so what you'll see is you'll see + +35:34.200 --> 35:37.760 + that for example, hardware manufacturers + +35:37.760 --> 35:40.920 + end up owning the hardware specific parts + +35:40.920 --> 35:44.520 + of their hardware, that's very common. + +35:45.560 --> 35:47.760 + Leaders in the community that have done really good work + +35:47.760 --> 35:50.920 + naturally become the de facto owner of something. + +35:50.920 --> 35:53.440 + And then usually somebody else is like, + +35:53.440 --> 35:55.520 + how about we make them the official code owner? + +35:55.520 --> 35:58.600 + And then we'll have somebody to make sure + +35:58.600 --> 36:00.320 + that all the patches get reviewed in a timely manner. + +36:00.320 --> 36:02.080 + And then everybody's like, yes, that's obvious. + +36:02.080 --> 36:03.240 + And then it happens, right? + +36:03.240 --> 36:06.080 + And usually this is a very organic thing, which is great. + +36:06.080 --> 36:08.720 + And so I'm nominally the top of that stack still, + +36:08.720 --> 36:11.560 + but I don't spend a lot of time reviewing patches. + +36:11.560 --> 36:16.520 + What I do is I help negotiate a lot of the technical + +36:16.520 --> 36:18.080 + disagreements that end up happening + +36:18.080 --> 36:19.680 + and making sure that the community as a whole + +36:19.680 --> 36:22.080 + makes progress and is moving in the right direction + +36:22.080 --> 36:23.960 + and doing that. + +36:23.960 --> 36:28.280 + So we also started a nonprofit six years ago, + +36:28.280 --> 36:30.880 + seven years ago, time's gone away. + +36:30.880 --> 36:34.640 + And the LVM Foundation nonprofit helps oversee + +36:34.640 --> 36:36.480 + all the business sides of things and make sure + +36:36.480 --> 36:39.680 + that the events that the LVM community has are funded + +36:39.680 --> 36:42.840 + and set up and run correctly and stuff like that. + +36:42.840 --> 36:45.200 + But the foundation is very much stays out + +36:45.200 --> 36:49.080 + of the technical side of where the project is going. + +36:49.080 --> 36:53.200 + Right, so it sounds like a lot of it is just organic, just. + +36:53.200 --> 36:55.720 + Yeah, well, and this is LVM is almost 20 years old, + +36:55.720 --> 36:56.640 + which is hard to believe. + +36:56.640 --> 37:00.360 + Somebody pointed out to me recently that LVM is now older + +37:00.360 --> 37:04.640 + than GCC was when LVM started, right? + +37:04.640 --> 37:06.880 + So time has a way of getting away from you. + +37:06.880 --> 37:10.440 + But the good thing about that is it has a really robust, + +37:10.440 --> 37:13.560 + really amazing community of people that are + +37:13.560 --> 37:14.720 + in their professional lives, + +37:14.720 --> 37:16.320 + spread across lots of different companies, + +37:16.320 --> 37:19.320 + but it's a community of people + +37:19.320 --> 37:21.160 + that are interested in similar kinds of problems + +37:21.160 --> 37:23.720 + and have been working together effectively for years + +37:23.720 --> 37:26.480 + and have a lot of trust and respect for each other. + +37:26.480 --> 37:28.960 + And even if they don't always agree that, you know, + +37:28.960 --> 37:31.200 + we're able to find a path forward. + +37:31.200 --> 37:34.520 + So then in a slightly different flavor of effort, + +37:34.520 --> 37:38.920 + you started at Apple in 2005 with the task of making, + +37:38.920 --> 37:41.840 + I guess, LVM production ready. + +37:41.840 --> 37:44.680 + And then eventually 2013 through 2017, + +37:44.680 --> 37:48.400 + leading the entire developer tools department. + +37:48.400 --> 37:53.000 + We're talking about LLVM, Xcode, Objective C to Swift. + +37:53.960 --> 37:58.600 + So in a quick overview of your time there, + +37:58.600 --> 37:59.640 + what were the challenges? + +37:59.640 --> 38:03.280 + First of all, leading such a huge group of developers. + +38:03.280 --> 38:06.560 + What was the big motivator dream mission + +38:06.560 --> 38:11.440 + behind creating Swift, the early birth of it + +38:11.440 --> 38:13.440 + from Objective C and so on and Xcode? + +38:13.440 --> 38:14.280 + What are some challenges? + +38:14.280 --> 38:15.920 + So these are different questions. + +38:15.920 --> 38:16.760 + Yeah, I know. + +38:16.760 --> 38:19.560 + But I want to talk about the other stuff too. + +38:19.560 --> 38:21.240 + I'll stay on the technical side, + +38:21.240 --> 38:23.440 + then we can talk about the big team pieces. + +38:23.440 --> 38:24.280 + That's okay? + +38:24.280 --> 38:25.120 + Sure. + +38:25.120 --> 38:27.760 + So it's to really oversimplify many years of hard work. + +38:27.760 --> 38:32.440 + LVM started, joined Apple, became a thing, + +38:32.440 --> 38:34.600 + became successful and became deployed. + +38:34.600 --> 38:36.760 + But then there was a question about + +38:36.760 --> 38:38.880 + how do we actually parse the source code? + +38:38.880 --> 38:40.320 + So LVM is that back part, + +38:40.320 --> 38:42.320 + the optimizer and the code generator. + +38:42.320 --> 38:44.640 + And LVM is really good for Apple as it went through + +38:44.640 --> 38:46.040 + a couple of hardware transitions. + +38:46.040 --> 38:47.920 + I joined right at the time of the Intel transition, + +38:47.920 --> 38:51.800 + for example, and 64 bit transitions + +38:51.800 --> 38:53.480 + and then the transition to ARM with the iPhone. + +38:53.480 --> 38:54.680 + And so LVM was very useful + +38:54.680 --> 38:56.920 + for some of these kinds of things. + +38:56.920 --> 38:57.760 + But at the same time, + +38:57.760 --> 39:00.080 + there's a lot of questions around developer experience. + +39:00.080 --> 39:01.880 + And so if you're a programmer pounding out + +39:01.880 --> 39:03.400 + at the time Objective C code, + +39:04.400 --> 39:06.440 + the error message you get, the compile time, + +39:06.440 --> 39:09.680 + the turnaround cycle, the tooling and the IDE + +39:09.680 --> 39:12.960 + were not great, were not as good as they could be. + +39:12.960 --> 39:17.960 + And so, as I occasionally do, I'm like, + +39:17.960 --> 39:20.080 + well, okay, how hard is it to write a C compiler? + +39:20.080 --> 39:20.920 + Right. + +39:20.920 --> 39:22.520 + And so I'm not gonna commit to anybody. + +39:22.520 --> 39:23.360 + I'm not gonna tell anybody. + +39:23.360 --> 39:25.960 + I'm just gonna just do it on nights and weekends + +39:25.960 --> 39:27.400 + and start working on it. + +39:27.400 --> 39:30.120 + And then I built up and see there's this thing + +39:30.120 --> 39:32.960 + called the preprocessor, which people don't like, + +39:32.960 --> 39:35.440 + but it's actually really hard and complicated + +39:35.440 --> 39:37.640 + and includes a bunch of really weird things + +39:37.640 --> 39:39.240 + like try graphs and other stuff like that + +39:39.240 --> 39:40.880 + that are really nasty. + +39:40.880 --> 39:44.000 + And it's the crux of a bunch of the performance issues + +39:44.000 --> 39:46.560 + in the compiler, start working on the parser + +39:46.560 --> 39:47.720 + and kind of got to the point where I'm like, + +39:47.720 --> 39:49.840 + oh, you know what, we could actually do this. + +39:49.840 --> 39:51.400 + Everybody's saying that this is impossible to do, + +39:51.400 --> 39:52.800 + but it's actually just hard. + +39:52.800 --> 39:53.880 + It's not impossible. + +39:53.880 --> 39:57.520 + And eventually told my manager about it + +39:57.520 --> 39:59.160 + and he's like, oh, wow, this is great. + +39:59.160 --> 40:00.280 + We do need to solve this problem. + +40:00.280 --> 40:01.120 + Oh, this is great. + +40:01.120 --> 40:04.360 + We can get you one other person to work with you on this. + +40:04.360 --> 40:08.240 + And so the team is formed and it starts taking off. + +40:08.240 --> 40:11.960 + And C++, for example, huge complicated language. + +40:11.960 --> 40:14.280 + People always assume that it's impossible to implement + +40:14.280 --> 40:16.160 + and it's very nearly impossible, + +40:16.160 --> 40:18.640 + but it's just really, really hard. + +40:18.640 --> 40:20.760 + And the way to get there is to build it + +40:20.760 --> 40:22.360 + one piece at a time incrementally. + +40:22.360 --> 40:26.360 + And that was only possible because we were lucky + +40:26.360 --> 40:28.080 + to hire some really exceptional engineers + +40:28.080 --> 40:30.280 + that knew various parts of it very well + +40:30.280 --> 40:32.600 + and could do great things. + +40:32.600 --> 40:34.360 + Swift was kind of a similar thing. + +40:34.360 --> 40:39.080 + So Swift came from, we were just finishing off + +40:39.080 --> 40:42.520 + the first version of C++ support in Clang. + +40:42.520 --> 40:47.160 + And C++ is a very formidable and very important language, + +40:47.160 --> 40:49.240 + but it's also ugly in lots of ways. + +40:49.240 --> 40:52.280 + And you can't implement C++ without thinking + +40:52.280 --> 40:54.320 + there has to be a better thing, right? + +40:54.320 --> 40:56.080 + And so I started working on Swift again + +40:56.080 --> 40:58.520 + with no hope or ambition that would go anywhere. + +40:58.520 --> 41:00.760 + Just let's see what could be done. + +41:00.760 --> 41:02.560 + Let's play around with this thing. + +41:02.560 --> 41:04.800 + It was me in my spare time, + +41:04.800 --> 41:08.160 + not telling anybody about it kind of a thing. + +41:08.160 --> 41:09.360 + And it made some good progress. + +41:09.360 --> 41:11.240 + I'm like, actually, it would make sense to do this. + +41:11.240 --> 41:14.760 + At the same time, I started talking with the senior VP + +41:14.760 --> 41:17.680 + of software at the time, a guy named Bertrand Sirle, + +41:17.680 --> 41:19.240 + and Bertrand was very encouraging. + +41:19.240 --> 41:22.040 + He was like, well, let's have fun, let's talk about this. + +41:22.040 --> 41:23.400 + And he was a little bit of a language guy. + +41:23.400 --> 41:26.120 + And so he helped guide some of the early work + +41:26.120 --> 41:30.360 + and encouraged me and got things off the ground. + +41:30.360 --> 41:34.240 + And eventually, I told my manager and told other people. + +41:34.240 --> 41:38.760 + And it started making progress. + +41:38.760 --> 41:40.920 + The complicating thing with Swift + +41:40.920 --> 41:43.840 + was that the idea of doing a new language + +41:43.840 --> 41:47.760 + is not obvious to anybody, including myself. + +41:47.760 --> 41:50.160 + And the tone at the time was that the iPhone + +41:50.160 --> 41:53.360 + was successful because of Objective C, right? + +41:53.360 --> 41:54.360 + Oh, interesting. + +41:54.360 --> 41:55.200 + In Objective C. + +41:55.200 --> 41:57.080 + Not despite of or just because of. + +41:57.080 --> 42:01.080 + And you have to understand that at the time, + +42:01.080 --> 42:05.360 + Apple was hiring software people that loved Objective C, right? + +42:05.360 --> 42:07.920 + And it wasn't that they came despite Objective C. + +42:07.920 --> 42:10.160 + They loved Objective C, and that's why they got hired. + +42:10.160 --> 42:13.680 + And so you had a software team that the leadership in many cases + +42:13.680 --> 42:18.440 + went all the way back to Next, where Objective C really became + +42:18.440 --> 42:19.320 + real. + +42:19.320 --> 42:23.200 + And so they, quote unquote, grew up writing Objective C. + +42:23.200 --> 42:25.680 + And many of the individual engineers + +42:25.680 --> 42:28.280 + all were hired because they loved Objective C. + +42:28.280 --> 42:30.520 + And so this notion of, OK, let's do new language + +42:30.520 --> 42:34.040 + was kind of heretical in many ways, right? + +42:34.040 --> 42:36.960 + Meanwhile, my sense was that the outside community wasn't really + +42:36.960 --> 42:38.520 + in love with Objective C. Some people were. + +42:38.520 --> 42:40.200 + And some of the most outspoken people were. + +42:40.200 --> 42:42.600 + But other people were hitting challenges + +42:42.600 --> 42:46.760 + because it has very sharp corners and it's difficult to learn. + +42:46.760 --> 42:50.040 + And so one of the challenges of making Swift happen + +42:50.040 --> 42:54.640 + that was totally non technical is the social part + +42:54.640 --> 42:57.760 + of what do we do? + +42:57.760 --> 43:00.280 + If we do a new language, which at Apple, many things + +43:00.280 --> 43:02.200 + happen that don't ship, right? + +43:02.200 --> 43:05.520 + So if we ship it, what is the metrics of success? + +43:05.520 --> 43:06.360 + Why would we do this? + +43:06.360 --> 43:07.920 + Why wouldn't we make Objective C better? + +43:07.920 --> 43:09.760 + If Objective C has problems, let's + +43:09.760 --> 43:12.120 + file off those rough corners and edges. + +43:12.120 --> 43:15.600 + And one of the major things that became the reason to do this + +43:15.600 --> 43:18.960 + was this notion of safety, memory safety. + +43:18.960 --> 43:22.880 + And the way Objective C works is that a lot of the object + +43:22.880 --> 43:26.440 + system and everything else is built on top of pointers + +43:26.440 --> 43:29.920 + in C. Objective C is an extension on top of C. + +43:29.920 --> 43:32.640 + And so pointers are unsafe. + +43:32.640 --> 43:34.600 + And if you get rid of the pointers, + +43:34.600 --> 43:36.400 + it's not Objective C anymore. + +43:36.400 --> 43:39.040 + And so fundamentally, that was an issue + +43:39.040 --> 43:42.160 + that you could not fix safety or memory safety + +43:42.160 --> 43:45.560 + without fundamentally changing the language. + +43:45.560 --> 43:49.880 + And so once we got through that part of the mental process + +43:49.880 --> 43:53.480 + and the thought process, it became a design process of saying, + +43:53.480 --> 43:56.240 + OK, well, if we're going to do something new, what is good? + +43:56.240 --> 43:57.400 + Like, how do we think about this? + +43:57.400 --> 43:59.960 + And what are we like, and what are we looking for? + +43:59.960 --> 44:02.400 + And that was a very different phase of it. + +44:02.400 --> 44:05.880 + So what are some design choices early on in Swift? + +44:05.880 --> 44:09.720 + Like, we're talking about braces. + +44:09.720 --> 44:12.040 + Are you making a type language or not? + +44:12.040 --> 44:13.200 + All those kinds of things. + +44:13.200 --> 44:16.000 + Yeah, so some of those were obvious given the context. + +44:16.000 --> 44:18.240 + So a type language, for example, Objective C + +44:18.240 --> 44:22.480 + is a type language, and going with an untyped language + +44:22.480 --> 44:24.280 + wasn't really seriously considered. + +44:24.280 --> 44:26.920 + We wanted the performance, and we wanted refactoring tools + +44:26.920 --> 44:29.600 + and other things like that that go with type languages. + +44:29.600 --> 44:30.800 + Quick dumb question. + +44:30.800 --> 44:31.400 + Yeah. + +44:31.400 --> 44:32.920 + Was it obvious? + +44:32.920 --> 44:34.600 + I think this would be a dumb question. + +44:34.600 --> 44:36.520 + But was it obvious that the language has + +44:36.520 --> 44:38.920 + to be a compiled language? + +44:38.920 --> 44:40.120 + Not an? + +44:40.120 --> 44:42.040 + Yes, that's not a dumb question. + +44:42.040 --> 44:44.000 + Earlier, I think late 90s, Apple + +44:44.000 --> 44:48.960 + had seriously considered moving its development experience to Java. + +44:48.960 --> 44:53.120 + But Swift started in 2010, which was several years + +44:53.120 --> 44:53.800 + after the iPhone. + +44:53.800 --> 44:56.600 + It was when the iPhone was definitely on an upper trajectory. + +44:56.600 --> 44:58.680 + And the iPhone was still extremely + +44:58.680 --> 45:01.760 + and is still a bit memory constrained. + +45:01.760 --> 45:05.480 + And so being able to compile the code and then ship it + +45:05.480 --> 45:09.720 + and then having standalone code that is not JIT compiled + +45:09.720 --> 45:11.320 + is a very big deal. + +45:11.320 --> 45:15.200 + And it's very much part of the Apple value system. + +45:15.200 --> 45:17.520 + Now, JavaScript's also a thing. + +45:17.520 --> 45:19.360 + I mean, it's not that this is exclusive, + +45:19.360 --> 45:23.880 + and technologies are good, depending on how they're applied. + +45:23.880 --> 45:27.200 + But in the design of Swift, saying how can we make + +45:27.200 --> 45:29.560 + Objective C better, Objective C was statically compiled, + +45:29.560 --> 45:32.480 + and that was the contiguous natural thing to do. + +45:32.480 --> 45:34.640 + Just skip ahead a little bit. + +45:34.640 --> 45:37.600 + Right back, just as a question, as you think about today + +45:37.600 --> 45:42.400 + in 2019, in your work at Google, TensorFlow, and so on, + +45:42.400 --> 45:47.480 + is, again, compilation, static compilation, + +45:47.480 --> 45:49.480 + still the right thing. + +45:49.480 --> 45:52.560 + Yeah, so the funny thing after working on compilers + +45:52.560 --> 45:56.480 + for a really long time is that, and this + +45:56.480 --> 45:59.080 + is one of the things that LLVM has helped with, + +45:59.080 --> 46:01.480 + is that I don't look at compilations + +46:01.480 --> 46:05.320 + being static or dynamic or interpreted or not. + +46:05.320 --> 46:09.160 + This is a spectrum, and one of the cool things about Swift + +46:09.160 --> 46:12.200 + is that Swift is not just statically compiled. + +46:12.200 --> 46:14.160 + It's actually dynamically compiled as well. + +46:14.160 --> 46:16.000 + And it can also be interpreted, though nobody's actually + +46:16.000 --> 46:17.560 + done that. + +46:17.560 --> 46:20.360 + And so what ends up happening when + +46:20.360 --> 46:22.760 + you use Swift in a workbook, for example, + +46:22.760 --> 46:25.320 + in Colab or in Jupyter, is it's actually dynamically + +46:25.320 --> 46:28.320 + compiling the statements as you execute them. + +46:28.320 --> 46:32.840 + And so this gets back to the software engineering problems, + +46:32.840 --> 46:34.960 + where if you layer the stack properly, + +46:34.960 --> 46:37.280 + you can actually completely change + +46:37.280 --> 46:39.320 + how and when things get compiled because you + +46:39.320 --> 46:41.120 + have the right abstractions there. + +46:41.120 --> 46:44.800 + And so the way that a Colab workbook works with Swift + +46:44.800 --> 46:47.720 + is that when you start typing into it, + +46:47.720 --> 46:50.320 + it creates a process, a UNIX process. + +46:50.320 --> 46:52.240 + And then each line of code you type in, + +46:52.240 --> 46:56.240 + it compiles it through the Swift compiler, the front end part, + +46:56.240 --> 46:58.400 + and then sends it through the optimizer, + +46:58.400 --> 47:01.120 + JIT compiles machine code, and then + +47:01.120 --> 47:03.920 + injects it into that process. + +47:03.920 --> 47:06.560 + And so as you're typing new stuff, + +47:06.560 --> 47:09.360 + it's like squirting in new code and overwriting and replacing + +47:09.360 --> 47:11.240 + and updating code in place. + +47:11.240 --> 47:13.520 + And the fact that it can do this is not an accident. + +47:13.520 --> 47:15.560 + Like Swift was designed for this. + +47:15.560 --> 47:18.120 + But it's an important part of how the language was set up + +47:18.120 --> 47:18.960 + and how it's layered. + +47:18.960 --> 47:21.360 + And this is a non obvious piece. + +47:21.360 --> 47:24.640 + And one of the things with Swift that was, for me, + +47:24.640 --> 47:27.040 + a very strong design point is to make it so that you + +47:27.040 --> 47:29.680 + can learn it very quickly. + +47:29.680 --> 47:32.080 + And so from a language design perspective, + +47:32.080 --> 47:34.520 + the thing that I always come back to is this UI principle + +47:34.520 --> 47:37.880 + of progressive disclosure of complexity. + +47:37.880 --> 47:41.680 + And so in Swift, you can start by saying print, quote, + +47:41.680 --> 47:43.960 + hello world, quote. + +47:43.960 --> 47:47.160 + And there's no slash n, just like Python, one line of code, + +47:47.160 --> 47:51.560 + no main, no header files, no public static class void, + +47:51.560 --> 47:55.600 + blah, blah, blah string, like Java has, one line of code. + +47:55.600 --> 47:58.280 + And you can teach that and it works great. + +47:58.280 --> 48:00.280 + Then you can say, well, let's introduce variables. + +48:00.280 --> 48:02.400 + And so you can declare a variable with var. + +48:02.400 --> 48:03.760 + So var x equals four. + +48:03.760 --> 48:04.680 + What is a variable? + +48:04.680 --> 48:06.280 + You can use x, x plus one. + +48:06.280 --> 48:07.720 + This is what it means. + +48:07.720 --> 48:09.480 + Then you can say, well, how about control flow? + +48:09.480 --> 48:10.840 + Well, this is one if statement is. + +48:10.840 --> 48:12.240 + This is what a for statement is. + +48:12.240 --> 48:15.320 + This is what a while statement is. + +48:15.320 --> 48:17.280 + Then you can say, let's introduce functions. + +48:17.280 --> 48:20.000 + And many languages like Python have + +48:20.000 --> 48:22.800 + had this kind of notion of let's introduce small things. + +48:22.800 --> 48:24.360 + And then you can add complexity. + +48:24.360 --> 48:25.720 + Then you can introduce classes. + +48:25.720 --> 48:28.040 + And then you can add generics in the case of Swift. + +48:28.040 --> 48:30.600 + And then you can build in modules and build out in terms + +48:30.600 --> 48:32.200 + of the things that you're expressing. + +48:32.200 --> 48:35.800 + But this is not very typical for compiled languages. + +48:35.800 --> 48:38.000 + And so this was a very strong design point. + +48:38.000 --> 48:40.960 + And one of the reasons that Swift in general + +48:40.960 --> 48:43.480 + is designed with this factoring of complexity in mind + +48:43.480 --> 48:46.440 + so that the language can express powerful things. + +48:46.440 --> 48:49.240 + You can write firmware in Swift if you want to. + +48:49.240 --> 48:52.800 + But it has a very high level feel, which is really + +48:52.800 --> 48:53.760 + this perfect blend. + +48:53.760 --> 48:57.440 + Because often you have very advanced library writers + +48:57.440 --> 49:00.520 + that want to be able to use the nitty gritty details. + +49:00.520 --> 49:02.960 + But then other people just want to use the libraries + +49:02.960 --> 49:04.880 + and work at a higher abstraction level. + +49:04.880 --> 49:07.200 + It's kind of cool that I saw that you can just + +49:07.200 --> 49:09.200 + enter a probability. + +49:09.200 --> 49:11.320 + I don't think I pronounced that word enough. + +49:11.320 --> 49:14.920 + But you can just drag in Python. + +49:14.920 --> 49:15.960 + It's just a string. + +49:15.960 --> 49:18.840 + You can import like, I saw this in the demo, + +49:18.840 --> 49:19.600 + import number. + +49:19.600 --> 49:20.760 + How do you make that happen? + +49:20.760 --> 49:21.240 + Yeah, well. + +49:21.240 --> 49:22.520 + What's up with that? + +49:22.520 --> 49:23.240 + Yeah. + +49:23.240 --> 49:24.960 + Is that as easy as it looks? + +49:24.960 --> 49:25.520 + Or is it? + +49:25.520 --> 49:26.560 + Yes, as easy as it looks. + +49:26.560 --> 49:29.440 + That's not a stage magic hack or anything like that. + +49:29.440 --> 49:31.400 + I don't mean from the user perspective. + +49:31.400 --> 49:33.200 + I mean from the implementation perspective + +49:33.200 --> 49:34.120 + to make it happen. + +49:34.120 --> 49:37.000 + So it's easy once all the pieces are in place. + +49:37.000 --> 49:37.920 + The way it works. + +49:37.920 --> 49:39.560 + So if you think about a dynamically typed language + +49:39.560 --> 49:42.160 + like Python, you can think about it in two different ways. + +49:42.160 --> 49:45.800 + You can say it has no types, which + +49:45.800 --> 49:47.480 + is what most people would say. + +49:47.480 --> 49:50.440 + Or you can say it has one type. + +49:50.440 --> 49:53.360 + And you can say it has one type and it's the Python object. + +49:53.360 --> 49:55.040 + And the Python object is passed around. + +49:55.040 --> 49:56.280 + And because there's only one type, + +49:56.280 --> 49:58.240 + it's implicit. + +49:58.240 --> 50:01.320 + And so what happens with Swift and Python talking to each other, + +50:01.320 --> 50:03.320 + Swift has lots of types, has arrays, + +50:03.320 --> 50:07.040 + and it has strings and all classes and that kind of stuff. + +50:07.040 --> 50:11.120 + But it now has a Python object type. + +50:11.120 --> 50:12.800 + So there is one Python object type. + +50:12.800 --> 50:16.440 + And so when you say import numpy, what you get + +50:16.440 --> 50:19.880 + is a Python object, which is the numpy module. + +50:19.880 --> 50:22.160 + And then you say np.array. + +50:22.160 --> 50:24.960 + It says, OK, hey Python object, I have no idea what you are. + +50:24.960 --> 50:27.280 + Give me your array member. + +50:27.280 --> 50:27.960 + OK, cool. + +50:27.960 --> 50:31.160 + And it just uses dynamic stuff, talks to the Python interpreter + +50:31.160 --> 50:33.680 + and says, hey Python, what's the dot array member + +50:33.680 --> 50:35.680 + in that Python object? + +50:35.680 --> 50:37.400 + It gives you back another Python object. + +50:37.400 --> 50:39.480 + And now you say, parentheses for the call + +50:39.480 --> 50:40.960 + and the arguments are going to pass. + +50:40.960 --> 50:43.640 + And so then it says, hey, a Python object that + +50:43.640 --> 50:48.040 + is the result of np.array, call with these arguments. + +50:48.040 --> 50:50.320 + Again, calling into the Python interpreter to do that work. + +50:50.320 --> 50:53.680 + And so right now, this is all really simple. + +50:53.680 --> 50:55.960 + And if you dive into the code, what you'll see + +50:55.960 --> 50:58.440 + is that the Python module in Swift + +50:58.440 --> 51:01.400 + is something like 1,200 lines of code or something. + +51:01.400 --> 51:02.360 + It's written in pure Swift. + +51:02.360 --> 51:03.560 + It's super simple. + +51:03.560 --> 51:06.560 + And it's built on top of the C interoperability + +51:06.560 --> 51:09.520 + because it just talks to the Python interpreter. + +51:09.520 --> 51:11.200 + But making that possible required us + +51:11.200 --> 51:13.480 + to add two major language features to Swift + +51:13.480 --> 51:15.400 + to be able to express these dynamic calls + +51:15.400 --> 51:17.200 + and the dynamic member lookups. + +51:17.200 --> 51:19.480 + And so what we've done over the last year + +51:19.480 --> 51:23.080 + is we've proposed, implement, standardized, + +51:23.080 --> 51:26.160 + and contributed new language features to the Swift language + +51:26.160 --> 51:29.560 + in order to make it so it is really trivial. + +51:29.560 --> 51:31.360 + And this is one of the things about Swift + +51:31.360 --> 51:35.000 + that is critical to the Swift for TensorFlow work, which + +51:35.000 --> 51:37.200 + is that we can actually add new language features. + +51:37.200 --> 51:39.160 + And the bar for adding those is high, + +51:39.160 --> 51:42.160 + but it's what makes it possible. + +51:42.160 --> 51:45.240 + So you're now at Google doing incredible work + +51:45.240 --> 51:47.680 + on several things, including TensorFlow. + +51:47.680 --> 51:52.240 + So TensorFlow 2.0 or whatever leading up to 2.0 + +51:52.240 --> 51:57.360 + has, by default, in 2.0, has eager execution in yet + +51:57.360 --> 52:00.480 + in order to make code optimized for GPU or GPU + +52:00.480 --> 52:04.080 + or some of these systems computation + +52:04.080 --> 52:05.960 + needs to be converted to a graph. + +52:05.960 --> 52:07.400 + So what's that process like? + +52:07.400 --> 52:08.920 + What are the challenges there? + +52:08.920 --> 52:11.680 + Yeah, so I'm tangentially involved in this. + +52:11.680 --> 52:15.240 + But the way that it works with Autograph + +52:15.240 --> 52:21.600 + is that you mark your function with a decorator. + +52:21.600 --> 52:24.280 + And when Python calls it, that decorator is invoked. + +52:24.280 --> 52:28.240 + And then it says, before I call this function, + +52:28.240 --> 52:29.480 + you can transform it. + +52:29.480 --> 52:32.400 + And so the way Autograph works is, as far as I understand, + +52:32.400 --> 52:34.440 + is it actually uses the Python parser + +52:34.440 --> 52:37.160 + to go parse that, turn into a syntax tree, + +52:37.160 --> 52:39.400 + and now apply compiler techniques to, again, + +52:39.400 --> 52:42.320 + transform this down into TensorFlow graphs. + +52:42.320 --> 52:45.880 + And so you can think of it as saying, hey, I have an if statement. + +52:45.880 --> 52:48.800 + I'm going to create an if node in the graph, like you say, + +52:48.800 --> 52:51.080 + tf.cond. + +52:51.080 --> 52:53.000 + You have a multiply. + +52:53.000 --> 52:55.320 + Well, I'll turn that into a multiply node in the graph. + +52:55.320 --> 52:57.720 + And it becomes this tree transformation. + +52:57.720 --> 53:01.280 + So where does the Swift for TensorFlow come in? + +53:01.280 --> 53:04.720 + Which is parallels. + +53:04.720 --> 53:06.960 + For one, Swift is an interface. + +53:06.960 --> 53:09.200 + Like Python is an interface with TensorFlow. + +53:09.200 --> 53:11.200 + But it seems like there's a lot more going on + +53:11.200 --> 53:13.120 + than just a different language interface. + +53:13.120 --> 53:15.240 + There's optimization methodology. + +53:15.240 --> 53:19.560 + So the TensorFlow world has a couple of different, what + +53:19.560 --> 53:21.240 + I'd call front end technologies. + +53:21.240 --> 53:25.400 + And so Swift, and Python, and Go, and Rust, and Julian, + +53:25.400 --> 53:29.360 + all these things share the TensorFlow graphs + +53:29.360 --> 53:32.760 + and all the runtime and everything that's later. + +53:32.760 --> 53:36.680 + And so Swift for TensorFlow is merely another front end + +53:36.680 --> 53:40.680 + for TensorFlow, just like any of these other systems are. + +53:40.680 --> 53:43.120 + There's a major difference between, I would say, + +53:43.120 --> 53:44.640 + three camps of technologies here. + +53:44.640 --> 53:46.920 + There's Python, which is a special case, + +53:46.920 --> 53:49.280 + because the vast majority of the community efforts + +53:49.280 --> 53:51.160 + go into the Python interface. + +53:51.160 --> 53:53.000 + And Python has its own approaches + +53:53.000 --> 53:55.800 + for automatic differentiation, has its own APIs, + +53:55.800 --> 53:58.200 + and all this kind of stuff. + +53:58.200 --> 54:00.240 + There's Swift, which I'll talk about in a second. + +54:00.240 --> 54:02.080 + And then there's kind of everything else. + +54:02.080 --> 54:05.440 + And so the everything else are effectively language bindings. + +54:05.440 --> 54:08.000 + So they call into the TensorFlow runtime. + +54:08.000 --> 54:10.960 + But they usually don't have automatic differentiation, + +54:10.960 --> 54:14.760 + or they usually don't provide anything other than APIs that + +54:14.760 --> 54:16.480 + call the C APIs in TensorFlow. + +54:16.480 --> 54:18.400 + And so they're kind of wrappers for that. + +54:18.400 --> 54:19.840 + Swift is really kind of special. + +54:19.840 --> 54:22.800 + And it's a very different approach. + +54:22.800 --> 54:25.360 + Swift for TensorFlow, that is, is a very different approach, + +54:25.360 --> 54:26.920 + because there we're saying, let's + +54:26.920 --> 54:28.440 + look at all the problems that need + +54:28.440 --> 54:34.120 + to be solved in the full stack of the TensorFlow compilation + +54:34.120 --> 54:35.680 + process, if you think about it that way. + +54:35.680 --> 54:38.200 + Because TensorFlow is fundamentally a compiler. + +54:38.200 --> 54:42.760 + It takes models, and then it makes them go fast on hardware. + +54:42.760 --> 54:43.800 + That's what a compiler does. + +54:43.800 --> 54:47.560 + And it has a front end, it has an optimizer, + +54:47.560 --> 54:49.320 + and it has many back ends. + +54:49.320 --> 54:51.680 + And so if you think about it the right way, + +54:51.680 --> 54:54.760 + or if you look at it in a particular way, + +54:54.760 --> 54:55.800 + it is a compiler. + +54:59.280 --> 55:02.120 + And so Swift is merely another front end. + +55:02.120 --> 55:05.560 + But it's saying, and the design principle is saying, + +55:05.560 --> 55:08.200 + let's look at all the problems that we face as machine + +55:08.200 --> 55:11.200 + learning practitioners, and what is the best possible way + +55:11.200 --> 55:13.840 + we can do that, given the fact that we can change literally + +55:13.840 --> 55:15.920 + anything in this entire stack. + +55:15.920 --> 55:18.440 + And Python, for example, where the vast majority + +55:18.440 --> 55:22.600 + of the engineering and effort has gone into, + +55:22.600 --> 55:25.280 + is constrained by being the best possible thing you can do + +55:25.280 --> 55:27.280 + with a Python library. + +55:27.280 --> 55:29.280 + There are no Python language features + +55:29.280 --> 55:32.520 + that are added because of machine learning that I'm aware of. + +55:32.520 --> 55:35.080 + They added a matrix multiplication operator with that, + +55:35.080 --> 55:38.280 + but that's as close as you get. + +55:38.280 --> 55:41.400 + And so with Swift, it's hard, but you + +55:41.400 --> 55:43.800 + can add language features to the language, + +55:43.800 --> 55:46.080 + and there's a community process for that. + +55:46.080 --> 55:48.000 + And so we look at these things and say, + +55:48.000 --> 55:49.680 + well, what is the right division of labor + +55:49.680 --> 55:52.000 + between the human programmer and the compiler? + +55:52.000 --> 55:55.280 + And Swift has a number of things that shift that balance. + +55:55.280 --> 56:00.520 + So because it has a type system, for example, + +56:00.520 --> 56:03.280 + it makes certain things possible for analysis of the code, + +56:03.280 --> 56:05.520 + and the compiler can automatically + +56:05.520 --> 56:08.800 + build graphs for you without you thinking about them. + +56:08.800 --> 56:10.520 + That's a big deal for a programmer. + +56:10.520 --> 56:11.640 + You just get free performance. + +56:11.640 --> 56:14.360 + You get clustering and fusion and optimization, + +56:14.360 --> 56:17.440 + things like that, without you as a programmer having + +56:17.440 --> 56:20.040 + to manually do it because the compiler can do it for you. + +56:20.040 --> 56:22.200 + Automatic differentiation is another big deal, + +56:22.200 --> 56:25.440 + and I think one of the key contributions of the Swift + +56:25.440 --> 56:29.600 + for TensorFlow project is that there's + +56:29.600 --> 56:32.200 + this entire body of work on automatic differentiation that + +56:32.200 --> 56:34.240 + dates back to the Fortran days. + +56:34.240 --> 56:36.360 + People doing a tremendous amount of numerical computing + +56:36.360 --> 56:39.800 + in Fortran used to write what they call source to source + +56:39.800 --> 56:43.600 + translators, where you take a bunch of code, shove it + +56:43.600 --> 56:47.280 + into a mini compiler, and it would push out more Fortran + +56:47.280 --> 56:50.200 + code, but it would generate the backwards passes + +56:50.200 --> 56:53.000 + for your functions for you, the derivatives. + +56:53.000 --> 56:57.840 + And so in that work in the 70s, a tremendous number + +56:57.840 --> 57:01.160 + of optimizations, a tremendous number of techniques + +57:01.160 --> 57:02.920 + for fixing numerical instability, + +57:02.920 --> 57:05.080 + and other kinds of problems were developed, + +57:05.080 --> 57:07.600 + but they're very difficult to port into a world + +57:07.600 --> 57:11.280 + where in eager execution, you get an op by op at a time. + +57:11.280 --> 57:13.280 + Like, you need to be able to look at an entire function + +57:13.280 --> 57:15.600 + and be able to reason about what's going on. + +57:15.600 --> 57:18.360 + And so when you have a language integrated + +57:18.360 --> 57:20.480 + automatic differentiation, which is one of the things + +57:20.480 --> 57:22.760 + that the Swift project is focusing on, + +57:22.760 --> 57:25.720 + you can open all these techniques and reuse them + +57:25.720 --> 57:30.160 + in familiar ways, but the language integration piece + +57:30.160 --> 57:33.280 + has a bunch of design room in it, and it's also complicated. + +57:33.280 --> 57:34.920 + The other piece of the puzzle here, + +57:34.920 --> 57:37.040 + this kind of interesting is TPUs at Google. + +57:37.040 --> 57:37.880 + Yes. + +57:37.880 --> 57:40.200 + So, you know, we're in a new world with deep learning. + +57:40.200 --> 57:43.000 + It's constantly changing, and I imagine + +57:43.000 --> 57:46.400 + without disclosing anything, I imagine, you know, + +57:46.400 --> 57:48.480 + you're still innovating on the TPU front too. + +57:48.480 --> 57:49.320 + Indeed. + +57:49.320 --> 57:52.280 + So how much sort of interplays there are + +57:52.280 --> 57:54.440 + between software and hardware and trying to figure out + +57:54.440 --> 57:56.760 + how to gather, move towards an optimized solution. + +57:56.760 --> 57:57.800 + There's an incredible amount. + +57:57.800 --> 57:59.520 + So we're on our third generation of TPUs, + +57:59.520 --> 58:02.800 + which are now 100 petaflops in a very large + +58:02.800 --> 58:07.800 + liquid cooled box, virtual box with no cover. + +58:07.800 --> 58:11.320 + And as you might imagine, we're not out of ideas yet. + +58:11.320 --> 58:14.400 + The great thing about TPUs is that they're + +58:14.400 --> 58:17.640 + a perfect example of hardware software co design. + +58:17.640 --> 58:19.840 + And so it's about saying, what hardware + +58:19.840 --> 58:23.280 + do we build to solve certain classes of machine learning + +58:23.280 --> 58:23.920 + problems? + +58:23.920 --> 58:26.360 + Well, the algorithms are changing. + +58:26.360 --> 58:30.480 + Like the hardware takes some cases years to produce, right? + +58:30.480 --> 58:34.160 + And so you have to make bets and decide what is going to happen. + +58:34.160 --> 58:37.280 + And so what is the best way to spend the transistors + +58:37.280 --> 58:41.560 + to get the maximum performance per watt or area per cost + +58:41.560 --> 58:44.120 + or whatever it is that you're optimizing for? + +58:44.120 --> 58:46.600 + And so one of the amazing things about TPUs + +58:46.600 --> 58:50.040 + is this numeric format called B Float 16. + +58:50.040 --> 58:54.160 + B Float 16 is a compressed 16 bit floating point format, + +58:54.160 --> 58:56.120 + but it puts the bits in different places. + +58:56.120 --> 58:59.000 + In numeric terms, it has a smaller mantissa + +58:59.000 --> 59:00.480 + and a larger exponent. + +59:00.480 --> 59:03.000 + That means that it's less precise, + +59:03.000 --> 59:06.120 + but it can represent larger ranges of values, which + +59:06.120 --> 59:08.640 + in the machine learning context is really important and useful. + +59:08.640 --> 59:13.120 + Because sometimes you have very small gradients + +59:13.120 --> 59:17.600 + you want to accumulate and very, very small numbers that + +59:17.600 --> 59:20.520 + are important to move things as you're learning. + +59:20.520 --> 59:23.240 + But sometimes you have very large magnitude numbers as well. + +59:23.240 --> 59:26.880 + And B Float 16 is not as precise. + +59:26.880 --> 59:28.240 + The mantissa is small. + +59:28.240 --> 59:30.360 + But it turns out the machine learning algorithms actually + +59:30.360 --> 59:31.640 + want to generalize. + +59:31.640 --> 59:34.320 + And so there's theories that this actually + +59:34.320 --> 59:36.440 + increases the ability for the network + +59:36.440 --> 59:38.040 + to generalize across data sets. + +59:38.040 --> 59:41.160 + And regardless of whether it's good or bad, + +59:41.160 --> 59:42.640 + it's much cheaper at the hardware level + +59:42.640 --> 59:48.160 + to implement because the area and time of a multiplier + +59:48.160 --> 59:50.880 + is n squared in the number of bits in the mantissa, + +59:50.880 --> 59:53.360 + but it's linear with size of the exponent. + +59:53.360 --> 59:56.360 + And you're connected to both efforts here, both on the hardware + +59:56.360 --> 59:57.200 + and the software side? + +59:57.200 --> 59:59.240 + Yeah, and so that was a breakthrough coming + +59:59.240 --> 1:00:01.800 + from the research side and people working + +1:00:01.800 --> 1:00:06.000 + on optimizing network transport of weights + +1:00:06.000 --> 1:00:08.280 + across a network originally and trying + +1:00:08.280 --> 1:00:10.160 + to find ways to compress that. + +1:00:10.160 --> 1:00:12.160 + But then it got burned into silicon. + +1:00:12.160 --> 1:00:15.320 + And it's a key part of what makes TPU performance so amazing. + +1:00:15.320 --> 1:00:17.880 + And great. + +1:00:17.880 --> 1:00:20.640 + Now, TPUs have many different aspects that are important. + +1:00:20.640 --> 1:00:25.080 + But the co design between the low level compiler bits + +1:00:25.080 --> 1:00:27.360 + and the software bits and the algorithms + +1:00:27.360 --> 1:00:28.640 + is all super important. + +1:00:28.640 --> 1:00:32.880 + And it's this amazing trifecta that only Google can do. + +1:00:32.880 --> 1:00:34.360 + Yeah, that's super exciting. + +1:00:34.360 --> 1:00:38.440 + So can you tell me about MLIR project, + +1:00:38.440 --> 1:00:41.400 + previously the secretive one? + +1:00:41.400 --> 1:00:43.000 + Yeah, so MLIR is a project that we + +1:00:43.000 --> 1:00:46.960 + announced at a compiler conference three weeks ago + +1:00:46.960 --> 1:00:50.880 + or something at the Compilers for Machine Learning Conference. + +1:00:50.880 --> 1:00:53.280 + Basically, again, if you look at TensorFlow as a compiler + +1:00:53.280 --> 1:00:55.040 + stack, it has a number of compiler algorithms + +1:00:55.040 --> 1:00:56.000 + within it. + +1:00:56.000 --> 1:00:57.480 + It also has a number of compilers + +1:00:57.480 --> 1:00:58.880 + that get embedded into it. + +1:00:58.880 --> 1:01:00.320 + And they're made by different vendors. + +1:01:00.320 --> 1:01:04.640 + For example, Google has XLA, which is a great compiler system. + +1:01:04.640 --> 1:01:08.680 + NVIDIA has TensorFlow RT, Intel has NGraph. + +1:01:08.680 --> 1:01:10.640 + There's a number of these different compiler systems. + +1:01:10.640 --> 1:01:13.600 + And they're very hardware specific. + +1:01:13.600 --> 1:01:16.280 + And they're trying to solve different parts of the problems. + +1:01:16.280 --> 1:01:18.920 + But they're all kind of similar in a sense + +1:01:18.920 --> 1:01:20.680 + of they want to integrate with TensorFlow. + +1:01:20.680 --> 1:01:22.720 + Now, TensorFlow has an optimizer. + +1:01:22.720 --> 1:01:25.480 + And it has these different code generation technologies + +1:01:25.480 --> 1:01:26.360 + built in. + +1:01:26.360 --> 1:01:28.680 + The idea of MLIR is to build a common infrastructure + +1:01:28.680 --> 1:01:31.040 + to support all these different subsystems. + +1:01:31.040 --> 1:01:34.120 + And initially, it's to be able to make it so that they all + +1:01:34.120 --> 1:01:34.840 + plug in together. + +1:01:34.840 --> 1:01:37.800 + And they can share a lot more code and can be reusable. + +1:01:37.800 --> 1:01:40.960 + But over time, we hope that the industry will start + +1:01:40.960 --> 1:01:42.440 + collaborating and sharing code. + +1:01:42.440 --> 1:01:45.200 + And instead of reinventing the same things over and over again, + +1:01:45.200 --> 1:01:49.240 + that we can actually foster some of that working together + +1:01:49.240 --> 1:01:51.520 + to solve common problem energy that + +1:01:51.520 --> 1:01:54.440 + has been useful in the compiler field before. + +1:01:54.440 --> 1:01:57.000 + Beyond that, MLIR is some people have + +1:01:57.000 --> 1:01:59.240 + joked that it's kind of LLVM2. + +1:01:59.240 --> 1:02:01.760 + It learns a lot about what LLVM has been good + +1:02:01.760 --> 1:02:04.280 + and what LLVM has done wrong. + +1:02:04.280 --> 1:02:06.800 + And it's a chance to fix that. + +1:02:06.800 --> 1:02:09.320 + And also, there are challenges in the LLVM ecosystem + +1:02:09.320 --> 1:02:11.840 + as well, where LLVM is very good at the thing + +1:02:11.840 --> 1:02:12.680 + it was designed to do. + +1:02:12.680 --> 1:02:15.480 + But 20 years later, the world has changed. + +1:02:15.480 --> 1:02:17.560 + And people are trying to solve higher level problems. + +1:02:17.560 --> 1:02:20.280 + And we need some new technology. + +1:02:20.280 --> 1:02:24.680 + And what's the future of open source in this context? + +1:02:24.680 --> 1:02:25.680 + Very soon. + +1:02:25.680 --> 1:02:27.440 + So it is not yet open source. + +1:02:27.440 --> 1:02:29.360 + But it will be, hopefully, the next couple of months. + +1:02:29.360 --> 1:02:30.960 + So you still believe in the value of open source + +1:02:30.960 --> 1:02:31.560 + and these kinds of kinds? + +1:02:31.560 --> 1:02:32.400 + Oh, yeah, absolutely. + +1:02:32.400 --> 1:02:36.080 + And I think that the TensorFlow community at large + +1:02:36.080 --> 1:02:37.640 + fully believes in open source. + +1:02:37.640 --> 1:02:40.080 + So I mean, there is a difference between Apple, + +1:02:40.080 --> 1:02:43.480 + where you were previously in Google, now in spirit and culture. + +1:02:43.480 --> 1:02:45.440 + And I would say the open sourcing of TensorFlow + +1:02:45.440 --> 1:02:48.360 + was a seminal moment in the history of software. + +1:02:48.360 --> 1:02:51.640 + Because here's this large company releasing + +1:02:51.640 --> 1:02:55.880 + a very large code base that's open sourcing. + +1:02:55.880 --> 1:02:57.880 + What are your thoughts on that? + +1:02:57.880 --> 1:03:00.800 + How happy or not were you to see that kind + +1:03:00.800 --> 1:03:02.880 + of degree of open sourcing? + +1:03:02.880 --> 1:03:05.320 + So between the two, I prefer the Google approach, + +1:03:05.320 --> 1:03:07.800 + if that's what you're saying. + +1:03:07.800 --> 1:03:12.360 + The Apple approach makes sense given the historical context + +1:03:12.360 --> 1:03:13.360 + that Apple came from. + +1:03:13.360 --> 1:03:15.720 + But that's been 35 years ago. + +1:03:15.720 --> 1:03:18.160 + And I think that Apple is definitely adapting. + +1:03:18.160 --> 1:03:20.240 + And the way I look at it is that there's + +1:03:20.240 --> 1:03:23.120 + different kinds of concerns in the space, right? + +1:03:23.120 --> 1:03:24.840 + It is very rational for a business + +1:03:24.840 --> 1:03:28.680 + to care about making money. + +1:03:28.680 --> 1:03:31.600 + That fundamentally is what a business is about, right? + +1:03:31.600 --> 1:03:34.280 + But I think it's also incredibly realistic + +1:03:34.280 --> 1:03:36.320 + to say it's not your string library that's + +1:03:36.320 --> 1:03:38.040 + the thing that's going to make you money. + +1:03:38.040 --> 1:03:41.440 + It's going to be the amazing UI product differentiating + +1:03:41.440 --> 1:03:42.880 + features and other things like that + +1:03:42.880 --> 1:03:45.200 + that you build on top of your string library. + +1:03:45.200 --> 1:03:49.480 + And so keeping your string library proprietary and secret + +1:03:49.480 --> 1:03:53.480 + and things like that isn't maybe not the important thing + +1:03:53.480 --> 1:03:54.680 + anymore, right? + +1:03:54.680 --> 1:03:57.720 + Or before, platforms were different, right? + +1:03:57.720 --> 1:04:01.480 + And even 15 years ago, things were a little bit different. + +1:04:01.480 --> 1:04:02.880 + But the world is changing. + +1:04:02.880 --> 1:04:05.280 + So Google strikes a very good balance, I think. + +1:04:05.280 --> 1:04:08.680 + And I think that TensorFlow being open source + +1:04:08.680 --> 1:04:12.000 + really changed the entire machine learning field + +1:04:12.000 --> 1:04:14.080 + and it caused a revolution in its own right. + +1:04:14.080 --> 1:04:17.560 + And so I think it's amazingly forward looking + +1:04:17.560 --> 1:04:21.520 + because I could have imagined, and I wasn't at Google at the time, + +1:04:21.520 --> 1:04:23.760 + but I could imagine a different context in a different world + +1:04:23.760 --> 1:04:26.520 + where a company says, machine learning is critical + +1:04:26.520 --> 1:04:27.960 + to what we're doing, we're not going + +1:04:27.960 --> 1:04:29.600 + to give it to other people, right? + +1:04:29.600 --> 1:04:35.840 + And so that decision is a profoundly brilliant insight + +1:04:35.840 --> 1:04:38.320 + that I think has really led to the world being better + +1:04:38.320 --> 1:04:40.160 + and better for Google as well. + +1:04:40.160 --> 1:04:42.200 + And has all kinds of ripple effects. + +1:04:42.200 --> 1:04:45.400 + I think it is really, I mean, you can't + +1:04:45.400 --> 1:04:49.800 + understate Google deciding how profound that is for software. + +1:04:49.800 --> 1:04:50.840 + It's awesome. + +1:04:50.840 --> 1:04:54.880 + Well, and again, I can understand the concern + +1:04:54.880 --> 1:04:57.640 + about if we release our machine learning software, + +1:04:57.640 --> 1:05:00.400 + our competitors could go faster. + +1:05:00.400 --> 1:05:02.480 + But on the other hand, I think that open sourcing TensorFlow + +1:05:02.480 --> 1:05:03.960 + has been fantastic for Google. + +1:05:03.960 --> 1:05:09.080 + And I'm sure that decision was very nonobvious at the time, + +1:05:09.080 --> 1:05:11.480 + but I think it's worked out very well. + +1:05:11.480 --> 1:05:13.200 + So let's try this real quick. + +1:05:13.200 --> 1:05:15.600 + You were at Tesla for five months + +1:05:15.600 --> 1:05:17.600 + as the VP of autopilot software. + +1:05:17.600 --> 1:05:20.480 + You led the team during the transition from H Hardware + +1:05:20.480 --> 1:05:22.320 + 1 to Hardware 2. + +1:05:22.320 --> 1:05:23.480 + I have a couple of questions. + +1:05:23.480 --> 1:05:26.320 + So one, first of all, to me, that's + +1:05:26.320 --> 1:05:28.520 + one of the bravest engineering decisions + +1:05:28.520 --> 1:05:33.320 + undertaking sort of like, undertaking really ever + +1:05:33.320 --> 1:05:36.000 + in the automotive industry to me, software wise, + +1:05:36.000 --> 1:05:37.440 + starting from scratch. + +1:05:37.440 --> 1:05:39.320 + It's a really brave engineering decision. + +1:05:39.320 --> 1:05:42.760 + So my one question is there is, what was that like? + +1:05:42.760 --> 1:05:43.960 + What was the challenge of that? + +1:05:43.960 --> 1:05:45.760 + Do you mean the career decision of jumping + +1:05:45.760 --> 1:05:48.880 + from a comfortable good job into the unknown? + +1:05:48.880 --> 1:05:51.560 + That combined, so at the individual level, + +1:05:51.560 --> 1:05:54.640 + you making that decision. + +1:05:54.640 --> 1:05:58.040 + And then when you show up, it's a really hard engineering + +1:05:58.040 --> 1:05:58.840 + problem. + +1:05:58.840 --> 1:06:04.880 + So you could just stay, maybe slow down, say, Hardware 1, + +1:06:04.880 --> 1:06:06.560 + or those kinds of decisions. + +1:06:06.560 --> 1:06:10.160 + So just taking it full on, let's do this from scratch. + +1:06:10.160 --> 1:06:11.080 + What was that like? + +1:06:11.080 --> 1:06:12.680 + Well, so I mean, I don't think Tesla + +1:06:12.680 --> 1:06:15.720 + has a culture of taking things slow and seeing how it goes. + +1:06:15.720 --> 1:06:18.080 + So one of the things that attracted me about Tesla + +1:06:18.080 --> 1:06:19.240 + is it's very much a gung ho. + +1:06:19.240 --> 1:06:20.200 + Let's change the world. + +1:06:20.200 --> 1:06:21.640 + Let's figure it out kind of a place. + +1:06:21.640 --> 1:06:25.680 + And so I have a huge amount of respect for that. + +1:06:25.680 --> 1:06:28.720 + Tesla has done very smart things with Hardware 1 + +1:06:28.720 --> 1:06:29.440 + in particular. + +1:06:29.440 --> 1:06:32.760 + And the Hardware 1 design was originally designed + +1:06:32.760 --> 1:06:37.280 + to be very simple automation features in the car + +1:06:37.280 --> 1:06:39.840 + for like traffic aware cruise control and things like that. + +1:06:39.840 --> 1:06:42.680 + And the fact that they were able to effectively feature + +1:06:42.680 --> 1:06:47.760 + creep it into lane holding and a very useful driver assistance + +1:06:47.760 --> 1:06:50.120 + feature is pretty astounding, particularly given + +1:06:50.120 --> 1:06:52.560 + the details of the hardware. + +1:06:52.560 --> 1:06:54.640 + Hardware 2 built on that in a lot of ways. + +1:06:54.640 --> 1:06:56.800 + And the challenge there was that they were transitioning + +1:06:56.800 --> 1:07:00.080 + from a third party provided vision stack + +1:07:00.080 --> 1:07:01.760 + to an in house built vision stack. + +1:07:01.760 --> 1:07:05.680 + And so for the first step, which I mostly helped with, + +1:07:05.680 --> 1:07:08.520 + was getting onto that new vision stack. + +1:07:08.520 --> 1:07:10.880 + And that was very challenging. + +1:07:10.880 --> 1:07:14.000 + And it was time critical for various reasons. + +1:07:14.000 --> 1:07:15.000 + And it was a big leap. + +1:07:15.000 --> 1:07:17.560 + But it was fortunate that it built on a lot of the knowledge + +1:07:17.560 --> 1:07:20.880 + and expertise in the team that had built Hardware 1's + +1:07:20.880 --> 1:07:22.920 + driver assistance features. + +1:07:22.920 --> 1:07:25.400 + So you spoke in a collected and kind way + +1:07:25.400 --> 1:07:26.720 + about your time at Tesla. + +1:07:26.720 --> 1:07:30.280 + But it was ultimately not a good fit Elon Musk. + +1:07:30.280 --> 1:07:33.440 + We've talked on this podcast, several guests of the course. + +1:07:33.440 --> 1:07:36.480 + Elon Musk continues to do some of the most bold and innovative + +1:07:36.480 --> 1:07:38.800 + engineering work in the world at times + +1:07:38.800 --> 1:07:41.320 + at the cost to some of the members of the Tesla team. + +1:07:41.320 --> 1:07:45.120 + What did you learn about this working in this chaotic world + +1:07:45.120 --> 1:07:46.720 + with Elon? + +1:07:46.720 --> 1:07:50.560 + Yeah, so I guess I would say that when I was at Tesla, + +1:07:50.560 --> 1:07:54.480 + I experienced and saw the highest degree of turnover + +1:07:54.480 --> 1:07:58.280 + I'd ever seen in a company, which was a bit of a shock. + +1:07:58.280 --> 1:08:00.520 + But one of the things I learned and I came to respect + +1:08:00.520 --> 1:08:03.400 + is that Elon's able to attract amazing talent + +1:08:03.400 --> 1:08:05.640 + because he has a very clear vision of the future. + +1:08:05.640 --> 1:08:07.200 + And he can get people to buy into it + +1:08:07.200 --> 1:08:09.840 + because they want that future to happen. + +1:08:09.840 --> 1:08:11.840 + And the power of vision is something + +1:08:11.840 --> 1:08:14.200 + that I have a tremendous amount of respect for. + +1:08:14.200 --> 1:08:17.600 + And I think that Elon is fairly singular in the world + +1:08:17.600 --> 1:08:22.320 + in terms of the things he's able to get people to believe in. + +1:08:22.320 --> 1:08:27.360 + And there are many people that stand in the street corner + +1:08:27.360 --> 1:08:29.320 + and say, ah, we're going to go to Mars, right? + +1:08:29.320 --> 1:08:31.600 + But then there are a few people that + +1:08:31.600 --> 1:08:35.200 + can get others to buy into it and believe in, build the path + +1:08:35.200 --> 1:08:36.160 + and make it happen. + +1:08:36.160 --> 1:08:39.120 + And so I respect that. + +1:08:39.120 --> 1:08:41.000 + I don't respect all of his methods, + +1:08:41.000 --> 1:08:44.960 + but I have a huge amount of respect for that. + +1:08:44.960 --> 1:08:46.840 + You've mentioned in a few places, + +1:08:46.840 --> 1:08:50.400 + including in this context, working hard. + +1:08:50.400 --> 1:08:51.960 + What does it mean to work hard? + +1:08:51.960 --> 1:08:53.480 + And when you look back at your life, + +1:08:53.480 --> 1:08:59.040 + what were some of the most brutal periods of having + +1:08:59.040 --> 1:09:03.360 + to really put everything you have into something? + +1:09:03.360 --> 1:09:05.040 + Yeah, good question. + +1:09:05.040 --> 1:09:07.480 + So working hard can be defined a lot of different ways. + +1:09:07.480 --> 1:09:08.680 + So a lot of hours. + +1:09:08.680 --> 1:09:12.440 + And so that is true. + +1:09:12.440 --> 1:09:14.480 + The thing to me that's the hardest + +1:09:14.480 --> 1:09:18.720 + is both being short term focused on delivering and executing + +1:09:18.720 --> 1:09:21.080 + and making a thing happen, while also thinking + +1:09:21.080 --> 1:09:24.360 + about the longer term and trying to balance that, right? + +1:09:24.360 --> 1:09:28.480 + Because if you are myopically focused on solving a task + +1:09:28.480 --> 1:09:31.920 + and getting that done and only think about that incremental + +1:09:31.920 --> 1:09:34.640 + next step, you will miss the next big hill + +1:09:34.640 --> 1:09:36.360 + you should jump over to, right? + +1:09:36.360 --> 1:09:38.000 + And so I've been really fortunate + +1:09:38.000 --> 1:09:42.080 + that I've been able to kind of oscillate between the two. + +1:09:42.080 --> 1:09:45.600 + And historically at Apple, for example, + +1:09:45.600 --> 1:09:47.080 + that was made possible because I was + +1:09:47.080 --> 1:09:49.080 + able to work with some really amazing people and build up + +1:09:49.080 --> 1:09:53.760 + teams and leadership structures and allow + +1:09:53.760 --> 1:09:57.120 + them to grow in their careers and take on responsibility, + +1:09:57.120 --> 1:10:00.080 + thereby freeing up me to be a little bit crazy + +1:10:00.080 --> 1:10:02.960 + and thinking about the next thing. + +1:10:02.960 --> 1:10:04.640 + And so it's a lot of that. + +1:10:04.640 --> 1:10:06.760 + But it's also about with the experience + +1:10:06.760 --> 1:10:10.120 + you make connections that other people don't necessarily make. + +1:10:10.120 --> 1:10:12.960 + And so I think that's a big part as well. + +1:10:12.960 --> 1:10:16.040 + But the bedrock is just a lot of hours. + +1:10:16.040 --> 1:10:19.720 + And that's OK with me. + +1:10:19.720 --> 1:10:21.480 + There's different theories on work life balance. + +1:10:21.480 --> 1:10:25.160 + And my theory for myself, which I do not project onto the team, + +1:10:25.160 --> 1:10:28.480 + but my theory for myself is that I + +1:10:28.480 --> 1:10:30.400 + want to love what I'm doing and work really hard. + +1:10:30.400 --> 1:10:33.960 + And my purpose, I feel like, and my goal + +1:10:33.960 --> 1:10:36.240 + is to change the world and make it a better place. + +1:10:36.240 --> 1:10:40.000 + And that's what I'm really motivated to do. + +1:10:40.000 --> 1:10:44.760 + So last question, LLVM logo is a dragon. + +1:10:44.760 --> 1:10:46.760 + You explained that this is because dragons + +1:10:46.760 --> 1:10:50.320 + have connotations of power, speed, intelligence. + +1:10:50.320 --> 1:10:53.320 + It can also be sleek, elegant, and modular, + +1:10:53.320 --> 1:10:56.280 + though you remove the modular part. + +1:10:56.280 --> 1:10:58.920 + What is your favorite dragon related character + +1:10:58.920 --> 1:11:01.480 + from fiction, video, or movies? + +1:11:01.480 --> 1:11:03.840 + So those are all very kind ways of explaining it. + +1:11:03.840 --> 1:11:06.200 + Do you want to know the real reason it's a dragon? + +1:11:06.200 --> 1:11:07.000 + Yeah. + +1:11:07.000 --> 1:11:07.920 + Is that better? + +1:11:07.920 --> 1:11:11.040 + So there is a seminal book on compiler design + +1:11:11.040 --> 1:11:12.480 + called The Dragon Book. + +1:11:12.480 --> 1:11:16.280 + And so this is a really old now book on compilers. + +1:11:16.280 --> 1:11:22.040 + And so the Dragon logo for LLVM came about because at Apple, + +1:11:22.040 --> 1:11:24.720 + we kept talking about LLVM related technologies, + +1:11:24.720 --> 1:11:26.960 + and there's no logo to put on a slide. + +1:11:26.960 --> 1:11:28.440 + And we're like, what do we do? + +1:11:28.440 --> 1:11:30.000 + And somebody's like, well, what kind of logo + +1:11:30.000 --> 1:11:32.160 + should a compiler technology have? + +1:11:32.160 --> 1:11:33.320 + And I'm like, I don't know. + +1:11:33.320 --> 1:11:37.280 + I mean, the dragon is the best thing that we've got. + +1:11:37.280 --> 1:11:40.600 + And Apple somehow magically came up with the logo. + +1:11:40.600 --> 1:11:43.240 + And it was a great thing, and the whole community + +1:11:43.240 --> 1:11:44.000 + rallied around it. + +1:11:44.000 --> 1:11:46.840 + And then it got better as other graphic designers got + +1:11:46.840 --> 1:11:47.320 + involved. + +1:11:47.320 --> 1:11:49.280 + But that's originally where it came from. + +1:11:49.280 --> 1:11:50.080 + The story. + +1:11:50.080 --> 1:11:53.960 + Is there dragons from fiction that you connect with? + +1:11:53.960 --> 1:11:58.000 + That Game of Thrones, Lord of the Rings, that kind of thing? + +1:11:58.000 --> 1:11:59.120 + Lord of the Rings is great. + +1:11:59.120 --> 1:12:01.440 + I also like role playing games and things like computer + +1:12:01.440 --> 1:12:02.160 + role playing games. + +1:12:02.160 --> 1:12:03.600 + And so dragons often show up in there. + +1:12:03.600 --> 1:12:07.080 + But it really comes back to the book. + +1:12:07.080 --> 1:12:08.480 + Oh, no, we need a thing. + +1:12:08.480 --> 1:12:09.880 + We need a lot to do. + +1:12:09.880 --> 1:12:13.640 + And hilariously, one of the funny things about LLVM + +1:12:13.640 --> 1:12:19.400 + is that my wife, who's amazing, runs the LLVM foundation. + +1:12:19.400 --> 1:12:21.040 + And she goes to Grace Hopper, and is + +1:12:21.040 --> 1:12:22.480 + trying to get more women involved. + +1:12:22.480 --> 1:12:24.600 + And she's also a compiler engineer. + +1:12:24.600 --> 1:12:26.040 + So she's trying to get other women + +1:12:26.040 --> 1:12:28.120 + to get interested in compilers and things like this. + +1:12:28.120 --> 1:12:29.960 + And so she hands out the stickers. + +1:12:29.960 --> 1:12:34.240 + And people like the LLVM sticker because of Game of Thrones. + +1:12:34.240 --> 1:12:36.800 + And so sometimes culture has this helpful effect + +1:12:36.800 --> 1:12:41.040 + to get the next generation of compiler engineers engaged + +1:12:41.040 --> 1:12:42.320 + with the cause. + +1:12:42.320 --> 1:12:43.240 + OK, awesome. + +1:12:43.240 --> 1:12:44.680 + Grace, thanks so much for talking with us. + +1:12:44.680 --> 1:13:07.440 + It's been great talking with you. +