diff --git "a/vtt/episode_038_large.vtt" "b/vtt/episode_038_large.vtt" deleted file mode 100644--- "a/vtt/episode_038_large.vtt" +++ /dev/null @@ -1,6905 +0,0 @@ -WEBVTT - -00:00.000 --> 00:03.720 - The following is a conversation with Francois Chollet. - -00:03.720 --> 00:05.760 - He's the creator of Keras, - -00:05.760 --> 00:08.080 - which is an open source deep learning library - -00:08.080 --> 00:11.480 - that is designed to enable fast, user friendly experimentation - -00:11.480 --> 00:13.600 - with deep neural networks. - -00:13.600 --> 00:16.680 - It serves as an interface to several deep learning libraries, - -00:16.680 --> 00:19.040 - most popular of which is TensorFlow, - -00:19.040 --> 00:22.600 - and it was integrated into the TensorFlow main code base - -00:22.600 --> 00:24.080 - a while ago. - -00:24.080 --> 00:27.000 - Meaning, if you want to create, train, - -00:27.000 --> 00:28.640 - and use neural networks, - -00:28.640 --> 00:31.040 - probably the easiest and most popular option - -00:31.040 --> 00:33.840 - is to use Keras inside TensorFlow. - -00:34.840 --> 00:37.240 - Aside from creating an exceptionally useful - -00:37.240 --> 00:38.680 - and popular library, - -00:38.680 --> 00:41.920 - Francois is also a world class AI researcher - -00:41.920 --> 00:43.680 - and software engineer at Google. - -00:44.560 --> 00:46.960 - And he's definitely an outspoken, - -00:46.960 --> 00:50.560 - if not controversial personality in the AI world, - -00:50.560 --> 00:52.920 - especially in the realm of ideas - -00:52.920 --> 00:55.920 - around the future of artificial intelligence. - -00:55.920 --> 00:58.600 - This is the Artificial Intelligence Podcast. - -00:58.600 --> 01:01.000 - If you enjoy it, subscribe on YouTube, - -01:01.000 --> 01:02.760 - give it five stars on iTunes, - -01:02.760 --> 01:04.160 - support it on Patreon, - -01:04.160 --> 01:06.120 - or simply connect with me on Twitter - -01:06.120 --> 01:09.960 - at Lex Friedman, spelled F R I D M A N. - -01:09.960 --> 01:13.840 - And now, here's my conversation with Francois Chollet. - -01:14.880 --> 01:17.320 - You're known for not sugarcoating your opinions - -01:17.320 --> 01:19.160 - and speaking your mind about ideas in AI, - -01:19.160 --> 01:21.160 - especially on Twitter. - -01:21.160 --> 01:22.760 - It's one of my favorite Twitter accounts. - -01:22.760 --> 01:26.320 - So what's one of the more controversial ideas - -01:26.320 --> 01:29.360 - you've expressed online and gotten some heat for? - -01:30.440 --> 01:31.360 - How do you pick? - -01:33.080 --> 01:33.920 - How do I pick? - -01:33.920 --> 01:36.880 - Yeah, no, I think if you go through the trouble - -01:36.880 --> 01:39.640 - of maintaining a Twitter account, - -01:39.640 --> 01:41.840 - you might as well speak your mind, you know? - -01:41.840 --> 01:44.600 - Otherwise, what's even the point of having a Twitter account? - -01:44.600 --> 01:45.480 - It's like having a nice car - -01:45.480 --> 01:47.560 - and just leaving it in the garage. - -01:48.600 --> 01:50.840 - Yeah, so what's one thing for which I got - -01:50.840 --> 01:53.600 - a lot of pushback? - -01:53.600 --> 01:56.640 - Perhaps, you know, that time I wrote something - -01:56.640 --> 02:00.920 - about the idea of intelligence explosion, - -02:00.920 --> 02:04.520 - and I was questioning the idea - -02:04.520 --> 02:06.840 - and the reasoning behind this idea. - -02:06.840 --> 02:09.640 - And I got a lot of pushback on that. - -02:09.640 --> 02:11.840 - I got a lot of flak for it. - -02:11.840 --> 02:13.600 - So yeah, so intelligence explosion, - -02:13.600 --> 02:14.960 - I'm sure you're familiar with the idea, - -02:14.960 --> 02:18.800 - but it's the idea that if you were to build - -02:18.800 --> 02:22.920 - general AI problem solving algorithms, - -02:22.920 --> 02:26.000 - well, the problem of building such an AI, - -02:27.480 --> 02:30.520 - that itself is a problem that could be solved by your AI, - -02:30.520 --> 02:31.880 - and maybe it could be solved better - -02:31.880 --> 02:33.760 - than what humans can do. - -02:33.760 --> 02:36.840 - So your AI could start tweaking its own algorithm, - -02:36.840 --> 02:39.520 - could start making a better version of itself, - -02:39.520 --> 02:43.240 - and so on iteratively in a recursive fashion. - -02:43.240 --> 02:47.320 - And so you would end up with an AI - -02:47.320 --> 02:50.080 - with exponentially increasing intelligence. - -02:50.080 --> 02:50.920 - That's right. - -02:50.920 --> 02:55.880 - And I was basically questioning this idea, - -02:55.880 --> 02:59.040 - first of all, because the notion of intelligence explosion - -02:59.040 --> 03:02.200 - uses an implicit definition of intelligence - -03:02.200 --> 03:05.360 - that doesn't sound quite right to me. - -03:05.360 --> 03:10.360 - It considers intelligence as a property of a brain - -03:11.200 --> 03:13.680 - that you can consider in isolation, - -03:13.680 --> 03:16.640 - like the height of a building, for instance. - -03:16.640 --> 03:19.040 - But that's not really what intelligence is. - -03:19.040 --> 03:22.200 - Intelligence emerges from the interaction - -03:22.200 --> 03:25.240 - between a brain, a body, - -03:25.240 --> 03:28.320 - like embodied intelligence, and an environment. - -03:28.320 --> 03:30.720 - And if you're missing one of these pieces, - -03:30.720 --> 03:33.800 - then you cannot really define intelligence anymore. - -03:33.800 --> 03:36.800 - So just tweaking a brain to make it smaller and smaller - -03:36.800 --> 03:39.120 - doesn't actually make any sense to me. - -03:39.120 --> 03:39.960 - So first of all, - -03:39.960 --> 03:43.000 - you're crushing the dreams of many people, right? - -03:43.000 --> 03:46.000 - So there's a, let's look at like Sam Harris. - -03:46.000 --> 03:48.680 - Actually, a lot of physicists, Max Tegmark, - -03:48.680 --> 03:52.120 - people who think the universe - -03:52.120 --> 03:54.640 - is an information processing system, - -03:54.640 --> 03:57.680 - our brain is kind of an information processing system. - -03:57.680 --> 03:59.400 - So what's the theoretical limit? - -03:59.400 --> 04:03.160 - Like, it doesn't make sense that there should be some, - -04:04.800 --> 04:07.520 - it seems naive to think that our own brain - -04:07.520 --> 04:10.000 - is somehow the limit of the capabilities - -04:10.000 --> 04:11.600 - of this information system. - -04:11.600 --> 04:13.600 - I'm playing devil's advocate here. - -04:13.600 --> 04:15.600 - This information processing system. - -04:15.600 --> 04:17.760 - And then if you just scale it, - -04:17.760 --> 04:19.360 - if you're able to build something - -04:19.360 --> 04:20.920 - that's on par with the brain, - -04:20.920 --> 04:24.040 - you just, the process that builds it just continues - -04:24.040 --> 04:26.400 - and it'll improve exponentially. - -04:26.400 --> 04:30.160 - So that's the logic that's used actually - -04:30.160 --> 04:32.560 - by almost everybody - -04:32.560 --> 04:36.920 - that is worried about super human intelligence. - -04:36.920 --> 04:39.120 - So you're trying to make, - -04:39.120 --> 04:40.960 - so most people who are skeptical of that - -04:40.960 --> 04:43.000 - are kind of like, this doesn't, - -04:43.000 --> 04:46.520 - their thought process, this doesn't feel right. - -04:46.520 --> 04:47.680 - Like that's for me as well. - -04:47.680 --> 04:49.760 - So I'm more like, it doesn't, - -04:51.440 --> 04:52.800 - the whole thing is shrouded in mystery - -04:52.800 --> 04:55.840 - where you can't really say anything concrete, - -04:55.840 --> 04:57.880 - but you could say this doesn't feel right. - -04:57.880 --> 05:00.640 - This doesn't feel like that's how the brain works. - -05:00.640 --> 05:02.400 - And you're trying to with your blog posts - -05:02.400 --> 05:05.680 - and now making it a little more explicit. - -05:05.680 --> 05:10.680 - So one idea is that the brain isn't exist alone. - -05:10.680 --> 05:13.200 - It exists within the environment. - -05:13.200 --> 05:15.680 - So you can't exponentially, - -05:15.680 --> 05:18.000 - you would have to somehow exponentially improve - -05:18.000 --> 05:20.920 - the environment and the brain together almost. - -05:20.920 --> 05:25.920 - Yeah, in order to create something that's much smarter - -05:25.960 --> 05:27.840 - in some kind of, - -05:27.840 --> 05:29.960 - of course we don't have a definition of intelligence. - -05:29.960 --> 05:31.280 - That's correct, that's correct. - -05:31.280 --> 05:34.280 - I don't think, you should look at very smart people today, - -05:34.280 --> 05:37.280 - even humans, not even talking about AIs. - -05:37.280 --> 05:38.640 - I don't think their brain - -05:38.640 --> 05:41.960 - and the performance of their brain is the bottleneck - -05:41.960 --> 05:45.200 - to their expressed intelligence, to their achievements. - -05:46.600 --> 05:49.960 - You cannot just tweak one part of this system, - -05:49.960 --> 05:52.840 - like of this brain, body, environment system - -05:52.840 --> 05:55.960 - and expect that capabilities like what emerges - -05:55.960 --> 06:00.280 - out of this system to just explode exponentially. - -06:00.280 --> 06:04.200 - Because anytime you improve one part of a system - -06:04.200 --> 06:06.760 - with many interdependencies like this, - -06:06.760 --> 06:09.520 - there's a new bottleneck that arises, right? - -06:09.520 --> 06:12.280 - And I don't think even today for very smart people, - -06:12.280 --> 06:15.000 - their brain is not the bottleneck - -06:15.000 --> 06:17.560 - to the sort of problems they can solve, right? - -06:17.560 --> 06:19.800 - In fact, many very smart people today, - -06:20.760 --> 06:22.520 - you know, they are not actually solving - -06:22.520 --> 06:24.800 - any big scientific problems, they're not Einstein. - -06:24.800 --> 06:28.280 - They're like Einstein, but you know, the patent clerk days. - -06:29.800 --> 06:31.920 - Like Einstein became Einstein - -06:31.920 --> 06:36.080 - because this was a meeting of a genius - -06:36.080 --> 06:39.480 - with a big problem at the right time, right? - -06:39.480 --> 06:42.480 - But maybe this meeting could have never happened - -06:42.480 --> 06:44.960 - and then Einstein would have just been a patent clerk, right? - -06:44.960 --> 06:48.400 - And in fact, many people today are probably like - -06:49.760 --> 06:52.240 - genius level smart, but you wouldn't know - -06:52.240 --> 06:54.800 - because they're not really expressing any of that. - -06:54.800 --> 06:55.640 - Wow, that's brilliant. - -06:55.640 --> 06:58.520 - So we can think of the world, Earth, - -06:58.520 --> 07:02.720 - but also the universe as just as a space of problems. - -07:02.720 --> 07:05.160 - So all these problems and tasks are roaming it - -07:05.160 --> 07:06.880 - of various difficulty. - -07:06.880 --> 07:10.120 - And there's agents, creatures like ourselves - -07:10.120 --> 07:13.360 - and animals and so on that are also roaming it. - -07:13.360 --> 07:16.480 - And then you get coupled with a problem - -07:16.480 --> 07:17.640 - and then you solve it. - -07:17.640 --> 07:19.880 - But without that coupling, - -07:19.880 --> 07:22.560 - you can't demonstrate your quote unquote intelligence. - -07:22.560 --> 07:24.480 - Exactly, intelligence is the meeting - -07:24.480 --> 07:27.480 - of great problem solving capabilities - -07:27.480 --> 07:28.760 - with a great problem. - -07:28.760 --> 07:30.560 - And if you don't have the problem, - -07:30.560 --> 07:32.280 - you don't really express any intelligence. - -07:32.280 --> 07:34.760 - All you're left with is potential intelligence, - -07:34.760 --> 07:36.240 - like the performance of your brain - -07:36.240 --> 07:38.680 - or how high your IQ is, - -07:38.680 --> 07:42.080 - which in itself is just a number, right? - -07:42.080 --> 07:46.520 - So you mentioned problem solving capacity. - -07:46.520 --> 07:47.360 - Yeah. - -07:47.360 --> 07:51.800 - What do you think of as problem solving capacity? - -07:51.800 --> 07:55.160 - Can you try to define intelligence? - -07:56.640 --> 08:00.000 - Like what does it mean to be more or less intelligent? - -08:00.000 --> 08:03.000 - Is it completely coupled to a particular problem - -08:03.000 --> 08:05.720 - or is there something a little bit more universal? - -08:05.720 --> 08:07.440 - Yeah, I do believe all intelligence - -08:07.440 --> 08:09.080 - is specialized intelligence. - -08:09.080 --> 08:12.200 - Even human intelligence has some degree of generality. - -08:12.200 --> 08:15.320 - Well, all intelligent systems have some degree of generality - -08:15.320 --> 08:19.400 - but they're always specialized in one category of problems. - -08:19.400 --> 08:21.880 - So the human intelligence is specialized - -08:21.880 --> 08:23.560 - in the human experience. - -08:23.560 --> 08:25.560 - And that shows at various levels, - -08:25.560 --> 08:30.200 - that shows in some prior knowledge that's innate - -08:30.200 --> 08:32.040 - that we have at birth. - -08:32.040 --> 08:35.360 - Knowledge about things like agents, - -08:35.360 --> 08:38.080 - goal driven behavior, visual priors - -08:38.080 --> 08:43.080 - about what makes an object, priors about time and so on. - -08:43.520 --> 08:45.360 - That shows also in the way we learn. - -08:45.360 --> 08:47.160 - For instance, it's very, very easy for us - -08:47.160 --> 08:48.600 - to pick up language. - -08:49.560 --> 08:52.080 - It's very, very easy for us to learn certain things - -08:52.080 --> 08:54.920 - because we are basically hard coded to learn them. - -08:54.920 --> 08:58.280 - And we are specialized in solving certain kinds of problem - -08:58.280 --> 08:59.720 - and we are quite useless - -08:59.720 --> 09:01.440 - when it comes to other kinds of problems. - -09:01.440 --> 09:06.160 - For instance, we are not really designed - -09:06.160 --> 09:08.800 - to handle very long term problems. - -09:08.800 --> 09:12.880 - We have no capability of seeing the very long term. - -09:12.880 --> 09:16.880 - We don't have very much working memory. - -09:18.000 --> 09:20.080 - So how do you think about long term? - -09:20.080 --> 09:21.360 - Do you think long term planning, - -09:21.360 --> 09:24.880 - are we talking about scale of years, millennia? - -09:24.880 --> 09:26.400 - What do you mean by long term? - -09:26.400 --> 09:28.120 - We're not very good. - -09:28.120 --> 09:29.760 - Well, human intelligence is specialized - -09:29.760 --> 09:30.720 - in the human experience. - -09:30.720 --> 09:32.800 - And human experience is very short. - -09:32.800 --> 09:34.240 - One lifetime is short. - -09:34.240 --> 09:35.880 - Even within one lifetime, - -09:35.880 --> 09:40.000 - we have a very hard time envisioning things - -09:40.000 --> 09:41.360 - on a scale of years. - -09:41.360 --> 09:43.240 - It's very difficult to project yourself - -09:43.240 --> 09:46.960 - at a scale of five years, at a scale of 10 years and so on. - -09:46.960 --> 09:50.000 - We can solve only fairly narrowly scoped problems. - -09:50.000 --> 09:52.320 - So when it comes to solving bigger problems, - -09:52.320 --> 09:53.760 - larger scale problems, - -09:53.760 --> 09:56.360 - we are not actually doing it on an individual level. - -09:56.360 --> 09:59.280 - So it's not actually our brain doing it. - -09:59.280 --> 10:03.040 - We have this thing called civilization, right? - -10:03.040 --> 10:06.600 - Which is itself a sort of problem solving system, - -10:06.600 --> 10:10.000 - a sort of artificially intelligent system, right? - -10:10.000 --> 10:12.120 - And it's not running on one brain, - -10:12.120 --> 10:14.080 - it's running on a network of brains. - -10:14.080 --> 10:15.640 - In fact, it's running on much more - -10:15.640 --> 10:16.760 - than a network of brains. - -10:16.760 --> 10:20.080 - It's running on a lot of infrastructure, - -10:20.080 --> 10:23.040 - like books and computers and the internet - -10:23.040 --> 10:25.800 - and human institutions and so on. - -10:25.800 --> 10:30.240 - And that is capable of handling problems - -10:30.240 --> 10:33.760 - on a much greater scale than any individual human. - -10:33.760 --> 10:37.600 - If you look at computer science, for instance, - -10:37.600 --> 10:39.840 - that's an institution that solves problems - -10:39.840 --> 10:42.560 - and it is superhuman, right? - -10:42.560 --> 10:44.200 - It operates on a greater scale. - -10:44.200 --> 10:46.880 - It can solve much bigger problems - -10:46.880 --> 10:49.080 - than an individual human could. - -10:49.080 --> 10:52.160 - And science itself, science as a system, as an institution, - -10:52.160 --> 10:57.120 - is a kind of artificially intelligent problem solving - -10:57.120 --> 10:59.360 - algorithm that is superhuman. - -10:59.360 --> 11:02.800 - Yeah, it's, at least computer science - -11:02.800 --> 11:07.720 - is like a theorem prover at a scale of thousands, - -11:07.720 --> 11:10.400 - maybe hundreds of thousands of human beings. - -11:10.400 --> 11:14.680 - At that scale, what do you think is an intelligent agent? - -11:14.680 --> 11:18.280 - So there's us humans at the individual level, - -11:18.280 --> 11:22.400 - there is millions, maybe billions of bacteria in our skin. - -11:23.880 --> 11:26.400 - There is, that's at the smaller scale. - -11:26.400 --> 11:29.160 - You can even go to the particle level - -11:29.160 --> 11:31.000 - as systems that behave, - -11:31.840 --> 11:34.360 - you can say intelligently in some ways. - -11:35.440 --> 11:37.840 - And then you can look at the earth as a single organism, - -11:37.840 --> 11:39.200 - you can look at our galaxy - -11:39.200 --> 11:42.160 - and even the universe as a single organism. - -11:42.160 --> 11:44.680 - Do you think, how do you think about scale - -11:44.680 --> 11:46.280 - in defining intelligent systems? - -11:46.280 --> 11:50.440 - And we're here at Google, there is millions of devices - -11:50.440 --> 11:53.360 - doing computation just in a distributed way. - -11:53.360 --> 11:55.880 - How do you think about intelligence versus scale? - -11:55.880 --> 11:59.400 - You can always characterize anything as a system. - -12:00.640 --> 12:03.600 - I think people who talk about things - -12:03.600 --> 12:05.320 - like intelligence explosion, - -12:05.320 --> 12:08.760 - tend to focus on one agent is basically one brain, - -12:08.760 --> 12:10.960 - like one brain considered in isolation, - -12:10.960 --> 12:13.200 - like a brain, a jaw that's controlling a body - -12:13.200 --> 12:16.280 - in a very like top to bottom kind of fashion. - -12:16.280 --> 12:19.480 - And that body is pursuing goals into an environment. - -12:19.480 --> 12:20.720 - So it's a very hierarchical view. - -12:20.720 --> 12:22.880 - You have the brain at the top of the pyramid, - -12:22.880 --> 12:25.960 - then you have the body just plainly receiving orders. - -12:25.960 --> 12:27.640 - And then the body is manipulating objects - -12:27.640 --> 12:28.920 - in the environment and so on. - -12:28.920 --> 12:32.920 - So everything is subordinate to this one thing, - -12:32.920 --> 12:34.720 - this epicenter, which is the brain. - -12:34.720 --> 12:37.120 - But in real life, intelligent agents - -12:37.120 --> 12:39.240 - don't really work like this, right? - -12:39.240 --> 12:40.920 - There is no strong delimitation - -12:40.920 --> 12:43.400 - between the brain and the body to start with. - -12:43.400 --> 12:45.000 - You have to look not just at the brain, - -12:45.000 --> 12:46.560 - but at the nervous system. - -12:46.560 --> 12:48.840 - But then the nervous system and the body - -12:48.840 --> 12:50.760 - are naturally two separate entities. - -12:50.760 --> 12:53.960 - So you have to look at an entire animal as one agent. - -12:53.960 --> 12:57.000 - But then you start realizing as you observe an animal - -12:57.000 --> 13:00.200 - over any length of time, - -13:00.200 --> 13:03.160 - that a lot of the intelligence of an animal - -13:03.160 --> 13:04.600 - is actually externalized. - -13:04.600 --> 13:06.240 - That's especially true for humans. - -13:06.240 --> 13:08.880 - A lot of our intelligence is externalized. - -13:08.880 --> 13:10.360 - When you write down some notes, - -13:10.360 --> 13:11.960 - that is externalized intelligence. - -13:11.960 --> 13:14.000 - When you write a computer program, - -13:14.000 --> 13:16.000 - you are externalizing cognition. - -13:16.000 --> 13:19.720 - So it's externalizing books, it's externalized in computers, - -13:19.720 --> 13:21.520 - the internet, in other humans. - -13:23.080 --> 13:25.400 - It's externalizing language and so on. - -13:25.400 --> 13:30.400 - So there is no hard delimitation - -13:30.480 --> 13:32.640 - of what makes an intelligent agent. - -13:32.640 --> 13:33.880 - It's all about context. - -13:34.960 --> 13:38.800 - Okay, but AlphaGo is better at Go - -13:38.800 --> 13:40.200 - than the best human player. - -13:42.520 --> 13:45.000 - There's levels of skill here. - -13:45.000 --> 13:48.600 - So do you think there's such a ability, - -13:48.600 --> 13:52.800 - such a concept as intelligence explosion - -13:52.800 --> 13:54.760 - in a specific task? - -13:54.760 --> 13:57.360 - And then, well, yeah. - -13:57.360 --> 14:00.120 - Do you think it's possible to have a category of tasks - -14:00.120 --> 14:02.080 - on which you do have something - -14:02.080 --> 14:05.040 - like an exponential growth of ability - -14:05.040 --> 14:07.440 - to solve that particular problem? - -14:07.440 --> 14:10.320 - I think if you consider a specific vertical, - -14:10.320 --> 14:13.720 - it's probably possible to some extent. - -14:15.320 --> 14:18.320 - I also don't think we have to speculate about it - -14:18.320 --> 14:22.280 - because we have real world examples - -14:22.280 --> 14:26.920 - of recursively self improving intelligent systems, right? - -14:26.920 --> 14:30.920 - So for instance, science is a problem solving system, - -14:30.920 --> 14:32.600 - a knowledge generation system, - -14:32.600 --> 14:36.240 - like a system that experiences the world in some sense - -14:36.240 --> 14:40.160 - and then gradually understands it and can act on it. - -14:40.160 --> 14:42.120 - And that system is superhuman - -14:42.120 --> 14:45.600 - and it is clearly recursively self improving - -14:45.600 --> 14:47.560 - because science feeds into technology. - -14:47.560 --> 14:50.200 - Technology can be used to build better tools, - -14:50.200 --> 14:52.880 - better computers, better instrumentation and so on, - -14:52.880 --> 14:56.720 - which in turn can make science faster, right? - -14:56.720 --> 15:00.560 - So science is probably the closest thing we have today - -15:00.560 --> 15:04.760 - to a recursively self improving superhuman AI. - -15:04.760 --> 15:08.520 - And you can just observe is science, - -15:08.520 --> 15:10.320 - is scientific progress to the exploding, - -15:10.320 --> 15:12.800 - which itself is an interesting question. - -15:12.800 --> 15:15.560 - You can use that as a basis to try to understand - -15:15.560 --> 15:17.920 - what will happen with a superhuman AI - -15:17.920 --> 15:21.000 - that has a science like behavior. - -15:21.000 --> 15:23.320 - Let me linger on it a little bit more. - -15:23.320 --> 15:27.600 - What is your intuition why an intelligence explosion - -15:27.600 --> 15:28.560 - is not possible? - -15:28.560 --> 15:30.920 - Like taking the scientific, - -15:30.920 --> 15:33.240 - all the semi scientific revolutions, - -15:33.240 --> 15:38.080 - why can't we slightly accelerate that process? - -15:38.080 --> 15:41.200 - So you can absolutely accelerate - -15:41.200 --> 15:43.120 - any problem solving process. - -15:43.120 --> 15:46.720 - So a recursively self improvement - -15:46.720 --> 15:48.640 - is absolutely a real thing. - -15:48.640 --> 15:51.880 - But what happens with a recursively self improving system - -15:51.880 --> 15:53.680 - is typically not explosion - -15:53.680 --> 15:56.520 - because no system exists in isolation. - -15:56.520 --> 15:58.640 - And so tweaking one part of the system - -15:58.640 --> 16:00.880 - means that suddenly another part of the system - -16:00.880 --> 16:02.200 - becomes a bottleneck. - -16:02.200 --> 16:03.800 - And if you look at science, for instance, - -16:03.800 --> 16:06.800 - which is clearly a recursively self improving, - -16:06.800 --> 16:09.040 - clearly a problem solving system, - -16:09.040 --> 16:12.000 - scientific progress is not actually exploding. - -16:12.000 --> 16:13.520 - If you look at science, - -16:13.520 --> 16:16.480 - what you see is the picture of a system - -16:16.480 --> 16:19.240 - that is consuming an exponentially increasing - -16:19.240 --> 16:20.520 - amount of resources, - -16:20.520 --> 16:23.960 - but it's having a linear output - -16:23.960 --> 16:26.000 - in terms of scientific progress. - -16:26.000 --> 16:28.960 - And maybe that will seem like a very strong claim. - -16:28.960 --> 16:31.160 - Many people are actually saying that, - -16:31.160 --> 16:34.560 - scientific progress is exponential, - -16:34.560 --> 16:36.120 - but when they're claiming this, - -16:36.120 --> 16:38.400 - they're actually looking at indicators - -16:38.400 --> 16:43.080 - of resource consumption by science. - -16:43.080 --> 16:46.680 - For instance, the number of papers being published, - -16:47.560 --> 16:49.960 - the number of patents being filed and so on, - -16:49.960 --> 16:53.600 - which are just completely correlated - -16:53.600 --> 16:58.480 - with how many people are working on science today. - -16:58.480 --> 17:00.640 - So it's actually an indicator of resource consumption, - -17:00.640 --> 17:03.200 - but what you should look at is the output, - -17:03.200 --> 17:06.680 - is progress in terms of the knowledge - -17:06.680 --> 17:08.040 - that science generates, - -17:08.040 --> 17:10.640 - in terms of the scope and significance - -17:10.640 --> 17:12.520 - of the problems that we solve. - -17:12.520 --> 17:16.720 - And some people have actually been trying to measure that. - -17:16.720 --> 17:20.160 - Like Michael Nielsen, for instance, - -17:20.160 --> 17:21.920 - he had a very nice paper, - -17:21.920 --> 17:23.720 - I think that was last year about it. - -17:25.200 --> 17:28.360 - So his approach to measure scientific progress - -17:28.360 --> 17:33.360 - was to look at the timeline of scientific discoveries - -17:33.720 --> 17:37.160 - over the past, you know, 100, 150 years. - -17:37.160 --> 17:41.360 - And for each major discovery, - -17:41.360 --> 17:44.360 - ask a panel of experts to rate - -17:44.360 --> 17:46.760 - the significance of the discovery. - -17:46.760 --> 17:49.600 - And if the output of science as an institution - -17:49.600 --> 17:50.440 - were exponential, - -17:50.440 --> 17:55.440 - you would expect the temporal density of significance - -17:56.600 --> 17:58.160 - to go up exponentially. - -17:58.160 --> 18:00.960 - Maybe because there's a faster rate of discoveries, - -18:00.960 --> 18:02.960 - maybe because the discoveries are, you know, - -18:02.960 --> 18:04.920 - increasingly more important. - -18:04.920 --> 18:06.800 - And what actually happens - -18:06.800 --> 18:10.040 - if you plot this temporal density of significance - -18:10.040 --> 18:11.320 - measured in this way, - -18:11.320 --> 18:14.520 - is that you see very much a flat graph. - -18:14.520 --> 18:16.600 - You see a flat graph across all disciplines, - -18:16.600 --> 18:19.720 - across physics, biology, medicine, and so on. - -18:19.720 --> 18:22.480 - And it actually makes a lot of sense - -18:22.480 --> 18:23.320 - if you think about it, - -18:23.320 --> 18:26.000 - because think about the progress of physics - -18:26.000 --> 18:28.000 - 110 years ago, right? - -18:28.000 --> 18:30.080 - It was a time of crazy change. - -18:30.080 --> 18:31.960 - Think about the progress of technology, - -18:31.960 --> 18:34.360 - you know, 170 years ago, - -18:34.360 --> 18:35.400 - when we started having, you know, - -18:35.400 --> 18:37.560 - replacing horses with cars, - -18:37.560 --> 18:40.000 - when we started having electricity and so on. - -18:40.000 --> 18:41.520 - It was a time of incredible change. - -18:41.520 --> 18:44.600 - And today is also a time of very, very fast change, - -18:44.600 --> 18:48.040 - but it would be an unfair characterization - -18:48.040 --> 18:50.560 - to say that today technology and science - -18:50.560 --> 18:52.920 - are moving way faster than they did 50 years ago - -18:52.920 --> 18:54.360 - or 100 years ago. - -18:54.360 --> 18:59.360 - And if you do try to rigorously plot - -18:59.520 --> 19:04.520 - the temporal density of the significance, - -19:04.880 --> 19:07.320 - yeah, of significance, sorry, - -19:07.320 --> 19:09.720 - you do see very flat curves. - -19:09.720 --> 19:12.040 - And you can check out the paper - -19:12.040 --> 19:16.000 - that Michael Nielsen had about this idea. - -19:16.000 --> 19:20.000 - And so the way I interpret it is, - -19:20.000 --> 19:24.160 - as you make progress in a given field, - -19:24.160 --> 19:26.120 - or in a given subfield of science, - -19:26.120 --> 19:28.680 - it becomes exponentially more difficult - -19:28.680 --> 19:30.440 - to make further progress. - -19:30.440 --> 19:35.000 - Like the very first person to work on information theory. - -19:35.000 --> 19:36.440 - If you enter a new field, - -19:36.440 --> 19:37.920 - and it's still the very early years, - -19:37.920 --> 19:41.160 - there's a lot of low hanging fruit you can pick. - -19:41.160 --> 19:42.000 - That's right, yeah. - -19:42.000 --> 19:43.960 - But the next generation of researchers - -19:43.960 --> 19:48.160 - is gonna have to dig much harder, actually, - -19:48.160 --> 19:50.640 - to make smaller discoveries, - -19:50.640 --> 19:52.640 - probably larger number of smaller discoveries, - -19:52.640 --> 19:54.640 - and to achieve the same amount of impact, - -19:54.640 --> 19:57.480 - you're gonna need a much greater head count. - -19:57.480 --> 20:00.040 - And that's exactly the picture you're seeing with science, - -20:00.040 --> 20:03.760 - that the number of scientists and engineers - -20:03.760 --> 20:06.520 - is in fact increasing exponentially. - -20:06.520 --> 20:08.400 - The amount of computational resources - -20:08.400 --> 20:10.040 - that are available to science - -20:10.040 --> 20:11.880 - is increasing exponentially and so on. - -20:11.880 --> 20:15.560 - So the resource consumption of science is exponential, - -20:15.560 --> 20:18.200 - but the output in terms of progress, - -20:18.200 --> 20:21.000 - in terms of significance, is linear. - -20:21.000 --> 20:23.120 - And the reason why is because, - -20:23.120 --> 20:26.000 - and even though science is regressively self improving, - -20:26.000 --> 20:28.440 - meaning that scientific progress - -20:28.440 --> 20:30.240 - turns into technological progress, - -20:30.240 --> 20:32.960 - which in turn helps science. - -20:32.960 --> 20:35.280 - If you look at computers, for instance, - -20:35.280 --> 20:38.480 - our products of science and computers - -20:38.480 --> 20:41.560 - are tremendously useful in speeding up science. - -20:41.560 --> 20:43.840 - The internet, same thing, the internet is a technology - -20:43.840 --> 20:47.480 - that's made possible by very recent scientific advances. - -20:47.480 --> 20:52.400 - And itself, because it enables scientists to network, - -20:52.400 --> 20:55.520 - to communicate, to exchange papers and ideas much faster, - -20:55.520 --> 20:57.440 - it is a way to speed up scientific progress. - -20:57.440 --> 20:58.440 - So even though you're looking - -20:58.440 --> 21:01.440 - at a regressively self improving system, - -21:01.440 --> 21:04.080 - it is consuming exponentially more resources - -21:04.080 --> 21:09.080 - to produce the same amount of problem solving, very much. - -21:09.200 --> 21:11.080 - So that's a fascinating way to paint it, - -21:11.080 --> 21:14.960 - and certainly that holds for the deep learning community. - -21:14.960 --> 21:18.120 - If you look at the temporal, what did you call it, - -21:18.120 --> 21:21.240 - the temporal density of significant ideas, - -21:21.240 --> 21:23.920 - if you look at in deep learning, - -21:24.840 --> 21:26.960 - I think, I'd have to think about that, - -21:26.960 --> 21:29.040 - but if you really look at significant ideas - -21:29.040 --> 21:32.400 - in deep learning, they might even be decreasing. - -21:32.400 --> 21:37.400 - So I do believe the per paper significance is decreasing, - -21:39.600 --> 21:41.240 - but the amount of papers - -21:41.240 --> 21:43.440 - is still today exponentially increasing. - -21:43.440 --> 21:45.880 - So I think if you look at an aggregate, - -21:45.880 --> 21:48.840 - my guess is that you would see a linear progress. - -21:48.840 --> 21:53.840 - If you were to sum the significance of all papers, - -21:56.120 --> 21:58.640 - you would see roughly in your progress. - -21:58.640 --> 22:03.640 - And in my opinion, it is not a coincidence - -22:03.880 --> 22:05.800 - that you're seeing linear progress in science - -22:05.800 --> 22:07.720 - despite exponential resource consumption. - -22:07.720 --> 22:10.280 - I think the resource consumption - -22:10.280 --> 22:15.280 - is dynamically adjusting itself to maintain linear progress - -22:15.880 --> 22:18.560 - because we as a community expect linear progress, - -22:18.560 --> 22:21.240 - meaning that if we start investing less - -22:21.240 --> 22:23.600 - and seeing less progress, it means that suddenly - -22:23.600 --> 22:26.800 - there are some lower hanging fruits that become available - -22:26.800 --> 22:31.280 - and someone's gonna step up and pick them, right? - -22:31.280 --> 22:36.280 - So it's very much like a market for discoveries and ideas. - -22:36.920 --> 22:38.720 - But there's another fundamental part - -22:38.720 --> 22:41.800 - which you're highlighting, which as a hypothesis - -22:41.800 --> 22:45.160 - as science or like the space of ideas, - -22:45.160 --> 22:48.160 - any one path you travel down, - -22:48.160 --> 22:51.080 - it gets exponentially more difficult - -22:51.080 --> 22:54.720 - to get a new way to develop new ideas. - -22:54.720 --> 22:57.640 - And your sense is that's gonna hold - -22:57.640 --> 23:01.520 - across our mysterious universe. - -23:01.520 --> 23:03.360 - Yes, well, exponential progress - -23:03.360 --> 23:05.480 - triggers exponential friction. - -23:05.480 --> 23:07.440 - So that if you tweak one part of the system, - -23:07.440 --> 23:10.680 - suddenly some other part becomes a bottleneck, right? - -23:10.680 --> 23:14.880 - For instance, let's say you develop some device - -23:14.880 --> 23:17.160 - that measures its own acceleration - -23:17.160 --> 23:18.720 - and then it has some engine - -23:18.720 --> 23:20.800 - and it outputs even more acceleration - -23:20.800 --> 23:22.360 - in proportion of its own acceleration - -23:22.360 --> 23:23.320 - and you drop it somewhere, - -23:23.320 --> 23:25.240 - it's not gonna reach infinite speed - -23:25.240 --> 23:27.880 - because it exists in a certain context. - -23:29.080 --> 23:31.000 - So the air around it is gonna generate friction - -23:31.000 --> 23:34.320 - and it's gonna block it at some top speed. - -23:34.320 --> 23:37.480 - And even if you were to consider the broader context - -23:37.480 --> 23:39.840 - and lift the bottleneck there, - -23:39.840 --> 23:42.240 - like the bottleneck of friction, - -23:43.120 --> 23:45.120 - then some other part of the system - -23:45.120 --> 23:48.120 - would start stepping in and creating exponential friction, - -23:48.120 --> 23:49.920 - maybe the speed of flight or whatever. - -23:49.920 --> 23:51.920 - And this definitely holds true - -23:51.920 --> 23:54.960 - when you look at the problem solving algorithm - -23:54.960 --> 23:58.160 - that is being run by science as an institution, - -23:58.160 --> 23:59.720 - science as a system. - -23:59.720 --> 24:01.720 - As you make more and more progress, - -24:01.720 --> 24:05.800 - despite having this recursive self improvement component, - -24:06.760 --> 24:09.840 - you are encountering exponential friction. - -24:09.840 --> 24:13.480 - The more researchers you have working on different ideas, - -24:13.480 --> 24:14.880 - the more overhead you have - -24:14.880 --> 24:18.040 - in terms of communication across researchers. - -24:18.040 --> 24:22.920 - If you look at, you were mentioning quantum mechanics, right? - -24:22.920 --> 24:26.880 - Well, if you want to start making significant discoveries - -24:26.880 --> 24:29.680 - today, significant progress in quantum mechanics, - -24:29.680 --> 24:33.000 - there is an amount of knowledge you have to ingest, - -24:33.000 --> 24:34.080 - which is huge. - -24:34.080 --> 24:36.520 - So there's a very large overhead - -24:36.520 --> 24:39.240 - to even start to contribute. - -24:39.240 --> 24:40.680 - There's a large amount of overhead - -24:40.680 --> 24:44.040 - to synchronize across researchers and so on. - -24:44.040 --> 24:47.440 - And of course, the significant practical experiments - -24:48.600 --> 24:52.160 - are going to require exponentially expensive equipment - -24:52.160 --> 24:56.480 - because the easier ones have already been run, right? - -24:56.480 --> 25:00.480 - So in your senses, there's no way escaping, - -25:00.480 --> 25:04.480 - there's no way of escaping this kind of friction - -25:04.480 --> 25:08.600 - with artificial intelligence systems. - -25:08.600 --> 25:11.520 - Yeah, no, I think science is a very good way - -25:11.520 --> 25:14.280 - to model what would happen with a superhuman - -25:14.280 --> 25:16.440 - recursive research improving AI. - -25:16.440 --> 25:18.240 - That's your sense, I mean, the... - -25:18.240 --> 25:19.680 - That's my intuition. - -25:19.680 --> 25:23.400 - It's not like a mathematical proof of anything. - -25:23.400 --> 25:24.400 - That's not my point. - -25:24.400 --> 25:26.600 - Like, I'm not trying to prove anything. - -25:26.600 --> 25:27.920 - I'm just trying to make an argument - -25:27.920 --> 25:31.160 - to question the narrative of intelligence explosion, - -25:31.160 --> 25:32.880 - which is quite a dominant narrative. - -25:32.880 --> 25:35.840 - And you do get a lot of pushback if you go against it. - -25:35.840 --> 25:39.320 - Because, so for many people, right, - -25:39.320 --> 25:42.200 - AI is not just a subfield of computer science. - -25:42.200 --> 25:44.120 - It's more like a belief system. - -25:44.120 --> 25:48.640 - Like this belief that the world is headed towards an event, - -25:48.640 --> 25:55.040 - the singularity, past which, you know, AI will become... - -25:55.040 --> 25:57.080 - will go exponential very much, - -25:57.080 --> 25:58.600 - and the world will be transformed, - -25:58.600 --> 26:00.840 - and humans will become obsolete. - -26:00.840 --> 26:03.880 - And if you go against this narrative, - -26:03.880 --> 26:06.920 - because it is not really a scientific argument, - -26:06.920 --> 26:08.880 - but more of a belief system, - -26:08.880 --> 26:11.240 - it is part of the identity of many people. - -26:11.240 --> 26:12.600 - If you go against this narrative, - -26:12.600 --> 26:14.400 - it's like you're attacking the identity - -26:14.400 --> 26:15.560 - of people who believe in it. - -26:15.560 --> 26:17.640 - It's almost like saying God doesn't exist, - -26:17.640 --> 26:19.000 - or something. - -26:19.000 --> 26:21.880 - So you do get a lot of pushback - -26:21.880 --> 26:24.040 - if you try to question these ideas. - -26:24.040 --> 26:26.520 - First of all, I believe most people, - -26:26.520 --> 26:29.240 - they might not be as eloquent or explicit as you're being, - -26:29.240 --> 26:30.920 - but most people in computer science - -26:30.920 --> 26:33.000 - are most people who actually have built - -26:33.000 --> 26:36.360 - anything that you could call AI, quote, unquote, - -26:36.360 --> 26:38.080 - would agree with you. - -26:38.080 --> 26:40.560 - They might not be describing in the same kind of way. - -26:40.560 --> 26:43.960 - It's more, so the pushback you're getting - -26:43.960 --> 26:48.080 - is from people who get attached to the narrative - -26:48.080 --> 26:51.000 - from, not from a place of science, - -26:51.000 --> 26:53.400 - but from a place of imagination. - -26:53.400 --> 26:54.760 - That's correct, that's correct. - -26:54.760 --> 26:56.920 - So why do you think that's so appealing? - -26:56.920 --> 27:01.920 - Because the usual dreams that people have - -27:02.120 --> 27:03.960 - when you create a superintelligence system - -27:03.960 --> 27:05.120 - past the singularity, - -27:05.120 --> 27:08.600 - that what people imagine is somehow always destructive. - -27:09.440 --> 27:12.240 - Do you have, if you were put on your psychology hat, - -27:12.240 --> 27:17.240 - what's, why is it so appealing to imagine - -27:17.400 --> 27:20.760 - the ways that all of human civilization will be destroyed? - -27:20.760 --> 27:22.080 - I think it's a good story. - -27:22.080 --> 27:23.120 - You know, it's a good story. - -27:23.120 --> 27:28.120 - And very interestingly, it mirrors a religious stories, - -27:28.160 --> 27:30.560 - right, religious mythology. - -27:30.560 --> 27:34.360 - If you look at the mythology of most civilizations, - -27:34.360 --> 27:38.280 - it's about the world being headed towards some final events - -27:38.280 --> 27:40.480 - in which the world will be destroyed - -27:40.480 --> 27:42.800 - and some new world order will arise - -27:42.800 --> 27:44.920 - that will be mostly spiritual, - -27:44.920 --> 27:49.400 - like the apocalypse followed by a paradise probably, right? - -27:49.400 --> 27:52.600 - It's a very appealing story on a fundamental level. - -27:52.600 --> 27:54.560 - And we all need stories. - -27:54.560 --> 27:58.160 - We all need stories to structure the way we see the world, - -27:58.160 --> 27:59.960 - especially at timescales - -27:59.960 --> 28:04.520 - that are beyond our ability to make predictions, right? - -28:04.520 --> 28:08.840 - So on a more serious non exponential explosion, - -28:08.840 --> 28:13.840 - question, do you think there will be a time - -28:15.000 --> 28:19.800 - when we'll create something like human level intelligence - -28:19.800 --> 28:23.800 - or intelligent systems that will make you sit back - -28:23.800 --> 28:28.520 - and be just surprised at damn how smart this thing is? - -28:28.520 --> 28:30.160 - That doesn't require exponential growth - -28:30.160 --> 28:32.120 - or an exponential improvement, - -28:32.120 --> 28:35.600 - but what's your sense of the timeline and so on - -28:35.600 --> 28:40.600 - that you'll be really surprised at certain capabilities? - -28:41.080 --> 28:42.560 - And we'll talk about limitations and deep learning. - -28:42.560 --> 28:44.480 - So do you think in your lifetime, - -28:44.480 --> 28:46.600 - you'll be really damn surprised? - -28:46.600 --> 28:51.440 - Around 2013, 2014, I was many times surprised - -28:51.440 --> 28:53.960 - by the capabilities of deep learning actually. - -28:53.960 --> 28:55.920 - That was before we had assessed exactly - -28:55.920 --> 28:57.880 - what deep learning could do and could not do. - -28:57.880 --> 29:00.600 - And it felt like a time of immense potential. - -29:00.600 --> 29:03.080 - And then we started narrowing it down, - -29:03.080 --> 29:04.360 - but I was very surprised. - -29:04.360 --> 29:07.120 - I would say it has already happened. - -29:07.120 --> 29:10.800 - Was there a moment, there must've been a day in there - -29:10.800 --> 29:14.360 - where your surprise was almost bordering - -29:14.360 --> 29:19.360 - on the belief of the narrative that we just discussed. - -29:19.440 --> 29:20.800 - Was there a moment, - -29:20.800 --> 29:22.400 - because you've written quite eloquently - -29:22.400 --> 29:23.960 - about the limits of deep learning, - -29:23.960 --> 29:25.760 - was there a moment that you thought - -29:25.760 --> 29:27.720 - that maybe deep learning is limitless? - -29:30.000 --> 29:32.400 - No, I don't think I've ever believed this. - -29:32.400 --> 29:35.560 - What was really shocking is that it worked. - -29:35.560 --> 29:37.640 - It worked at all, yeah. - -29:37.640 --> 29:40.520 - But there's a big jump between being able - -29:40.520 --> 29:43.400 - to do really good computer vision - -29:43.400 --> 29:44.920 - and human level intelligence. - -29:44.920 --> 29:49.520 - So I don't think at any point I wasn't under the impression - -29:49.520 --> 29:51.280 - that the results we got in computer vision - -29:51.280 --> 29:54.080 - meant that we were very close to human level intelligence. - -29:54.080 --> 29:56.040 - I don't think we're very close to human level intelligence. - -29:56.040 --> 29:58.520 - I do believe that there's no reason - -29:58.520 --> 30:01.760 - why we won't achieve it at some point. - -30:01.760 --> 30:06.400 - I also believe that it's the problem - -30:06.400 --> 30:08.560 - with talking about human level intelligence - -30:08.560 --> 30:11.240 - that implicitly you're considering - -30:11.240 --> 30:14.360 - like an axis of intelligence with different levels, - -30:14.360 --> 30:16.720 - but that's not really how intelligence works. - -30:16.720 --> 30:19.480 - Intelligence is very multi dimensional. - -30:19.480 --> 30:22.480 - And so there's the question of capabilities, - -30:22.480 --> 30:25.560 - but there's also the question of being human like, - -30:25.560 --> 30:27.040 - and it's two very different things. - -30:27.040 --> 30:28.280 - Like you can build potentially - -30:28.280 --> 30:30.640 - very advanced intelligent agents - -30:30.640 --> 30:32.640 - that are not human like at all. - -30:32.640 --> 30:35.240 - And you can also build very human like agents. - -30:35.240 --> 30:37.840 - And these are two very different things, right? - -30:37.840 --> 30:38.760 - Right. - -30:38.760 --> 30:42.240 - Let's go from the philosophical to the practical. - -30:42.240 --> 30:44.240 - Can you give me a history of Keras - -30:44.240 --> 30:46.440 - and all the major deep learning frameworks - -30:46.440 --> 30:48.480 - that you kind of remember in relation to Keras - -30:48.480 --> 30:52.040 - and in general, TensorFlow, Theano, the old days. - -30:52.040 --> 30:55.400 - Can you give a brief overview Wikipedia style history - -30:55.400 --> 30:59.120 - and your role in it before we return to AGI discussions? - -30:59.120 --> 31:00.720 - Yeah, that's a broad topic. - -31:00.720 --> 31:04.040 - So I started working on Keras. - -31:04.920 --> 31:06.240 - It was the name Keras at the time. - -31:06.240 --> 31:08.320 - I actually picked the name like - -31:08.320 --> 31:10.200 - just the day I was going to release it. - -31:10.200 --> 31:14.800 - So I started working on it in February, 2015. - -31:14.800 --> 31:17.240 - And so at the time there weren't too many people - -31:17.240 --> 31:20.320 - working on deep learning, maybe like fewer than 10,000. - -31:20.320 --> 31:22.840 - The software tooling was not really developed. - -31:25.320 --> 31:28.800 - So the main deep learning library was Cafe, - -31:28.800 --> 31:30.840 - which was mostly C++. - -31:30.840 --> 31:32.760 - Why do you say Cafe was the main one? - -31:32.760 --> 31:36.000 - Cafe was vastly more popular than Theano - -31:36.000 --> 31:38.920 - in late 2014, early 2015. - -31:38.920 --> 31:42.400 - Cafe was the one library that everyone was using - -31:42.400 --> 31:43.400 - for computer vision. - -31:43.400 --> 31:46.120 - And computer vision was the most popular problem - -31:46.120 --> 31:46.960 - in deep learning at the time. - -31:46.960 --> 31:47.800 - Absolutely. - -31:47.800 --> 31:50.440 - Like ConvNets was like the subfield of deep learning - -31:50.440 --> 31:53.160 - that everyone was working on. - -31:53.160 --> 31:57.680 - So myself, so in late 2014, - -31:57.680 --> 32:00.600 - I was actually interested in RNNs, - -32:00.600 --> 32:01.760 - in recurrent neural networks, - -32:01.760 --> 32:05.800 - which was a very niche topic at the time, right? - -32:05.800 --> 32:08.640 - It really took off around 2016. - -32:08.640 --> 32:11.080 - And so I was looking for good tools. - -32:11.080 --> 32:14.800 - I had used Torch 7, I had used Theano, - -32:14.800 --> 32:17.640 - used Theano a lot in Kaggle competitions. - -32:19.320 --> 32:20.840 - I had used Cafe. - -32:20.840 --> 32:25.840 - And there was no like good solution for RNNs at the time. - -32:25.840 --> 32:28.640 - Like there was no reusable open source implementation - -32:28.640 --> 32:30.000 - of an LSTM, for instance. - -32:30.000 --> 32:32.920 - So I decided to build my own. - -32:32.920 --> 32:35.440 - And at first, the pitch for that was, - -32:35.440 --> 32:39.960 - it was gonna be mostly around LSTM recurrent neural networks. - -32:39.960 --> 32:41.360 - It was gonna be in Python. - -32:42.280 --> 32:44.280 - An important decision at the time - -32:44.280 --> 32:45.440 - that was kind of not obvious - -32:45.440 --> 32:50.360 - is that the models would be defined via Python code, - -32:50.360 --> 32:54.400 - which was kind of like going against the mainstream - -32:54.400 --> 32:58.000 - at the time because Cafe, Pylon 2, and so on, - -32:58.000 --> 33:00.600 - like all the big libraries were actually going - -33:00.600 --> 33:03.520 - with the approach of setting configuration files - -33:03.520 --> 33:05.560 - in YAML to define models. - -33:05.560 --> 33:08.840 - So some libraries were using code to define models, - -33:08.840 --> 33:12.280 - like Torch 7, obviously, but that was not Python. - -33:12.280 --> 33:16.680 - Lasagne was like a Theano based very early library - -33:16.680 --> 33:18.640 - that was, I think, developed, I don't remember exactly, - -33:18.640 --> 33:20.240 - probably late 2014. - -33:20.240 --> 33:21.200 - It's Python as well. - -33:21.200 --> 33:22.040 - It's Python as well. - -33:22.040 --> 33:24.320 - It was like on top of Theano. - -33:24.320 --> 33:28.320 - And so I started working on something - -33:29.480 --> 33:32.520 - and the value proposition at the time was that - -33:32.520 --> 33:36.240 - not only what I think was the first - -33:36.240 --> 33:38.800 - reusable open source implementation of LSTM, - -33:40.400 --> 33:44.440 - you could combine RNNs and covenants - -33:44.440 --> 33:45.440 - with the same library, - -33:45.440 --> 33:46.920 - which is not really possible before, - -33:46.920 --> 33:49.080 - like Cafe was only doing covenants. - -33:50.440 --> 33:52.560 - And it was kind of easy to use - -33:52.560 --> 33:54.440 - because, so before I was using Theano, - -33:54.440 --> 33:55.680 - I was actually using scikitlin - -33:55.680 --> 33:58.320 - and I loved scikitlin for its usability. - -33:58.320 --> 34:01.560 - So I drew a lot of inspiration from scikitlin - -34:01.560 --> 34:02.400 - when I made Keras. - -34:02.400 --> 34:05.600 - It's almost like scikitlin for neural networks. - -34:05.600 --> 34:06.680 - The fit function. - -34:06.680 --> 34:07.960 - Exactly, the fit function, - -34:07.960 --> 34:10.800 - like reducing a complex string loop - -34:10.800 --> 34:12.880 - to a single function call, right? - -34:12.880 --> 34:14.880 - And of course, some people will say, - -34:14.880 --> 34:16.320 - this is hiding a lot of details, - -34:16.320 --> 34:18.680 - but that's exactly the point, right? - -34:18.680 --> 34:20.280 - The magic is the point. - -34:20.280 --> 34:22.680 - So it's magical, but in a good way. - -34:22.680 --> 34:24.960 - It's magical in the sense that it's delightful. - -34:24.960 --> 34:26.160 - Yeah, yeah. - -34:26.160 --> 34:27.640 - I'm actually quite surprised. - -34:27.640 --> 34:29.600 - I didn't know that it was born out of desire - -34:29.600 --> 34:32.480 - to implement RNNs and LSTMs. - -34:32.480 --> 34:33.320 - It was. - -34:33.320 --> 34:34.160 - That's fascinating. - -34:34.160 --> 34:36.040 - So you were actually one of the first people - -34:36.040 --> 34:37.960 - to really try to attempt - -34:37.960 --> 34:41.000 - to get the major architectures together. - -34:41.000 --> 34:42.760 - And it's also interesting. - -34:42.760 --> 34:45.160 - You made me realize that that was a design decision at all - -34:45.160 --> 34:47.360 - is defining the model and code. - -34:47.360 --> 34:49.920 - Just, I'm putting myself in your shoes, - -34:49.920 --> 34:53.200 - whether the YAML, especially if cafe was the most popular. - -34:53.200 --> 34:54.720 - It was the most popular by far. - -34:54.720 --> 34:58.480 - If I was, if I were, yeah, I don't, - -34:58.480 --> 34:59.560 - I didn't like the YAML thing, - -34:59.560 --> 35:02.840 - but it makes more sense that you will put - -35:02.840 --> 35:05.720 - in a configuration file, the definition of a model. - -35:05.720 --> 35:07.200 - That's an interesting gutsy move - -35:07.200 --> 35:10.040 - to stick with defining it in code. - -35:10.040 --> 35:11.600 - Just if you look back. - -35:11.600 --> 35:13.480 - Other libraries were doing it as well, - -35:13.480 --> 35:16.320 - but it was definitely the more niche option. - -35:16.320 --> 35:17.160 - Yeah. - -35:17.160 --> 35:18.360 - Okay, Keras and then. - -35:18.360 --> 35:21.520 - So I released Keras in March, 2015, - -35:21.520 --> 35:24.160 - and it got users pretty much from the start. - -35:24.160 --> 35:25.800 - So the deep learning community was very, very small - -35:25.800 --> 35:27.240 - at the time. - -35:27.240 --> 35:30.600 - Lots of people were starting to be interested in LSTM. - -35:30.600 --> 35:32.440 - So it was gonna release it at the right time - -35:32.440 --> 35:35.560 - because it was offering an easy to use LSTM implementation. - -35:35.560 --> 35:37.680 - Exactly at the time where lots of people started - -35:37.680 --> 35:42.280 - to be intrigued by the capabilities of RNN, RNNs for NLP. - -35:42.280 --> 35:43.920 - So it grew from there. - -35:43.920 --> 35:48.920 - Then I joined Google about six months later, - -35:51.480 --> 35:54.920 - and that was actually completely unrelated to Keras. - -35:54.920 --> 35:57.080 - So I actually joined a research team - -35:57.080 --> 35:59.520 - working on image classification, - -35:59.520 --> 36:00.680 - mostly like computer vision. - -36:00.680 --> 36:02.320 - So I was doing computer vision research - -36:02.320 --> 36:03.640 - at Google initially. - -36:03.640 --> 36:05.520 - And immediately when I joined Google, - -36:05.520 --> 36:10.520 - I was exposed to the early internal version of TensorFlow. - -36:10.520 --> 36:13.920 - And the way it appeared to me at the time, - -36:13.920 --> 36:15.720 - and it was definitely the way it was at the time - -36:15.720 --> 36:20.760 - is that this was an improved version of Theano. - -36:20.760 --> 36:24.720 - So I immediately knew I had to port Keras - -36:24.720 --> 36:26.800 - to this new TensorFlow thing. - -36:26.800 --> 36:29.800 - And I was actually very busy as a noobler, - -36:29.800 --> 36:30.720 - as a new Googler. - -36:31.600 --> 36:34.520 - So I had not time to work on that. - -36:34.520 --> 36:38.680 - But then in November, I think it was November, 2015, - -36:38.680 --> 36:41.240 - TensorFlow got released. - -36:41.240 --> 36:44.560 - And it was kind of like my wake up call - -36:44.560 --> 36:47.320 - that, hey, I had to actually go and make it happen. - -36:47.320 --> 36:52.200 - So in December, I ported Keras to run on top of TensorFlow, - -36:52.200 --> 36:53.320 - but it was not exactly a port. - -36:53.320 --> 36:55.280 - It was more like a refactoring - -36:55.280 --> 36:57.920 - where I was abstracting away - -36:57.920 --> 37:00.480 - all the backend functionality into one module - -37:00.480 --> 37:02.320 - so that the same code base - -37:02.320 --> 37:05.080 - could run on top of multiple backends. - -37:05.080 --> 37:07.440 - So on top of TensorFlow or Theano. - -37:07.440 --> 37:09.760 - And for the next year, - -37:09.760 --> 37:14.760 - Theano stayed as the default option. - -37:15.400 --> 37:20.400 - It was easier to use, somewhat less buggy. - -37:20.640 --> 37:23.360 - It was much faster, especially when it came to audience. - -37:23.360 --> 37:26.360 - But eventually, TensorFlow overtook it. - -37:27.480 --> 37:30.200 - And TensorFlow, the early TensorFlow, - -37:30.200 --> 37:33.960 - has similar architectural decisions as Theano, right? - -37:33.960 --> 37:37.440 - So it was a natural transition. - -37:37.440 --> 37:38.320 - Yeah, absolutely. - -37:38.320 --> 37:42.960 - So what, I mean, that still Keras is a side, - -37:42.960 --> 37:45.280 - almost fun project, right? - -37:45.280 --> 37:49.040 - Yeah, so it was not my job assignment. - -37:49.040 --> 37:50.360 - It was not. - -37:50.360 --> 37:52.240 - I was doing it on the side. - -37:52.240 --> 37:55.840 - And even though it grew to have a lot of users - -37:55.840 --> 37:59.600 - for a deep learning library at the time, like Stroud 2016, - -37:59.600 --> 38:02.480 - but I wasn't doing it as my main job. - -38:02.480 --> 38:04.760 - So things started changing in, - -38:04.760 --> 38:09.760 - I think it must have been maybe October, 2016. - -38:10.200 --> 38:11.320 - So one year later. - -38:12.360 --> 38:15.240 - So Rajat, who was the lead on TensorFlow, - -38:15.240 --> 38:19.240 - basically showed up one day in our building - -38:19.240 --> 38:20.080 - where I was doing like, - -38:20.080 --> 38:21.640 - so I was doing research and things like, - -38:21.640 --> 38:24.640 - so I did a lot of computer vision research, - -38:24.640 --> 38:27.560 - also collaborations with Christian Zighetti - -38:27.560 --> 38:29.640 - and deep learning for theorem proving. - -38:29.640 --> 38:32.920 - It was a really interesting research topic. - -38:34.520 --> 38:37.640 - And so Rajat was saying, - -38:37.640 --> 38:41.040 - hey, we saw Keras, we like it. - -38:41.040 --> 38:42.440 - We saw that you're at Google. - -38:42.440 --> 38:45.280 - Why don't you come over for like a quarter - -38:45.280 --> 38:47.280 - and work with us? - -38:47.280 --> 38:49.240 - And I was like, yeah, that sounds like a great opportunity. - -38:49.240 --> 38:50.400 - Let's do it. - -38:50.400 --> 38:55.400 - And so I started working on integrating the Keras API - -38:55.720 --> 38:57.320 - into TensorFlow more tightly. - -38:57.320 --> 39:02.320 - So what followed up is a sort of like temporary - -39:02.640 --> 39:05.480 - TensorFlow only version of Keras - -39:05.480 --> 39:09.320 - that was in TensorFlow.com Trib for a while. - -39:09.320 --> 39:12.200 - And finally moved to TensorFlow Core. - -39:12.200 --> 39:15.360 - And I've never actually gotten back - -39:15.360 --> 39:17.600 - to my old team doing research. - -39:17.600 --> 39:22.320 - Well, it's kind of funny that somebody like you - -39:22.320 --> 39:27.320 - who dreams of, or at least sees the power of AI systems - -39:28.960 --> 39:31.680 - that reason and theorem proving we'll talk about - -39:31.680 --> 39:36.520 - has also created a system that makes the most basic - -39:36.520 --> 39:40.400 - kind of Lego building that is deep learning - -39:40.400 --> 39:42.640 - super accessible, super easy. - -39:42.640 --> 39:43.800 - So beautifully so. - -39:43.800 --> 39:47.720 - It's a funny irony that you're both, - -39:47.720 --> 39:49.120 - you're responsible for both things, - -39:49.120 --> 39:54.000 - but so TensorFlow 2.0 is kind of, there's a sprint. - -39:54.000 --> 39:55.080 - I don't know how long it'll take, - -39:55.080 --> 39:56.960 - but there's a sprint towards the finish. - -39:56.960 --> 40:01.040 - What do you look, what are you working on these days? - -40:01.040 --> 40:02.160 - What are you excited about? - -40:02.160 --> 40:04.280 - What are you excited about in 2.0? - -40:04.280 --> 40:05.760 - I mean, eager execution. - -40:05.760 --> 40:08.440 - There's so many things that just make it a lot easier - -40:08.440 --> 40:09.760 - to work. - -40:09.760 --> 40:13.640 - What are you excited about and what's also really hard? - -40:13.640 --> 40:15.800 - What are the problems you have to kind of solve? - -40:15.800 --> 40:19.080 - So I've spent the past year and a half working on - -40:19.080 --> 40:22.920 - TensorFlow 2.0 and it's been a long journey. - -40:22.920 --> 40:25.080 - I'm actually extremely excited about it. - -40:25.080 --> 40:26.440 - I think it's a great product. - -40:26.440 --> 40:29.360 - It's a delightful product compared to TensorFlow 1.0. - -40:29.360 --> 40:31.440 - We've made huge progress. - -40:32.640 --> 40:37.400 - So on the Keras side, what I'm really excited about is that, - -40:37.400 --> 40:42.400 - so previously Keras has been this very easy to use - -40:42.400 --> 40:45.840 - high level interface to do deep learning. - -40:45.840 --> 40:47.280 - But if you wanted to, - -40:50.520 --> 40:53.040 - if you wanted a lot of flexibility, - -40:53.040 --> 40:57.520 - the Keras framework was probably not the optimal way - -40:57.520 --> 40:59.760 - to do things compared to just writing everything - -40:59.760 --> 41:00.600 - from scratch. - -41:01.800 --> 41:04.680 - So in some way, the framework was getting in the way. - -41:04.680 --> 41:07.960 - And in TensorFlow 2.0, you don't have this at all, actually. - -41:07.960 --> 41:11.040 - You have the usability of the high level interface, - -41:11.040 --> 41:14.480 - but you have the flexibility of this lower level interface. - -41:14.480 --> 41:16.800 - And you have this spectrum of workflows - -41:16.800 --> 41:21.560 - where you can get more or less usability - -41:21.560 --> 41:26.560 - and flexibility trade offs depending on your needs, right? - -41:26.640 --> 41:29.680 - You can write everything from scratch - -41:29.680 --> 41:32.320 - and you get a lot of help doing so - -41:32.320 --> 41:36.400 - by subclassing models and writing some train loops - -41:36.400 --> 41:38.200 - using ego execution. - -41:38.200 --> 41:40.160 - It's very flexible, it's very easy to debug, - -41:40.160 --> 41:41.400 - it's very powerful. - -41:42.280 --> 41:45.000 - But all of this integrates seamlessly - -41:45.000 --> 41:49.440 - with higher level features up to the classic Keras workflows, - -41:49.440 --> 41:51.560 - which are very scikit learn like - -41:51.560 --> 41:56.040 - and are ideal for a data scientist, - -41:56.040 --> 41:58.240 - machine learning engineer type of profile. - -41:58.240 --> 42:00.840 - So now you can have the same framework - -42:00.840 --> 42:02.880 - offering the same set of APIs - -42:02.880 --> 42:05.000 - that enable a spectrum of workflows - -42:05.000 --> 42:08.560 - that are more or less low level, more or less high level - -42:08.560 --> 42:13.520 - that are suitable for profiles ranging from researchers - -42:13.520 --> 42:15.560 - to data scientists and everything in between. - -42:15.560 --> 42:16.960 - Yeah, so that's super exciting. - -42:16.960 --> 42:18.400 - I mean, it's not just that, - -42:18.400 --> 42:21.680 - it's connected to all kinds of tooling. - -42:21.680 --> 42:24.520 - You can go on mobile, you can go with TensorFlow Lite, - -42:24.520 --> 42:27.240 - you can go in the cloud or serving and so on. - -42:27.240 --> 42:28.960 - It all is connected together. - -42:28.960 --> 42:31.880 - Now some of the best software written ever - -42:31.880 --> 42:36.880 - is often done by one person, sometimes two. - -42:36.880 --> 42:40.800 - So with a Google, you're now seeing sort of Keras - -42:40.800 --> 42:42.840 - having to be integrated in TensorFlow, - -42:42.840 --> 42:46.800 - I'm sure has a ton of engineers working on. - -42:46.800 --> 42:51.040 - And there's, I'm sure a lot of tricky design decisions - -42:51.040 --> 42:52.200 - to be made. - -42:52.200 --> 42:54.440 - How does that process usually happen - -42:54.440 --> 42:56.800 - from at least your perspective? - -42:56.800 --> 42:59.800 - What are the debates like? - -43:00.720 --> 43:04.200 - Is there a lot of thinking, - -43:04.200 --> 43:06.880 - considering different options and so on? - -43:06.880 --> 43:08.160 - Yes. - -43:08.160 --> 43:12.640 - So a lot of the time I spend at Google - -43:12.640 --> 43:17.280 - is actually discussing design discussions, right? - -43:17.280 --> 43:20.480 - Writing design docs, participating in design review meetings - -43:20.480 --> 43:22.080 - and so on. - -43:22.080 --> 43:25.240 - This is as important as actually writing a code. - -43:25.240 --> 43:26.080 - Right. - -43:26.080 --> 43:28.120 - So there's a lot of thought, there's a lot of thought - -43:28.120 --> 43:32.280 - and a lot of care that is taken - -43:32.280 --> 43:34.160 - in coming up with these decisions - -43:34.160 --> 43:37.160 - and taking into account all of our users - -43:37.160 --> 43:40.680 - because TensorFlow has this extremely diverse user base, - -43:40.680 --> 43:41.520 - right? - -43:41.520 --> 43:43.120 - It's not like just one user segment - -43:43.120 --> 43:45.480 - where everyone has the same needs. - -43:45.480 --> 43:47.640 - We have small scale production users, - -43:47.640 --> 43:49.520 - large scale production users. - -43:49.520 --> 43:52.800 - We have startups, we have researchers, - -43:53.720 --> 43:55.080 - you know, it's all over the place. - -43:55.080 --> 43:57.560 - And we have to cater to all of their needs. - -43:57.560 --> 44:00.040 - If I just look at the standard debates - -44:00.040 --> 44:04.000 - of C++ or Python, there's some heated debates. - -44:04.000 --> 44:06.000 - Do you have those at Google? - -44:06.000 --> 44:08.080 - I mean, they're not heated in terms of emotionally, - -44:08.080 --> 44:10.800 - but there's probably multiple ways to do it, right? - -44:10.800 --> 44:14.040 - So how do you arrive through those design meetings - -44:14.040 --> 44:15.440 - at the best way to do it? - -44:15.440 --> 44:19.280 - Especially in deep learning where the field is evolving - -44:19.280 --> 44:20.880 - as you're doing it. - -44:21.880 --> 44:23.600 - Is there some magic to it? - -44:23.600 --> 44:26.240 - Is there some magic to the process? - -44:26.240 --> 44:28.280 - I don't know if there's magic to the process, - -44:28.280 --> 44:30.640 - but there definitely is a process. - -44:30.640 --> 44:33.760 - So making design decisions - -44:33.760 --> 44:36.080 - is about satisfying a set of constraints, - -44:36.080 --> 44:39.920 - but also trying to do so in the simplest way possible, - -44:39.920 --> 44:42.240 - because this is what can be maintained, - -44:42.240 --> 44:44.920 - this is what can be expanded in the future. - -44:44.920 --> 44:49.120 - So you don't want to naively satisfy the constraints - -44:49.120 --> 44:51.880 - by just, you know, for each capability you need available, - -44:51.880 --> 44:53.960 - you're gonna come up with one argument in your API - -44:53.960 --> 44:54.800 - and so on. - -44:54.800 --> 44:59.800 - You want to design APIs that are modular and hierarchical - -45:00.640 --> 45:04.080 - so that they have an API surface - -45:04.080 --> 45:07.040 - that is as small as possible, right? - -45:07.040 --> 45:11.640 - And you want this modular hierarchical architecture - -45:11.640 --> 45:14.560 - to reflect the way that domain experts - -45:14.560 --> 45:16.400 - think about the problem. - -45:16.400 --> 45:17.880 - Because as a domain expert, - -45:17.880 --> 45:19.840 - when you are reading about a new API, - -45:19.840 --> 45:24.760 - you're reading a tutorial or some docs pages, - -45:24.760 --> 45:28.200 - you already have a way that you're thinking about the problem. - -45:28.200 --> 45:32.320 - You already have like certain concepts in mind - -45:32.320 --> 45:35.680 - and you're thinking about how they relate together. - -45:35.680 --> 45:37.200 - And when you're reading docs, - -45:37.200 --> 45:40.280 - you're trying to build as quickly as possible - -45:40.280 --> 45:45.280 - a mapping between the concepts featured in your API - -45:45.280 --> 45:46.800 - and the concepts in your mind. - -45:46.800 --> 45:48.880 - So you're trying to map your mental model - -45:48.880 --> 45:53.600 - as a domain expert to the way things work in the API. - -45:53.600 --> 45:57.040 - So you need an API and an underlying implementation - -45:57.040 --> 46:00.120 - that are reflecting the way people think about these things. - -46:00.120 --> 46:02.880 - So in minimizing the time it takes to do the mapping. - -46:02.880 --> 46:04.680 - Yes, minimizing the time, - -46:04.680 --> 46:06.560 - the cognitive load there is - -46:06.560 --> 46:10.920 - in ingesting this new knowledge about your API. - -46:10.920 --> 46:13.160 - An API should not be self referential - -46:13.160 --> 46:15.520 - or referring to implementation details. - -46:15.520 --> 46:19.160 - It should only be referring to domain specific concepts - -46:19.160 --> 46:21.360 - that people already understand. - -46:23.240 --> 46:24.480 - Brilliant. - -46:24.480 --> 46:27.560 - So what's the future of Keras and TensorFlow look like? - -46:27.560 --> 46:29.640 - What does TensorFlow 3.0 look like? - -46:30.600 --> 46:33.720 - So that's kind of too far in the future for me to answer, - -46:33.720 --> 46:37.800 - especially since I'm not even the one making these decisions. - -46:37.800 --> 46:39.080 - Okay. - -46:39.080 --> 46:41.240 - But so from my perspective, - -46:41.240 --> 46:43.200 - which is just one perspective - -46:43.200 --> 46:46.040 - among many different perspectives on the TensorFlow team, - -46:47.200 --> 46:52.200 - I'm really excited by developing even higher level APIs, - -46:52.360 --> 46:53.560 - higher level than Keras. - -46:53.560 --> 46:56.480 - I'm really excited by hyperparameter tuning, - -46:56.480 --> 46:59.240 - by automated machine learning, AutoML. - -47:01.120 --> 47:03.200 - I think the future is not just, you know, - -47:03.200 --> 47:07.600 - defining a model like you were assembling Lego blocks - -47:07.600 --> 47:09.200 - and then collect fit on it. - -47:09.200 --> 47:13.680 - It's more like an automagical model - -47:13.680 --> 47:16.080 - that would just look at your data - -47:16.080 --> 47:19.040 - and optimize the objective you're after, right? - -47:19.040 --> 47:23.040 - So that's what I'm looking into. - -47:23.040 --> 47:26.480 - Yeah, so you put the baby into a room with the problem - -47:26.480 --> 47:28.760 - and come back a few hours later - -47:28.760 --> 47:30.960 - with a fully solved problem. - -47:30.960 --> 47:33.560 - Exactly, it's not like a box of Legos. - -47:33.560 --> 47:35.920 - It's more like the combination of a kid - -47:35.920 --> 47:38.800 - that's really good at Legos and a box of Legos. - -47:38.800 --> 47:41.520 - It's just building the thing on its own. - -47:41.520 --> 47:42.680 - Very nice. - -47:42.680 --> 47:44.160 - So that's an exciting future. - -47:44.160 --> 47:46.080 - I think there's a huge amount of applications - -47:46.080 --> 47:48.560 - and revolutions to be had - -47:49.920 --> 47:52.640 - under the constraints of the discussion we previously had. - -47:52.640 --> 47:57.480 - But what do you think of the current limits of deep learning? - -47:57.480 --> 48:02.480 - If we look specifically at these function approximators - -48:03.840 --> 48:06.160 - that tries to generalize from data. - -48:06.160 --> 48:10.160 - You've talked about local versus extreme generalization. - -48:11.120 --> 48:13.280 - You mentioned that neural networks don't generalize well - -48:13.280 --> 48:14.560 - and humans do. - -48:14.560 --> 48:15.760 - So there's this gap. - -48:17.640 --> 48:20.840 - And you've also mentioned that extreme generalization - -48:20.840 --> 48:23.960 - requires something like reasoning to fill those gaps. - -48:23.960 --> 48:27.560 - So how can we start trying to build systems like that? - -48:27.560 --> 48:30.600 - Right, yeah, so this is by design, right? - -48:30.600 --> 48:37.080 - Deep learning models are like huge parametric models, - -48:37.080 --> 48:39.280 - differentiable, so continuous, - -48:39.280 --> 48:42.680 - that go from an input space to an output space. - -48:42.680 --> 48:44.120 - And they're trained with gradient descent. - -48:44.120 --> 48:47.160 - So they're trained pretty much point by point. - -48:47.160 --> 48:50.520 - They are learning a continuous geometric morphing - -48:50.520 --> 48:55.320 - from an input vector space to an output vector space. - -48:55.320 --> 48:58.960 - And because this is done point by point, - -48:58.960 --> 49:02.200 - a deep neural network can only make sense - -49:02.200 --> 49:05.880 - of points in experience space that are very close - -49:05.880 --> 49:08.520 - to things that it has already seen in string data. - -49:08.520 --> 49:12.520 - At best, it can do interpolation across points. - -49:13.840 --> 49:17.360 - But that means in order to train your network, - -49:17.360 --> 49:21.680 - you need a dense sampling of the input cross output space, - -49:22.880 --> 49:25.240 - almost a point by point sampling, - -49:25.240 --> 49:27.160 - which can be very expensive if you're dealing - -49:27.160 --> 49:29.320 - with complex real world problems, - -49:29.320 --> 49:33.240 - like autonomous driving, for instance, or robotics. - -49:33.240 --> 49:36.000 - It's doable if you're looking at the subset - -49:36.000 --> 49:37.120 - of the visual space. - -49:37.120 --> 49:38.800 - But even then, it's still fairly expensive. - -49:38.800 --> 49:40.920 - You still need millions of examples. - -49:40.920 --> 49:44.240 - And it's only going to be able to make sense of things - -49:44.240 --> 49:46.880 - that are very close to what it has seen before. - -49:46.880 --> 49:49.160 - And in contrast to that, well, of course, - -49:49.160 --> 49:50.160 - you have human intelligence. - -49:50.160 --> 49:53.240 - But even if you're not looking at human intelligence, - -49:53.240 --> 49:56.800 - you can look at very simple rules, algorithms. - -49:56.800 --> 49:58.080 - If you have a symbolic rule, - -49:58.080 --> 50:03.080 - it can actually apply to a very, very large set of inputs - -50:03.120 --> 50:04.880 - because it is abstract. - -50:04.880 --> 50:09.560 - It is not obtained by doing a point by point mapping. - -50:10.720 --> 50:14.000 - For instance, if you try to learn a sorting algorithm - -50:14.000 --> 50:15.520 - using a deep neural network, - -50:15.520 --> 50:18.520 - well, you're very much limited to learning point by point - -50:20.080 --> 50:24.360 - what the sorted representation of this specific list is like. - -50:24.360 --> 50:29.360 - But instead, you could have a very, very simple - -50:29.400 --> 50:31.920 - sorting algorithm written in a few lines. - -50:31.920 --> 50:34.520 - Maybe it's just two nested loops. - -50:35.560 --> 50:40.560 - And it can process any list at all because it is abstract, - -50:41.040 --> 50:42.240 - because it is a set of rules. - -50:42.240 --> 50:45.160 - So deep learning is really like point by point - -50:45.160 --> 50:48.640 - geometric morphings, train with good and decent. - -50:48.640 --> 50:53.640 - And meanwhile, abstract rules can generalize much better. - -50:53.640 --> 50:56.160 - And I think the future is we need to combine the two. - -50:56.160 --> 50:59.160 - So how do we, do you think, combine the two? - -50:59.160 --> 51:03.040 - How do we combine good point by point functions - -51:03.040 --> 51:08.040 - with programs, which is what the symbolic AI type systems? - -51:08.920 --> 51:11.600 - At which levels the combination happen? - -51:11.600 --> 51:14.680 - I mean, obviously we're jumping into the realm - -51:14.680 --> 51:16.880 - of where there's no good answers. - -51:16.880 --> 51:20.280 - It's just kind of ideas and intuitions and so on. - -51:20.280 --> 51:23.080 - Well, if you look at the really successful AI systems - -51:23.080 --> 51:26.320 - today, I think they are already hybrid systems - -51:26.320 --> 51:29.520 - that are combining symbolic AI with deep learning. - -51:29.520 --> 51:32.520 - For instance, successful robotics systems - -51:32.520 --> 51:36.400 - are already mostly model based, rule based, - -51:37.400 --> 51:39.400 - things like planning algorithms and so on. - -51:39.400 --> 51:42.200 - At the same time, they're using deep learning - -51:42.200 --> 51:43.840 - as perception modules. - -51:43.840 --> 51:46.000 - Sometimes they're using deep learning as a way - -51:46.000 --> 51:50.920 - to inject fuzzy intuition into a rule based process. - -51:50.920 --> 51:54.560 - If you look at the system like in a self driving car, - -51:54.560 --> 51:57.240 - it's not just one big end to end neural network. - -51:57.240 --> 51:59.000 - You know, that wouldn't work at all. - -51:59.000 --> 52:00.760 - Precisely because in order to train that, - -52:00.760 --> 52:05.160 - you would need a dense sampling of experience base - -52:05.160 --> 52:06.200 - when it comes to driving, - -52:06.200 --> 52:08.880 - which is completely unrealistic, obviously. - -52:08.880 --> 52:12.440 - Instead, the self driving car is mostly - -52:13.920 --> 52:18.360 - symbolic, you know, it's software, it's programmed by hand. - -52:18.360 --> 52:21.640 - So it's mostly based on explicit models. - -52:21.640 --> 52:25.840 - In this case, mostly 3D models of the environment - -52:25.840 --> 52:29.520 - around the car, but it's interfacing with the real world - -52:29.520 --> 52:31.440 - using deep learning modules, right? - -52:31.440 --> 52:33.440 - So the deep learning there serves as a way - -52:33.440 --> 52:36.080 - to convert the raw sensory information - -52:36.080 --> 52:38.320 - to something usable by symbolic systems. - -52:39.760 --> 52:42.400 - Okay, well, let's linger on that a little more. - -52:42.400 --> 52:45.440 - So dense sampling from input to output. - -52:45.440 --> 52:48.240 - You said it's obviously very difficult. - -52:48.240 --> 52:50.120 - Is it possible? - -52:50.120 --> 52:51.800 - In the case of self driving, you mean? - -52:51.800 --> 52:53.040 - Let's say self driving, right? - -52:53.040 --> 52:55.760 - Self driving for many people, - -52:57.560 --> 52:59.520 - let's not even talk about self driving, - -52:59.520 --> 53:03.880 - let's talk about steering, so staying inside the lane. - -53:05.040 --> 53:07.080 - Lane following, yeah, it's definitely a problem - -53:07.080 --> 53:08.880 - you can solve with an end to end deep learning model, - -53:08.880 --> 53:10.600 - but that's like one small subset. - -53:10.600 --> 53:11.440 - Hold on a second. - -53:11.440 --> 53:12.760 - Yeah, I don't know why you're jumping - -53:12.760 --> 53:14.480 - from the extreme so easily, - -53:14.480 --> 53:16.280 - because I disagree with you on that. - -53:16.280 --> 53:21.000 - I think, well, it's not obvious to me - -53:21.000 --> 53:23.400 - that you can solve lane following. - -53:23.400 --> 53:25.840 - No, it's not obvious, I think it's doable. - -53:25.840 --> 53:30.840 - I think in general, there is no hard limitations - -53:31.200 --> 53:33.680 - to what you can learn with a deep neural network, - -53:33.680 --> 53:38.680 - as long as the search space is rich enough, - -53:40.320 --> 53:42.240 - is flexible enough, and as long as you have - -53:42.240 --> 53:45.360 - this dense sampling of the input cross output space. - -53:45.360 --> 53:47.720 - The problem is that this dense sampling - -53:47.720 --> 53:51.120 - could mean anything from 10,000 examples - -53:51.120 --> 53:52.840 - to like trillions and trillions. - -53:52.840 --> 53:54.360 - So that's my question. - -53:54.360 --> 53:56.200 - So what's your intuition? - -53:56.200 --> 53:58.720 - And if you could just give it a chance - -53:58.720 --> 54:01.880 - and think what kind of problems can be solved - -54:01.880 --> 54:04.240 - by getting a huge amounts of data - -54:04.240 --> 54:08.000 - and thereby creating a dense mapping. - -54:08.000 --> 54:12.480 - So let's think about natural language dialogue, - -54:12.480 --> 54:14.000 - the Turing test. - -54:14.000 --> 54:17.000 - Do you think the Turing test can be solved - -54:17.000 --> 54:21.120 - with a neural network alone? - -54:21.120 --> 54:24.440 - Well, the Turing test is all about tricking people - -54:24.440 --> 54:26.880 - into believing they're talking to a human. - -54:26.880 --> 54:29.040 - And I don't think that's actually very difficult - -54:29.040 --> 54:34.040 - because it's more about exploiting human perception - -54:35.600 --> 54:37.520 - and not so much about intelligence. - -54:37.520 --> 54:39.680 - There's a big difference between mimicking - -54:39.680 --> 54:42.080 - intelligent behavior and actual intelligent behavior. - -54:42.080 --> 54:45.360 - So, okay, let's look at maybe the Alexa prize and so on. - -54:45.360 --> 54:47.480 - The different formulations of the natural language - -54:47.480 --> 54:50.520 - conversation that are less about mimicking - -54:50.520 --> 54:52.800 - and more about maintaining a fun conversation - -54:52.800 --> 54:54.720 - that lasts for 20 minutes. - -54:54.720 --> 54:56.200 - That's a little less about mimicking - -54:56.200 --> 54:59.080 - and that's more about, I mean, it's still mimicking, - -54:59.080 --> 55:01.440 - but it's more about being able to carry forward - -55:01.440 --> 55:03.640 - a conversation with all the tangents that happen - -55:03.640 --> 55:05.080 - in dialogue and so on. - -55:05.080 --> 55:08.320 - Do you think that problem is learnable - -55:08.320 --> 55:13.320 - with a neural network that does the point to point mapping? - -55:14.520 --> 55:16.280 - So I think it would be very, very challenging - -55:16.280 --> 55:17.800 - to do this with deep learning. - -55:17.800 --> 55:21.480 - I don't think it's out of the question either. - -55:21.480 --> 55:23.240 - I wouldn't rule it out. - -55:23.240 --> 55:25.400 - The space of problems that can be solved - -55:25.400 --> 55:26.920 - with a large neural network. - -55:26.920 --> 55:30.080 - What's your sense about the space of those problems? - -55:30.080 --> 55:32.560 - So useful problems for us. - -55:32.560 --> 55:34.800 - In theory, it's infinite, right? - -55:34.800 --> 55:36.200 - You can solve any problem. - -55:36.200 --> 55:39.800 - In practice, well, deep learning is a great fit - -55:39.800 --> 55:41.800 - for perception problems. - -55:41.800 --> 55:46.800 - In general, any problem which is naturally amenable - -55:47.640 --> 55:52.200 - to explicit handcrafted rules or rules that you can generate - -55:52.200 --> 55:54.960 - by exhaustive search over some program space. - -55:56.080 --> 55:59.320 - So perception, artificial intuition, - -55:59.320 --> 56:03.240 - as long as you have a sufficient training dataset. - -56:03.240 --> 56:05.360 - And that's the question, I mean, perception, - -56:05.360 --> 56:08.400 - there's interpretation and understanding of the scene, - -56:08.400 --> 56:10.280 - which seems to be outside the reach - -56:10.280 --> 56:12.960 - of current perception systems. - -56:12.960 --> 56:15.920 - So do you think larger networks will be able - -56:15.920 --> 56:18.280 - to start to understand the physics - -56:18.280 --> 56:21.080 - and the physics of the scene, - -56:21.080 --> 56:23.400 - the three dimensional structure and relationships - -56:23.400 --> 56:25.560 - of objects in the scene and so on? - -56:25.560 --> 56:28.320 - Or really that's where symbolic AI has to step in? - -56:28.320 --> 56:34.480 - Well, it's always possible to solve these problems - -56:34.480 --> 56:36.800 - with deep learning. - -56:36.800 --> 56:38.560 - It's just extremely inefficient. - -56:38.560 --> 56:42.000 - A model would be an explicit rule based abstract model - -56:42.000 --> 56:45.240 - would be a far better, more compressed - -56:45.240 --> 56:46.840 - representation of physics. - -56:46.840 --> 56:49.080 - Then learning just this mapping between - -56:49.080 --> 56:50.960 - in this situation, this thing happens. - -56:50.960 --> 56:52.720 - If you change the situation slightly, - -56:52.720 --> 56:54.760 - then this other thing happens and so on. - -56:54.760 --> 56:57.440 - Do you think it's possible to automatically generate - -56:57.440 --> 57:02.200 - the programs that would require that kind of reasoning? - -57:02.200 --> 57:05.360 - Or does it have to, so the way the expert systems fail, - -57:05.360 --> 57:07.120 - there's so many facts about the world - -57:07.120 --> 57:08.960 - had to be hand coded in. - -57:08.960 --> 57:14.600 - Do you think it's possible to learn those logical statements - -57:14.600 --> 57:18.200 - that are true about the world and their relationships? - -57:18.200 --> 57:20.360 - Do you think, I mean, that's kind of what theorem proving - -57:20.360 --> 57:22.680 - at a basic level is trying to do, right? - -57:22.680 --> 57:26.160 - Yeah, except it's much harder to formulate statements - -57:26.160 --> 57:28.480 - about the world compared to formulating - -57:28.480 --> 57:30.320 - mathematical statements. - -57:30.320 --> 57:34.200 - Statements about the world tend to be subjective. - -57:34.200 --> 57:39.600 - So can you learn rule based models? - -57:39.600 --> 57:40.920 - Yes, definitely. - -57:40.920 --> 57:43.640 - That's the field of program synthesis. - -57:43.640 --> 57:48.040 - However, today we just don't really know how to do it. - -57:48.040 --> 57:52.400 - So it's very much a grass search or tree search problem. - -57:52.400 --> 57:56.800 - And so we are limited to the sort of tree session grass - -57:56.800 --> 57:58.560 - search algorithms that we have today. - -57:58.560 --> 58:02.760 - Personally, I think genetic algorithms are very promising. - -58:02.760 --> 58:04.360 - So almost like genetic programming. - -58:04.360 --> 58:05.560 - Genetic programming, exactly. - -58:05.560 --> 58:08.840 - Can you discuss the field of program synthesis? - -58:08.840 --> 58:14.560 - Like how many people are working and thinking about it? - -58:14.560 --> 58:17.960 - Where we are in the history of program synthesis - -58:17.960 --> 58:20.720 - and what are your hopes for it? - -58:20.720 --> 58:24.600 - Well, if it were deep learning, this is like the 90s. - -58:24.600 --> 58:29.120 - So meaning that we already have existing solutions. - -58:29.120 --> 58:34.280 - We are starting to have some basic understanding - -58:34.280 --> 58:35.480 - of what this is about. - -58:35.480 --> 58:38.000 - But it's still a field that is in its infancy. - -58:38.000 --> 58:40.440 - There are very few people working on it. - -58:40.440 --> 58:44.480 - There are very few real world applications. - -58:44.480 --> 58:47.640 - So the one real world application I'm aware of - -58:47.640 --> 58:51.680 - is Flash Fill in Excel. - -58:51.680 --> 58:55.080 - It's a way to automatically learn very simple programs - -58:55.080 --> 58:58.200 - to format cells in an Excel spreadsheet - -58:58.200 --> 59:00.240 - from a few examples. - -59:00.240 --> 59:02.800 - For instance, learning a way to format a date, things like that. - -59:02.800 --> 59:03.680 - Oh, that's fascinating. - -59:03.680 --> 59:04.560 - Yeah. - -59:04.560 --> 59:06.280 - You know, OK, that's a fascinating topic. - -59:06.280 --> 59:10.480 - I always wonder when I provide a few samples to Excel, - -59:10.480 --> 59:12.600 - what it's able to figure out. - -59:12.600 --> 59:15.960 - Like just giving it a few dates, what - -59:15.960 --> 59:18.480 - are you able to figure out from the pattern I just gave you? - -59:18.480 --> 59:19.760 - That's a fascinating question. - -59:19.760 --> 59:23.320 - And it's fascinating whether that's learnable patterns. - -59:23.320 --> 59:25.520 - And you're saying they're working on that. - -59:25.520 --> 59:28.200 - How big is the toolbox currently? - -59:28.200 --> 59:29.520 - Are we completely in the dark? - -59:29.520 --> 59:30.440 - So if you said the 90s. - -59:30.440 --> 59:31.720 - In terms of program synthesis? - -59:31.720 --> 59:32.360 - No. - -59:32.360 --> 59:37.720 - So I would say, so maybe 90s is even too optimistic. - -59:37.720 --> 59:41.080 - Because by the 90s, we already understood back prop. - -59:41.080 --> 59:43.960 - We already understood the engine of deep learning, - -59:43.960 --> 59:47.280 - even though we couldn't really see its potential quite. - -59:47.280 --> 59:48.520 - Today, I don't think we have found - -59:48.520 --> 59:50.400 - the engine of program synthesis. - -59:50.400 --> 59:52.880 - So we're in the winter before back prop. - -59:52.880 --> 59:54.160 - Yeah. - -59:54.160 --> 59:55.720 - In a way, yes. - -59:55.720 --> 1:00:00.120 - So I do believe program synthesis and general discrete search - -1:00:00.120 --> 1:00:02.760 - over rule based models is going to be - -1:00:02.760 --> 1:00:06.640 - a cornerstone of AI research in the next century. - -1:00:06.640 --> 1:00:10.200 - And that doesn't mean we are going to drop deep learning. - -1:00:10.200 --> 1:00:11.880 - Deep learning is immensely useful. - -1:00:11.880 --> 1:00:17.200 - Like, being able to learn is a very flexible, adaptable, - -1:00:17.200 --> 1:00:18.120 - parametric model. - -1:00:18.120 --> 1:00:20.720 - So it's got to understand that's actually immensely useful. - -1:00:20.720 --> 1:00:23.040 - All it's doing is pattern cognition. - -1:00:23.040 --> 1:00:25.640 - But being good at pattern cognition, given lots of data, - -1:00:25.640 --> 1:00:27.920 - is just extremely powerful. - -1:00:27.920 --> 1:00:30.320 - So we are still going to be working on deep learning. - -1:00:30.320 --> 1:00:31.840 - We are going to be working on program synthesis. - -1:00:31.840 --> 1:00:34.680 - We are going to be combining the two in increasingly automated - -1:00:34.680 --> 1:00:36.400 - ways. - -1:00:36.400 --> 1:00:38.520 - So let's talk a little bit about data. - -1:00:38.520 --> 1:00:44.600 - You've tweeted, about 10,000 deep learning papers - -1:00:44.600 --> 1:00:47.080 - have been written about hard coding priors - -1:00:47.080 --> 1:00:49.600 - about a specific task in a neural network architecture - -1:00:49.600 --> 1:00:52.440 - works better than a lack of a prior. - -1:00:52.440 --> 1:00:55.120 - Basically, summarizing all these efforts, - -1:00:55.120 --> 1:00:56.920 - they put a name to an architecture. - -1:00:56.920 --> 1:00:59.280 - But really, what they're doing is hard coding some priors - -1:00:59.280 --> 1:01:01.560 - that improve the performance of the system. - -1:01:01.560 --> 1:01:06.880 - But which gets straight to the point is probably true. - -1:01:06.880 --> 1:01:09.800 - So you say that you can always buy performance by, - -1:01:09.800 --> 1:01:12.920 - in quotes, performance by either training on more data, - -1:01:12.920 --> 1:01:15.480 - better data, or by injecting task information - -1:01:15.480 --> 1:01:18.400 - to the architecture of the preprocessing. - -1:01:18.400 --> 1:01:21.280 - However, this isn't informative about the generalization power - -1:01:21.280 --> 1:01:23.080 - the techniques use, the fundamental ability - -1:01:23.080 --> 1:01:24.200 - to generalize. - -1:01:24.200 --> 1:01:26.800 - Do you think we can go far by coming up - -1:01:26.800 --> 1:01:29.920 - with better methods for this kind of cheating, - -1:01:29.920 --> 1:01:33.520 - for better methods of large scale annotation of data? - -1:01:33.520 --> 1:01:34.960 - So building better priors. - -1:01:34.960 --> 1:01:37.280 - If you automate it, it's not cheating anymore. - -1:01:37.280 --> 1:01:38.360 - Right. - -1:01:38.360 --> 1:01:41.600 - I'm joking about the cheating, but large scale. - -1:01:41.600 --> 1:01:46.560 - So basically, I'm asking about something - -1:01:46.560 --> 1:01:48.280 - that hasn't, from my perspective, - -1:01:48.280 --> 1:01:53.360 - been researched too much is exponential improvement - -1:01:53.360 --> 1:01:55.960 - in annotation of data. - -1:01:55.960 --> 1:01:58.120 - Do you often think about? - -1:01:58.120 --> 1:02:00.840 - I think it's actually been researched quite a bit. - -1:02:00.840 --> 1:02:02.720 - You just don't see publications about it. - -1:02:02.720 --> 1:02:05.840 - Because people who publish papers - -1:02:05.840 --> 1:02:07.920 - are going to publish about known benchmarks. - -1:02:07.920 --> 1:02:09.800 - Sometimes they're going to read a new benchmark. - -1:02:09.800 --> 1:02:12.200 - People who actually have real world large scale - -1:02:12.200 --> 1:02:13.880 - depending on problems, they're going - -1:02:13.880 --> 1:02:16.960 - to spend a lot of resources into data annotation - -1:02:16.960 --> 1:02:18.400 - and good data annotation pipelines, - -1:02:18.400 --> 1:02:19.640 - but you don't see any papers about it. - -1:02:19.640 --> 1:02:20.400 - That's interesting. - -1:02:20.400 --> 1:02:22.720 - So do you think, certainly resources, - -1:02:22.720 --> 1:02:24.840 - but do you think there's innovation happening? - -1:02:24.840 --> 1:02:25.880 - Oh, yeah. - -1:02:25.880 --> 1:02:28.880 - To clarify the point in the tweet. - -1:02:28.880 --> 1:02:31.160 - So machine learning in general is - -1:02:31.160 --> 1:02:33.840 - the science of generalization. - -1:02:33.840 --> 1:02:37.800 - You want to generate knowledge that - -1:02:37.800 --> 1:02:40.440 - can be reused across different data sets, - -1:02:40.440 --> 1:02:42.000 - across different tasks. - -1:02:42.000 --> 1:02:45.280 - And if instead you're looking at one data set - -1:02:45.280 --> 1:02:50.000 - and then you are hard coding knowledge about this task - -1:02:50.000 --> 1:02:54.040 - into your architecture, this is no more useful - -1:02:54.040 --> 1:02:56.760 - than training a network and then saying, oh, I - -1:02:56.760 --> 1:03:01.920 - found these weight values perform well. - -1:03:01.920 --> 1:03:05.680 - So David Ha, I don't know if you know David, - -1:03:05.680 --> 1:03:08.760 - he had a paper the other day about weight - -1:03:08.760 --> 1:03:10.400 - agnostic neural networks. - -1:03:10.400 --> 1:03:12.120 - And this is a very interesting paper - -1:03:12.120 --> 1:03:14.400 - because it really illustrates the fact - -1:03:14.400 --> 1:03:17.400 - that an architecture, even without weights, - -1:03:17.400 --> 1:03:21.360 - an architecture is knowledge about a task. - -1:03:21.360 --> 1:03:23.640 - It encodes knowledge. - -1:03:23.640 --> 1:03:25.840 - And when it comes to architectures - -1:03:25.840 --> 1:03:30.440 - that are uncrafted by researchers, in some cases, - -1:03:30.440 --> 1:03:34.160 - it is very, very clear that all they are doing - -1:03:34.160 --> 1:03:38.880 - is artificially reencoding the template that - -1:03:38.880 --> 1:03:44.400 - corresponds to the proper way to solve the task encoding - -1:03:44.400 --> 1:03:45.200 - a given data set. - -1:03:45.200 --> 1:03:48.120 - For instance, I know if you looked - -1:03:48.120 --> 1:03:52.280 - at the baby data set, which is about natural language - -1:03:52.280 --> 1:03:55.520 - question answering, it is generated by an algorithm. - -1:03:55.520 --> 1:03:57.680 - So this is a question answer pairs - -1:03:57.680 --> 1:03:59.280 - that are generated by an algorithm. - -1:03:59.280 --> 1:04:01.520 - The algorithm is solving a certain template. - -1:04:01.520 --> 1:04:04.400 - Turns out, if you craft a network that - -1:04:04.400 --> 1:04:06.360 - literally encodes this template, you - -1:04:06.360 --> 1:04:09.640 - can solve this data set with nearly 100% accuracy. - -1:04:09.640 --> 1:04:11.160 - But that doesn't actually tell you - -1:04:11.160 --> 1:04:14.640 - anything about how to solve question answering - -1:04:14.640 --> 1:04:17.680 - in general, which is the point. - -1:04:17.680 --> 1:04:19.400 - The question is just to linger on it, - -1:04:19.400 --> 1:04:21.560 - whether it's from the data side or from the size - -1:04:21.560 --> 1:04:23.280 - of the network. - -1:04:23.280 --> 1:04:25.920 - I don't know if you've read the blog post by Rich Sutton, - -1:04:25.920 --> 1:04:28.400 - The Bitter Lesson, where he says, - -1:04:28.400 --> 1:04:31.480 - the biggest lesson that we can read from 70 years of AI - -1:04:31.480 --> 1:04:34.720 - research is that general methods that leverage computation - -1:04:34.720 --> 1:04:37.160 - are ultimately the most effective. - -1:04:37.160 --> 1:04:39.720 - So as opposed to figuring out methods - -1:04:39.720 --> 1:04:41.840 - that can generalize effectively, do you - -1:04:41.840 --> 1:04:47.720 - think we can get pretty far by just having something - -1:04:47.720 --> 1:04:51.520 - that leverages computation and the improvement of computation? - -1:04:51.520 --> 1:04:54.960 - Yeah, so I think Rich is making a very good point, which - -1:04:54.960 --> 1:04:57.560 - is that a lot of these papers, which are actually - -1:04:57.560 --> 1:05:02.800 - all about manually hardcoding prior knowledge about a task - -1:05:02.800 --> 1:05:04.720 - into some system, it doesn't have - -1:05:04.720 --> 1:05:08.600 - to be deep learning architecture, but into some system. - -1:05:08.600 --> 1:05:11.920 - These papers are not actually making any impact. - -1:05:11.920 --> 1:05:14.800 - Instead, what's making really long term impact - -1:05:14.800 --> 1:05:18.520 - is very simple, very general systems - -1:05:18.520 --> 1:05:21.280 - that are really agnostic to all these tricks. - -1:05:21.280 --> 1:05:23.320 - Because these tricks do not generalize. - -1:05:23.320 --> 1:05:27.480 - And of course, the one general and simple thing - -1:05:27.480 --> 1:05:33.160 - that you should focus on is that which leverages computation. - -1:05:33.160 --> 1:05:36.200 - Because computation, the availability - -1:05:36.200 --> 1:05:39.400 - of large scale computation has been increasing exponentially - -1:05:39.400 --> 1:05:40.560 - following Moore's law. - -1:05:40.560 --> 1:05:44.080 - So if your algorithm is all about exploiting this, - -1:05:44.080 --> 1:05:47.440 - then your algorithm is suddenly exponentially improving. - -1:05:47.440 --> 1:05:52.400 - So I think Rich is definitely right. - -1:05:52.400 --> 1:05:57.120 - However, he's right about the past 70 years. - -1:05:57.120 --> 1:05:59.440 - He's like assessing the past 70 years. - -1:05:59.440 --> 1:06:02.360 - I am not sure that this assessment will still - -1:06:02.360 --> 1:06:04.880 - hold true for the next 70 years. - -1:06:04.880 --> 1:06:07.160 - It might to some extent. - -1:06:07.160 --> 1:06:08.560 - I suspect it will not. - -1:06:08.560 --> 1:06:11.560 - Because the truth of his assessment - -1:06:11.560 --> 1:06:16.800 - is a function of the context in which this research took place. - -1:06:16.800 --> 1:06:18.600 - And the context is changing. - -1:06:18.600 --> 1:06:21.440 - Moore's law might not be applicable anymore, - -1:06:21.440 --> 1:06:23.760 - for instance, in the future. - -1:06:23.760 --> 1:06:31.200 - And I do believe that when you tweak one aspect of a system, - -1:06:31.200 --> 1:06:32.920 - when you exploit one aspect of a system, - -1:06:32.920 --> 1:06:36.480 - some other aspect starts becoming the bottleneck. - -1:06:36.480 --> 1:06:38.800 - Let's say you have unlimited computation. - -1:06:38.800 --> 1:06:41.440 - Well, then data is the bottleneck. - -1:06:41.440 --> 1:06:43.560 - And I think we are already starting - -1:06:43.560 --> 1:06:45.720 - to be in a regime where our systems are - -1:06:45.720 --> 1:06:48.120 - so large in scale and so data ingrained - -1:06:48.120 --> 1:06:50.360 - that data today and the quality of data - -1:06:50.360 --> 1:06:53.040 - and the scale of data is the bottleneck. - -1:06:53.040 --> 1:06:58.160 - And in this environment, the bitter lesson from Rich - -1:06:58.160 --> 1:07:00.800 - is not going to be true anymore. - -1:07:00.800 --> 1:07:03.960 - So I think we are going to move from a focus - -1:07:03.960 --> 1:07:09.840 - on a computation scale to focus on data efficiency. - -1:07:09.840 --> 1:07:10.720 - Data efficiency. - -1:07:10.720 --> 1:07:13.120 - So that's getting to the question of symbolic AI. - -1:07:13.120 --> 1:07:16.280 - But to linger on the deep learning approaches, - -1:07:16.280 --> 1:07:19.240 - do you have hope for either unsupervised learning - -1:07:19.240 --> 1:07:23.280 - or reinforcement learning, which are - -1:07:23.280 --> 1:07:28.120 - ways of being more data efficient in terms - -1:07:28.120 --> 1:07:31.560 - of the amount of data they need that required human annotation? - -1:07:31.560 --> 1:07:34.280 - So unsupervised learning and reinforcement learning - -1:07:34.280 --> 1:07:36.640 - are frameworks for learning, but they are not - -1:07:36.640 --> 1:07:39.000 - like any specific technique. - -1:07:39.000 --> 1:07:41.200 - So usually when people say reinforcement learning, - -1:07:41.200 --> 1:07:43.320 - what they really mean is deep reinforcement learning, - -1:07:43.320 --> 1:07:47.440 - which is like one approach which is actually very questionable. - -1:07:47.440 --> 1:07:50.920 - The question I was asking was unsupervised learning - -1:07:50.920 --> 1:07:54.680 - with deep neural networks and deep reinforcement learning. - -1:07:54.680 --> 1:07:56.840 - Well, these are not really data efficient - -1:07:56.840 --> 1:08:00.520 - because you're still leveraging these huge parametric models - -1:08:00.520 --> 1:08:03.720 - point by point with gradient descent. - -1:08:03.720 --> 1:08:08.000 - It is more efficient in terms of the number of annotations, - -1:08:08.000 --> 1:08:09.520 - the density of annotations you need. - -1:08:09.520 --> 1:08:13.840 - So the idea being to learn the latent space around which - -1:08:13.840 --> 1:08:17.960 - the data is organized and then map the sparse annotations - -1:08:17.960 --> 1:08:18.760 - into it. - -1:08:18.760 --> 1:08:23.560 - And sure, I mean, that's clearly a very good idea. - -1:08:23.560 --> 1:08:26.080 - It's not really a topic I would be working on, - -1:08:26.080 --> 1:08:28.040 - but it's clearly a good idea. - -1:08:28.040 --> 1:08:31.760 - So it would get us to solve some problems that? - -1:08:31.760 --> 1:08:34.880 - It will get us to incremental improvements - -1:08:34.880 --> 1:08:38.240 - in labeled data efficiency. - -1:08:38.240 --> 1:08:43.520 - Do you have concerns about short term or long term threats - -1:08:43.520 --> 1:08:47.800 - from AI, from artificial intelligence? - -1:08:47.800 --> 1:08:50.640 - Yes, definitely to some extent. - -1:08:50.640 --> 1:08:52.800 - And what's the shape of those concerns? - -1:08:52.800 --> 1:08:56.880 - This is actually something I've briefly written about. - -1:08:56.880 --> 1:09:02.680 - But the capabilities of deep learning technology - -1:09:02.680 --> 1:09:05.200 - can be used in many ways that are - -1:09:05.200 --> 1:09:09.760 - concerning from mass surveillance with things - -1:09:09.760 --> 1:09:11.880 - like facial recognition. - -1:09:11.880 --> 1:09:15.440 - In general, tracking lots of data about everyone - -1:09:15.440 --> 1:09:18.920 - and then being able to making sense of this data - -1:09:18.920 --> 1:09:22.240 - to do identification, to do prediction. - -1:09:22.240 --> 1:09:23.160 - That's concerning. - -1:09:23.160 --> 1:09:26.560 - That's something that's being very aggressively pursued - -1:09:26.560 --> 1:09:31.440 - by totalitarian states like China. - -1:09:31.440 --> 1:09:34.000 - One thing I am very much concerned about - -1:09:34.000 --> 1:09:40.640 - is that our lives are increasingly online, - -1:09:40.640 --> 1:09:43.280 - are increasingly digital, made of information, - -1:09:43.280 --> 1:09:48.080 - made of information consumption and information production, - -1:09:48.080 --> 1:09:51.800 - our digital footprint, I would say. - -1:09:51.800 --> 1:09:56.280 - And if you absorb all of this data - -1:09:56.280 --> 1:10:01.440 - and you are in control of where you consume information, - -1:10:01.440 --> 1:10:06.960 - social networks and so on, recommendation engines, - -1:10:06.960 --> 1:10:10.200 - then you can build a sort of reinforcement - -1:10:10.200 --> 1:10:13.760 - loop for human behavior. - -1:10:13.760 --> 1:10:18.360 - You can observe the state of your mind at time t. - -1:10:18.360 --> 1:10:21.080 - You can predict how you would react - -1:10:21.080 --> 1:10:23.800 - to different pieces of content, how - -1:10:23.800 --> 1:10:27.000 - to get you to move your mind in a certain direction. - -1:10:27.000 --> 1:10:33.160 - And then you can feed you the specific piece of content - -1:10:33.160 --> 1:10:35.680 - that would move you in a specific direction. - -1:10:35.680 --> 1:10:41.800 - And you can do this at scale in terms - -1:10:41.800 --> 1:10:44.960 - of doing it continuously in real time. - -1:10:44.960 --> 1:10:46.440 - You can also do it at scale in terms - -1:10:46.440 --> 1:10:50.480 - of scaling this to many, many people, to entire populations. - -1:10:50.480 --> 1:10:53.840 - So potentially, artificial intelligence, - -1:10:53.840 --> 1:10:57.440 - even in its current state, if you combine it - -1:10:57.440 --> 1:11:01.760 - with the internet, with the fact that all of our lives - -1:11:01.760 --> 1:11:05.120 - are moving to digital devices and digital information - -1:11:05.120 --> 1:11:08.720 - consumption and creation, what you get - -1:11:08.720 --> 1:11:14.480 - is the possibility to achieve mass manipulation of behavior - -1:11:14.480 --> 1:11:16.840 - and mass psychological control. - -1:11:16.840 --> 1:11:18.520 - And this is a very real possibility. - -1:11:18.520 --> 1:11:22.080 - Yeah, so you're talking about any kind of recommender system. - -1:11:22.080 --> 1:11:26.160 - Let's look at the YouTube algorithm, Facebook, - -1:11:26.160 --> 1:11:29.720 - anything that recommends content you should watch next. - -1:11:29.720 --> 1:11:32.960 - And it's fascinating to think that there's - -1:11:32.960 --> 1:11:41.120 - some aspects of human behavior that you can say a problem of, - -1:11:41.120 --> 1:11:45.400 - is this person hold Republican beliefs or Democratic beliefs? - -1:11:45.400 --> 1:11:50.240 - And this is a trivial, that's an objective function. - -1:11:50.240 --> 1:11:52.600 - And you can optimize, and you can measure, - -1:11:52.600 --> 1:11:54.360 - and you can turn everybody into a Republican - -1:11:54.360 --> 1:11:56.080 - or everybody into a Democrat. - -1:11:56.080 --> 1:11:57.840 - I do believe it's true. - -1:11:57.840 --> 1:12:03.680 - So the human mind is very, if you look at the human mind - -1:12:03.680 --> 1:12:05.320 - as a kind of computer program, it - -1:12:05.320 --> 1:12:07.560 - has a very large exploit surface. - -1:12:07.560 --> 1:12:09.360 - It has many, many vulnerabilities. - -1:12:09.360 --> 1:12:10.840 - Exploit surfaces, yeah. - -1:12:10.840 --> 1:12:13.520 - Ways you can control it. - -1:12:13.520 --> 1:12:16.680 - For instance, when it comes to your political beliefs, - -1:12:16.680 --> 1:12:19.400 - this is very much tied to your identity. - -1:12:19.400 --> 1:12:23.040 - So for instance, if I'm in control of your news feed - -1:12:23.040 --> 1:12:26.000 - on your favorite social media platforms, - -1:12:26.000 --> 1:12:29.360 - this is actually where you're getting your news from. - -1:12:29.360 --> 1:12:32.960 - And of course, I can choose to only show you - -1:12:32.960 --> 1:12:37.120 - news that will make you see the world in a specific way. - -1:12:37.120 --> 1:12:41.920 - But I can also create incentives for you - -1:12:41.920 --> 1:12:44.720 - to post about some political beliefs. - -1:12:44.720 --> 1:12:47.960 - And then when I get you to express a statement, - -1:12:47.960 --> 1:12:51.840 - if it's a statement that me as the controller, - -1:12:51.840 --> 1:12:53.800 - I want to reinforce. - -1:12:53.800 --> 1:12:55.560 - I can just show it to people who will agree, - -1:12:55.560 --> 1:12:56.880 - and they will like it. - -1:12:56.880 --> 1:12:59.280 - And that will reinforce the statement in your mind. - -1:12:59.280 --> 1:13:02.760 - If this is a statement I want you to, - -1:13:02.760 --> 1:13:05.320 - this is a belief I want you to abandon, - -1:13:05.320 --> 1:13:09.600 - I can, on the other hand, show it to opponents. - -1:13:09.600 --> 1:13:10.640 - We'll attack you. - -1:13:10.640 --> 1:13:12.840 - And because they attack you, at the very least, - -1:13:12.840 --> 1:13:16.840 - next time you will think twice about posting it. - -1:13:16.840 --> 1:13:20.280 - But maybe you will even start believing this - -1:13:20.280 --> 1:13:22.840 - because you got pushback. - -1:13:22.840 --> 1:13:28.440 - So there are many ways in which social media platforms - -1:13:28.440 --> 1:13:30.520 - can potentially control your opinions. - -1:13:30.520 --> 1:13:35.040 - And today, so all of these things - -1:13:35.040 --> 1:13:38.240 - are already being controlled by AI algorithms. - -1:13:38.240 --> 1:13:41.880 - These algorithms do not have any explicit political goal - -1:13:41.880 --> 1:13:42.880 - today. - -1:13:42.880 --> 1:13:48.680 - Well, potentially they could, like if some totalitarian - -1:13:48.680 --> 1:13:52.720 - government takes over social media platforms - -1:13:52.720 --> 1:13:55.360 - and decides that now we are going to use this not just - -1:13:55.360 --> 1:13:58.040 - for mass surveillance, but also for mass opinion control - -1:13:58.040 --> 1:13:59.360 - and behavior control. - -1:13:59.360 --> 1:14:01.840 - Very bad things could happen. - -1:14:01.840 --> 1:14:06.480 - But what's really fascinating and actually quite concerning - -1:14:06.480 --> 1:14:11.280 - is that even without an explicit intent to manipulate, - -1:14:11.280 --> 1:14:14.760 - you're already seeing very dangerous dynamics - -1:14:14.760 --> 1:14:18.160 - in terms of how these content recommendation - -1:14:18.160 --> 1:14:19.800 - algorithms behave. - -1:14:19.800 --> 1:14:24.920 - Because right now, the goal, the objective function - -1:14:24.920 --> 1:14:28.640 - of these algorithms is to maximize engagement, - -1:14:28.640 --> 1:14:32.520 - which seems fairly innocuous at first. - -1:14:32.520 --> 1:14:36.480 - However, it is not because content - -1:14:36.480 --> 1:14:42.000 - that will maximally engage people, get people to react - -1:14:42.000 --> 1:14:44.720 - in an emotional way, get people to click on something. - -1:14:44.720 --> 1:14:52.200 - It is very often content that is not - -1:14:52.200 --> 1:14:54.400 - healthy to the public discourse. - -1:14:54.400 --> 1:14:58.200 - For instance, fake news are far more - -1:14:58.200 --> 1:15:01.320 - likely to get you to click on them than real news - -1:15:01.320 --> 1:15:06.960 - simply because they are not constrained to reality. - -1:15:06.960 --> 1:15:11.360 - So they can be as outrageous, as surprising, - -1:15:11.360 --> 1:15:15.880 - as good stories as you want because they're artificial. - -1:15:15.880 --> 1:15:18.880 - To me, that's an exciting world because so much good - -1:15:18.880 --> 1:15:19.560 - can come. - -1:15:19.560 --> 1:15:24.520 - So there's an opportunity to educate people. - -1:15:24.520 --> 1:15:31.200 - You can balance people's worldview with other ideas. - -1:15:31.200 --> 1:15:33.800 - So there's so many objective functions. - -1:15:33.800 --> 1:15:35.840 - The space of objective functions that - -1:15:35.840 --> 1:15:40.720 - create better civilizations is large, arguably infinite. - -1:15:40.720 --> 1:15:43.720 - But there's also a large space that - -1:15:43.720 --> 1:15:51.480 - creates division and destruction, civil war, - -1:15:51.480 --> 1:15:53.160 - a lot of bad stuff. - -1:15:53.160 --> 1:15:56.920 - And the worry is, naturally, probably that space - -1:15:56.920 --> 1:15:59.160 - is bigger, first of all. - -1:15:59.160 --> 1:16:04.920 - And if we don't explicitly think about what kind of effects - -1:16:04.920 --> 1:16:08.320 - are going to be observed from different objective functions, - -1:16:08.320 --> 1:16:10.160 - then we're going to get into trouble. - -1:16:10.160 --> 1:16:14.480 - But the question is, how do we get into rooms - -1:16:14.480 --> 1:16:18.560 - and have discussions, so inside Google, inside Facebook, - -1:16:18.560 --> 1:16:21.840 - inside Twitter, and think about, OK, - -1:16:21.840 --> 1:16:24.840 - how can we drive up engagement and, at the same time, - -1:16:24.840 --> 1:16:28.200 - create a good society? - -1:16:28.200 --> 1:16:29.560 - Is it even possible to have that kind - -1:16:29.560 --> 1:16:31.720 - of philosophical discussion? - -1:16:31.720 --> 1:16:33.080 - I think you can definitely try. - -1:16:33.080 --> 1:16:37.280 - So from my perspective, I would feel rather uncomfortable - -1:16:37.280 --> 1:16:41.560 - with companies that are uncomfortable with these new - -1:16:41.560 --> 1:16:47.120 - student algorithms, with them making explicit decisions - -1:16:47.120 --> 1:16:50.440 - to manipulate people's opinions or behaviors, - -1:16:50.440 --> 1:16:53.480 - even if the intent is good, because that's - -1:16:53.480 --> 1:16:55.200 - a very totalitarian mindset. - -1:16:55.200 --> 1:16:57.440 - So instead, what I would like to see - -1:16:57.440 --> 1:16:58.880 - is probably never going to happen, - -1:16:58.880 --> 1:17:00.360 - because it's not super realistic, - -1:17:00.360 --> 1:17:02.520 - but that's actually something I really care about. - -1:17:02.520 --> 1:17:06.280 - I would like all these algorithms - -1:17:06.280 --> 1:17:10.560 - to present configuration settings to their users, - -1:17:10.560 --> 1:17:14.600 - so that the users can actually make the decision about how - -1:17:14.600 --> 1:17:19.000 - they want to be impacted by these information - -1:17:19.000 --> 1:17:21.960 - recommendation, content recommendation algorithms. - -1:17:21.960 --> 1:17:24.240 - For instance, as a user of something - -1:17:24.240 --> 1:17:26.520 - like YouTube or Twitter, maybe I want - -1:17:26.520 --> 1:17:30.280 - to maximize learning about a specific topic. - -1:17:30.280 --> 1:17:36.800 - So I want the algorithm to feed my curiosity, - -1:17:36.800 --> 1:17:38.760 - which is in itself a very interesting problem. - -1:17:38.760 --> 1:17:41.200 - So instead of maximizing my engagement, - -1:17:41.200 --> 1:17:44.600 - it will maximize how fast and how much I'm learning. - -1:17:44.600 --> 1:17:47.360 - And it will also take into account the accuracy, - -1:17:47.360 --> 1:17:50.680 - hopefully, of the information I'm learning. - -1:17:50.680 --> 1:17:55.680 - So yeah, the user should be able to determine exactly - -1:17:55.680 --> 1:17:58.560 - how these algorithms are affecting their lives. - -1:17:58.560 --> 1:18:03.520 - I don't want actually any entity making decisions - -1:18:03.520 --> 1:18:09.480 - about in which direction they're going to try to manipulate me. - -1:18:09.480 --> 1:18:11.680 - I want technology. - -1:18:11.680 --> 1:18:14.280 - So AI, these algorithms are increasingly - -1:18:14.280 --> 1:18:18.560 - going to be our interface to a world that is increasingly - -1:18:18.560 --> 1:18:19.960 - made of information. - -1:18:19.960 --> 1:18:25.840 - And I want everyone to be in control of this interface, - -1:18:25.840 --> 1:18:29.160 - to interface with the world on their own terms. - -1:18:29.160 --> 1:18:32.840 - So if someone wants these algorithms - -1:18:32.840 --> 1:18:37.640 - to serve their own personal growth goals, - -1:18:37.640 --> 1:18:40.640 - they should be able to configure these algorithms - -1:18:40.640 --> 1:18:41.800 - in such a way. - -1:18:41.800 --> 1:18:46.680 - Yeah, but so I know it's painful to have explicit decisions. - -1:18:46.680 --> 1:18:51.080 - But there is underlying explicit decisions, - -1:18:51.080 --> 1:18:53.360 - which is some of the most beautiful fundamental - -1:18:53.360 --> 1:18:57.400 - philosophy that we have before us, - -1:18:57.400 --> 1:19:01.120 - which is personal growth. - -1:19:01.120 --> 1:19:05.680 - If I want to watch videos from which I can learn, - -1:19:05.680 --> 1:19:08.080 - what does that mean? - -1:19:08.080 --> 1:19:11.800 - So if I have a checkbox that wants to emphasize learning, - -1:19:11.800 --> 1:19:15.480 - there's still an algorithm with explicit decisions in it - -1:19:15.480 --> 1:19:17.800 - that would promote learning. - -1:19:17.800 --> 1:19:19.200 - What does that mean for me? - -1:19:19.200 --> 1:19:22.800 - For example, I've watched a documentary on flat Earth - -1:19:22.800 --> 1:19:23.640 - theory, I guess. - -1:19:27.280 --> 1:19:28.240 - I learned a lot. - -1:19:28.240 --> 1:19:29.800 - I'm really glad I watched it. - -1:19:29.800 --> 1:19:32.560 - It was a friend recommended it to me. - -1:19:32.560 --> 1:19:35.800 - Because I don't have such an allergic reaction to crazy - -1:19:35.800 --> 1:19:37.640 - people, as my fellow colleagues do. - -1:19:37.640 --> 1:19:40.360 - But it was very eye opening. - -1:19:40.360 --> 1:19:42.120 - And for others, it might not be. - -1:19:42.120 --> 1:19:45.560 - From others, they might just get turned off from that, same - -1:19:45.560 --> 1:19:47.160 - with Republican and Democrat. - -1:19:47.160 --> 1:19:50.200 - And it's a non trivial problem. - -1:19:50.200 --> 1:19:52.880 - And first of all, if it's done well, - -1:19:52.880 --> 1:19:56.560 - I don't think it's something that wouldn't happen, - -1:19:56.560 --> 1:19:59.280 - that YouTube wouldn't be promoting, - -1:19:59.280 --> 1:20:00.200 - or Twitter wouldn't be. - -1:20:00.200 --> 1:20:02.280 - It's just a really difficult problem, - -1:20:02.280 --> 1:20:05.520 - how to give people control. - -1:20:05.520 --> 1:20:08.960 - Well, it's mostly an interface design problem. - -1:20:08.960 --> 1:20:11.080 - The way I see it, you want to create technology - -1:20:11.080 --> 1:20:16.400 - that's like a mentor, or a coach, or an assistant, - -1:20:16.400 --> 1:20:20.520 - so that it's not your boss. - -1:20:20.520 --> 1:20:22.560 - You are in control of it. - -1:20:22.560 --> 1:20:25.760 - You are telling it what to do for you. - -1:20:25.760 --> 1:20:27.840 - And if you feel like it's manipulating you, - -1:20:27.840 --> 1:20:31.760 - it's not actually doing what you want. - -1:20:31.760 --> 1:20:34.920 - You should be able to switch to a different algorithm. - -1:20:34.920 --> 1:20:36.440 - So that's fine tune control. - -1:20:36.440 --> 1:20:38.840 - You kind of learn that you're trusting - -1:20:38.840 --> 1:20:40.080 - the human collaboration. - -1:20:40.080 --> 1:20:41.920 - I mean, that's how I see autonomous vehicles too, - -1:20:41.920 --> 1:20:44.480 - is giving as much information as possible, - -1:20:44.480 --> 1:20:47.240 - and you learn that dance yourself. - -1:20:47.240 --> 1:20:50.280 - Yeah, Adobe, I don't know if you use Adobe product - -1:20:50.280 --> 1:20:52.280 - for like Photoshop. - -1:20:52.280 --> 1:20:55.040 - They're trying to see if they can inject YouTube - -1:20:55.040 --> 1:20:57.120 - into their interface, but basically allow you - -1:20:57.120 --> 1:20:59.840 - to show you all these videos, - -1:20:59.840 --> 1:21:03.320 - that everybody's confused about what to do with features. - -1:21:03.320 --> 1:21:07.120 - So basically teach people by linking to, - -1:21:07.120 --> 1:21:10.280 - in that way, it's an assistant that uses videos - -1:21:10.280 --> 1:21:13.440 - as a basic element of information. - -1:21:13.440 --> 1:21:18.240 - Okay, so what practically should people do - -1:21:18.240 --> 1:21:24.000 - to try to fight against abuses of these algorithms, - -1:21:24.000 --> 1:21:27.400 - or algorithms that manipulate us? - -1:21:27.400 --> 1:21:29.280 - Honestly, it's a very, very difficult problem, - -1:21:29.280 --> 1:21:32.800 - because to start with, there is very little public awareness - -1:21:32.800 --> 1:21:35.040 - of these issues. - -1:21:35.040 --> 1:21:38.520 - Very few people would think there's anything wrong - -1:21:38.520 --> 1:21:39.720 - with the unused algorithm, - -1:21:39.720 --> 1:21:42.040 - even though there is actually something wrong already, - -1:21:42.040 --> 1:21:44.480 - which is that it's trying to maximize engagement - -1:21:44.480 --> 1:21:49.880 - most of the time, which has very negative side effects. - -1:21:49.880 --> 1:21:56.160 - So ideally, so the very first thing is to stop - -1:21:56.160 --> 1:21:59.560 - trying to purely maximize engagement, - -1:21:59.560 --> 1:22:06.560 - try to propagate content based on popularity, right? - -1:22:06.560 --> 1:22:11.040 - Instead, take into account the goals - -1:22:11.040 --> 1:22:13.560 - and the profiles of each user. - -1:22:13.560 --> 1:22:16.920 - So you will be, one example is, for instance, - -1:22:16.920 --> 1:22:20.800 - when I look at topic recommendations on Twitter, - -1:22:20.800 --> 1:22:24.480 - it's like, you know, they have this news tab - -1:22:24.480 --> 1:22:25.480 - with switch recommendations. - -1:22:25.480 --> 1:22:28.480 - It's always the worst coverage, - -1:22:28.480 --> 1:22:30.360 - because it's content that appeals - -1:22:30.360 --> 1:22:34.080 - to the smallest common denominator - -1:22:34.080 --> 1:22:37.080 - to all Twitter users, because they're trying to optimize. - -1:22:37.080 --> 1:22:39.040 - They're purely trying to optimize popularity. - -1:22:39.040 --> 1:22:41.320 - They're purely trying to optimize engagement. - -1:22:41.320 --> 1:22:42.960 - But that's not what I want. - -1:22:42.960 --> 1:22:46.080 - So they should put me in control of some setting - -1:22:46.080 --> 1:22:50.360 - so that I define what's the objective function - -1:22:50.360 --> 1:22:52.200 - that Twitter is going to be following - -1:22:52.200 --> 1:22:54.120 - to show me this content. - -1:22:54.120 --> 1:22:57.360 - And honestly, so this is all about interface design. - -1:22:57.360 --> 1:22:59.440 - And we are not, it's not realistic - -1:22:59.440 --> 1:23:01.760 - to give users control of a bunch of knobs - -1:23:01.760 --> 1:23:03.400 - that define algorithm. - -1:23:03.400 --> 1:23:06.760 - Instead, we should purely put them in charge - -1:23:06.760 --> 1:23:09.400 - of defining the objective function. - -1:23:09.400 --> 1:23:13.240 - Like, let the user tell us what they want to achieve, - -1:23:13.240 --> 1:23:15.280 - how they want this algorithm to impact their lives. - -1:23:15.280 --> 1:23:16.680 - So do you think it is that, - -1:23:16.680 --> 1:23:19.360 - or do they provide individual article by article - -1:23:19.360 --> 1:23:21.600 - reward structure where you give a signal, - -1:23:21.600 --> 1:23:24.720 - I'm glad I saw this, or I'm glad I didn't? - -1:23:24.720 --> 1:23:28.480 - So like a Spotify type feedback mechanism, - -1:23:28.480 --> 1:23:30.680 - it works to some extent. - -1:23:30.680 --> 1:23:32.000 - I'm kind of skeptical about it - -1:23:32.000 --> 1:23:34.880 - because the only way the algorithm, - -1:23:34.880 --> 1:23:39.120 - the algorithm will attempt to relate your choices - -1:23:39.120 --> 1:23:41.040 - with the choices of everyone else, - -1:23:41.040 --> 1:23:45.000 - which might, you know, if you have an average profile - -1:23:45.000 --> 1:23:47.880 - that works fine, I'm sure Spotify accommodations work fine - -1:23:47.880 --> 1:23:49.560 - if you just like mainstream stuff. - -1:23:49.560 --> 1:23:53.960 - If you don't, it can be, it's not optimal at all actually. - -1:23:53.960 --> 1:23:56.040 - It'll be in an efficient search - -1:23:56.040 --> 1:24:00.800 - for the part of the Spotify world that represents you. - -1:24:00.800 --> 1:24:02.960 - So it's a tough problem, - -1:24:02.960 --> 1:24:07.960 - but do note that even a feedback system - -1:24:07.960 --> 1:24:10.880 - like what Spotify has does not give me control - -1:24:10.880 --> 1:24:15.000 - over what the algorithm is trying to optimize for. - -1:24:16.320 --> 1:24:19.360 - Well, public awareness, which is what we're doing now, - -1:24:19.360 --> 1:24:21.360 - is a good place to start. - -1:24:21.360 --> 1:24:25.960 - Do you have concerns about longterm existential threats - -1:24:25.960 --> 1:24:27.360 - of artificial intelligence? - -1:24:28.280 --> 1:24:31.040 - Well, as I was saying, - -1:24:31.040 --> 1:24:33.360 - our world is increasingly made of information. - -1:24:33.360 --> 1:24:36.240 - AI algorithms are increasingly going to be our interface - -1:24:36.240 --> 1:24:37.880 - to this world of information, - -1:24:37.880 --> 1:24:41.480 - and somebody will be in control of these algorithms. - -1:24:41.480 --> 1:24:45.920 - And that puts us in any kind of a bad situation, right? - -1:24:45.920 --> 1:24:46.880 - It has risks. - -1:24:46.880 --> 1:24:50.840 - It has risks coming from potentially large companies - -1:24:50.840 --> 1:24:53.760 - wanting to optimize their own goals, - -1:24:53.760 --> 1:24:55.960 - maybe profit, maybe something else. - -1:24:55.960 --> 1:25:00.720 - Also from governments who might want to use these algorithms - -1:25:00.720 --> 1:25:03.520 - as a means of control of the population. - -1:25:03.520 --> 1:25:05.000 - Do you think there's existential threat - -1:25:05.000 --> 1:25:06.320 - that could arise from that? - -1:25:06.320 --> 1:25:09.120 - So existential threat. - -1:25:09.120 --> 1:25:13.240 - So maybe you're referring to the singularity narrative - -1:25:13.240 --> 1:25:15.560 - where robots just take over. - -1:25:15.560 --> 1:25:18.320 - Well, I don't, I'm not terminating robots, - -1:25:18.320 --> 1:25:21.000 - and I don't believe it has to be a singularity. - -1:25:21.000 --> 1:25:24.800 - We're just talking to, just like you said, - -1:25:24.800 --> 1:25:27.920 - the algorithm controlling masses of populations. - -1:25:28.920 --> 1:25:31.120 - The existential threat being, - -1:25:32.640 --> 1:25:36.760 - hurt ourselves much like a nuclear war would hurt ourselves. - -1:25:36.760 --> 1:25:37.600 - That kind of thing. - -1:25:37.600 --> 1:25:39.480 - I don't think that requires a singularity. - -1:25:39.480 --> 1:25:42.560 - That requires a loss of control over AI algorithm. - -1:25:42.560 --> 1:25:43.560 - Yes. - -1:25:43.560 --> 1:25:47.000 - So I do agree there are concerning trends. - -1:25:47.000 --> 1:25:52.000 - Honestly, I wouldn't want to make any longterm predictions. - -1:25:52.960 --> 1:25:56.000 - I don't think today we really have the capability - -1:25:56.000 --> 1:25:58.560 - to see what the dangers of AI - -1:25:58.560 --> 1:26:01.360 - are going to be in 50 years, in 100 years. - -1:26:01.360 --> 1:26:04.800 - I do see that we are already faced - -1:26:04.800 --> 1:26:08.840 - with concrete and present dangers - -1:26:08.840 --> 1:26:11.560 - surrounding the negative side effects - -1:26:11.560 --> 1:26:14.960 - of content recombination systems, of newsfeed algorithms - -1:26:14.960 --> 1:26:17.640 - concerning algorithmic bias as well. - -1:26:18.640 --> 1:26:21.200 - So we are delegating more and more - -1:26:22.240 --> 1:26:25.080 - decision processes to algorithms. - -1:26:25.080 --> 1:26:26.760 - Some of these algorithms are uncrafted, - -1:26:26.760 --> 1:26:29.360 - some are learned from data, - -1:26:29.360 --> 1:26:31.920 - but we are delegating control. - -1:26:32.920 --> 1:26:36.280 - Sometimes it's a good thing, sometimes not so much. - -1:26:36.280 --> 1:26:39.480 - And there is in general very little supervision - -1:26:39.480 --> 1:26:41.000 - of this process, right? - -1:26:41.000 --> 1:26:45.400 - So we are still in this period of very fast change, - -1:26:45.400 --> 1:26:50.400 - even chaos, where society is restructuring itself, - -1:26:50.920 --> 1:26:53.160 - turning into an information society, - -1:26:53.160 --> 1:26:54.520 - which itself is turning into - -1:26:54.520 --> 1:26:58.360 - an increasingly automated information passing society. - -1:26:58.360 --> 1:27:02.520 - And well, yeah, I think the best we can do today - -1:27:02.520 --> 1:27:06.040 - is try to raise awareness around some of these issues. - -1:27:06.040 --> 1:27:07.680 - And I think we're actually making good progress. - -1:27:07.680 --> 1:27:11.720 - If you look at algorithmic bias, for instance, - -1:27:12.760 --> 1:27:14.760 - three years ago, even two years ago, - -1:27:14.760 --> 1:27:17.040 - very, very few people were talking about it. - -1:27:17.040 --> 1:27:20.320 - And now all the big companies are talking about it. - -1:27:20.320 --> 1:27:22.360 - They are often not in a very serious way, - -1:27:22.360 --> 1:27:24.560 - but at least it is part of the public discourse. - -1:27:24.560 --> 1:27:27.080 - You see people in Congress talking about it. - -1:27:27.080 --> 1:27:31.960 - And it all started from raising awareness. - -1:27:31.960 --> 1:27:32.800 - Right. - -1:27:32.800 --> 1:27:36.080 - So in terms of alignment problem, - -1:27:36.080 --> 1:27:39.400 - trying to teach as we allow algorithms, - -1:27:39.400 --> 1:27:41.520 - just even recommender systems on Twitter, - -1:27:43.640 --> 1:27:47.080 - encoding human values and morals, - -1:27:48.280 --> 1:27:50.200 - decisions that touch on ethics, - -1:27:50.200 --> 1:27:52.600 - how hard do you think that problem is? - -1:27:52.600 --> 1:27:57.240 - How do we have lost functions in neural networks - -1:27:57.240 --> 1:27:58.640 - that have some component, - -1:27:58.640 --> 1:28:01.080 - some fuzzy components of human morals? - -1:28:01.080 --> 1:28:06.080 - Well, I think this is really all about objective function engineering, - -1:28:06.080 --> 1:28:10.520 - which is probably going to be increasingly a topic of concern in the future. - -1:28:10.520 --> 1:28:14.640 - Like for now, we're just using very naive loss functions - -1:28:14.640 --> 1:28:17.760 - because the hard part is not actually what you're trying to minimize. - -1:28:17.760 --> 1:28:19.040 - It's everything else. - -1:28:19.040 --> 1:28:22.840 - But as the everything else is going to be increasingly automated, - -1:28:22.840 --> 1:28:27.040 - we're going to be focusing our human attention - -1:28:27.040 --> 1:28:30.240 - on increasingly high level components, - -1:28:30.240 --> 1:28:32.680 - like what's actually driving the whole learning system, - -1:28:32.680 --> 1:28:33.960 - like the objective function. - -1:28:33.960 --> 1:28:36.920 - So loss function engineering is going to be, - -1:28:36.920 --> 1:28:40.640 - loss function engineer is probably going to be a job title in the future. - -1:28:40.640 --> 1:28:44.520 - And then the tooling you're creating with Keras essentially - -1:28:44.520 --> 1:28:47.040 - takes care of all the details underneath. - -1:28:47.040 --> 1:28:52.720 - And basically the human expert is needed for exactly that. - -1:28:52.720 --> 1:28:53.920 - That's the idea. - -1:28:53.920 --> 1:28:57.640 - Keras is the interface between the data you're collecting - -1:28:57.640 --> 1:28:59.080 - and the business goals. - -1:28:59.080 --> 1:29:03.480 - And your job as an engineer is going to be to express your business goals - -1:29:03.480 --> 1:29:06.720 - and your understanding of your business or your product, - -1:29:06.720 --> 1:29:11.840 - your system as a kind of loss function or a kind of set of constraints. - -1:29:11.840 --> 1:29:19.480 - Does the possibility of creating an AGI system excite you or scare you or bore you? - -1:29:19.480 --> 1:29:22.080 - So intelligence can never really be general. - -1:29:22.080 --> 1:29:26.400 - You know, at best it can have some degree of generality like human intelligence. - -1:29:26.400 --> 1:29:30.640 - It also always has some specialization in the same way that human intelligence - -1:29:30.640 --> 1:29:33.440 - is specialized in a certain category of problems, - -1:29:33.440 --> 1:29:35.440 - is specialized in the human experience. - -1:29:35.440 --> 1:29:37.280 - And when people talk about AGI, - -1:29:37.280 --> 1:29:42.520 - I'm never quite sure if they're talking about very, very smart AI, - -1:29:42.520 --> 1:29:45.080 - so smart that it's even smarter than humans, - -1:29:45.080 --> 1:29:48.000 - or they're talking about human like intelligence, - -1:29:48.000 --> 1:29:49.680 - because these are different things. - -1:29:49.680 --> 1:29:54.760 - Let's say, presumably I'm oppressing you today with my humanness. - -1:29:54.760 --> 1:29:59.240 - So imagine that I was in fact a robot. - -1:29:59.240 --> 1:30:01.920 - So what does that mean? - -1:30:01.920 --> 1:30:04.920 - That I'm impressing you with natural language processing. - -1:30:04.920 --> 1:30:07.840 - Maybe if you weren't able to see me, maybe this is a phone call. - -1:30:07.840 --> 1:30:10.000 - So that kind of system. - -1:30:10.000 --> 1:30:11.120 - Companion. - -1:30:11.120 --> 1:30:15.040 - So that's very much about building human like AI. - -1:30:15.040 --> 1:30:18.200 - And you're asking me, you know, is this an exciting perspective? - -1:30:18.200 --> 1:30:19.440 - Yes. - -1:30:19.440 --> 1:30:21.760 - I think so, yes. - -1:30:21.760 --> 1:30:28.000 - Not so much because of what artificial human like intelligence could do, - -1:30:28.000 --> 1:30:30.880 - but, you know, from an intellectual perspective, - -1:30:30.880 --> 1:30:34.120 - I think if you could build truly human like intelligence, - -1:30:34.120 --> 1:30:37.240 - that means you could actually understand human intelligence, - -1:30:37.240 --> 1:30:39.880 - which is fascinating, right? - -1:30:39.880 --> 1:30:42.680 - Human like intelligence is going to require emotions. - -1:30:42.680 --> 1:30:44.400 - It's going to require consciousness, - -1:30:44.400 --> 1:30:49.720 - which is not things that would normally be required by an intelligent system. - -1:30:49.720 --> 1:30:53.160 - If you look at, you know, we were mentioning earlier like science - -1:30:53.160 --> 1:30:59.600 - as a superhuman problem solving agent or system, - -1:30:59.600 --> 1:31:02.120 - it does not have consciousness, it doesn't have emotions. - -1:31:02.120 --> 1:31:04.320 - In general, so emotions, - -1:31:04.320 --> 1:31:07.640 - I see consciousness as being on the same spectrum as emotions. - -1:31:07.640 --> 1:31:12.280 - It is a component of the subjective experience - -1:31:12.280 --> 1:31:18.800 - that is meant very much to guide behavior generation, right? - -1:31:18.800 --> 1:31:20.800 - It's meant to guide your behavior. - -1:31:20.800 --> 1:31:24.520 - In general, human intelligence and animal intelligence - -1:31:24.520 --> 1:31:29.280 - has evolved for the purpose of behavior generation, right? - -1:31:29.280 --> 1:31:30.680 - Including in a social context. - -1:31:30.680 --> 1:31:32.480 - So that's why we actually need emotions. - -1:31:32.480 --> 1:31:34.920 - That's why we need consciousness. - -1:31:34.920 --> 1:31:38.360 - An artificial intelligence system developed in a different context - -1:31:38.360 --> 1:31:42.800 - may well never need them, may well never be conscious like science. - -1:31:42.800 --> 1:31:47.960 - Well, on that point, I would argue it's possible to imagine - -1:31:47.960 --> 1:31:51.480 - that there's echoes of consciousness in science - -1:31:51.480 --> 1:31:55.480 - when viewed as an organism, that science is consciousness. - -1:31:55.480 --> 1:31:59.160 - So, I mean, how would you go about testing this hypothesis? - -1:31:59.160 --> 1:32:07.000 - How do you probe the subjective experience of an abstract system like science? - -1:32:07.000 --> 1:32:10.400 - Well, the point of probing any subjective experience is impossible - -1:32:10.400 --> 1:32:13.200 - because I'm not science, I'm Lex. - -1:32:13.200 --> 1:32:20.520 - So I can't probe another entity, it's no more than bacteria on my skin. - -1:32:20.520 --> 1:32:24.160 - You're Lex, I can ask you questions about your subjective experience - -1:32:24.160 --> 1:32:28.440 - and you can answer me, and that's how I know you're conscious. - -1:32:28.440 --> 1:32:31.840 - Yes, but that's because we speak the same language. - -1:32:31.840 --> 1:32:35.520 - You perhaps, we have to speak the language of science in order to ask it. - -1:32:35.520 --> 1:32:40.320 - Honestly, I don't think consciousness, just like emotions of pain and pleasure, - -1:32:40.320 --> 1:32:44.160 - is not something that inevitably arises - -1:32:44.160 --> 1:32:47.920 - from any sort of sufficiently intelligent information processing. - -1:32:47.920 --> 1:32:53.920 - It is a feature of the mind, and if you've not implemented it explicitly, it is not there. - -1:32:53.920 --> 1:32:58.960 - So you think it's an emergent feature of a particular architecture. - -1:32:58.960 --> 1:33:00.320 - So do you think... - -1:33:00.320 --> 1:33:02.000 - It's a feature in the same sense. - -1:33:02.000 --> 1:33:08.240 - So, again, the subjective experience is all about guiding behavior. - -1:33:08.240 --> 1:33:15.120 - If the problems you're trying to solve don't really involve an embodied agent, - -1:33:15.120 --> 1:33:19.520 - maybe in a social context, generating behavior and pursuing goals like this. - -1:33:19.520 --> 1:33:22.160 - And if you look at science, that's not really what's happening. - -1:33:22.160 --> 1:33:27.920 - Even though it is, it is a form of artificial AI, artificial intelligence, - -1:33:27.920 --> 1:33:31.920 - in the sense that it is solving problems, it is accumulating knowledge, - -1:33:31.920 --> 1:33:35.040 - accumulating solutions and so on. - -1:33:35.040 --> 1:33:39.440 - So if you're not explicitly implementing a subjective experience, - -1:33:39.440 --> 1:33:44.000 - implementing certain emotions and implementing consciousness, - -1:33:44.000 --> 1:33:47.360 - it's not going to just spontaneously emerge. - -1:33:47.360 --> 1:33:48.080 - Yeah. - -1:33:48.080 --> 1:33:53.200 - But so for a system like, human like intelligence system that has consciousness, - -1:33:53.200 --> 1:33:55.840 - do you think it needs to have a body? - -1:33:55.840 --> 1:33:56.720 - Yes, definitely. - -1:33:56.720 --> 1:33:59.600 - I mean, it doesn't have to be a physical body, right? - -1:33:59.600 --> 1:34:03.440 - And there's not that much difference between a realistic simulation in the real world. - -1:34:03.440 --> 1:34:06.400 - So there has to be something you have to preserve kind of thing. - -1:34:06.400 --> 1:34:11.840 - Yes, but human like intelligence can only arise in a human like context. - -1:34:11.840 --> 1:34:16.800 - Intelligence needs other humans in order for you to demonstrate - -1:34:16.800 --> 1:34:19.040 - that you have human like intelligence, essentially. - -1:34:19.040 --> 1:34:19.540 - Yes. - -1:34:20.320 --> 1:34:28.080 - So what kind of tests and demonstration would be sufficient for you - -1:34:28.080 --> 1:34:30.960 - to demonstrate human like intelligence? - -1:34:30.960 --> 1:34:31.360 - Yeah. - -1:34:31.360 --> 1:34:35.600 - Just out of curiosity, you've talked about in terms of theorem proving - -1:34:35.600 --> 1:34:38.000 - and program synthesis, I think you've written about - -1:34:38.000 --> 1:34:40.480 - that there's no good benchmarks for this. - -1:34:40.480 --> 1:34:40.720 - Yeah. - -1:34:40.720 --> 1:34:42.000 - That's one of the problems. - -1:34:42.000 --> 1:34:46.320 - So let's talk program synthesis. - -1:34:46.320 --> 1:34:47.760 - So what do you imagine is a good... - -1:34:48.800 --> 1:34:51.360 - I think it's related questions for human like intelligence - -1:34:51.360 --> 1:34:52.560 - and for program synthesis. - -1:34:53.360 --> 1:34:56.080 - What's a good benchmark for either or both? - -1:34:56.080 --> 1:34:56.480 - Right. - -1:34:56.480 --> 1:34:59.200 - So I mean, you're actually asking two questions, - -1:34:59.200 --> 1:35:02.480 - which is one is about quantifying intelligence - -1:35:02.480 --> 1:35:06.880 - and comparing the intelligence of an artificial system - -1:35:06.880 --> 1:35:08.480 - to the intelligence for human. - -1:35:08.480 --> 1:35:13.440 - And the other is about the degree to which this intelligence is human like. - -1:35:13.440 --> 1:35:15.120 - It's actually two different questions. - -1:35:16.560 --> 1:35:18.960 - So you mentioned earlier the Turing test. - -1:35:19.680 --> 1:35:23.200 - Well, I actually don't like the Turing test because it's very lazy. - -1:35:23.200 --> 1:35:28.720 - It's all about completely bypassing the problem of defining and measuring intelligence - -1:35:28.720 --> 1:35:34.160 - and instead delegating to a human judge or a panel of human judges. - -1:35:34.160 --> 1:35:37.120 - So it's a total copout, right? - -1:35:38.160 --> 1:35:43.200 - If you want to measure how human like an agent is, - -1:35:43.760 --> 1:35:46.640 - I think you have to make it interact with other humans. - -1:35:47.600 --> 1:35:53.760 - Maybe it's not necessarily a good idea to have these other humans be the judges. - -1:35:53.760 --> 1:35:59.280 - Maybe you should just observe behavior and compare it to what a human would actually have done. - -1:36:00.560 --> 1:36:05.120 - When it comes to measuring how smart, how clever an agent is - -1:36:05.120 --> 1:36:11.120 - and comparing that to the degree of human intelligence. - -1:36:11.120 --> 1:36:12.960 - So we're already talking about two things, right? - -1:36:13.520 --> 1:36:20.320 - The degree, kind of like the magnitude of an intelligence and its direction, right? - -1:36:20.320 --> 1:36:23.280 - Like the norm of a vector and its direction. - -1:36:23.280 --> 1:36:32.000 - And the direction is like human likeness and the magnitude, the norm is intelligence. - -1:36:32.720 --> 1:36:34.080 - You could call it intelligence, right? - -1:36:34.080 --> 1:36:41.040 - So the direction, your sense, the space of directions that are human like is very narrow. - -1:36:41.040 --> 1:36:41.200 - Yeah. - -1:36:42.240 --> 1:36:48.880 - So the way you would measure the magnitude of intelligence in a system - -1:36:48.880 --> 1:36:54.640 - in a way that also enables you to compare it to that of a human. - -1:36:54.640 --> 1:36:59.200 - Well, if you look at different benchmarks for intelligence today, - -1:36:59.200 --> 1:37:04.160 - they're all too focused on skill at a given task. - -1:37:04.160 --> 1:37:08.720 - Like skill at playing chess, skill at playing Go, skill at playing Dota. - -1:37:10.720 --> 1:37:15.600 - And I think that's not the right way to go about it because you can always - -1:37:15.600 --> 1:37:18.240 - beat a human at one specific task. - -1:37:19.200 --> 1:37:23.920 - The reason why our skill at playing Go or juggling or anything is impressive - -1:37:23.920 --> 1:37:28.400 - is because we are expressing this skill within a certain set of constraints. - -1:37:28.400 --> 1:37:32.320 - If you remove the constraints, the constraints that we have one lifetime, - -1:37:32.320 --> 1:37:36.080 - that we have this body and so on, if you remove the context, - -1:37:36.080 --> 1:37:40.480 - if you have unlimited string data, if you can have access to, you know, - -1:37:40.480 --> 1:37:44.640 - for instance, if you look at juggling, if you have no restriction on the hardware, - -1:37:44.640 --> 1:37:48.400 - then achieving arbitrary levels of skill is not very interesting - -1:37:48.400 --> 1:37:52.400 - and says nothing about the amount of intelligence you've achieved. - -1:37:52.400 --> 1:37:57.440 - So if you want to measure intelligence, you need to rigorously define what - -1:37:57.440 --> 1:38:02.960 - intelligence is, which in itself, you know, it's a very challenging problem. - -1:38:02.960 --> 1:38:04.320 - And do you think that's possible? - -1:38:04.320 --> 1:38:06.000 - To define intelligence? Yes, absolutely. - -1:38:06.000 --> 1:38:09.760 - I mean, you can provide, many people have provided, you know, some definition. - -1:38:10.560 --> 1:38:12.000 - I have my own definition. - -1:38:12.000 --> 1:38:13.440 - Where does your definition begin? - -1:38:13.440 --> 1:38:16.240 - Where does your definition begin if it doesn't end? - -1:38:16.240 --> 1:38:21.680 - Well, I think intelligence is essentially the efficiency - -1:38:22.320 --> 1:38:29.760 - with which you turn experience into generalizable programs. - -1:38:29.760 --> 1:38:32.800 - So what that means is it's the efficiency with which - -1:38:32.800 --> 1:38:36.720 - you turn a sampling of experience space into - -1:38:36.720 --> 1:38:46.000 - the ability to process a larger chunk of experience space. - -1:38:46.000 --> 1:38:52.560 - So measuring skill can be one proxy across many different tasks, - -1:38:52.560 --> 1:38:54.480 - can be one proxy for measuring intelligence. - -1:38:54.480 --> 1:38:58.720 - But if you want to only measure skill, you should control for two things. - -1:38:58.720 --> 1:39:04.960 - You should control for the amount of experience that your system has - -1:39:04.960 --> 1:39:08.080 - and the priors that your system has. - -1:39:08.080 --> 1:39:13.120 - But if you look at two agents and you give them the same priors - -1:39:13.120 --> 1:39:16.160 - and you give them the same amount of experience, - -1:39:16.160 --> 1:39:21.360 - there is one of the agents that is going to learn programs, - -1:39:21.360 --> 1:39:25.440 - representations, something, a model that will perform well - -1:39:25.440 --> 1:39:28.720 - on the larger chunk of experience space than the other. - -1:39:28.720 --> 1:39:30.960 - And that is the smaller agent. - -1:39:30.960 --> 1:39:36.960 - Yeah. So if you fix the experience, which generate better programs, - -1:39:37.680 --> 1:39:39.520 - better meaning more generalizable. - -1:39:39.520 --> 1:39:40.560 - That's really interesting. - -1:39:40.560 --> 1:39:42.400 - That's a very nice, clean definition of... - -1:39:42.400 --> 1:39:47.280 - Oh, by the way, in this definition, it is already very obvious - -1:39:47.280 --> 1:39:49.440 - that intelligence has to be specialized - -1:39:49.440 --> 1:39:51.680 - because you're talking about experience space - -1:39:51.680 --> 1:39:54.080 - and you're talking about segments of experience space. - -1:39:54.080 --> 1:39:57.200 - You're talking about priors and you're talking about experience. - -1:39:57.200 --> 1:40:02.480 - All of these things define the context in which intelligence emerges. - -1:40:04.480 --> 1:40:08.640 - And you can never look at the totality of experience space, right? - -1:40:09.760 --> 1:40:12.160 - So intelligence has to be specialized. - -1:40:12.160 --> 1:40:14.960 - But it can be sufficiently large, the experience space, - -1:40:14.960 --> 1:40:16.080 - even though it's specialized. - -1:40:16.080 --> 1:40:19.120 - There's a certain point when the experience space is large enough - -1:40:19.120 --> 1:40:21.440 - to where it might as well be general. - -1:40:22.000 --> 1:40:23.920 - It feels general. It looks general. - -1:40:23.920 --> 1:40:25.680 - Sure. I mean, it's very relative. - -1:40:25.680 --> 1:40:29.360 - Like, for instance, many people would say human intelligence is general. - -1:40:29.360 --> 1:40:31.200 - In fact, it is quite specialized. - -1:40:32.800 --> 1:40:37.120 - We can definitely build systems that start from the same innate priors - -1:40:37.120 --> 1:40:39.120 - as what humans have at birth. - -1:40:39.120 --> 1:40:42.320 - Because we already understand fairly well - -1:40:42.320 --> 1:40:44.480 - what sort of priors we have as humans. - -1:40:44.480 --> 1:40:46.080 - Like many people have worked on this problem. - -1:40:46.800 --> 1:40:51.040 - Most notably, Elisabeth Spelke from Harvard. - -1:40:51.040 --> 1:40:52.240 - I don't know if you know her. - -1:40:52.240 --> 1:40:56.000 - She's worked a lot on what she calls core knowledge. - -1:40:56.000 --> 1:41:00.640 - And it is very much about trying to determine and describe - -1:41:00.640 --> 1:41:02.320 - what priors we are born with. - -1:41:02.320 --> 1:41:04.720 - Like language skills and so on, all that kind of stuff. - -1:41:04.720 --> 1:41:05.220 - Exactly. - -1:41:06.880 --> 1:41:11.440 - So we have some pretty good understanding of what priors we are born with. - -1:41:11.440 --> 1:41:12.560 - So we could... - -1:41:13.760 --> 1:41:17.760 - So I've actually been working on a benchmark for the past couple years, - -1:41:17.760 --> 1:41:18.640 - you know, on and off. - -1:41:18.640 --> 1:41:20.480 - I hope to be able to release it at some point. - -1:41:20.480 --> 1:41:21.760 - That's exciting. - -1:41:21.760 --> 1:41:25.680 - The idea is to measure the intelligence of systems - -1:41:26.800 --> 1:41:28.640 - by countering for priors, - -1:41:28.640 --> 1:41:30.480 - countering for amount of experience, - -1:41:30.480 --> 1:41:34.800 - and by assuming the same priors as what humans are born with. - -1:41:34.800 --> 1:41:39.520 - So that you can actually compare these scores to human intelligence. - -1:41:39.520 --> 1:41:43.280 - You can actually have humans pass the same test in a way that's fair. - -1:41:43.280 --> 1:41:52.320 - Yeah. And so importantly, such a benchmark should be such that any amount - -1:41:52.960 --> 1:41:55.920 - of practicing does not increase your score. - -1:41:56.480 --> 1:42:00.560 - So try to picture a game where no matter how much you play this game, - -1:42:01.600 --> 1:42:05.040 - that does not change your skill at the game. - -1:42:05.040 --> 1:42:05.920 - Can you picture that? - -1:42:05.920 --> 1:42:11.040 - As a person who deeply appreciates practice, I cannot actually. - -1:42:11.040 --> 1:42:16.560 - There's actually a very simple trick. - -1:42:16.560 --> 1:42:19.440 - So in order to come up with a task, - -1:42:19.440 --> 1:42:21.760 - so the only thing you can measure is skill at the task. - -1:42:21.760 --> 1:42:22.320 - Yes. - -1:42:22.320 --> 1:42:24.800 - All tasks are going to involve priors. - -1:42:24.800 --> 1:42:25.600 - Yes. - -1:42:25.600 --> 1:42:29.920 - The trick is to know what they are and to describe that. - -1:42:29.920 --> 1:42:33.760 - And then you make sure that this is the same set of priors as what humans start with. - -1:42:33.760 --> 1:42:38.560 - So you create a task that assumes these priors, that exactly documents these priors, - -1:42:38.560 --> 1:42:42.240 - so that the priors are made explicit and there are no other priors involved. - -1:42:42.240 --> 1:42:48.960 - And then you generate a certain number of samples in experience space for this task, right? - -1:42:49.840 --> 1:42:56.320 - And this, for one task, assuming that the task is new for the agent passing it, - -1:42:56.320 --> 1:43:04.320 - that's one test of this definition of intelligence that we set up. - -1:43:04.320 --> 1:43:06.880 - And now you can scale that to many different tasks, - -1:43:06.880 --> 1:43:10.480 - that each task should be new to the agent passing it, right? - -1:43:11.360 --> 1:43:14.480 - And also it should be human interpretable and understandable - -1:43:14.480 --> 1:43:16.880 - so that you can actually have a human pass the same test. - -1:43:16.880 --> 1:43:19.760 - And then you can compare the score of your machine and the score of your human. - -1:43:19.760 --> 1:43:20.720 - Which could be a lot of stuff. - -1:43:20.720 --> 1:43:23.040 - You could even start a task like MNIST. - -1:43:23.040 --> 1:43:28.800 - Just as long as you start with the same set of priors. - -1:43:28.800 --> 1:43:34.080 - So the problem with MNIST, humans are already trying to recognize digits, right? - -1:43:35.600 --> 1:43:40.960 - But let's say we're considering objects that are not digits, - -1:43:42.400 --> 1:43:43.920 - some completely arbitrary patterns. - -1:43:44.480 --> 1:43:48.880 - Well, humans already come with visual priors about how to process that. - -1:43:48.880 --> 1:43:54.080 - So in order to make the game fair, you would have to isolate these priors - -1:43:54.080 --> 1:43:57.280 - and describe them and then express them as computational rules. - -1:43:57.280 --> 1:44:01.680 - Having worked a lot with vision science people, that's exceptionally difficult. - -1:44:01.680 --> 1:44:03.120 - A lot of progress has been made. - -1:44:03.120 --> 1:44:08.080 - There's been a lot of good tests and basically reducing all of human vision into some good priors. - -1:44:08.640 --> 1:44:10.960 - We're still probably far away from that perfectly, - -1:44:10.960 --> 1:44:14.640 - but as a start for a benchmark, that's an exciting possibility. - -1:44:14.640 --> 1:44:24.240 - Yeah, so Elisabeth Spelke actually lists objectness as one of the core knowledge priors. - -1:44:24.800 --> 1:44:25.920 - Objectness, cool. - -1:44:25.920 --> 1:44:26.880 - Objectness, yeah. - -1:44:27.440 --> 1:44:31.520 - So we have priors about objectness, like about the visual space, about time, - -1:44:31.520 --> 1:44:34.240 - about agents, about goal oriented behavior. - -1:44:35.280 --> 1:44:39.280 - We have many different priors, but what's interesting is that, - -1:44:39.280 --> 1:44:43.920 - sure, we have this pretty diverse and rich set of priors, - -1:44:43.920 --> 1:44:46.880 - but it's also not that diverse, right? - -1:44:46.880 --> 1:44:50.800 - We are not born into this world with a ton of knowledge about the world, - -1:44:50.800 --> 1:44:57.840 - with only a small set of core knowledge. - -1:44:58.640 --> 1:45:05.040 - Yeah, sorry, do you have a sense of how it feels to us humans that that set is not that large? - -1:45:05.040 --> 1:45:09.600 - But just even the nature of time that we kind of integrate pretty effectively - -1:45:09.600 --> 1:45:11.600 - through all of our perception, all of our reasoning, - -1:45:12.640 --> 1:45:17.680 - maybe how, you know, do you have a sense of how easy it is to encode those priors? - -1:45:17.680 --> 1:45:24.560 - Maybe it requires building a universe and then the human brain in order to encode those priors. - -1:45:25.440 --> 1:45:28.640 - Or do you have a hope that it can be listed like an axiomatic? - -1:45:28.640 --> 1:45:29.280 - I don't think so. - -1:45:29.280 --> 1:45:33.040 - So you have to keep in mind that any knowledge about the world that we are - -1:45:33.040 --> 1:45:41.120 - born with is something that has to have been encoded into our DNA by evolution at some point. - -1:45:41.120 --> 1:45:41.440 - Right. - -1:45:41.440 --> 1:45:45.440 - And DNA is a very, very low bandwidth medium. - -1:45:46.000 --> 1:45:51.200 - Like it's extremely long and expensive to encode anything into DNA because first of all, - -1:45:52.560 --> 1:45:57.440 - you need some sort of evolutionary pressure to guide this writing process. - -1:45:57.440 --> 1:46:03.440 - And then, you know, the higher level of information you're trying to write, the longer it's going to take. - -1:46:04.480 --> 1:46:13.520 - And the thing in the environment that you're trying to encode knowledge about has to be stable - -1:46:13.520 --> 1:46:15.280 - over this duration. - -1:46:15.280 --> 1:46:20.960 - So you can only encode into DNA things that constitute an evolutionary advantage. - -1:46:20.960 --> 1:46:25.280 - So this is actually a very small subset of all possible knowledge about the world. - -1:46:25.280 --> 1:46:32.080 - You can only encode things that are stable, that are true, over very, very long periods of time, - -1:46:32.080 --> 1:46:33.680 - typically millions of years. - -1:46:33.680 --> 1:46:38.720 - For instance, we might have some visual prior about the shape of snakes, right? - -1:46:38.720 --> 1:46:43.920 - But what makes a face, what's the difference between a face and an art face? - -1:46:44.560 --> 1:46:48.080 - But consider this interesting question. - -1:46:48.080 --> 1:46:56.640 - Do we have any innate sense of the visual difference between a male face and a female face? - -1:46:56.640 --> 1:46:57.600 - What do you think? - -1:46:58.640 --> 1:46:59.840 - For a human, I mean. - -1:46:59.840 --> 1:47:04.000 - I would have to look back into evolutionary history when the genders emerged. - -1:47:04.000 --> 1:47:06.240 - But yeah, most... - -1:47:06.240 --> 1:47:09.840 - I mean, the faces of humans are quite different from the faces of great apes. - -1:47:10.640 --> 1:47:11.600 - Great apes, right? - -1:47:12.880 --> 1:47:13.600 - Yeah. - -1:47:13.600 --> 1:47:14.800 - That's interesting. - -1:47:14.800 --> 1:47:22.800 - Yeah, you couldn't tell the face of a female chimpanzee from the face of a male chimpanzee, - -1:47:22.800 --> 1:47:23.440 - probably. - -1:47:23.440 --> 1:47:26.160 - Yeah, and I don't think most humans have all that ability. - -1:47:26.160 --> 1:47:33.280 - So we do have innate knowledge of what makes a face, but it's actually impossible for us to - -1:47:33.280 --> 1:47:40.320 - have any DNA encoded knowledge of the difference between a female human face and a male human face - -1:47:40.320 --> 1:47:50.560 - because that knowledge, that information came up into the world actually very recently. - -1:47:50.560 --> 1:47:56.400 - If you look at the slowness of the process of encoding knowledge into DNA. - -1:47:56.400 --> 1:47:57.360 - Yeah, so that's interesting. - -1:47:57.360 --> 1:48:02.080 - That's a really powerful argument that DNA is a low bandwidth and it takes a long time to encode. - -1:48:02.800 --> 1:48:05.200 - That naturally creates a very efficient encoding. - -1:48:05.200 --> 1:48:12.800 - But one important consequence of this is that, so yes, we are born into this world with a bunch of - -1:48:12.800 --> 1:48:17.600 - knowledge, sometimes high level knowledge about the world, like the shape, the rough shape of a - -1:48:17.600 --> 1:48:19.520 - snake, of the rough shape of a face. - -1:48:20.480 --> 1:48:26.960 - But importantly, because this knowledge takes so long to write, almost all of this innate - -1:48:26.960 --> 1:48:32.080 - knowledge is shared with our cousins, with great apes, right? - -1:48:32.080 --> 1:48:35.600 - So it is not actually this innate knowledge that makes us special. - -1:48:36.320 --> 1:48:42.000 - But to throw it right back at you from the earlier on in our discussion, it's that encoding - -1:48:42.960 --> 1:48:48.320 - might also include the entirety of the environment of Earth. - -1:48:49.360 --> 1:48:49.920 - To some extent. - -1:48:49.920 --> 1:48:56.480 - So it can include things that are important to survival and production, so for which there is - -1:48:56.480 --> 1:49:02.880 - some evolutionary pressure, and things that are stable, constant over very, very, very long time - -1:49:02.880 --> 1:49:03.380 - periods. - -1:49:04.160 --> 1:49:06.320 - And honestly, it's not that much information. - -1:49:06.320 --> 1:49:14.400 - There's also, besides the bandwidths constraint and the constraints of the writing process, - -1:49:14.400 --> 1:49:21.440 - there's also memory constraints, like DNA, the part of DNA that deals with the human brain, - -1:49:21.440 --> 1:49:22.640 - it's actually fairly small. - -1:49:22.640 --> 1:49:25.520 - It's like, you know, on the order of megabytes, right? - -1:49:25.520 --> 1:49:29.600 - There's not that much high level knowledge about the world you can encode. - -1:49:31.600 --> 1:49:38.880 - That's quite brilliant and hopeful for a benchmark that you're referring to of encoding - -1:49:38.880 --> 1:49:39.360 - priors. - -1:49:39.360 --> 1:49:43.120 - I actually look forward to, I'm skeptical whether you can do it in the next couple of - -1:49:43.120 --> 1:49:44.320 - years, but hopefully. - -1:49:45.040 --> 1:49:45.760 - I've been working. - -1:49:45.760 --> 1:49:49.920 - So honestly, it's a very simple benchmark, and it's not like a big breakthrough or anything. - -1:49:49.920 --> 1:49:53.200 - It's more like a fun side project, right? - -1:49:53.200 --> 1:49:55.680 - But these fun, so is ImageNet. - -1:49:56.480 --> 1:50:04.080 - These fun side projects could launch entire groups of efforts towards creating reasoning - -1:50:04.080 --> 1:50:04.960 - systems and so on. - -1:50:04.960 --> 1:50:05.440 - And I think... - -1:50:05.440 --> 1:50:06.160 - Yeah, that's the goal. - -1:50:06.160 --> 1:50:12.080 - It's trying to measure strong generalization, to measure the strength of abstraction in - -1:50:12.080 --> 1:50:16.960 - our minds, well, in our minds and in artificial intelligence agencies. - -1:50:16.960 --> 1:50:24.800 - And if there's anything true about this science organism is its individual cells love competition. - -1:50:24.800 --> 1:50:26.800 - So and benchmarks encourage competition. - -1:50:26.800 --> 1:50:29.520 - So that's an exciting possibility. - -1:50:29.520 --> 1:50:32.640 - If you, do you think an AI winter is coming? - -1:50:33.520 --> 1:50:34.640 - And how do we prevent it? - -1:50:35.440 --> 1:50:36.080 - Not really. - -1:50:36.080 --> 1:50:42.160 - So an AI winter is something that would occur when there's a big mismatch between how we - -1:50:42.160 --> 1:50:47.280 - are selling the capabilities of AI and the actual capabilities of AI. - -1:50:47.280 --> 1:50:50.560 - And today, some deep learning is creating a lot of value. - -1:50:50.560 --> 1:50:56.240 - And it will keep creating a lot of value in the sense that these models are applicable - -1:50:56.240 --> 1:51:00.000 - to a very wide range of problems that are relevant today. - -1:51:00.000 --> 1:51:05.120 - And we are only just getting started with applying these algorithms to every problem - -1:51:05.120 --> 1:51:06.320 - they could be solving. - -1:51:06.320 --> 1:51:10.160 - So deep learning will keep creating a lot of value for the time being. - -1:51:10.160 --> 1:51:15.920 - What's concerning, however, is that there's a lot of hype around deep learning and around - -1:51:15.920 --> 1:51:16.240 - AI. - -1:51:16.240 --> 1:51:22.000 - There are lots of people are overselling the capabilities of these systems, not just - -1:51:22.000 --> 1:51:27.760 - the capabilities, but also overselling the fact that they might be more or less, you - -1:51:27.760 --> 1:51:36.640 - know, brain like, like given the kind of a mystical aspect, these technologies and also - -1:51:36.640 --> 1:51:43.840 - overselling the pace of progress, which, you know, it might look fast in the sense that - -1:51:43.840 --> 1:51:46.480 - we have this exponentially increasing number of papers. - -1:51:47.760 --> 1:51:52.960 - But again, that's just a simple consequence of the fact that we have ever more people - -1:51:52.960 --> 1:51:53.840 - coming into the field. - -1:51:54.400 --> 1:51:57.440 - It doesn't mean the progress is actually exponentially fast. - -1:51:58.640 --> 1:52:02.720 - Let's say you're trying to raise money for your startup or your research lab. - -1:52:02.720 --> 1:52:09.120 - You might want to tell, you know, a grandiose story to investors about how deep learning - -1:52:09.120 --> 1:52:14.240 - is just like the brain and how it can solve all these incredible problems like self driving - -1:52:14.240 --> 1:52:15.760 - and robotics and so on. - -1:52:15.760 --> 1:52:19.440 - And maybe you can tell them that the field is progressing so fast and we are going to - -1:52:19.440 --> 1:52:23.040 - have AGI within 15 years or even 10 years. - -1:52:23.040 --> 1:52:25.920 - And none of this is true. - -1:52:25.920 --> 1:52:32.800 - And every time you're like saying these things and an investor or, you know, a decision maker - -1:52:32.800 --> 1:52:41.680 - believes them, well, this is like the equivalent of taking on credit card debt, but for trust, - -1:52:41.680 --> 1:52:42.480 - right? - -1:52:42.480 --> 1:52:50.160 - And maybe this will, you know, this will be what enables you to raise a lot of money, - -1:52:50.160 --> 1:52:54.320 - but ultimately you are creating damage, you are damaging the field. - -1:52:54.320 --> 1:53:00.160 - So that's the concern is that that debt, that's what happens with the other AI winters is - -1:53:00.160 --> 1:53:04.160 - the concern is you actually tweeted about this with autonomous vehicles, right? - -1:53:04.160 --> 1:53:08.960 - There's almost every single company now have promised that they will have full autonomous - -1:53:08.960 --> 1:53:11.760 - vehicles by 2021, 2022. - -1:53:11.760 --> 1:53:18.080 - That's a good example of the consequences of over hyping the capabilities of AI and - -1:53:18.080 --> 1:53:19.280 - the pace of progress. - -1:53:19.280 --> 1:53:25.200 - So because I work especially a lot recently in this area, I have a deep concern of what - -1:53:25.200 --> 1:53:30.400 - happens when all of these companies after I've invested billions have a meeting and - -1:53:30.400 --> 1:53:33.600 - say, how much do we actually, first of all, do we have an autonomous vehicle? - -1:53:33.600 --> 1:53:35.280 - The answer will definitely be no. - -1:53:35.840 --> 1:53:40.560 - And second will be, wait a minute, we've invested one, two, three, four billion dollars - -1:53:40.560 --> 1:53:43.120 - into this and we made no profit. - -1:53:43.120 --> 1:53:49.200 - And the reaction to that may be going very hard in other directions that might impact - -1:53:49.200 --> 1:53:50.400 - even other industries. - -1:53:50.400 --> 1:53:55.520 - And that's what we call an AI winter is when there is backlash where no one believes any - -1:53:55.520 --> 1:53:59.360 - of these promises anymore because they've turned that to be big lies the first time - -1:53:59.360 --> 1:54:00.240 - around. - -1:54:00.240 --> 1:54:06.000 - And this will definitely happen to some extent for autonomous vehicles because the public - -1:54:06.000 --> 1:54:13.360 - and decision makers have been convinced that around 2015, they've been convinced by these - -1:54:13.360 --> 1:54:19.600 - people who are trying to raise money for their startups and so on, that L5 driving was coming - -1:54:19.600 --> 1:54:22.880 - in maybe 2016, maybe 2017, maybe 2018. - -1:54:22.880 --> 1:54:26.080 - Now we're in 2019, we're still waiting for it. - -1:54:27.600 --> 1:54:32.800 - And so I don't believe we are going to have a full on AI winter because we have these - -1:54:32.800 --> 1:54:36.640 - technologies that are producing a tremendous amount of real value. - -1:54:37.680 --> 1:54:39.920 - But there is also too much hype. - -1:54:39.920 --> 1:54:43.520 - So there will be some backlash, especially there will be backlash. - -1:54:44.960 --> 1:54:53.040 - So some startups are trying to sell the dream of AGI and the fact that AGI is going to create - -1:54:53.040 --> 1:54:53.760 - infinite value. - -1:54:53.760 --> 1:54:55.680 - Like AGI is like a free lunch. - -1:54:55.680 --> 1:55:02.800 - Like if you can develop an AI system that passes a certain threshold of IQ or something, - -1:55:02.800 --> 1:55:04.400 - then suddenly you have infinite value. - -1:55:04.400 --> 1:55:14.160 - And well, there are actually lots of investors buying into this idea and they will wait maybe - -1:55:14.160 --> 1:55:17.760 - 10, 15 years and nothing will happen. - -1:55:17.760 --> 1:55:22.560 - And the next time around, well, maybe there will be a new generation of investors. - -1:55:22.560 --> 1:55:23.360 - No one will care. - -1:55:24.800 --> 1:55:27.280 - Human memory is fairly short after all. - -1:55:27.280 --> 1:55:34.320 - I don't know about you, but because I've spoken about AGI sometimes poetically, I get a lot - -1:55:34.320 --> 1:55:42.000 - of emails from people giving me, they're usually like a large manifestos of they've, they say - -1:55:42.000 --> 1:55:47.200 - to me that they have created an AGI system or they know how to do it. - -1:55:47.200 --> 1:55:48.880 - And there's a long write up of how to do it. - -1:55:48.880 --> 1:55:50.560 - I get a lot of these emails, yeah. - -1:55:50.560 --> 1:55:57.760 - They're a little bit feel like it's generated by an AI system actually, but there's usually - -1:55:57.760 --> 1:56:06.640 - no diagram, you have a transformer generating crank papers about AGI. - -1:56:06.640 --> 1:56:12.160 - So the question is about, because you've been such a good, you have a good radar for crank - -1:56:12.160 --> 1:56:16.720 - papers, how do we know they're not onto something? - -1:56:16.720 --> 1:56:24.240 - How do I, so when you start to talk about AGI or anything like the reasoning benchmarks - -1:56:24.240 --> 1:56:28.160 - and so on, so something that doesn't have a benchmark, it's really difficult to know. - -1:56:29.120 --> 1:56:34.560 - I mean, I talked to Jeff Hawkins, who's really looking at neuroscience approaches to how, - -1:56:35.200 --> 1:56:41.520 - and there's some, there's echoes of really interesting ideas in at least Jeff's case, - -1:56:41.520 --> 1:56:42.320 - which he's showing. - -1:56:43.280 --> 1:56:45.040 - How do you usually think about this? - -1:56:46.640 --> 1:56:52.880 - Like preventing yourself from being too narrow minded and elitist about deep learning, it - -1:56:52.880 --> 1:56:56.720 - has to work on these particular benchmarks, otherwise it's trash. - -1:56:56.720 --> 1:57:05.280 - Well, you know, the thing is, intelligence does not exist in the abstract. - -1:57:05.280 --> 1:57:07.200 - Intelligence has to be applied. - -1:57:07.200 --> 1:57:11.040 - So if you don't have a benchmark, if you have an improvement in some benchmark, maybe it's - -1:57:11.040 --> 1:57:12.400 - a new benchmark, right? - -1:57:12.400 --> 1:57:16.640 - Maybe it's not something we've been looking at before, but you do need a problem that - -1:57:16.640 --> 1:57:17.360 - you're trying to solve. - -1:57:17.360 --> 1:57:20.000 - You're not going to come up with a solution without a problem. - -1:57:20.000 --> 1:57:25.520 - So you, general intelligence, I mean, you've clearly highlighted generalization. - -1:57:26.320 --> 1:57:31.200 - If you want to claim that you have an intelligence system, it should come with a benchmark. - -1:57:31.200 --> 1:57:35.760 - It should, yes, it should display capabilities of some kind. - -1:57:35.760 --> 1:57:41.840 - It should show that it can create some form of value, even if it's a very artificial form - -1:57:41.840 --> 1:57:42.800 - of value. - -1:57:42.800 --> 1:57:48.800 - And that's also the reason why you don't actually need to care about telling which papers have - -1:57:48.800 --> 1:57:52.000 - actually some hidden potential and which do not. - -1:57:53.120 --> 1:57:59.200 - Because if there is a new technique that's actually creating value, this is going to - -1:57:59.200 --> 1:58:02.480 - be brought to light very quickly because it's actually making a difference. - -1:58:02.480 --> 1:58:08.160 - So it's the difference between something that is ineffectual and something that is actually - -1:58:08.160 --> 1:58:08.800 - useful. - -1:58:08.800 --> 1:58:14.080 - And ultimately usefulness is our guide, not just in this field, but if you look at science - -1:58:14.080 --> 1:58:18.720 - in general, maybe there are many, many people over the years that have had some really interesting - -1:58:19.440 --> 1:58:22.800 - theories of everything, but they were just completely useless. - -1:58:22.800 --> 1:58:27.280 - And you don't actually need to tell the interesting theories from the useless theories. - -1:58:28.000 --> 1:58:34.080 - All you need is to see, is this actually having an effect on something else? - -1:58:34.080 --> 1:58:35.360 - Is this actually useful? - -1:58:35.360 --> 1:58:36.800 - Is this making an impact or not? - -1:58:37.600 --> 1:58:38.640 - That's beautifully put. - -1:58:38.640 --> 1:58:43.680 - I mean, the same applies to quantum mechanics, to string theory, to the holographic principle. - -1:58:43.680 --> 1:58:45.280 - We are doing deep learning because it works. - -1:58:46.960 --> 1:58:52.720 - Before it started working, people considered people working on neural networks as cranks - -1:58:52.720 --> 1:58:53.120 - very much. - -1:58:54.560 --> 1:58:56.320 - No one was working on this anymore. - -1:58:56.320 --> 1:58:59.120 - And now it's working, which is what makes it valuable. - -1:58:59.120 --> 1:59:00.320 - It's not about being right. - -1:59:01.120 --> 1:59:02.560 - It's about being effective. - -1:59:02.560 --> 1:59:08.080 - And nevertheless, the individual entities of this scientific mechanism, just like Yoshua - -1:59:08.080 --> 1:59:12.480 - Banjo or Jan Lekun, they, while being called cranks, stuck with it. - -1:59:12.480 --> 1:59:12.880 - Right? - -1:59:12.880 --> 1:59:13.280 - Yeah. - -1:59:13.280 --> 1:59:17.840 - And so us individual agents, even if everyone's laughing at us, just stick with it. - -1:59:18.880 --> 1:59:21.840 - If you believe you have something, you should stick with it and see it through. - -1:59:23.520 --> 1:59:25.920 - That's a beautiful inspirational message to end on. - -1:59:25.920 --> 1:59:27.600 - Francois, thank you so much for talking today. - -1:59:27.600 --> 1:59:28.640 - That was amazing. - -1:59:28.640 --> 1:59:44.000 - Thank you. -