WEBVTT 00:00.000 --> 00:05.040 The following is a conversation from Stuart Russell. He's a professor of computer science at UC 00:05.040 --> 00:11.360 Berkeley and a coauthor of a book that introduced me and millions of other people to the amazing world 00:11.360 --> 00:18.320 of AI called Artificial Intelligence The Modern Approach. So it was an honor for me to have this 00:18.320 --> 00:24.480 conversation as part of MIT course on artificial journal intelligence and the artificial intelligence 00:24.480 --> 00:30.800 podcast. If you enjoy it, please subscribe on YouTube, iTunes or your podcast provider of choice 00:31.360 --> 00:37.600 or simply connect with me on Twitter at Lex Freedman spelled F R I D. And now here's my 00:37.600 --> 00:46.160 conversation with Stuart Russell. So you've mentioned in 1975 in high school you've created 00:46.160 --> 00:54.160 one of your first AI programs that played chess. Were you ever able to build a program that 00:54.160 --> 01:02.080 beat you at chess or another board game? So my program never beat me at chess. 01:03.520 --> 01:10.480 I actually wrote the program at Imperial College. So I used to take the bus every Wednesday with a 01:10.480 --> 01:17.200 box of cards this big and shove them into the card reader and they gave us eight seconds of CPU time. 01:17.200 --> 01:24.720 It took about five seconds to read the cards in and compile the code. So we had three seconds of 01:24.720 --> 01:30.960 CPU time, which was enough to make one move, you know, with a not very deep search. And then we 01:30.960 --> 01:34.960 would print that move out and then we'd have to go to the back of the queue and wait to feed the 01:34.960 --> 01:40.480 cards in again. How deep was the search? Well, are we talking about two moves? So no, I think we've 01:40.480 --> 01:48.000 got we got an eight move, eight, you know, depth eight with alpha beta. And we had some tricks of 01:48.000 --> 01:54.480 our own about move ordering and some pruning of the tree. And we were still able to beat that 01:54.480 --> 02:00.960 program. Yeah, yeah, I was a reasonable chess player in my youth. I did an Othello program 02:01.680 --> 02:05.920 and a backgammon program. So when I got to Berkeley, I worked a lot on 02:05.920 --> 02:12.560 what we call meta reasoning, which really means reasoning about reasoning. And in the case of 02:13.200 --> 02:18.320 a game playing program, you need to reason about what parts of the search tree you're actually 02:18.320 --> 02:23.440 going to explore, because the search tree is enormous, you know, bigger than the number of 02:23.440 --> 02:30.960 atoms in the universe. And the way programs succeed and the way humans succeed is by only 02:30.960 --> 02:36.160 looking at a small fraction of the search tree. And if you look at the right fraction, you play 02:36.160 --> 02:41.360 really well. If you look at the wrong fraction, if you waste your time thinking about things that 02:41.360 --> 02:45.840 are never going to happen, the moves that no one's ever going to make, then you're going to lose, 02:45.840 --> 02:53.760 because you won't be able to figure out the right decision. So that question of how machines can 02:53.760 --> 02:59.760 manage their own computation, how they decide what to think about is the meta reasoning question. 02:59.760 --> 03:05.920 We developed some methods for doing that. And very simply, a machine should think about 03:06.640 --> 03:11.920 whatever thoughts are going to improve its decision quality. We were able to show that 03:12.640 --> 03:18.240 both for a fellow, which is a standard two player game, and for backgammon, which includes 03:19.040 --> 03:24.000 dice rolls, so it's a two player game with uncertainty. For both of those cases, we could 03:24.000 --> 03:30.480 come up with algorithms that were actually much more efficient than the standard alpha beta search, 03:31.120 --> 03:36.560 which chess programs at the time were using. And that those programs could beat me. 03:38.080 --> 03:44.720 And I think you can see the same basic ideas in AlphaGo and AlphaZero today. 03:44.720 --> 03:52.560 The way they explore the tree is using a form of meta reasoning to select what to think about 03:52.560 --> 03:57.840 based on how useful it is to think about it. Is there any insights you can describe 03:57.840 --> 04:03.040 without Greek symbols of how do we select which paths to go down? 04:04.240 --> 04:10.560 There's really two kinds of learning going on. So as you say, AlphaGo learns to evaluate board 04:10.560 --> 04:17.680 to evaluate board position. So it can look at a go board. And it actually has probably a super 04:17.680 --> 04:25.760 human ability to instantly tell how promising that situation is. To me, the amazing thing about 04:25.760 --> 04:34.560 AlphaGo is not that it can be the world champion with its hands tied behind his back. But the fact that 04:34.560 --> 04:41.360 if you stop it from searching altogether, so you say, okay, you're not allowed to do 04:41.360 --> 04:46.480 any thinking ahead. You can just consider each of your legal moves and then look at the 04:47.120 --> 04:53.280 resulting situation and evaluate it. So what we call a depth one search. So just the immediate 04:53.280 --> 04:57.920 outcome of your moves and decide if that's good or bad. That version of AlphaGo 04:57.920 --> 05:05.200 can still play at a professional level. And human professionals are sitting there for 05:05.200 --> 05:13.440 five, 10 minutes deciding what to do. And AlphaGo in less than a second can instantly intuit what 05:13.440 --> 05:18.880 is the right move to make based on its ability to evaluate positions. And that is remarkable 05:19.760 --> 05:26.080 because we don't have that level of intuition about go. We actually have to think about the 05:26.080 --> 05:35.200 situation. So anyway, that capability that AlphaGo has is one big part of why it beats humans. 05:35.840 --> 05:44.560 The other big part is that it's able to look ahead 40, 50, 60 moves into the future. And 05:46.880 --> 05:51.200 if it was considering all possibilities, 40 or 50 or 60 moves into the future, 05:51.200 --> 06:02.240 that would be 10 to the 200 possibilities. So way more than atoms in the universe and so on. 06:02.240 --> 06:09.680 So it's very, very selective about what it looks at. So let me try to give you an intuition about 06:10.880 --> 06:15.360 how you decide what to think about. It's a combination of two things. One is 06:15.360 --> 06:22.000 how promising it is. So if you're already convinced that a move is terrible, 06:22.560 --> 06:26.560 there's no point spending a lot more time convincing yourself that it's terrible. 06:27.520 --> 06:33.600 Because it's probably not going to change your mind. So the real reason you think is because 06:33.600 --> 06:39.920 there's some possibility of changing your mind about what to do. And is that changing your mind 06:39.920 --> 06:46.000 that would result then in a better final action in the real world. So that's the purpose of thinking 06:46.800 --> 06:53.520 is to improve the final action in the real world. And so if you think about a move that is guaranteed 06:53.520 --> 06:58.000 to be terrible, you can convince yourself it's terrible, you're still not going to change your 06:58.000 --> 07:04.320 mind. But on the other hand, suppose you had a choice between two moves, one of them you've 07:04.320 --> 07:10.400 already figured out is guaranteed to be a draw, let's say. And then the other one looks a little 07:10.400 --> 07:14.000 bit worse. Like it looks fairly likely that if you make that move, you're going to lose. 07:14.640 --> 07:20.720 But there's still some uncertainty about the value of that move. There's still some possibility 07:20.720 --> 07:25.920 that it will turn out to be a win. Then it's worth thinking about that. So even though it's 07:25.920 --> 07:31.280 less promising on average than the other move, which is guaranteed to be a draw, there's still 07:31.280 --> 07:36.160 some purpose in thinking about it because there's a chance that you'll change your mind and discover 07:36.160 --> 07:42.080 that in fact it's a better move. So it's a combination of how good the move appears to be 07:42.080 --> 07:48.000 and how much uncertainty there is about its value. The more uncertainty, the more it's worth thinking 07:48.000 --> 07:52.800 about because there's a higher upside if you want to think of it that way. And of course in the 07:52.800 --> 07:59.920 beginning, especially in the AlphaGo Zero formulation, it's everything is shrouded in 07:59.920 --> 08:06.240 uncertainty. So you're really swimming in a sea of uncertainty. So it benefits you to 08:07.520 --> 08:11.120 I mean, actually falling in the same process as you described, but because you're so uncertain 08:11.120 --> 08:15.280 about everything, you basically have to try a lot of different directions. 08:15.280 --> 08:22.400 Yeah. So the early parts of the search tree are fairly bushy that it will look at a lot 08:22.400 --> 08:27.840 of different possibilities, but fairly quickly, the degree of certainty about some of the moves. 08:27.840 --> 08:31.760 I mean, if a move is really terrible, you'll pretty quickly find out, right? You'll lose 08:31.760 --> 08:37.200 half your pieces or half your territory. And then you'll say, okay, this is not worth thinking 08:37.200 --> 08:45.280 about anymore. And then so further down, the tree becomes very long and narrow. And you're following 08:45.280 --> 08:54.800 various lines of play, 10, 20, 30, 40, 50 moves into the future. And that again is something 08:54.800 --> 09:01.920 that human beings have a very hard time doing mainly because they just lack the short term memory. 09:02.480 --> 09:09.440 You just can't remember a sequence of moves. That's 50 moves long. And you can't imagine 09:09.440 --> 09:15.520 the board correctly for that many moves into the future. Of course, the top players, 09:16.400 --> 09:19.280 I'm much more familiar with chess, but the top players probably have, 09:19.280 --> 09:26.480 they have echoes of the same kind of intuition instinct that in a moment's time, AlphaGo applies 09:27.280 --> 09:31.760 when they see a board. I mean, they've seen those patterns, human beings have seen those 09:31.760 --> 09:37.680 patterns before at the top, at the grandmaster level. It seems that there is some 09:40.000 --> 09:45.920 similarities, or maybe it's our imagination creates a vision of those similarities, but it 09:45.920 --> 09:53.120 feels like this kind of pattern recognition that the AlphaGo approaches are using is similar to 09:53.120 --> 10:00.560 what human beings at the top level are using. I think there's some truth to that. 10:01.520 --> 10:08.960 But not entirely. Yeah, I mean, I think the extent to which a human grandmaster can reliably 10:10.080 --> 10:13.680 instantly recognize the right move and instantly recognize the values of position. 10:13.680 --> 10:19.120 I think that's a little bit overrated. But if you sacrifice a queen, for example, 10:19.120 --> 10:23.840 I mean, there's these, there's these beautiful games of chess with Bobby Fisher, somebody where 10:24.640 --> 10:32.720 it's seeming to make a bad move. And I'm not sure there's a perfect degree of calculation 10:32.720 --> 10:37.440 involved where they've calculated all the possible things that happen. But there's an 10:37.440 --> 10:45.440 instinct there, right, that somehow adds up to. Yeah, so I think what happens is you get a sense 10:45.440 --> 10:51.680 that there's some possibility in the position, even if you make a weird looking move, that it 10:51.680 --> 11:05.120 opens up some lines of calculation that otherwise would be definitely bad. And it's that intuition 11:05.120 --> 11:13.920 that there's something here in this position that might might yield a win down the set. And then 11:13.920 --> 11:20.640 you follow that. Right. And in some sense, when a chess player is following a line in his or her 11:20.640 --> 11:27.120 mind, they're they mentally simulating what the other person is going to do, what the opponent 11:27.120 --> 11:33.680 is going to do. And they can do that as long as the moves are kind of forced, right, as long as 11:33.680 --> 11:39.440 there's a we call a forcing variation where the opponent doesn't really have much choice how to 11:39.440 --> 11:45.200 respond. And then you see if you can force them into a situation where you win. We see plenty 11:45.200 --> 11:53.520 of mistakes, even in grandmaster games, where they just miss some simple three, four, five move 11:54.560 --> 12:00.400 combination that wasn't particularly apparent in the position, but was still there. 12:00.400 --> 12:07.360 That's the thing that makes us human. Yeah. So when you mentioned that in Othello, those games 12:07.360 --> 12:13.760 were after some meta reasoning improvements and research was able to beat you. How did that make 12:13.760 --> 12:21.280 you feel part of the meta reasoning capability that it had was based on learning. And, 12:23.280 --> 12:28.160 and you could sit down the next day and you could just feel that it had got a lot smarter. 12:28.160 --> 12:33.280 You know, and all of a sudden, you really felt like you're sort of pressed against 12:34.480 --> 12:40.800 the wall because it was it was much more aggressive and was totally unforgiving of any 12:40.800 --> 12:47.760 minor mistake that you might make. And actually, it seemed understood the game better than I did. 12:47.760 --> 12:55.520 And Gary Kasparov has this quote where during his match against Deep Blue, he said he suddenly 12:55.520 --> 13:01.680 felt that there was a new kind of intelligence across the board. Do you think that's a scary or 13:01.680 --> 13:10.240 an exciting possibility for Kasparov and for yourself in the context of chess purely sort of 13:10.240 --> 13:16.720 in this like that feeling, whatever that is, I think it's definitely an exciting feeling. 13:17.600 --> 13:23.680 You know, this is what made me work on AI in the first place was as soon as I really understood 13:23.680 --> 13:30.080 what a computer was, I wanted to make it smart. You know, I started out with the first program I 13:30.080 --> 13:38.640 wrote was for the Sinclair Programmable Calculator. And I think you could write a 21 step algorithm. 13:38.640 --> 13:44.160 That was the biggest program you could write something like that and do little arithmetic 13:44.160 --> 13:49.440 calculations. So I think I implemented Newton's method for square roots and a few other things 13:49.440 --> 13:56.640 like that. But then, you know, I thought, okay, if I just had more space, I could make this thing 13:56.640 --> 14:10.560 intelligent. And so I started thinking about AI. And I think the thing that's scary is not the 14:10.560 --> 14:19.520 chess program, because you know, chess programs, they're not in the taking over the world business. 14:19.520 --> 14:29.440 But if you extrapolate, you know, there are things about chess that don't resemble the real 14:29.440 --> 14:37.600 world, right? We know, we know the rules of chess. The chess board is completely visible 14:37.600 --> 14:43.280 to the program where, of course, the real world is not. Most the real world is not visible from 14:43.280 --> 14:52.400 wherever you're sitting, so to speak. And to overcome those kinds of problems, you need 14:52.400 --> 14:58.240 qualitatively different algorithms. Another thing about the real world is that, you know, we 14:58.240 --> 15:07.520 we regularly plan ahead on the timescales involving billions or trillions of steps. Now, 15:07.520 --> 15:13.760 we don't plan those in detail. But, you know, when you choose to do a PhD at Berkeley, 15:14.800 --> 15:20.480 that's a five year commitment that amounts to about a trillion motor control steps that you 15:20.480 --> 15:26.160 will eventually be committed to. Including going up the stairs, opening doors, 15:26.160 --> 15:32.880 a drinking water type. Yeah, I mean, every every finger movement while you're typing every character 15:32.880 --> 15:37.280 of every paper and the thesis and everything. So you're not committing in advance to the specific 15:37.280 --> 15:43.760 motor control steps, but you're still reasoning on a timescale that will eventually reduce to 15:44.400 --> 15:50.000 trillions of motor control actions. And so for all these reasons, 15:50.000 --> 15:58.160 you know, AlphaGo and Deep Blue and so on don't represent any kind of threat to humanity. But 15:58.160 --> 16:08.320 they are a step towards it, right? And progress in AI occurs by essentially removing one by one 16:08.320 --> 16:14.640 these assumptions that make problems easy, like the assumption of complete observability 16:14.640 --> 16:21.120 of the situation, right? We remove that assumption, you need a much more complicated kind of computing 16:22.160 --> 16:26.000 design and you need something that actually keeps track of all the things you can't see 16:26.000 --> 16:31.920 and tries to estimate what's going on. And there's inevitable uncertainty in that. So it becomes a 16:31.920 --> 16:38.160 much more complicated problem. But, you know, we are removing those assumptions, we are starting to 16:38.160 --> 16:44.400 have algorithms that can cope with much longer timescales, cope with uncertainty that can cope 16:44.400 --> 16:53.360 with partial observability. And so each of those steps sort of magnifies by a thousand the range 16:53.360 --> 16:58.400 of things that we can do with AI systems. So the way I started in AI, I wanted to be a psychiatrist 16:58.400 --> 17:03.840 for a long time and understand the mind in high school, and of course program and so on. And I 17:03.840 --> 17:10.640 showed up University of Illinois to an AI lab and they said, okay, I don't have time for you, but here 17:10.640 --> 17:18.480 is a book, AI Modern Approach, I think it was the first edition at the time. Here, go learn this. 17:18.480 --> 17:23.120 And I remember the lay of the land was, well, it's incredible that we solved chess, but we'll 17:23.120 --> 17:30.480 never solve go. I mean, it was pretty certain that go in the way we thought about systems that reason 17:31.520 --> 17:36.080 wasn't possible to solve. And now we've solved it. So it's a very... Well, I think I would have said 17:36.080 --> 17:44.080 that it's unlikely we could take the kind of algorithm that was used for chess and just get 17:44.080 --> 17:55.680 it to scale up and work well for go. And at the time, what we thought was that in order to solve 17:55.680 --> 18:01.600 go, we would have to do something similar to the way humans manage the complexity of go, 18:01.600 --> 18:06.960 which is to break it down into kind of sub games. So when a human thinks about a go board, 18:06.960 --> 18:12.480 they think about different parts of the board as sort of weakly connected to each other. 18:12.480 --> 18:16.880 And they think about, okay, within this part of the board, here's how things could go. 18:16.880 --> 18:20.560 In that part of board, here's how things could go. And then you try to sort of couple those 18:20.560 --> 18:26.400 two analyses together and deal with the interactions and maybe revise your views of how things are 18:26.400 --> 18:32.160 going to go in each part. And then you've got maybe five, six, seven, 10 parts of the board. And 18:33.440 --> 18:40.640 that actually resembles the real world much more than chess does. Because in the real world, 18:41.440 --> 18:49.200 we have work, we have home life, we have sport, whatever different kinds of activities, shopping, 18:49.200 --> 18:57.040 these all are connected to each other, but they're weakly connected. So when I'm typing a paper, 18:58.480 --> 19:03.600 I don't simultaneously have to decide which order I'm going to get the milk and the butter. 19:04.400 --> 19:09.760 That doesn't affect the typing. But I do need to realize, okay, better finish this 19:10.320 --> 19:14.080 before the shops close because I don't have anything, I don't have any food at home. 19:14.080 --> 19:20.560 So there's some weak connection, but not in the way that chess works, where everything is tied 19:20.560 --> 19:27.600 into a single stream of thought. So the thought was that go to solve go would have to make progress 19:27.600 --> 19:31.600 on stuff that would be useful for the real world. And in a way, AlphaGo is a little bit disappointing 19:32.480 --> 19:38.160 because the program designed for AlphaGo is actually not that different from 19:38.160 --> 19:45.840 from Deep Blue or even from Arthur Samuel's Jacob playing program from the 1950s. 19:48.160 --> 19:54.560 And in fact, the two things that make AlphaGo work is one is this amazing ability to evaluate 19:54.560 --> 19:59.200 the positions. And the other is the meta reasoning capability, which allows it to 19:59.200 --> 20:06.960 to explore some paths in the tree very deeply and to abandon other paths very quickly. 20:06.960 --> 20:14.640 So this word meta reasoning, while technically correct, inspires perhaps the wrong 20:16.000 --> 20:21.360 degree of power that AlphaGo has, for example, the word reasoning is a powerful word. So let me 20:21.360 --> 20:29.840 ask you sort of, do you were part of the symbolic AI world for a while, like where AI was, there's 20:29.840 --> 20:38.960 a lot of excellent interesting ideas there that unfortunately met a winter. And so do you think 20:38.960 --> 20:46.800 it reemerges? Oh, so I would say, yeah, it's not quite as simple as that. So the AI winter, 20:46.800 --> 20:54.400 the first winter that was actually named as such was the one in the late 80s. 20:56.400 --> 21:00.880 And that came about because in the mid 80s, there was 21:03.280 --> 21:10.480 really a concerted attempt to push AI out into the real world using what was called 21:10.480 --> 21:17.280 expert system technology. And for the most part, that technology was just not ready for prime 21:17.280 --> 21:27.200 time. They were trying in many cases to do a form of uncertain reasoning, judgment combinations of 21:27.200 --> 21:34.640 evidence diagnosis, those kinds of things, which was simply invalid. And when you try to apply 21:34.640 --> 21:40.960 invalid reasoning methods to real problems, you can fudge it for small versions of the problem. 21:40.960 --> 21:47.200 But when it starts to get larger, the thing just falls apart. So many companies found that 21:49.040 --> 21:53.440 the stuff just didn't work. And they were spending tons of money on consultants to 21:53.440 --> 21:59.600 try to make it work. And there were other practical reasons, like they were asking 21:59.600 --> 22:07.760 the companies to buy incredibly expensive Lisp machine workstations, which were literally 22:07.760 --> 22:17.680 between $50,000 and $100,000 in 1980s money, which would be between $150,000 and $300,000 per 22:17.680 --> 22:24.000 workstation in current prices. Then the bottom line, they weren't seeing a profit from it. 22:24.000 --> 22:29.840 Yeah. In many cases, I think there were some successes. There's no doubt about that. But 22:30.880 --> 22:37.760 people, I would say, over invested. Every major company was starting an AI department just like 22:37.760 --> 22:45.840 now. And I worry a bit that we might see similar disappointments, not because the 22:45.840 --> 22:57.600 current technology is invalid, but it's limited in its scope. And it's almost the dual of the 22:57.600 --> 23:03.360 scope problems that expert systems had. What have you learned from that hype cycle? And 23:03.360 --> 23:09.760 what can we do to prevent another winter, for example? Yeah. So when I'm giving talks these 23:09.760 --> 23:17.520 days, that's one of the warnings that I give. So there's two part warning slide. One is that 23:18.480 --> 23:24.000 rather than data being the new oil, data is the new snake oil. That's a good line. And then 23:26.000 --> 23:35.440 the other is that we might see a very visible failure in some of the major application areas. 23:35.440 --> 23:42.400 And I think self driving cars would be the flagship. And I think 23:43.600 --> 23:48.560 when you look at the history, so the first self driving car was on the freeway, 23:51.200 --> 24:00.400 driving itself, changing lanes, overtaking in 1987. And so it's more than 30 years. 24:00.400 --> 24:06.720 And that kind of looks like where we are today, right? Prototypes on the freeway, 24:06.720 --> 24:13.760 changing lanes and overtaking. Now, I think significant progress has been made, particularly 24:13.760 --> 24:20.560 on the perception side. So we worked a lot on autonomous vehicles in the early, mid 90s at 24:20.560 --> 24:29.040 Berkeley. And we had our own big demonstrations. We put congressmen into self driving cars and 24:29.040 --> 24:36.000 had them zooming along the freeway. And the problem was clearly perception. 24:37.520 --> 24:42.880 At the time, the problem was perception. Yeah. So in simulation, with perfect perception, 24:42.880 --> 24:47.200 you could actually show that you can drive safely for a long time, even if the other cars 24:47.200 --> 24:55.360 are misbehaving and so on. But simultaneously, we worked on machine vision for detecting cars and 24:55.360 --> 25:03.040 tracking pedestrians and so on. And we couldn't get the reliability of detection and tracking 25:03.040 --> 25:11.440 up to a high enough level, particularly in bad weather conditions, nighttime rainfall. 25:11.440 --> 25:16.000 Good enough for demos, but perhaps not good enough to cover the general operation. 25:16.000 --> 25:20.800 Yeah. So the thing about driving is, so suppose you're a taxi driver and you drive every day, 25:20.800 --> 25:27.360 eight hours a day for 10 years, that's 100 million seconds of driving. And any one of those 25:27.360 --> 25:33.280 seconds, you can make a fatal mistake. So you're talking about eight nines of reliability. 25:34.960 --> 25:43.840 Now, if your vision system only detects 98.3% of the vehicles, that's sort of one 25:43.840 --> 25:52.720 on a bit nine reliability. So you have another seven orders of magnitude to go. And this is 25:52.720 --> 25:57.920 what people don't understand. They think, oh, because I had a successful demo, I'm pretty much 25:57.920 --> 26:07.440 done. But you're not even within seven orders of magnitude of being done. And that's the difficulty. 26:07.440 --> 26:14.320 And it's not, can I follow a white line? That's not the problem. We follow a white line all the 26:14.320 --> 26:22.160 way across the country. But it's the weird stuff that happens. It's all the edge cases. Yeah. 26:22.160 --> 26:30.640 The edge case, other drivers doing weird things. So if you talk to Google, so they had actually 26:30.640 --> 26:36.560 a very classical architecture where you had machine vision, which would detect all the 26:36.560 --> 26:41.920 other cars and pedestrians and the white lines and the road signs. And then basically, 26:42.480 --> 26:49.680 that was fed into a logical database. And then you had a classical 1970s rule based expert system 26:52.000 --> 26:55.680 telling you, okay, if you're in the middle lane, and there's a bicyclist in the right lane, 26:55.680 --> 27:03.040 who is signaling this, then then do that, right? And what they found was that every day that go 27:03.040 --> 27:07.760 out and there'd be another situation that the rules didn't cover. So they come to a traffic 27:07.760 --> 27:11.680 circle and there's a little girl riding her bicycle the wrong way around the traffic circle. 27:11.680 --> 27:17.520 Okay, what do you do? We don't have a rule. Oh my God. Okay, stop. And then they come back 27:17.520 --> 27:24.400 and add more rules. And they just found that this was not really converging. And if you think about 27:24.400 --> 27:31.280 it, right, how do you deal with an unexpected situation, meaning one that you've never previously 27:31.280 --> 27:37.200 encountered and the sort of the reasoning required to figure out the solution for that 27:37.200 --> 27:42.800 situation has never been done. It doesn't match any previous situation in terms of the kind of 27:42.800 --> 27:49.520 reasoning you have to do. Well, in chess programs, this happens all the time. You're constantly 27:49.520 --> 27:54.560 coming up with situations you haven't seen before. And you have to reason about them and you have 27:54.560 --> 27:59.840 to think about, okay, here are the possible things I could do. Here are the outcomes. Here's how 27:59.840 --> 28:04.560 desirable the outcomes are and then pick the right one. In the 90s, we were saying, okay, 28:04.560 --> 28:08.160 this is how you're going to have to do automated vehicles. They're going to have to have a look 28:08.160 --> 28:14.400 ahead capability. But the look ahead for driving is more difficult than it is for chess. Because 28:14.400 --> 28:20.720 of humans. Right, there's humans and they're less predictable than chess pieces. Well, 28:20.720 --> 28:28.240 then you have an opponent in chess who's also somewhat unpredictable. But for example, in chess, 28:28.240 --> 28:33.600 you always know the opponent's intention. They're trying to beat you. Whereas in driving, you don't 28:33.600 --> 28:39.040 know, is this guy trying to turn left or has he just forgotten to turn off his turn signal? Or is 28:39.040 --> 28:45.680 he drunk? Or is he changing the channel on his radio or whatever it might be, you got to try and 28:45.680 --> 28:52.560 figure out the mental state, the intent of the other drivers to forecast the possible evolutions 28:52.560 --> 28:58.160 of their trajectories. And then you got to figure out, okay, which is the trajectory for me that's 28:58.160 --> 29:04.000 going to be safest. And those all interact with each other because the other drivers are going 29:04.000 --> 29:09.120 to react to your trajectory and so on. So, you know, they've got the classic merging onto the 29:09.120 --> 29:14.640 freeway problem where you're kind of racing a vehicle that's already on the freeway and you're 29:14.640 --> 29:17.680 are you going to pull ahead of them or are you going to let them go first and pull in behind 29:17.680 --> 29:23.680 and you get this sort of uncertainty about who's going first. So all those kinds of things 29:23.680 --> 29:34.720 mean that you need a decision making architecture that's very different from either a rule based 29:34.720 --> 29:41.360 system or it seems to me a kind of an end to end neural network system. You know, so just as Alpha 29:41.360 --> 29:47.360 Go is pretty good when it doesn't do any look ahead, but it's way, way, way, way better when it does. 29:47.360 --> 29:52.720 I think the same is going to be true for driving. You can have a driving system that's pretty good 29:54.080 --> 29:59.280 when it doesn't do any look ahead, but that's not good enough. You know, and we've already seen 29:59.920 --> 30:07.440 multiple deaths caused by poorly designed machine learning algorithms that don't really 30:07.440 --> 30:13.600 understand what they're doing. Yeah, and on several levels, I think it's on the perception side, 30:13.600 --> 30:19.520 there's mistakes being made by those algorithms where the perception is very shallow on the 30:19.520 --> 30:26.720 planning side, the look ahead, like you said, and the thing that we come up against that's 30:28.560 --> 30:32.080 really interesting when you try to deploy systems in the real world is 30:33.280 --> 30:37.680 you can't think of an artificial intelligence system as a thing that responds to the world always. 30:38.320 --> 30:41.600 You have to realize that it's an agent that others will respond to as well. 30:41.600 --> 30:47.200 Well, so in order to drive successfully, you can't just try to do obstacle avoidance. 30:47.840 --> 30:51.520 You can't pretend that you're invisible, right? You're the invisible car. 30:52.400 --> 30:57.280 It doesn't work that way. I mean, but you have to assert, yet others have to be scared of you, 30:57.280 --> 31:04.160 just there's this tension, there's this game. So we study a lot of work with pedestrians. 31:04.160 --> 31:09.360 If you approach pedestrians as purely an obstacle avoidance, so you're doing look 31:09.360 --> 31:15.040 ahead as in modeling the intent, they're not going to take advantage of you. 31:15.040 --> 31:20.080 They're not going to respect you at all. There has to be a tension, a fear, some amount of 31:20.080 --> 31:26.720 uncertainty. That's how we have created. Or at least just a kind of a resoluteness. 31:28.000 --> 31:32.000 You have to display a certain amount of resoluteness. You can't be too tentative. 31:32.000 --> 31:42.480 Yeah. So the solutions then become pretty complicated. You get into game theoretic 31:42.480 --> 31:50.960 analyses. So at Berkeley now, we're working a lot on this kind of interaction between machines 31:50.960 --> 32:03.600 and humans. And that's exciting. So my colleague, Anka Dragan, actually, if you formulate the problem 32:03.600 --> 32:08.800 game theoretically and you just let the system figure out the solution, it does interesting, 32:08.800 --> 32:16.640 unexpected things. Like sometimes at a stop sign, if no one is going first, the car will 32:16.640 --> 32:23.200 actually back up a little. It's just to indicate to the other cars that they should go. And that's 32:23.200 --> 32:28.480 something it invented entirely by itself. That's interesting. We didn't say this is the language 32:28.480 --> 32:36.240 of communication at stop signs. It figured it out. That's really interesting. So let me one just 32:36.240 --> 32:42.960 step back for a second. Just this beautiful philosophical notion. So Pamela McCordick in 32:42.960 --> 32:50.320 1979 wrote AI began with the ancient wish to forge the gods. So when you think about the 32:50.320 --> 32:57.520 history of our civilization, do you think that there is an inherent desire to create, 32:58.960 --> 33:05.680 let's not say gods, but to create superintelligence? Is it inherent to us? Is it in our genes, 33:05.680 --> 33:13.680 that the natural arc of human civilization is to create things that are of greater and greater 33:13.680 --> 33:21.680 power and perhaps echoes of ourselves? So to create the gods, as Pamela said. 33:21.680 --> 33:34.160 It may be. I mean, we're all individuals, but certainly we see over and over again in history 33:35.760 --> 33:42.320 individuals who thought about this possibility. Hopefully, I'm not being too philosophical here. 33:42.320 --> 33:48.560 But if you look at the arc of this, where this is going and we'll talk about AI safety, 33:48.560 --> 33:55.840 we'll talk about greater and greater intelligence, do you see that when you created the Othello 33:55.840 --> 34:01.680 program and you felt this excitement, what was that excitement? Was it the excitement of a tinkerer 34:01.680 --> 34:10.240 who created something cool, like a clock? Or was there a magic, or was it more like a child being 34:10.240 --> 34:17.520 born? Yeah. So I mean, I certainly understand that viewpoint. And if you look at the light 34:17.520 --> 34:26.640 hill report, so in the 70s, there was a lot of controversy in the UK about AI and whether it 34:26.640 --> 34:34.720 was for real and how much the money the government should invest. So it's a long story, but the 34:34.720 --> 34:43.280 government commissioned a report by Lighthill, who was a physicist, and he wrote a very damning 34:43.280 --> 34:53.920 report about AI, which I think was the point. And he said that these are frustrated men who 34:54.480 --> 35:05.760 unable to have children would like to create life as a kind of replacement, which I think is 35:05.760 --> 35:21.600 really pretty unfair. But there is a kind of magic, I would say, when you build something 35:25.680 --> 35:29.760 and what you're building in is really just you're building in some understanding of the 35:29.760 --> 35:37.120 principles of learning and decision making. And to see those principles actually then 35:37.840 --> 35:47.920 turn into intelligent behavior in specific situations, it's an incredible thing. And 35:47.920 --> 35:58.480 that is naturally going to make you think, okay, where does this end? 36:00.080 --> 36:08.240 And so there's a there's magical, optimistic views of word and whatever your view of optimism is, 36:08.240 --> 36:13.360 whatever your view of utopia is, it's probably different for everybody. But you've often talked 36:13.360 --> 36:26.080 about concerns you have of how things might go wrong. So I've talked to Max Tegmark. There's a 36:26.080 --> 36:33.360 lot of interesting ways to think about AI safety. You're one of the seminal people thinking about 36:33.360 --> 36:39.360 this problem amongst sort of being in the weeds of actually solving specific AI problems, 36:39.360 --> 36:44.080 you're also thinking about the big picture of where we're going. So can you talk about 36:44.080 --> 36:49.200 several elements of it? Let's just talk about maybe the control problem. So this idea of 36:50.800 --> 36:58.720 losing ability to control the behavior of our AI system. So how do you see that? How do you see 36:58.720 --> 37:04.480 that coming about? What do you think we can do to manage it? 37:04.480 --> 37:11.520 Well, so it doesn't take a genius to realize that if you make something that's smarter than you, 37:11.520 --> 37:20.320 you might have a problem. Alan Turing wrote about this and gave lectures about this, 37:21.600 --> 37:32.480 1951. He did a lecture on the radio. And he basically says, once the machine thinking method 37:32.480 --> 37:45.600 starts, very quickly, they'll outstrip humanity. And if we're lucky, we might be able to turn off 37:45.600 --> 37:52.160 the power at strategic moments, but even so, our species would be humbled. And actually, 37:52.160 --> 37:56.240 I think it was wrong about that. If it's a sufficiently intelligent machine, it's not 37:56.240 --> 38:00.160 going to let you switch it off. It's actually in competition with you. 38:00.160 --> 38:05.840 So what do you think is meant just for a quick tangent if we shut off this 38:05.840 --> 38:08.800 super intelligent machine that our species would be humbled? 38:11.840 --> 38:20.560 I think he means that we would realize that we are inferior, that we only survive by the skin 38:20.560 --> 38:27.440 of our teeth because we happen to get to the off switch just in time. And if we hadn't, 38:27.440 --> 38:34.400 then we would have lost control over the earth. So are you more worried when you think about 38:34.400 --> 38:41.600 this stuff about super intelligent AI or are you more worried about super powerful AI that's not 38:41.600 --> 38:49.760 aligned with our values? So the paperclip scenarios kind of... I think so the main problem I'm 38:49.760 --> 38:58.960 working on is the control problem, the problem of machines pursuing objectives that are, as you 38:58.960 --> 39:06.720 say, not aligned with human objectives. And this has been the way we've thought about AI 39:06.720 --> 39:15.120 since the beginning. You build a machine for optimizing and then you put in some objective 39:15.120 --> 39:25.520 and it optimizes. And we can think of this as the king Midas problem. Because if the king Midas 39:26.480 --> 39:32.640 put in this objective, everything I touch should turn to gold and the gods, that's like the machine, 39:32.640 --> 39:39.360 they said, okay, done. You now have this power and of course his food and his drink and his family 39:39.360 --> 39:50.080 all turned to gold and then he dies of misery and starvation. It's a warning, it's a failure mode that 39:50.080 --> 39:56.160 pretty much every culture in history has had some story along the same lines. There's the 39:56.160 --> 40:01.920 genie that gives you three wishes and third wish is always, please undo the first two wishes because 40:01.920 --> 40:11.920 I messed up. And when Arthur Samuel wrote his checker playing program, which learned to play 40:11.920 --> 40:16.800 checkers considerably better than Arthur Samuel could play and actually reached a pretty decent 40:16.800 --> 40:25.040 standard, Norbert Wiener, who was one of the major mathematicians of the 20th century, he's sort of 40:25.040 --> 40:32.560 the father of modern automation control systems. He saw this and he basically extrapolated 40:33.360 --> 40:43.680 as Turing did and said, okay, this is how we could lose control. And specifically that 40:45.520 --> 40:50.960 we have to be certain that the purpose we put into the machine is the purpose which we really 40:50.960 --> 40:59.760 desire. And the problem is, we can't do that. Right. You mean we're not, it's a very difficult 40:59.760 --> 41:05.440 to encode, to put our values on paper is really difficult, or you're just saying it's impossible? 41:09.120 --> 41:15.360 The line is great between the two. So theoretically, it's possible, but in practice, 41:15.360 --> 41:23.520 it's extremely unlikely that we could specify correctly in advance the full range of concerns 41:23.520 --> 41:29.360 of humanity. You talked about cultural transmission of values, I think is how humans to human 41:29.360 --> 41:36.320 transmission of values happens, right? Well, we learn, yeah, I mean, as we grow up, we learn about 41:36.320 --> 41:42.640 the values that matter, how things should go, what is reasonable to pursue and what isn't 41:42.640 --> 41:47.920 reasonable to pursue. I think machines can learn in the same kind of way. Yeah. So I think that 41:49.120 --> 41:54.480 what we need to do is to get away from this idea that you build an optimizing machine and then you 41:54.480 --> 42:03.200 put the objective into it. Because if it's possible that you might put in a wrong objective, and we 42:03.200 --> 42:08.880 already know this is possible because it's happened lots of times, right? That means that the machine 42:08.880 --> 42:17.760 should never take an objective that's given as gospel truth. Because once it takes the objective 42:17.760 --> 42:26.800 as gospel truth, then it believes that whatever actions it's taking in pursuit of that objective 42:26.800 --> 42:31.200 are the correct things to do. So you could be jumping up and down and saying, no, no, no, no, 42:31.200 --> 42:36.480 you're going to destroy the world, but the machine knows what the true objective is and is pursuing 42:36.480 --> 42:42.640 it and tough luck to you. And this is not restricted to AI, right? This is, I think, 42:43.360 --> 42:48.880 many of the 20th century technologies, right? So in statistics, you minimize a loss function, 42:48.880 --> 42:54.320 the loss function is exogenously specified in control theory, you minimize a cost function, 42:54.320 --> 42:59.840 in operations research, you maximize a reward function, and so on. So in all these disciplines, 42:59.840 --> 43:08.560 this is how we conceive of the problem. And it's the wrong problem. Because we cannot specify 43:08.560 --> 43:15.360 with certainty the correct objective, right? We need uncertainty, we need the machine to be 43:15.360 --> 43:19.440 uncertain about what it is that it's supposed to be maximizing. 43:19.440 --> 43:25.200 It's my favorite idea of yours. I've heard you say somewhere, well, I shouldn't pick favorites, 43:25.200 --> 43:32.640 but it just sounds beautiful. We need to teach machines humility. It's a beautiful way to put 43:32.640 --> 43:40.320 it. I love it. That they're humble. They know that they don't know what it is they're supposed 43:40.320 --> 43:48.240 to be doing. And that those objectives, I mean, they exist, they're within us, but we may not 43:48.240 --> 43:56.160 be able to explicate them. We may not even know how we want our future to go. 43:57.040 --> 44:06.800 So exactly. And a machine that's uncertain is going to be differential to us. So if we say, 44:06.800 --> 44:11.840 don't do that, well, now the machines learn something a bit more about our true objectives, 44:11.840 --> 44:16.480 because something that it thought was reasonable in pursuit of our objective, 44:16.480 --> 44:20.800 it turns out not to be so now it's learned something. So it's going to defer because it 44:20.800 --> 44:30.240 wants to be doing what we really want. And that point, I think, is absolutely central 44:30.240 --> 44:37.920 to solving the control problem. And it's a different kind of AI when you take away this 44:37.920 --> 44:44.560 idea that the objective is known, then in fact, a lot of the theoretical frameworks that we're so 44:44.560 --> 44:53.520 familiar with, you know, mark off decision processes, goal based planning, you know, 44:53.520 --> 44:59.280 standard game research, all of these techniques actually become inapplicable. 45:01.040 --> 45:11.120 And you get a more complicated problem because because now the interaction with the human becomes 45:11.120 --> 45:20.400 part of the problem. Because the human by making choices is giving you more information about 45:21.280 --> 45:25.360 the true objective and that information helps you achieve the objective better. 45:26.640 --> 45:32.000 And so that really means that you're mostly dealing with game theoretic problems where you've 45:32.000 --> 45:38.000 got the machine and the human and they're coupled together, rather than a machine going off by itself 45:38.000 --> 45:43.600 with a fixed objective. Which is fascinating on the machine and the human level that we, 45:44.400 --> 45:51.920 when you don't have an objective means you're together coming up with an objective. I mean, 45:51.920 --> 45:56.160 there's a lot of philosophy that, you know, you could argue that life doesn't really have meaning. 45:56.160 --> 46:01.680 We we together agree on what gives it meaning and we kind of culturally create 46:01.680 --> 46:08.560 things that give why the heck we are in this earth anyway. We together as a society create 46:08.560 --> 46:13.680 that meaning and you have to learn that objective. And one of the biggest, I thought that's where 46:13.680 --> 46:19.200 you were going to go for a second. One of the biggest troubles we run into outside of statistics 46:19.200 --> 46:26.240 and machine learning and AI in just human civilization is when you look at, I came from, 46:26.240 --> 46:32.160 I was born in the Soviet Union. And the history of the 20th century, we ran into the most trouble, 46:32.160 --> 46:40.160 us humans, when there was a certainty about the objective. And you do whatever it takes to achieve 46:40.160 --> 46:46.480 that objective, whether you're talking about Germany or communist Russia, you get into trouble 46:46.480 --> 46:52.960 with humans. And I would say with corporations, in fact, some people argue that we don't have 46:52.960 --> 46:57.840 to look forward to a time when AI systems take over the world, they already have. And they call 46:57.840 --> 47:04.880 corporations, right? That corporations happen to be using people as components right now. 47:05.920 --> 47:11.680 But they are effectively algorithmic machines, and they're optimizing an objective, which is 47:11.680 --> 47:18.080 quarterly profit that isn't aligned with overall well being of the human race. And they are 47:18.080 --> 47:24.160 destroying the world. They are primarily responsible for our inability to tackle climate change. 47:24.960 --> 47:30.400 So I think that's one way of thinking about what's going on with corporations. But 47:31.840 --> 47:39.680 I think the point you're making is valid, that there are many systems in the real world where 47:39.680 --> 47:48.480 we've sort of prematurely fixed on the objective and then decoupled the machine from those that 47:48.480 --> 47:54.720 are supposed to be serving. And I think you see this with government, right? Government is supposed 47:54.720 --> 48:02.720 to be a machine that serves people. But instead, it tends to be taken over by people who have their 48:02.720 --> 48:08.160 own objective and use government to optimize that objective, regardless of what people want. 48:08.160 --> 48:16.080 Do you find appealing the idea of almost arguing machines where you have multiple AI systems with 48:16.080 --> 48:22.400 a clear fixed objective? We have in government the red team and the blue team that are very fixed 48:22.400 --> 48:28.240 on their objectives. And they argue, and it kind of maybe would disagree, but it kind of seems to 48:28.240 --> 48:39.440 make it work somewhat that the duality of it, okay, let's go 100 years back when there was still 48:39.440 --> 48:44.480 was going on or at the founding of this country, there was disagreements and that disagreement is 48:44.480 --> 48:52.160 where so it was a balance between certainty and forced humility because the power was distributed. 48:52.160 --> 49:05.280 Yeah, I think that the nature of debate and disagreement argument takes as a premise the idea 49:05.280 --> 49:12.800 that you could be wrong, which means that you're not necessarily absolutely convinced that your 49:12.800 --> 49:19.520 objective is the correct one. If you were absolutely convinced, there'd be no point 49:19.520 --> 49:24.160 in having any discussion or argument because you would never change your mind. And there wouldn't 49:24.160 --> 49:32.080 be any sort of synthesis or anything like that. So I think you can think of argumentation as an 49:32.080 --> 49:44.640 implementation of a form of uncertain reasoning. I've been reading recently about utilitarianism 49:44.640 --> 49:54.960 and the history of efforts to define in a sort of clear mathematical way a if you like a formula for 49:54.960 --> 50:02.320 moral or political decision making. And it's really interesting that the parallels between 50:02.320 --> 50:08.720 the philosophical discussions going back 200 years and what you see now in discussions about 50:08.720 --> 50:15.040 existential risk because it's almost exactly the same. So someone would say, okay, well, 50:15.040 --> 50:21.600 here's a formula for how we should make decisions. So utilitarianism is roughly each person has a 50:21.600 --> 50:27.680 utility function and then we make decisions to maximize the sum of everybody's utility. 50:28.720 --> 50:36.480 And then people point out, well, in that case, the best policy is one that leads to 50:36.480 --> 50:42.480 the enormously vast population, all of whom are living a life that's barely worth living. 50:43.520 --> 50:50.640 And this is called the repugnant conclusion. And another version is that we should maximize 50:51.200 --> 50:57.680 pleasure and that's what we mean by utility. And then you'll get people effectively saying, 50:57.680 --> 51:02.480 well, in that case, we might as well just have everyone hooked up to a heroin drip. And they 51:02.480 --> 51:08.720 didn't use those words. But that debate was happening in the 19th century, as it is now 51:09.920 --> 51:17.600 about AI, that if we get the formula wrong, we're going to have AI systems working towards 51:17.600 --> 51:22.080 an outcome that in retrospect, would be exactly wrong. 51:22.080 --> 51:26.960 Do you think there's has beautifully put so the echoes are there. But do you think, 51:26.960 --> 51:34.640 I mean, if you look at Sam Harris, our imagination worries about the AI version of that, because 51:34.640 --> 51:44.640 of the speed at which the things going wrong in the utilitarian context could happen. 51:45.840 --> 51:47.280 Is that a worry for you? 51:47.280 --> 51:55.360 Yeah, I think that in most cases, not in all, but if we have a wrong political idea, 51:55.360 --> 52:01.200 we see it starting to go wrong. And we're not completely stupid. And so we said, okay, 52:02.000 --> 52:10.160 maybe that was a mistake. Let's try something different. And also, we're very slow and inefficient 52:10.160 --> 52:15.520 about implementing these things and so on. So you have to worry when you have corporations 52:15.520 --> 52:20.800 or political systems that are extremely efficient. But when we look at AI systems, 52:20.800 --> 52:27.840 or even just computers in general, right, they have this different characteristic 52:28.400 --> 52:35.040 from ordinary human activity in the past. So let's say you were a surgeon. You had some idea 52:35.040 --> 52:40.480 about how to do some operation, right? Well, and let's say you were wrong, right, that that way 52:40.480 --> 52:45.840 of doing the operation would mostly kill the patient. Well, you'd find out pretty quickly, 52:45.840 --> 52:56.000 like after three, maybe three or four tries, right? But that isn't true for pharmaceutical 52:56.000 --> 53:03.040 companies, because they don't do three or four operations. They manufacture three or four billion 53:03.040 --> 53:08.800 pills and they sell them. And then they find out maybe six months or a year later that, oh, 53:08.800 --> 53:14.880 people are dying of heart attacks or getting cancer from this drug. And so that's why we have the FDA, 53:14.880 --> 53:22.960 right? Because of the scalability of pharmaceutical production. And there have been some unbelievably 53:22.960 --> 53:34.320 bad episodes in the history of pharmaceuticals and adulteration of products and so on that have 53:34.320 --> 53:37.520 killed tens of thousands or paralyzed hundreds of thousands of people. 53:39.360 --> 53:43.280 Now, with computers, we have that same scalability problem that you can 53:43.280 --> 53:49.520 sit there and type for i equals one to five billion, two, right? And all of a sudden, 53:49.520 --> 53:55.360 you're having an impact on a global scale. And yet we have no FDA, right? There's absolutely no 53:55.360 --> 54:02.480 controls at all over what a bunch of undergraduates with too much caffeine can do to the world. 54:03.440 --> 54:09.600 And, you know, we look at what happened with Facebook, well, social media in general, and 54:09.600 --> 54:18.480 click through optimization. So you have a simple feedback algorithm that's trying to just optimize 54:18.480 --> 54:24.080 click through, right? That sounds reasonable, right? Because you don't want to be feeding people 54:24.080 --> 54:33.200 ads that they don't care about or not interested in. And you might even think of that process as 54:33.200 --> 54:42.160 simply adjusting the the feeding of ads or news articles or whatever it might be to match people's 54:42.160 --> 54:50.000 preferences, right? Which sounds like a good idea. But in fact, that isn't how the algorithm works, 54:50.880 --> 54:59.760 right? You make more money. The algorithm makes more money. If it can better predict what people 54:59.760 --> 55:06.400 are going to click on, because then it can feed them exactly that, right? So the way to maximize 55:06.400 --> 55:13.360 click through is actually to modify the people, to make them more predictable. And one way to do 55:13.360 --> 55:21.280 that is to feed them information which will change their behavior and preferences towards 55:21.920 --> 55:27.600 extremes that make them predictable. Whatever is the nearest extreme or the nearest predictable 55:27.600 --> 55:33.840 point, that's where you're going to end up. And the machines will force you there. 55:34.480 --> 55:40.400 Now, and I think there's a reasonable argument to say that this, among other things, is 55:40.400 --> 55:48.880 contributing to the destruction of democracy in the world. And where was the oversight 55:50.160 --> 55:55.600 of this process? Where were the people saying, okay, you would like to apply this algorithm to 55:55.600 --> 56:01.120 five billion people on the face of the earth? Can you show me that it's safe? Can you show 56:01.120 --> 56:07.040 me that it won't have various kinds of negative effects? No, there was no one asking that question. 56:07.040 --> 56:14.800 There was no one placed between the undergrads with too much caffeine and the human race. 56:15.520 --> 56:20.480 It's just they just did it. And some way outside the scope of my knowledge, 56:20.480 --> 56:27.120 so economists would argue that the invisible hand, so the capitalist system, it was the 56:27.120 --> 56:32.480 oversight. So if you're going to corrupt society with whatever decision you make as a company, 56:32.480 --> 56:38.640 then that's going to be reflected in people not using your product. That's one model of oversight. 56:39.280 --> 56:48.000 We shall see. But in the meantime, but you might even have broken the political system 56:48.000 --> 56:54.960 that enables capitalism to function. Well, you've changed it. So we should see. Yeah. 56:54.960 --> 57:01.360 Change is often painful. So my question is absolutely, it's fascinating. You're absolutely 57:01.360 --> 57:07.840 right that there was zero oversight on algorithms that can have a profound civilization changing 57:09.120 --> 57:15.440 effect. So do you think it's possible? I mean, I haven't, have you seen government? 57:15.440 --> 57:22.800 So do you think it's possible to create regulatory bodies oversight over AI algorithms, 57:22.800 --> 57:28.400 which are inherently such cutting edge set of ideas and technologies? 57:30.960 --> 57:37.520 Yeah, but I think it takes time to figure out what kind of oversight, what kinds of controls. 57:37.520 --> 57:42.960 I mean, it took time to design the FDA regime. Some people still don't like it and they want 57:42.960 --> 57:50.400 to fix it. And I think there are clear ways that it could be improved. But the whole notion that 57:50.400 --> 57:56.400 you have stage one, stage two, stage three, and here are the criteria for what you have to do 57:56.400 --> 58:02.000 to pass a stage one trial, right? We haven't even thought about what those would be for algorithms. 58:02.000 --> 58:10.320 So I mean, I think there are, there are things we could do right now with regard to bias, for 58:10.320 --> 58:19.040 example, we have a pretty good technical handle on how to detect algorithms that are propagating 58:19.040 --> 58:26.720 bias that exists in data sets, how to debias those algorithms, and even what it's going to cost you 58:26.720 --> 58:34.480 to do that. So I think we could start having some standards on that. I think there are things to do 58:34.480 --> 58:41.840 with impersonation and falsification that we could, we could work on. So I think, yeah. 58:43.360 --> 58:50.240 Or the very simple point. So impersonation is a machine acting as if it was a person. 58:51.440 --> 58:58.800 I can't see a real justification for why we shouldn't insist that machine self identify as 58:58.800 --> 59:08.480 machines. Where is the social benefit in fooling people into thinking that this is really a person 59:08.480 --> 59:15.120 when it isn't? I don't mind if it uses a human like voice that's easy to understand. That's fine. 59:15.120 --> 59:22.000 But it should just say, I'm a machine in some form. And now many people are speaking to that. 59:22.800 --> 59:26.400 I would think relatively obvious facts. So I think most people... Yeah. I mean, 59:26.400 --> 59:32.400 there is actually a law in California that bans impersonation, but only in certain 59:33.280 --> 59:41.280 restricted circumstances. So for the purpose of engaging in a fraudulent transaction and for 59:41.280 --> 59:48.160 the purpose of modifying someone's voting behavior. So those are the circumstances where 59:48.160 --> 59:55.520 machines have to self identify. But I think, arguably, it should be in all circumstances. 59:56.400 --> 1:00:03.280 And then when you talk about deep fakes, we're just at the beginning. But already, 1:00:03.280 --> 1:00:10.480 it's possible to make a movie of anybody saying anything in ways that are pretty hard to detect. 1:00:11.520 --> 1:00:15.040 Including yourself because you're on camera now and your voice is coming through with high 1:00:15.040 --> 1:00:19.360 resolution. Yeah. So you could take what I'm saying and replace it with pretty much anything 1:00:19.360 --> 1:00:24.400 else you wanted me to be saying. And even it would change my lips and facial expressions to fit. 1:00:26.960 --> 1:00:35.280 And there's actually not much in the way of real legal protection against that. 1:00:35.920 --> 1:00:38.640 I think in the commercial area, you could say, yeah, that's... 1:00:38.640 --> 1:00:46.000 You're using my brand and so on. There are rules about that. But in the political sphere, I think, 1:00:47.600 --> 1:00:52.560 at the moment, it's anything goes. So that could be really, really damaging. 1:00:53.920 --> 1:01:02.400 And let me just try to make not an argument, but try to look back at history and say something 1:01:02.400 --> 1:01:09.680 dark, in essence, is while regulation seems to be... Oversight seems to be exactly the 1:01:09.680 --> 1:01:14.480 right thing to do here. It seems that human beings, what they naturally do is they wait 1:01:14.480 --> 1:01:20.080 for something to go wrong. If you're talking about nuclear weapons, you can't talk about 1:01:20.080 --> 1:01:26.000 nuclear weapons being dangerous until somebody actually, like the United States drops the bomb, 1:01:26.000 --> 1:01:34.960 or Chernobyl melting. Do you think we will have to wait for things going wrong in a way that's 1:01:34.960 --> 1:01:39.840 obviously damaging to society, not an existential risk, but obviously damaging? 1:01:42.320 --> 1:01:48.000 Or do you have faith that... I hope not. But I mean, I think we do have to look at history. 1:01:48.000 --> 1:01:57.600 So the two examples you gave, nuclear weapons and nuclear power, are very, very interesting because 1:01:59.520 --> 1:02:07.840 nuclear weapons, we knew in the early years of the 20th century that atoms contained a huge 1:02:07.840 --> 1:02:13.280 amount of energy. We had E equals MC squared. We knew the mass differences between the different 1:02:13.280 --> 1:02:20.640 atoms and their components, and we knew that you might be able to make an incredibly powerful 1:02:20.640 --> 1:02:28.000 explosive. So H.G. Wells wrote science fiction book, I think, in 1912. Frederick Soddy, who was the 1:02:28.000 --> 1:02:35.760 guy who discovered isotopes as a Nobel Prize winner, he gave a speech in 1915 saying that 1:02:37.840 --> 1:02:42.160 one pound of this new explosive would be the equivalent of 150 tons of dynamite, 1:02:42.160 --> 1:02:50.720 which turns out to be about right. And this was in World War I, so he was imagining how much worse 1:02:51.360 --> 1:02:57.600 the world war would be if we were using that kind of explosive. But the physics establishment 1:02:57.600 --> 1:03:05.760 simply refused to believe that these things could be made. Including the people who were making it. 1:03:06.400 --> 1:03:11.280 Well, so they were doing the nuclear physics. I mean, eventually were the ones who made it. 1:03:11.280 --> 1:03:21.280 You talk about Fermi or whoever. Well, so up to the development was mostly theoretical. So it was 1:03:21.280 --> 1:03:28.320 people using sort of primitive kinds of particle acceleration and doing experiments at the level 1:03:28.320 --> 1:03:36.480 of single particles or collections of particles. They weren't yet thinking about how to actually 1:03:36.480 --> 1:03:40.160 make a bomb or anything like that. But they knew the energy was there and they figured if they 1:03:40.160 --> 1:03:46.720 understood it better, it might be possible. But the physics establishment, their view, and I think 1:03:46.720 --> 1:03:51.840 because they did not want it to be true, their view was that it could not be true. 1:03:53.360 --> 1:04:00.240 That this could not provide a way to make a super weapon. And there was this famous 1:04:01.120 --> 1:04:08.240 speech given by Rutherford, who was the sort of leader of nuclear physics. And it was on 1:04:08.240 --> 1:04:14.800 September 11, 1933. And he said, you know, anyone who talks about the possibility of 1:04:14.800 --> 1:04:21.600 obtaining energy from transformation of atoms is talking complete moonshine. And the next 1:04:22.800 --> 1:04:28.560 morning, Leo Zillard read about that speech and then invented the nuclear chain reaction. 1:04:28.560 --> 1:04:35.920 And so as soon as he invented, as soon as he had that idea, that you could make a chain reaction 1:04:35.920 --> 1:04:40.640 with neutrons because neutrons were not repelled by the nucleus so they could enter the nucleus 1:04:40.640 --> 1:04:48.480 and then continue the reaction. As soon as he has that idea, he instantly realized that the world 1:04:48.480 --> 1:04:58.160 was in deep doo doo. Because this is 1933, right? Hitler had recently come to power in Germany. 1:04:58.160 --> 1:05:09.280 Zillard was in London. He eventually became a refugee and he came to the US. And in the 1:05:09.280 --> 1:05:14.880 process of having the idea about the chain reaction, he figured out basically how to make 1:05:14.880 --> 1:05:22.800 a bomb and also how to make a reactor. And he patented the reactor in 1934. But because 1:05:22.800 --> 1:05:28.480 of the situation, the great power conflict situation that he could see happening, 1:05:29.200 --> 1:05:38.160 he kept that a secret. And so between then and the beginning of World War II, 1:05:39.600 --> 1:05:48.000 people were working, including the Germans, on how to actually create neutron sources, 1:05:48.000 --> 1:05:54.720 what specific fission reactions would produce neutrons of the right energy to continue the 1:05:54.720 --> 1:06:02.480 reaction. And that was demonstrated in Germany, I think in 1938, if I remember correctly. The first 1:06:03.760 --> 1:06:17.280 nuclear weapon patent was 1939 by the French. So this was actually going on well before World War 1:06:17.280 --> 1:06:22.720 II really got going. And then the British probably had the most advanced capability 1:06:22.720 --> 1:06:27.920 in this area. But for safety reasons, among others, and bless just sort of just resources, 1:06:28.720 --> 1:06:33.520 they moved the program from Britain to the US. And then that became Manhattan Project. 1:06:34.480 --> 1:06:37.920 So the reason why we couldn't 1:06:37.920 --> 1:06:48.320 have any kind of oversight of nuclear weapons and nuclear technology was because we were basically 1:06:48.320 --> 1:06:57.520 already in an arms race in a war. But you mentioned in the 20s and 30s, so what are the echoes? 1:07:00.000 --> 1:07:04.320 The way you've described the story, I mean, there's clearly echoes. What do you think most AI 1:07:04.320 --> 1:07:11.440 researchers, folks who are really close to the metal, they really are not concerned about AI, 1:07:11.440 --> 1:07:16.960 they don't think about it, whether it's they don't want to think about it. But what are the, 1:07:16.960 --> 1:07:23.760 yeah, why do you think that is? What are the echoes of the nuclear situation to the current AI 1:07:23.760 --> 1:07:33.120 situation? And what can we do about it? I think there is a kind of motivated cognition, which is 1:07:33.120 --> 1:07:40.640 a term in psychology means that you believe what you would like to be true, rather than what is 1:07:40.640 --> 1:07:50.240 true. And it's unsettling to think that what you're working on might be the end of the human race, 1:07:50.800 --> 1:07:58.400 obviously. So you would rather instantly deny it and come up with some reason why it couldn't be 1:07:58.400 --> 1:08:07.440 true. And I collected a long list of regions that extremely intelligent, competent AI scientists 1:08:08.080 --> 1:08:16.560 have come up with for why we shouldn't worry about this. For example, calculators are superhuman at 1:08:16.560 --> 1:08:21.680 arithmetic and they haven't taken over the world, so there's nothing to worry about. Well, okay, 1:08:21.680 --> 1:08:29.440 my five year old could have figured out why that was an unreasonable and really quite weak argument. 1:08:31.520 --> 1:08:40.800 Another one was, while it's theoretically possible that you could have superhuman AI 1:08:41.760 --> 1:08:46.480 destroy the world, it's also theoretically possible that a black hole could materialize 1:08:46.480 --> 1:08:51.760 right next to the earth and destroy humanity. I mean, yes, it's theoretically possible, 1:08:51.760 --> 1:08:56.720 quantum theoretically, extremely unlikely that it would just materialize right there. 1:08:58.400 --> 1:09:04.720 But that's a completely bogus analogy because if the whole physics community on earth was working 1:09:04.720 --> 1:09:11.680 to materialize a black hole in near earth orbit, wouldn't you ask them, is that a good idea? Is 1:09:11.680 --> 1:09:19.040 that going to be safe? What if you succeed? And that's the thing. The AI community is sort of 1:09:19.040 --> 1:09:26.240 refused to ask itself, what if you succeed? And initially, I think that was because it was too 1:09:26.240 --> 1:09:35.520 hard, but Alan Turing asked himself that and he said, we'd be toast. If we were lucky, we might 1:09:35.520 --> 1:09:40.000 be able to switch off the power but probably we'd be toast. But there's also an aspect 1:09:40.000 --> 1:09:49.680 that because we're not exactly sure what the future holds, it's not clear exactly so technically 1:09:49.680 --> 1:09:58.480 what to worry about, sort of how things go wrong. And so there is something it feels like, maybe 1:09:58.480 --> 1:10:04.400 you can correct me if I'm wrong, but there's something paralyzing about worrying about something 1:10:04.400 --> 1:10:10.000 that logically is inevitable. But you don't really know what that will look like. 1:10:10.720 --> 1:10:19.440 Yeah, I think that's a reasonable point. And it's certainly in terms of existential risks, 1:10:19.440 --> 1:10:26.800 it's different from asteroid collides with the earth, which again is quite possible. It's 1:10:26.800 --> 1:10:33.040 happened in the past. It'll probably happen again. We don't know right now. But if we did detect an 1:10:33.040 --> 1:10:39.200 asteroid that was going to hit the earth in 75 years time, we'd certainly be doing something 1:10:39.200 --> 1:10:43.600 about it. Well, it's clear there's got big rock and we'll probably have a meeting and see what 1:10:43.600 --> 1:10:49.040 do we do about the big rock with AI. Right, with AI. I mean, there are very few people who think it's 1:10:49.040 --> 1:10:54.400 not going to happen within the next 75 years. I know Rod Brooks doesn't think it's going to happen. 1:10:55.200 --> 1:11:00.880 Maybe Andrew Ng doesn't think it's happened. But a lot of the people who work day to day, 1:11:00.880 --> 1:11:07.360 you know, as you say, at the rock face, they think it's going to happen. I think the median 1:11:08.640 --> 1:11:14.320 estimate from AI researchers is somewhere in 40 to 50 years from now. Or maybe, you know, 1:11:14.320 --> 1:11:20.000 I think in Asia, they think it's going to be even faster than that. I'm a little bit 1:11:21.280 --> 1:11:25.840 more conservative. I think it probably take longer than that. But I think, you know, as 1:11:25.840 --> 1:11:32.400 happened with nuclear weapons, it can happen overnight that you have these breakthroughs. 1:11:32.400 --> 1:11:38.240 And we need more than one breakthrough. But, you know, it's on the order of half a dozen. 1:11:38.800 --> 1:11:43.920 This is a very rough scale. But so half a dozen breakthroughs of that nature 1:11:45.840 --> 1:11:53.600 would have to happen for us to reach superhuman AI. But the AI research community is 1:11:53.600 --> 1:12:00.640 vast now, the massive investments from governments, from corporations, tons of really, 1:12:00.640 --> 1:12:05.760 really smart people. You just have to look at the rate of progress in different areas of AI 1:12:05.760 --> 1:12:10.800 to see that things are moving pretty fast. So to say, oh, it's just going to be thousands of years. 1:12:11.920 --> 1:12:18.160 I don't see any basis for that. You know, I see, you know, for example, the 1:12:18.160 --> 1:12:28.640 Stanford 100 year AI project, which is supposed to be sort of, you know, the serious establishment view, 1:12:29.520 --> 1:12:33.440 their most recent report actually said it's probably not even possible. 1:12:34.160 --> 1:12:34.720 Oh, wow. 1:12:35.280 --> 1:12:42.960 Right. Which if you want a perfect example of people in denial, that's it. Because, you know, 1:12:42.960 --> 1:12:49.760 for the whole history of AI, we've been saying to philosophers who said it wasn't possible. Well, 1:12:49.760 --> 1:12:53.520 you have no idea what you're talking about. Of course, it's possible. Right. Give me an 1:12:53.520 --> 1:12:59.760 argument for why it couldn't happen. And there isn't one. Right. And now, because people are 1:12:59.760 --> 1:13:04.000 worried that maybe AI might get a bad name, or I just don't want to think about this, 1:13:05.200 --> 1:13:09.520 they're saying, okay, well, of course, it's not really possible. You know, imagine, right? Imagine 1:13:09.520 --> 1:13:16.000 if, you know, the leaders of the cancer biology community got up and said, well, you know, of 1:13:16.000 --> 1:13:23.680 course, curing cancer, it's not really possible. There'd be a complete outrage and dismay. And, 1:13:25.040 --> 1:13:30.800 you know, I find this really a strange phenomenon. So, 1:13:30.800 --> 1:13:38.640 okay, so if you accept it as possible, and if you accept that it's probably going to happen, 1:13:40.560 --> 1:13:44.400 the point that you're making that, you know, how does it go wrong? 1:13:46.320 --> 1:13:51.600 A valid question without that, without an answer to that question, then you're stuck with what I 1:13:51.600 --> 1:13:56.160 call the gorilla problem, which is, you know, the problem that the gorillas face, right? They 1:13:56.160 --> 1:14:02.800 made something more intelligent than them, namely us a few million years ago, and now they're in 1:14:02.800 --> 1:14:09.360 deep doo doo. So there's really nothing they can do. They've lost the control. They failed to solve 1:14:09.360 --> 1:14:16.240 the control problem of controlling humans. And so they've lost. So we don't want to be in that 1:14:16.240 --> 1:14:22.320 situation. And if the gorilla problem is the only formulation you have, there's not a lot you can do. 1:14:22.320 --> 1:14:28.240 Right. Other than to say, okay, we should try to stop. You know, we should just not make the humans 1:14:28.240 --> 1:14:33.120 or, or in this case, not make the AI. And I think that's really hard to do. 1:14:35.120 --> 1:14:42.480 To, I'm not actually proposing that that's a feasible course of action. And I also think that, 1:14:43.040 --> 1:14:46.080 you know, if properly controlled AI could be incredibly beneficial. 1:14:46.080 --> 1:14:54.960 So the, but it seems to me that there's a, there's a consensus that one of the major 1:14:54.960 --> 1:15:02.320 failure modes is this loss of control that we create AI systems that are pursuing incorrect 1:15:02.320 --> 1:15:11.040 objectives. And because the AI system believes it knows what the objective is, it has no incentive 1:15:11.040 --> 1:15:16.800 to listen to us anymore, so to speak, right? It's just carrying out the, 1:15:17.920 --> 1:15:22.160 the strategy that it, it has computed as being the optimal solution. 1:15:24.320 --> 1:15:31.040 And, you know, it may be that in the process, it needs to acquire more resources to increase the 1:15:31.600 --> 1:15:37.360 possibility of success or prevent various failure modes by defending itself against interference. 1:15:37.360 --> 1:15:42.640 And so that collection of problems, I think, is something we can address. 1:15:45.280 --> 1:15:55.360 The other problems are roughly speaking, you know, misuse, right? So even if we solve the control 1:15:55.360 --> 1:16:00.960 problem, we make perfectly safe controllable AI systems. Well, why, you know, why does Dr. 1:16:00.960 --> 1:16:05.600 Evil going to use those, right? He wants to just take over the world and he'll make unsafe AI systems 1:16:05.600 --> 1:16:12.000 that then get out of control. So that's one problem, which is sort of a, you know, partly a 1:16:12.000 --> 1:16:20.880 policing problem, partly a sort of a cultural problem for the profession of how we teach people 1:16:21.760 --> 1:16:26.560 what kinds of AI systems are safe. You talk about autonomous weapon system and how pretty much 1:16:26.560 --> 1:16:31.920 everybody agrees that there's too many ways that that can go horribly wrong. You have this great 1:16:31.920 --> 1:16:36.560 Slotabots movie that kind of illustrates that beautifully. Well, I want to talk about that. 1:16:36.560 --> 1:16:41.440 That's another, there's another topic I'm happy to talk about the, I just want to mention that 1:16:41.440 --> 1:16:48.080 what I see is the third major failure mode, which is overuse, not so much misuse, but overuse of AI, 1:16:49.680 --> 1:16:55.280 that we become overly dependent. So I call this the warly problems. If you've seen the warly, 1:16:55.280 --> 1:17:00.800 the movie, all right, all the humans are on the spaceship and the machines look after everything 1:17:00.800 --> 1:17:07.680 for them. And they just watch TV and drink big gulps. And they're all sort of obese and stupid. 1:17:07.680 --> 1:17:17.040 And they sort of totally lost any notion of human autonomy. And, you know, so in effect, right, 1:17:18.240 --> 1:17:23.520 this would happen like the slow boiling frog, right, we would gradually turn over 1:17:24.320 --> 1:17:28.480 more and more of the management of our civilization to machines as we are already doing. 1:17:28.480 --> 1:17:34.560 And this, you know, this, if this process continues, you know, we sort of gradually 1:17:34.560 --> 1:17:41.440 switch from sort of being the masters of technology to just being the guests, right? 1:17:41.440 --> 1:17:45.840 So, so we become guests on a cruise ship, you know, which is fine for a week, but not, 1:17:46.480 --> 1:17:53.520 not for the rest of eternity, right? You know, and it's almost irreversible, right? Once you, 1:17:53.520 --> 1:18:00.000 once you lose the incentive to, for example, you know, learn to be an engineer or a doctor 1:18:00.800 --> 1:18:08.000 or a sanitation operative or any other of the, the infinitely many ways that we 1:18:08.000 --> 1:18:13.200 maintain and propagate our civilization. You know, if you, if you don't have the 1:18:13.200 --> 1:18:18.400 incentive to do any of that, you won't. And then it's really hard to recover. 1:18:18.400 --> 1:18:23.440 And of course they add just one of the technologies that could, that third failure mode result in that. 1:18:23.440 --> 1:18:27.360 There's probably other technology in general detaches us from. 1:18:28.400 --> 1:18:33.440 It does a bit, but the, the, the difference is that in terms of the knowledge to, 1:18:34.080 --> 1:18:39.360 to run our civilization, you know, up to now we've had no alternative, but to put it into 1:18:39.360 --> 1:18:44.400 people's heads, right? And if you, if you, software with Google, I mean, so software in 1:18:44.400 --> 1:18:51.600 general, so computers in general, but, but the, you know, the knowledge of how, you know, how 1:18:51.600 --> 1:18:56.560 a sanitation system works, you know, that's an AI has to understand that it's no good putting it 1:18:56.560 --> 1:19:02.960 into Google. So, I mean, we, we've always put knowledge in on paper, but paper doesn't run 1:19:02.960 --> 1:19:07.520 our civilization. It only runs when it goes from the paper into people's heads again, right? So 1:19:07.520 --> 1:19:13.920 we've always propagated civilization through human minds and we've spent about a trillion 1:19:13.920 --> 1:19:19.440 person years doing that literally, right? You, you can work it out. It's about, right? There's 1:19:19.440 --> 1:19:25.120 about just over a hundred billion people who've ever lived and each of them has spent about 10 1:19:25.120 --> 1:19:30.640 years learning stuff to keep their civilization going. And so that's a trillion person years we 1:19:30.640 --> 1:19:35.760 put into this effort. Beautiful way to describe all of civilization. And now we're, you know, 1:19:35.760 --> 1:19:39.840 we're in danger of throwing that away. So this is a problem that AI can't solve. It's not a 1:19:39.840 --> 1:19:47.120 technical problem. It's a, you know, and if we do our job right, the AI systems will say, you know, 1:19:47.120 --> 1:19:52.800 the human race doesn't in the long run want to be passengers in a cruise ship. The human race 1:19:52.800 --> 1:19:59.840 wants autonomy. This is part of human preferences. So we, the AI systems are not going to do this 1:19:59.840 --> 1:20:05.440 stuff for you. You've got to do it for yourself, right? I'm not going to carry you to the top of 1:20:05.440 --> 1:20:11.840 Everest in an autonomous helicopter. You have to climb it if you want to get the benefit and so on. So 1:20:14.160 --> 1:20:20.160 but I'm afraid that because we are short sighted and lazy, we're going to override the AI systems. 1:20:20.880 --> 1:20:27.520 And, and there's an amazing short story that I recommend to everyone that I talk to about this 1:20:27.520 --> 1:20:36.080 called the machine stops written in 1909 by Ian Forster, who, you know, wrote novels about the 1:20:36.080 --> 1:20:41.200 British Empire and sort of things that became costume dramas on the BBC. But he wrote this one 1:20:41.200 --> 1:20:49.280 science fiction story, which is an amazing vision of the future. It has, it has basically iPads. 1:20:49.280 --> 1:20:57.680 It has video conferencing. It has MOOCs. It has computer and computer induced obesity. I mean, 1:20:57.680 --> 1:21:02.960 literally, the whole thing is what people spend their time doing is giving online courses or 1:21:02.960 --> 1:21:07.920 listening to online courses and talking about ideas. But they never get out there in the real 1:21:07.920 --> 1:21:13.680 world. They don't really have a lot of face to face contact. Everything is done online. 1:21:13.680 --> 1:21:19.680 You know, so all the things we're worrying about now were described in the story and and then the 1:21:19.680 --> 1:21:26.640 human race becomes more and more dependent on the machine loses knowledge of how things really run 1:21:27.600 --> 1:21:35.200 and then becomes vulnerable to collapse. And so it's a it's a pretty unbelievably amazing 1:21:35.200 --> 1:21:41.520 story for someone writing in 1909 to imagine all this. Plus, yeah. So there's very few people 1:21:41.520 --> 1:21:46.080 that represent artificial intelligence more than you, Stuart Russell. 1:21:46.960 --> 1:21:50.880 If you say it's okay, that's very kind. So it's all my fault. 1:21:50.880 --> 1:21:59.680 It's all your fault. No, right. You're often brought up as the person. Well, Stuart Russell, 1:22:00.560 --> 1:22:04.960 like the AI person is worried about this. That's why you should be worried about it. 1:22:06.080 --> 1:22:11.280 Do you feel the burden of that? I don't know if you feel that at all. But when I talk to people, 1:22:11.280 --> 1:22:16.800 like from you talk about people outside of computer science, when they think about this, 1:22:16.800 --> 1:22:22.560 Stuart Russell is worried about AI safety, you should be worried too. Do you feel the burden 1:22:22.560 --> 1:22:31.520 of that? I mean, in a practical sense, yeah, because I get, you know, a dozen, sometimes 1:22:31.520 --> 1:22:39.600 25 invitations a day to talk about it, to give interviews, to write press articles and so on. 1:22:39.600 --> 1:22:47.120 So in that very practical sense, I'm seeing that people are concerned and really interested about 1:22:47.120 --> 1:22:53.680 this. Are you worried that you could be wrong, as all good scientists are? Of course. I worry about 1:22:53.680 --> 1:23:00.400 that all the time. I mean, that's always been the way that I've worked, you know, is like I have an 1:23:00.400 --> 1:23:06.320 argument in my head with myself, right? So I have some idea. And then I think, okay, 1:23:06.320 --> 1:23:11.680 okay, how could that be wrong? Or did someone else already have that idea? So I'll go and 1:23:12.800 --> 1:23:18.240 search in as much literature as I can to see whether someone else already thought of that 1:23:18.240 --> 1:23:25.600 or even refuted it. So, you know, right now, I'm reading a lot of philosophy because, 1:23:25.600 --> 1:23:37.920 you know, in the form of the debates over utilitarianism and other kinds of moral formulas, 1:23:37.920 --> 1:23:44.320 shall we say, people have already thought through some of these issues. But, you know, 1:23:44.320 --> 1:23:51.280 what one of the things I'm not seeing in a lot of these debates is this specific idea about 1:23:51.280 --> 1:23:58.560 the importance of uncertainty in the objective, that this is the way we should think about machines 1:23:58.560 --> 1:24:06.800 that are beneficial to humans. So this idea of provably beneficial machines based on explicit 1:24:06.800 --> 1:24:15.200 uncertainty in the objective, you know, it seems to be, you know, my gut feeling is this is the core 1:24:15.200 --> 1:24:21.200 of it. It's going to have to be elaborated in a lot of different directions. And they're a lot 1:24:21.200 --> 1:24:27.440 of beneficial. Yeah, but they're, I mean, it has to be, right? We can't afford, you know, 1:24:27.440 --> 1:24:33.040 hand wavy beneficial. Because there are, you know, whenever we do hand wavy stuff, there are 1:24:33.040 --> 1:24:38.080 loopholes. And the thing about super intelligent machines is they find the loopholes. You know, 1:24:38.080 --> 1:24:44.320 just like, you know, tax evaders, if you don't write your tax law properly, people will find 1:24:44.320 --> 1:24:53.440 the loopholes and end up paying no tax. And so you should think of it this way. And getting those 1:24:53.440 --> 1:25:03.440 definitions right, you know, it is really a long process, you know, so you can you can define 1:25:03.440 --> 1:25:07.760 mathematical frameworks. And within that framework, you can prove mathematical theorems that, yes, 1:25:07.760 --> 1:25:12.800 this will, you know, this this theoretical entity will be provably beneficial to that theoretical 1:25:12.800 --> 1:25:20.160 entity. But that framework may not match the real world in some crucial way. So the long process 1:25:20.160 --> 1:25:27.120 thinking through it to iterating and so on. Last question. Yep. You have 10 seconds to answer it. 1:25:27.120 --> 1:25:34.480 What is your favorite sci fi movie about AI? I would say interstellar has my favorite robots. 1:25:34.480 --> 1:25:42.160 Oh, beats space. Yeah, yeah, yeah. So so Tars, the robots, one of the robots in interstellar is 1:25:42.160 --> 1:25:52.080 the way robots should behave. And I would say X Machina is in some ways the one, the one that 1:25:52.080 --> 1:25:58.000 makes you think in a nervous kind of way about about where we're going. 1:25:58.000 --> 1:26:13.920 Well, Stuart, thank you so much for talking today. Pleasure.