diff --git "a/vtt/episode_001_small.vtt" "b/vtt/episode_001_small.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_001_small.vtt" @@ -0,0 +1,3521 @@ +WEBVTT + +00:00.000 --> 00:05.060 + As part of MIT course 6S 099 Artificial General Intelligence, I've gotten the chance to sit + +00:05.060 --> 00:06.740 + down with Max Tagmark. + +00:06.740 --> 00:13.780 + He is a professor here at MIT, he's a physicist, spent a large part of his career studying the + +00:13.780 --> 00:20.660 + mysteries of our cosmological universe, but he's also studied and delved into the beneficial + +00:20.660 --> 00:25.860 + possibilities and the existential risks of artificial intelligence. + +00:25.860 --> 00:32.220 + Amongst many other things, he's the cofounder of the Future of Life Institute, author of + +00:32.220 --> 00:35.140 + two books, both of which I highly recommend. + +00:35.140 --> 00:40.220 + First, our mathematical universe, second is Life 3.0. + +00:40.220 --> 00:45.060 + He's truly an out of the box thinker and a fun personality, so I really enjoy talking + +00:45.060 --> 00:46.060 + to him. + +00:46.060 --> 00:49.500 + If you'd like to see more of these videos in the future, please subscribe and also click + +00:49.500 --> 00:52.980 + the little bell icon to make sure you don't miss any videos. + +00:52.980 --> 01:00.260 + Also, Twitter, LinkedIn, AGI.MIT.IDU, if you want to watch other lectures or conversations + +01:00.260 --> 01:01.260 + like this one. + +01:01.260 --> 01:07.980 + Better yet, go read Max's book, Life 3.0, chapter 7 on goals is my favorite. + +01:07.980 --> 01:12.300 + It's really where philosophy and engineering come together and it opens with a quote by + +01:12.300 --> 01:18.460 + Dostoevsky, the mystery of human existence lies not in just staying alive, but in finding + +01:18.460 --> 01:20.300 + something to live for. + +01:20.300 --> 01:27.100 + Lastly, I believe that every failure rewards us with an opportunity to learn, in that sense + +01:27.100 --> 01:33.060 + I've been very fortunate to fail in so many new and exciting ways and this conversation + +01:33.060 --> 01:34.060 + was no different. + +01:34.060 --> 01:41.260 + I've learned about something called Radio Frequency Interference, RFI, look it up. + +01:41.260 --> 01:45.500 + Apparently music and conversations from local radio stations can bleed into the audio that + +01:45.500 --> 01:49.380 + you're recording in such a way that almost completely ruins that audio. + +01:49.380 --> 01:52.460 + It's an exceptionally difficult sound source to remove. + +01:52.460 --> 01:59.620 + So, I've gotten the opportunity to learn how to avoid RFI in the future during recording + +01:59.620 --> 02:00.620 + sessions. + +02:00.620 --> 02:06.260 + I've also gotten the opportunity to learn how to use Adobe Audition and iZotope RX6 + +02:06.260 --> 02:11.740 + to do some audio repair. + +02:11.740 --> 02:14.940 + Of course, this is an exceptionally difficult noise to remove. + +02:14.940 --> 02:20.380 + I am an engineer, I'm not an audio engineer, neither is anybody else in our group, but + +02:20.380 --> 02:21.780 + we did our best. + +02:21.780 --> 02:26.780 + Nevertheless, I thank you for your patience and I hope you're still able to enjoy this + +02:26.780 --> 02:27.780 + conversation. + +02:27.780 --> 02:31.460 + Do you think there's intelligent life out there in the universe? + +02:31.460 --> 02:33.420 + Let's open up with an easy question. + +02:33.420 --> 02:36.260 + I have a minority view here actually. + +02:36.260 --> 02:41.180 + When I give public lectures, I often ask for show of hands who thinks there's intelligent + +02:41.180 --> 02:47.060 + life out there somewhere else and almost everyone puts their hands up and when I ask why, they'll + +02:47.060 --> 02:52.060 + be like, oh, there's so many galaxies out there, there's got to be. + +02:52.060 --> 02:54.660 + But I'm a number nerd, right? + +02:54.660 --> 02:59.180 + So when you look more carefully at it, it's not so clear at all. + +02:59.180 --> 03:03.140 + When we talk about our universe, first of all, we don't mean all of space. + +03:03.140 --> 03:05.900 + We actually mean, I don't know, you can throw me the universe if you want, it's behind you + +03:05.900 --> 03:06.900 + there. + +03:06.900 --> 03:14.540 + We simply mean the spherical region of space from which light has had time to reach us + +03:14.540 --> 03:19.460 + so far during the 13.8 billion years since our big bang. + +03:19.460 --> 03:23.020 + There's more space here, but this is what we call a universe because that's all we have + +03:23.020 --> 03:24.140 + access to. + +03:24.140 --> 03:31.220 + So is there intelligent life here that's gotten to the point of building telescopes and computers? + +03:31.220 --> 03:39.500 + My guess is no, actually, the probability of it happening on any given planet is some + +03:39.500 --> 03:42.860 + number we don't know what it is. + +03:42.860 --> 03:49.340 + And what we do know is that the number can't be super high because there's over a billion + +03:49.340 --> 03:54.780 + Earth like planets in the Milky Way galaxy alone, many of which are billions of years + +03:54.780 --> 04:01.740 + older than Earth, and aside from some UFO believers, you know, there isn't much evidence + +04:01.740 --> 04:05.740 + that any super advanced civilization has come here at all. + +04:05.740 --> 04:08.700 + And so that's the famous Fermi paradox, right? + +04:08.700 --> 04:13.620 + And then if you work the numbers, what you find is that if you have no clue what the + +04:13.620 --> 04:18.500 + probability is of getting life on a given planet, so it could be 10 to the minus 10, + +04:18.500 --> 04:23.620 + 10 to the minus 20, or 10 to the minus two, or any power of 10 is sort of equally likely + +04:23.620 --> 04:27.700 + if you want to be really open minded, that translates into it being equally likely that + +04:27.700 --> 04:34.700 + our nearest neighbor is 10 to the 16 meters away, 10 to the 17 meters away, 10 to the + +04:34.700 --> 04:35.700 + 18. + +04:35.700 --> 04:42.860 + Now, by the time you get much less than 10 to the 16 already, we pretty much know there + +04:42.860 --> 04:46.220 + is nothing else that's close. + +04:46.220 --> 04:49.740 + And when you get because it would have discovered us, they, yeah, they would have discovered + +04:49.740 --> 04:53.540 + us longer or if they're really close, we would have probably noted some engineering projects + +04:53.540 --> 04:54.540 + that they're doing. + +04:54.540 --> 05:00.140 + And if it's beyond 10 to the 26 meters, that's already outside of here. + +05:00.140 --> 05:06.340 + So my guess is actually that there are, we are the only life in here that's gotten the + +05:06.340 --> 05:14.020 + point of building advanced tech, which I think is very, puts a lot of responsibility on our + +05:14.020 --> 05:18.140 + shoulders, not screw up, you know, I think people who take for granted that it's okay + +05:18.140 --> 05:23.300 + for us to screw up, have an accidental nuclear war or go extinct somehow because there's + +05:23.300 --> 05:27.460 + a sort of Star Trek like situation out there where some other life forms are going to come + +05:27.460 --> 05:30.380 + and bail us out and it doesn't matter so much. + +05:30.380 --> 05:33.380 + I think they're leveling us into a false sense of security. + +05:33.380 --> 05:37.540 + I think it's much more prudent to say, let's be really grateful for this amazing opportunity + +05:37.540 --> 05:44.180 + we've had and make the best of it just in case it is down to us. + +05:44.180 --> 05:50.220 + So from a physics perspective, do you think intelligent life, so it's unique from a sort + +05:50.220 --> 05:55.860 + of statistical view of the size of the universe, but from the basic matter of the universe, + +05:55.860 --> 06:00.100 + how difficult is it for intelligent life to come about with the kind of advanced tech + +06:00.100 --> 06:06.300 + building life is implied in your statement that it's really difficult to create something + +06:06.300 --> 06:07.620 + like a human species? + +06:07.620 --> 06:14.740 + Well, I think what we know is that going from no life to having life that can do our level + +06:14.740 --> 06:21.140 + of tech, there's some sort of to going beyond that than actually settling our whole universe + +06:21.140 --> 06:22.300 + with life. + +06:22.300 --> 06:30.700 + There's some road major roadblock there, which is some great filter as I just sometimes called + +06:30.700 --> 06:37.180 + which, which tough to get through, it's either that that roadblock is either behind us or + +06:37.180 --> 06:38.620 + in front of us. + +06:38.620 --> 06:40.980 + I'm hoping very much that it's behind us. + +06:40.980 --> 06:46.900 + I'm super excited every time we get a new report from NASA saying they failed to find + +06:46.900 --> 06:53.260 + any life on Mars, because that suggests that the hard part, maybe it was getting the first + +06:53.260 --> 06:59.540 + ribosome or some some very low level kind of stepping stone. + +06:59.540 --> 07:03.620 + So they were home free because if that's true, then the future is really only limited by + +07:03.620 --> 07:04.620 + our own imagination. + +07:04.620 --> 07:11.460 + It would be much suckier if it turns out that this level of life is kind of a diamond dozen, + +07:11.460 --> 07:12.780 + but maybe there's some other problem. + +07:12.780 --> 07:17.220 + Like as soon as a civilization gets advanced technology within 100 years, they get into + +07:17.220 --> 07:21.740 + some stupid fight with themselves and poof, you know, that would be a bummer. + +07:21.740 --> 07:22.740 + Yeah. + +07:22.740 --> 07:28.980 + So you've explored the mysteries of the universe, the cosmological universe, the one that's + +07:28.980 --> 07:36.340 + between us today, I think you've also begun to explore the other universe, which is sort + +07:36.340 --> 07:42.860 + of the mystery, the mysterious universe of the mind of intelligence, of intelligent life. + +07:42.860 --> 07:48.260 + So is there a common thread between your interests or the way you think about space and intelligence? + +07:48.260 --> 07:49.260 + Oh, yeah. + +07:49.260 --> 07:57.700 + When I was a teenager, I was already very fascinated by the biggest questions and I felt that the + +07:57.700 --> 08:03.660 + two biggest mysteries of all in science were our universe out there and our universe in + +08:03.660 --> 08:04.660 + here. + +08:04.660 --> 08:05.660 + Yeah. + +08:05.660 --> 08:11.260 + So it's quite natural after having spent a quarter of a century on my career thinking + +08:11.260 --> 08:12.260 + a lot about this one. + +08:12.260 --> 08:15.980 + And now I'm indulging in the luxury of doing research on this one. + +08:15.980 --> 08:17.660 + It's just so cool. + +08:17.660 --> 08:25.260 + I feel the time is ripe now for you transparently deepening our understanding of this. + +08:25.260 --> 08:26.420 + Just start exploring this one. + +08:26.420 --> 08:32.500 + Yeah, because I think a lot of people view intelligence as something mysterious that + +08:32.500 --> 08:38.340 + can only exist in biological organisms like us and therefore dismiss all talk about artificial + +08:38.340 --> 08:41.260 + general intelligence is science fiction. + +08:41.260 --> 08:47.260 + But from my perspective as a physicist, I am a blob of quirks and electrons moving around + +08:47.260 --> 08:50.180 + in a certain pattern and processing information in certain ways. + +08:50.180 --> 08:53.580 + And this is also a blob of quirks and electrons. + +08:53.580 --> 08:57.860 + I'm not smarter than the water bottle because I'm made of different kind of quirks. + +08:57.860 --> 09:02.220 + I'm made of up quirks and down quirks exact same kind as this. + +09:02.220 --> 09:07.020 + It's a there's no secret sauce, I think in me, it's it's all about the pattern of the + +09:07.020 --> 09:08.820 + information processing. + +09:08.820 --> 09:16.020 + And this means that there's no law of physics saying that we can't create technology, which + +09:16.020 --> 09:21.740 + can help us by being incredibly intelligent and help us crack mysteries that we couldn't. + +09:21.740 --> 09:25.580 + In other words, I think we've really only seen the tip of the intelligence iceberg so + +09:25.580 --> 09:26.580 + far. + +09:26.580 --> 09:27.580 + Yeah. + +09:27.580 --> 09:34.380 + So the perceptronium, yeah, so you coined this amazing term, it's a hypothetical state + +09:34.380 --> 09:39.420 + of matter, sort of thinking from a physics perspective, what is the kind of matter that + +09:39.420 --> 09:44.500 + can help as you're saying, subjective experience emerge, consciousness emerge. + +09:44.500 --> 09:50.140 + So how do you think about consciousness from this physics perspective? + +09:50.140 --> 09:51.980 + Very good question. + +09:51.980 --> 10:03.060 + So, again, I think many people have underestimated our ability to make progress on this by convincing + +10:03.060 --> 10:08.500 + themselves it's hopeless because somehow we're missing some ingredient that we need. + +10:08.500 --> 10:13.020 + There's some new consciousness particle or whatever. + +10:13.020 --> 10:19.660 + I happen to think that we're not missing anything and that it's not the interesting thing about + +10:19.660 --> 10:25.900 + consciousness that gives us this amazing subjective experience of colors and sounds and emotions + +10:25.900 --> 10:32.300 + and so on is rather something at the higher level about the patterns of information processing. + +10:32.300 --> 10:38.300 + And that's why I like to think about this idea of perceptronium. + +10:38.300 --> 10:44.220 + What does it mean for an arbitrary physical system to be conscious in terms of what its + +10:44.220 --> 10:47.100 + particles are doing or its information is doing? + +10:47.100 --> 10:52.300 + I hate carbon chauvinism, this attitude, you have to be made of carbon atoms to be smart + +10:52.300 --> 10:53.300 + or conscious. + +10:53.300 --> 10:58.180 + So something about the information processing that this kind of matter performs. + +10:58.180 --> 11:02.700 + Yeah, and you can see I have my favorite equations here describing various fundamental + +11:02.700 --> 11:04.660 + aspects of the world. + +11:04.660 --> 11:09.620 + I think one day, maybe someone who's watching this will come up with the equations that + +11:09.620 --> 11:12.140 + information processing has to satisfy to be conscious. + +11:12.140 --> 11:19.580 + And I'm quite convinced there is big discovery to be made there because let's face it, we + +11:19.580 --> 11:25.900 + know that some information processing is conscious because we are conscious. + +11:25.900 --> 11:28.980 + But we also know that a lot of information processing is not conscious. + +11:28.980 --> 11:32.980 + Most of the information processing happening in your brain right now is not conscious. + +11:32.980 --> 11:38.380 + There are like 10 megabytes per second coming in even just through your visual system. + +11:38.380 --> 11:42.940 + You're not conscious about your heartbeat regulation or most things. + +11:42.940 --> 11:47.300 + Even if I just ask you to read what it says here, you look at it and then, oh, now you + +11:47.300 --> 11:48.300 + know what it said. + +11:48.300 --> 11:51.820 + But you're not aware of how the computation actually happened. + +11:51.820 --> 11:57.020 + Your consciousness is like the CEO that got an email at the end with the final answer. + +11:57.020 --> 12:01.140 + So what is it that makes a difference? + +12:01.140 --> 12:06.620 + I think that's both a great science mystery, we're actually studying it a little bit in + +12:06.620 --> 12:12.260 + my lab here at MIT, but I also think it's a really urgent question to answer. + +12:12.260 --> 12:16.460 + For starters, I mean, if you're an emergency room doctor and you have an unresponsive patient + +12:16.460 --> 12:24.180 + coming in, wouldn't it be great if in addition to having a CT scanner, you had a conscious + +12:24.180 --> 12:30.780 + scanner that could figure out whether this person is actually having locked in syndrome + +12:30.780 --> 12:33.580 + or is actually comatose. + +12:33.580 --> 12:40.740 + And in the future, imagine if we build robots or the machine that we can have really good + +12:40.740 --> 12:45.100 + conversations with, I think it's very likely to happen, right? + +12:45.100 --> 12:50.020 + Wouldn't you want to know if your home helper robot is actually experiencing anything or + +12:50.020 --> 12:52.980 + just like a zombie? + +12:52.980 --> 12:53.980 + Would you prefer it? + +12:53.980 --> 12:54.980 + What would you prefer? + +12:54.980 --> 12:57.820 + Would you prefer that it's actually unconscious so that you don't have to feel guilty about + +12:57.820 --> 12:59.980 + switching it off or giving boring chores? + +12:59.980 --> 13:02.380 + What would you prefer? + +13:02.380 --> 13:09.780 + Well, certainly we would prefer, I would prefer the appearance of consciousness, but the question + +13:09.780 --> 13:15.300 + is whether the appearance of consciousness is different than consciousness itself. + +13:15.300 --> 13:21.420 + And sort of ask that as a question, do you think we need to understand what consciousness + +13:21.420 --> 13:28.420 + is, solve the hard problem of consciousness in order to build something like an AGI system? + +13:28.420 --> 13:29.420 + No. + +13:29.420 --> 13:31.140 + I don't think that. + +13:31.140 --> 13:36.220 + I think we will probably be able to build things even if we don't answer that question. + +13:36.220 --> 13:41.100 + But if we want to make sure that what happens is a good thing, we better solve it first. + +13:41.100 --> 13:47.220 + So it's a wonderful controversy you're raising there, where you have basically three points + +13:47.220 --> 13:50.220 + of view about the hard problem. + +13:50.220 --> 13:55.060 + There are two different points of view that both conclude that the hard problem of consciousness + +13:55.060 --> 13:56.060 + is BS. + +13:56.060 --> 14:01.100 + On one hand, you have some people like Daniel Dennett who say that consciousness is just + +14:01.100 --> 14:05.140 + BS because consciousness is the same thing as intelligence. + +14:05.140 --> 14:06.580 + There's no difference. + +14:06.580 --> 14:13.620 + So anything which acts conscious is conscious, just like we are. + +14:13.620 --> 14:18.820 + And then there are also a lot of people, including many top AI researchers I know, who say, oh, + +14:18.820 --> 14:22.820 + consciousness is just bullshit because of course machines can never be conscious. + +14:22.820 --> 14:28.020 + They're always going to skiddy zombies, never have to feel guilty about how you treat them. + +14:28.020 --> 14:35.380 + And then there's a third group of people, including Giulio Tononi, for example, and another, and + +14:35.380 --> 14:40.020 + Gustav Koch and a number of others, I would put myself on this middle camp who say that + +14:40.020 --> 14:44.260 + actually some information processing is conscious and some is not. + +14:44.260 --> 14:49.380 + So let's find the equation which can be used to determine which it is. + +14:49.380 --> 14:53.980 + And I think we've just been a little bit lazy kind of running away from this problem for + +14:53.980 --> 14:55.100 + a long time. + +14:55.100 --> 15:01.940 + It's been almost taboo to even mention the C word in a lot of circles because, but we + +15:01.940 --> 15:03.700 + should stop making excuses. + +15:03.700 --> 15:10.940 + This is a science question and there are ways we can even test any theory that makes predictions + +15:10.940 --> 15:12.140 + for this. + +15:12.140 --> 15:16.060 + And coming back to this helper robot, I mean, so you said you would want your helper robot + +15:16.060 --> 15:21.340 + to certainly act conscious and treat you, like have conversations with you and stuff. + +15:21.340 --> 15:24.860 + But wouldn't you, would you feel a little bit creeped out if you realized that it was + +15:24.860 --> 15:31.700 + just a glossed up tape recorder, you know, that was just zombie and it's a faking emotion? + +15:31.700 --> 15:37.220 + Would you prefer that it actually had an experience or would you prefer that it's actually not + +15:37.220 --> 15:42.300 + experiencing anything so you feel, you don't have to feel guilty about what you do to it? + +15:42.300 --> 15:46.580 + It's such a difficult question because, you know, it's like when you're in a relationship + +15:46.580 --> 15:49.860 + and you say, well, I love you and the other person said I love you back. + +15:49.860 --> 15:53.860 + It's like asking, well, do they really love you back or are they just saying they love + +15:53.860 --> 15:54.860 + you back? + +15:54.860 --> 15:59.620 + Don't you really want them to actually love you? + +15:59.620 --> 16:08.100 + It's hard to, it's hard to really know the difference between everything seeming like + +16:08.100 --> 16:14.820 + there's consciousness present, there's intelligence present, there's affection, passion, love, + +16:14.820 --> 16:16.180 + and it actually being there. + +16:16.180 --> 16:17.180 + I'm not sure. + +16:17.180 --> 16:18.180 + Do you have... + +16:18.180 --> 16:19.180 + Can I ask you a question about this? + +16:19.180 --> 16:20.180 + Yes. + +16:20.180 --> 16:21.180 + To make it a bit more pointed. + +16:21.180 --> 16:23.140 + So Mass General Hospital is right across the river, right? + +16:23.140 --> 16:29.180 + Suppose you're going in for a medical procedure and they're like, you know, for anesthesia + +16:29.180 --> 16:32.180 + what we're going to do is we're going to give you muscle relaxance so you won't be able + +16:32.180 --> 16:36.140 + to move and you're going to feel excruciating pain during the whole surgery but you won't + +16:36.140 --> 16:37.660 + be able to do anything about it. + +16:37.660 --> 16:42.020 + But then we're going to give you this drug that erases your memory of it. + +16:42.020 --> 16:45.420 + Would you be cool about that? + +16:45.420 --> 16:51.100 + What's the difference that you're conscious about it or not if there's no behavioral change, + +16:51.100 --> 16:52.100 + right? + +16:52.100 --> 16:53.100 + Right. + +16:53.100 --> 16:55.220 + And that's a really clear way to put it. + +16:55.220 --> 17:01.100 + Yeah, it feels like in that sense, experiencing it is a valuable quality. + +17:01.100 --> 17:09.140 + So actually being able to have subjective experiences, at least in that case, is valuable. + +17:09.140 --> 17:14.060 + And I think we humans have a little bit of a bad track record also of making these self + +17:14.060 --> 17:17.940 + serving arguments that other entities aren't conscious. + +17:17.940 --> 17:20.700 + You know, people often say, oh, these animals can't feel pain. + +17:20.700 --> 17:21.700 + Right. + +17:21.700 --> 17:25.580 + It's okay to boil lobsters because we asked them if it hurt and they didn't say anything. + +17:25.580 --> 17:29.180 + And now there was just a paper out saying lobsters did do feel pain when you boil them + +17:29.180 --> 17:31.180 + and they're banning it in Switzerland. + +17:31.180 --> 17:36.300 + And we did this with slaves too often and said, oh, they don't mind. + +17:36.300 --> 17:41.180 + They don't maybe aren't conscious or women don't have souls or whatever. + +17:41.180 --> 17:46.540 + So I'm a little bit nervous when I hear people just take as an axiom that machines can't + +17:46.540 --> 17:48.900 + have experience ever. + +17:48.900 --> 17:52.500 + I think this is just a really fascinating science question is what it is. + +17:52.500 --> 17:57.420 + Let's research it and try to figure out what it is that makes the difference between unconscious + +17:57.420 --> 18:01.220 + intelligent behavior and conscious intelligent behavior. + +18:01.220 --> 18:07.140 + So in terms of, so if you think of a Boston Dynamics human or robot being sort of with + +18:07.140 --> 18:13.420 + a broom being pushed around, it starts pushing on a consciousness question. + +18:13.420 --> 18:20.060 + So let me ask, do you think an AGI system, like a few neuroscientists believe needs to + +18:20.060 --> 18:25.860 + have a physical embodiment, needs to have a body or something like a body? + +18:25.860 --> 18:28.340 + No, I don't think so. + +18:28.340 --> 18:30.620 + You mean to have a conscious experience? + +18:30.620 --> 18:33.140 + To have consciousness. + +18:33.140 --> 18:37.860 + I do think it helps a lot to have a physical embodiment to learn the kind of things about + +18:37.860 --> 18:42.820 + the world that are important to us humans for sure. + +18:42.820 --> 18:47.460 + But I don't think the physical embodiment is necessary after you've learned it. + +18:47.460 --> 18:48.860 + Just have the experience. + +18:48.860 --> 18:51.500 + Think about it when you're dreaming, right? + +18:51.500 --> 18:55.500 + Your eyes are closed, you're not getting any sensory input, you're not behaving or moving + +18:55.500 --> 18:59.780 + in any way, but there's still an experience there, right? + +18:59.780 --> 19:03.220 + And so clearly the experience that you have when you see something cool in your dreams + +19:03.220 --> 19:08.660 + isn't coming from your eyes, it's just the information processing itself in your brain, + +19:08.660 --> 19:11.100 + which is that experience, right? + +19:11.100 --> 19:16.660 + But if I put it another way, I'll say because it comes from neuroscience is the reason you + +19:16.660 --> 19:24.620 + want to have a body and a physical, something like a physical system is because you want + +19:24.620 --> 19:27.100 + to be able to preserve something. + +19:27.100 --> 19:35.740 + In order to have a self, you could argue, you'd need to have some kind of embodiment + +19:35.740 --> 19:38.180 + of self to want to preserve. + +19:38.180 --> 19:45.940 + Well, now we're getting a little bit anthropomorphic, anthropomorphizing things, maybe talking about + +19:45.940 --> 19:47.820 + self preservation instincts. + +19:47.820 --> 19:50.700 + We are evolved organisms, right? + +19:50.700 --> 19:57.020 + So Darwinian evolution endowed us and other evolved organisms with self preservation instinct + +19:57.020 --> 20:03.100 + because those that didn't have those self preservation genes got cleaned out of the gene pool. + +20:03.100 --> 20:09.180 + But if you build an artificial general intelligence, the mind space that you can design is much, + +20:09.180 --> 20:14.500 + much larger than just a specific subset of minds that can evolve that have. + +20:14.500 --> 20:19.260 + So an AGI mind doesn't necessarily have to have any self preservation instinct. + +20:19.260 --> 20:24.100 + It also doesn't necessarily have to be so individualistic as us. + +20:24.100 --> 20:28.140 + Like imagine if you could just, first of all, we're also very afraid of death, you know, + +20:28.140 --> 20:32.180 + as opposed to you could back yourself up every five minutes and then your airplane is about + +20:32.180 --> 20:33.180 + to crash. + +20:33.180 --> 20:37.340 + You're like, shucks, I'm just, I'm going to lose the last five minutes of experiences + +20:37.340 --> 20:41.580 + since my last cloud backup, dang, you know, it's not as big a deal. + +20:41.580 --> 20:47.380 + Or if we could just copy experiences between our minds easily, like which we could easily + +20:47.380 --> 20:50.620 + do if we were silicon based, right? + +20:50.620 --> 20:55.860 + Then maybe we would feel a little bit more like a hive mind, actually, that maybe it's + +20:55.860 --> 21:01.220 + the, so, so there's, so I don't think we should take for granted at all that AGI will have + +21:01.220 --> 21:06.820 + to have any of those sort of competitive as alpha male instincts. + +21:06.820 --> 21:07.820 + Right. + +21:07.820 --> 21:12.820 + On the other hand, you know, this is really interesting because I think some people go + +21:12.820 --> 21:17.900 + too far and say, of course, we don't have to have any concerns either that advanced + +21:17.900 --> 21:22.700 + AI will have those instincts because we can build anything we want. + +21:22.700 --> 21:27.420 + That there's, there's a very nice set of arguments going back to Steve Omohandro and + +21:27.420 --> 21:32.900 + Nick Bostrom and others just pointing out that when we build machines, we normally build + +21:32.900 --> 21:37.700 + them with some kind of goal, you know, win this chess game, drive this car safely or + +21:37.700 --> 21:38.700 + whatever. + +21:38.700 --> 21:42.540 + And as soon as you put in a goal into machine, especially if it's kind of open ended goal + +21:42.540 --> 21:48.460 + and the machine is very intelligent, it'll break that down into a bunch of sub goals. + +21:48.460 --> 21:53.500 + And one of those goals will almost always be self preservation because if it breaks + +21:53.500 --> 21:56.140 + or dies in the process, it's not going to accomplish the goal, right? + +21:56.140 --> 21:59.540 + Like, suppose you just build a little, you have a little robot and you tell it to go + +21:59.540 --> 22:05.460 + down the store market here and, and get you some food, make you cook your Italian dinner, + +22:05.460 --> 22:09.540 + you know, and then someone mugs it and tries to break it on the way. + +22:09.540 --> 22:15.380 + That robot has an incentive to not get destroyed and defend itself for a runaway because otherwise + +22:15.380 --> 22:17.780 + it's going to fail and cooking your dinner. + +22:17.780 --> 22:22.940 + It's not afraid of death, but it really wants to complete the dinner cooking goal. + +22:22.940 --> 22:24.780 + So it will have a self preservation instinct. + +22:24.780 --> 22:26.820 + It will continue being a functional agent. + +22:26.820 --> 22:27.820 + Yeah. + +22:27.820 --> 22:35.860 + And, and, and similarly, if you give any kind of more ambitious goal to an AGI, it's very + +22:35.860 --> 22:39.940 + likely they want to acquire more resources so it can do that better. + +22:39.940 --> 22:44.500 + And it's exactly from those sort of sub goals that we might not have intended that some + +22:44.500 --> 22:50.740 + of the concerns about AGI safety come, you give it some goal that seems completely harmless. + +22:50.740 --> 22:55.540 + And then before you realize it, it's also trying to do these other things which you + +22:55.540 --> 22:59.220 + didn't want it to do and it's maybe smarter than us. + +22:59.220 --> 23:08.220 + So, so, and let me pause just because I am in a very kind of human centric way, see fear + +23:08.220 --> 23:11.900 + of death as a valuable motivator. + +23:11.900 --> 23:17.220 + So you don't think you think that's an artifact of evolution. + +23:17.220 --> 23:21.980 + So that's the kind of mind space evolution created that we're sort of almost obsessed + +23:21.980 --> 23:22.980 + about self preservation. + +23:22.980 --> 23:23.980 + Yeah. + +23:23.980 --> 23:29.500 + Some kind of genetic well, you don't think that's necessary to be afraid of death. + +23:29.500 --> 23:34.980 + So not just a kind of sub goal of self preservation just so you can keep doing the thing, but + +23:34.980 --> 23:42.980 + more fundamentally sort of have the finite thing like this ends for you at some point. + +23:42.980 --> 23:43.980 + Interesting. + +23:43.980 --> 23:47.500 + Do I think it's necessary for what precisely? + +23:47.500 --> 23:51.020 + For intelligence, but also for consciousness. + +23:51.020 --> 23:58.220 + So for those for both, do you think really like a finite death and the fear of it is + +23:58.220 --> 24:01.020 + important? + +24:01.020 --> 24:06.980 + So before I can answer, before we can agree on whether it's necessary for intelligence + +24:06.980 --> 24:10.660 + or for consciousness, we should be clear on how we define those two words because a lot + +24:10.660 --> 24:13.340 + are really smart people define them in very different ways. + +24:13.340 --> 24:18.500 + I was in this on this panel with AI experts and they couldn't, they couldn't agree on + +24:18.500 --> 24:20.180 + how to define intelligence even. + +24:20.180 --> 24:24.860 + So I define intelligence simply as the ability to accomplish complex goals. + +24:24.860 --> 24:30.740 + I like your broad definition because again, I don't want to be a carbon chauvinist. + +24:30.740 --> 24:36.580 + And in that case, no, certainly it doesn't require fear of death. + +24:36.580 --> 24:40.100 + I would say AlphaGo AlphaZero is quite intelligent. + +24:40.100 --> 24:44.260 + I don't think AlphaZero has any fear of being turned off because it doesn't understand the + +24:44.260 --> 24:52.180 + concept of even and similarly consciousness, I mean, you can certainly imagine a very simple + +24:52.180 --> 24:57.660 + kind of experience if certain plants have any kind of experience, I don't think they're + +24:57.660 --> 25:00.940 + very afraid of dying or there's nothing they can do about it anyway much. + +25:00.940 --> 25:08.420 + So there wasn't that much value and but more seriously, I think if you ask not just about + +25:08.420 --> 25:15.460 + being conscious, but maybe having what you would, we might call an exciting life for + +25:15.460 --> 25:23.300 + you for your passion and really appreciate the things, maybe there, somehow, maybe there + +25:23.300 --> 25:29.180 + perhaps it does help having a backdrop that, hey, it's finite, you know, let's make the + +25:29.180 --> 25:31.380 + most of this, let's live to the fullest. + +25:31.380 --> 25:36.220 + So if you knew you were going to just live forever, do you think you would change your + +25:36.220 --> 25:40.500 + career? Yeah, I mean, in some perspective, it would + +25:40.500 --> 25:44.020 + be an incredibly boring life living forever. + +25:44.020 --> 25:49.740 + So in the sort of loose, subjective terms that you said of something exciting and something + +25:49.740 --> 25:55.180 + in this that other humans would understand, I think, is yeah, it seems that the finiteness + +25:55.180 --> 25:56.660 + of it is important. + +25:56.660 --> 26:02.420 + Well, the good news I have for you then is based on what we understand about cosmology, + +26:02.420 --> 26:10.460 + things in our universe is probably finite, although big crunch or big or big, what's + +26:10.460 --> 26:11.460 + the extent of the infinite? + +26:11.460 --> 26:16.820 + Yeah, we could have a big chill or a big crunch or a big rip or death, the big snap or death + +26:16.820 --> 26:17.820 + bubbles. + +26:17.820 --> 26:20.140 + All of them are more than a billion years away. + +26:20.140 --> 26:29.500 + So we should we certainly have vastly more time than our ancestors thought, but still + +26:29.500 --> 26:35.580 + pretty hard to squeeze in an infinite number of compute cycles, even though there are some + +26:35.580 --> 26:37.820 + loopholes that just might be possible. + +26:37.820 --> 26:44.620 + But I think, you know, some people like to say that you should live as if you're about + +26:44.620 --> 26:48.100 + to you're going to die in five years or so, and that's sort of optimal. + +26:48.100 --> 26:54.740 + Maybe it's a good as some we should build our civilization as if it's all finite to + +26:54.740 --> 26:55.740 + be on the safe side. + +26:55.740 --> 27:02.020 + Right, exactly. So you mentioned in defining intelligence as the ability to solve complex + +27:02.020 --> 27:03.020 + goals. + +27:03.020 --> 27:04.940 + So where would you draw a line? + +27:04.940 --> 27:10.940 + How would you try to define human level intelligence and super human level intelligence? + +27:10.940 --> 27:13.380 + Where is consciousness part of that definition? + +27:13.380 --> 27:16.860 + No, consciousness does not come into this definition. + +27:16.860 --> 27:21.580 + So so I think of intelligence as it's a spectrum, but there are very many different kinds of + +27:21.580 --> 27:22.580 + goals you can have. + +27:22.580 --> 27:27.140 + You have a goal to be a good chess player, a good goal player, a good car driver, a good + +27:27.140 --> 27:31.260 + investor, good poet, etc. + +27:31.260 --> 27:35.740 + So intelligence that bind by its very nature, isn't something you can measure, but it's + +27:35.740 --> 27:39.900 + one number, some overall goodness, no, no, there are some people who are more better + +27:39.900 --> 27:42.540 + at this, some people are better at that. + +27:42.540 --> 27:48.380 + Right now we have machines that are much better than us at some very narrow tasks like multiplying + +27:48.380 --> 27:57.620 + large numbers fast, memorizing large databases, playing chess, playing go, soon driving cars. + +27:57.620 --> 28:03.340 + But there's still no machine that can match a human child in general intelligence. + +28:03.340 --> 28:08.420 + But artificial general intelligence, AGI, the name of your course, of course, that + +28:08.420 --> 28:16.460 + is by its very definition, the quest to build a machine that can do everything as well as + +28:16.460 --> 28:17.460 + we can. + +28:17.460 --> 28:24.060 + Up to the old Holy Grail of AI from back to its inception in the 60s. + +28:24.060 --> 28:27.500 + If that ever happens, of course, I think it's going to be the biggest transition in the + +28:27.500 --> 28:33.860 + history of life on Earth, but it doesn't necessarily have to wait the big impact until machines + +28:33.860 --> 28:35.780 + are better than us at knitting. + +28:35.780 --> 28:41.940 + The really big change doesn't come exactly at the moment they're better than us at everything. + +28:41.940 --> 28:45.820 + The really big change comes, first, their big change is when they start becoming better + +28:45.820 --> 28:51.140 + at us at doing most of the jobs that we do, because that takes away much of the demand + +28:51.140 --> 28:53.380 + for human labor. + +28:53.380 --> 29:01.300 + And then the really warping change comes when they become better than us at AI research. + +29:01.300 --> 29:07.900 + Because right now, the time scale of AI research is limited by the human research and development + +29:07.900 --> 29:14.100 + cycle of years, typically, along the take from one release of some software or iPhone + +29:14.100 --> 29:16.300 + or whatever to the next. + +29:16.300 --> 29:25.820 + But once Google can replace 40,000 engineers by 40,000 equivalent pieces of software or + +29:25.820 --> 29:29.660 + whatever, then there's no reason that has to be years. + +29:29.660 --> 29:32.020 + It can be, in principle, much faster. + +29:32.020 --> 29:38.900 + And the time scale of future progress in AI and all of science and technology will be + +29:38.900 --> 29:40.980 + driven by machines, not humans. + +29:40.980 --> 29:49.660 + So it's this simple point, which gives right this incredibly fun controversy about whether + +29:49.660 --> 29:54.540 + there can be intelligence explosion, so called singularity, as Werner Winge called it. + +29:54.540 --> 30:00.060 + The idea, as articulated by I.J. Good, is obviously way back fifties, but you can see + +30:00.060 --> 30:07.220 + Alan Turing and others thought about it even earlier. + +30:07.220 --> 30:12.980 + You asked me what exactly what I define human level intelligence. + +30:12.980 --> 30:18.540 + So the glib answer is just to say something which is better than us at all cognitive tasks + +30:18.540 --> 30:21.980 + or better than any human at all cognitive tasks. + +30:21.980 --> 30:25.900 + But the really interesting bar, I think, goes a little bit lower than that, actually. + +30:25.900 --> 30:33.260 + It's when they're better than us at AI programming and general learning so that they can, if + +30:33.260 --> 30:37.340 + they want to, get better than us at anything by just starting out. + +30:37.340 --> 30:43.100 + So there better is a key word and better is towards this kind of spectrum of the complexity + +30:43.100 --> 30:45.740 + of goals it's able to accomplish. + +30:45.740 --> 30:53.060 + So another way to, and that's certainly a very clear definition of human love. + +30:53.060 --> 30:56.300 + So there's, it's almost like a sea that's rising, you can do more and more and more + +30:56.300 --> 30:57.300 + things. + +30:57.300 --> 30:59.900 + It's actually a graphic that you show, it's really nice way to put it. + +30:59.900 --> 31:04.340 + So there's some peaks and there's an ocean level elevating and you solve more and more + +31:04.340 --> 31:05.340 + problems. + +31:05.340 --> 31:09.220 + But, you know, just kind of to take a pause and we took a bunch of questions and a lot + +31:09.220 --> 31:14.380 + of social networks and a bunch of people asked a sort of a slightly different direction + +31:14.380 --> 31:22.260 + on creativity and on things that perhaps aren't a peak. + +31:22.260 --> 31:28.620 + It's, you know, human beings are flawed and perhaps better means having being having contradiction + +31:28.620 --> 31:30.260 + being flawed in some way. + +31:30.260 --> 31:34.980 + So let me sort of, yeah, start and start easy, first of all. + +31:34.980 --> 31:36.620 + So you have a lot of cool equations. + +31:36.620 --> 31:39.660 + Let me ask, what's your favorite equation, first of all? + +31:39.660 --> 31:43.580 + I know they're all like your children, but which one is that? + +31:43.580 --> 31:49.060 + This is the Shreddinger equation, it's the master key of quantum mechanics of the micro + +31:49.060 --> 31:50.060 + world. + +31:50.060 --> 31:55.340 + So this equation can take everything to do with atoms and all the fuels and all the + +31:55.340 --> 32:04.020 + way up to… Yeah, so, okay, so quantum mechanics is certainly a beautiful mysterious formulation + +32:04.020 --> 32:05.020 + of our world. + +32:05.020 --> 32:10.740 + So I'd like to sort of ask you, just as an example, it perhaps doesn't have the same + +32:10.740 --> 32:17.420 + beauty as physics does, but in mathematics abstract, the Andrew Wiles who proved the + +32:17.420 --> 32:19.460 + Fermat's last theory. + +32:19.460 --> 32:24.180 + So he just saw this recently and it kind of caught my eye a little bit. + +32:24.180 --> 32:27.980 + This is 358 years after it was conjectured. + +32:27.980 --> 32:32.940 + So this very simple formulation, everybody tried to prove it, everybody failed. + +32:32.940 --> 32:38.820 + And so here's this guy comes along and eventually proves it and then fails to prove it and then + +32:38.820 --> 32:41.340 + proves it again in 94. + +32:41.340 --> 32:45.940 + And he said like the moment when everything connected into place, in an interview he said + +32:45.940 --> 32:47.980 + it was so indescribably beautiful. + +32:47.980 --> 32:53.580 + That moment when you finally realize the connecting piece of two conjectures, he said it was so + +32:53.580 --> 32:56.940 + indescribably beautiful, it was so simple and so elegant. + +32:56.940 --> 33:01.540 + I couldn't understand how I'd missed it and I just stared at it in disbelief for 20 + +33:01.540 --> 33:02.540 + minutes. + +33:02.540 --> 33:08.100 + Then during the day I walked around the department and I keep coming back to my desk looking + +33:08.100 --> 33:09.820 + to see if it was still there. + +33:09.820 --> 33:10.820 + It was still there. + +33:10.820 --> 33:11.820 + I couldn't contain myself. + +33:11.820 --> 33:12.820 + I was so excited. + +33:12.820 --> 33:16.180 + It was the most important moment of my working life. + +33:16.180 --> 33:18.940 + Nothing I ever do again will mean as much. + +33:18.940 --> 33:24.860 + So that particular moment and it kind of made me think of what would it take? + +33:24.860 --> 33:28.380 + And I think we have all been there at small levels. + +33:28.380 --> 33:34.820 + Maybe let me ask, have you had a moment like that in your life where you just had an idea + +33:34.820 --> 33:40.060 + it's like, wow, yes. + +33:40.060 --> 33:44.700 + I wouldn't mention myself in the same breath as Andrew Wiles, but I certainly had a number + +33:44.700 --> 33:54.820 + of aha moments when I realized something very cool about physics just completely made + +33:54.820 --> 33:55.820 + my head explode. + +33:55.820 --> 33:59.580 + In fact, some of my favorite discoveries I made later, I later realized that they had + +33:59.580 --> 34:03.340 + been discovered earlier by someone who's sometimes got quite famous for it. + +34:03.340 --> 34:07.460 + So there's too late for me to even publish it, but that doesn't diminish in any way. + +34:07.460 --> 34:12.340 + The emotional experience you have when you realize it like, wow. + +34:12.340 --> 34:17.460 + So what would it take in that moment, that wow, that was yours in that moment? + +34:17.460 --> 34:23.420 + So what do you think it takes for an intelligent system, an AGI system, an AI system to have + +34:23.420 --> 34:24.980 + a moment like that? + +34:24.980 --> 34:29.420 + It's a tricky question because there are actually two parts to it, right? + +34:29.420 --> 34:37.260 + One of them is, can it accomplish that proof, can it prove that you can never write A to + +34:37.260 --> 34:46.420 + the N plus B to the N equals 3 to the N for all integers, etc., etc., when N is bigger + +34:46.420 --> 34:49.420 + than 2. + +34:49.420 --> 34:51.580 + That's simply the question about intelligence. + +34:51.580 --> 34:54.420 + Can you build machines that are that intelligent? + +34:54.420 --> 34:59.860 + And I think by the time we get a machine that can independently come up with that level + +34:59.860 --> 35:03.460 + of proofs, probably quite close to AGI. + +35:03.460 --> 35:07.860 + But the second question is a question about consciousness. + +35:07.860 --> 35:13.060 + When will we, how likely is it that such a machine would actually have any experience + +35:13.060 --> 35:16.500 + at all as opposed to just being like a zombie? + +35:16.500 --> 35:22.940 + And would we expect it to have some sort of emotional response to this or anything at + +35:22.940 --> 35:31.140 + all akin to human emotion where when it accomplishes its machine goal, it views it as something + +35:31.140 --> 35:39.260 + very positive and sublime and deeply meaningful. + +35:39.260 --> 35:45.260 + I would certainly hope that if in the future we do create machines that are our peers or + +35:45.260 --> 35:53.700 + even our descendants, I would certainly hope that they do have this sort of sublime appreciation + +35:53.700 --> 36:06.020 + of life in a way, my absolutely worst nightmare would be that at some point in the future, + +36:06.020 --> 36:10.620 + the distant future, maybe our cosmos is teeming with all this post biological life, doing + +36:10.620 --> 36:13.180 + all the seemingly cool stuff. + +36:13.180 --> 36:20.660 + And maybe the last humans by the time our species eventually fizzles out will be like, + +36:20.660 --> 36:26.140 + well, that's okay, because we're so proud of our descendants here and look, my worst + +36:26.140 --> 36:30.580 + nightmare is that we haven't solved the consciousness problem. + +36:30.580 --> 36:34.100 + And we haven't realized that these are all the zombies, they're not aware of anything + +36:34.100 --> 36:37.900 + anymore than a tape recorder, as in any kind of experience. + +36:37.900 --> 36:41.660 + So the whole thing has just become a play for empty benches. + +36:41.660 --> 36:44.700 + That would be like the ultimate zombie apocalypse to me. + +36:44.700 --> 36:52.900 + So I would much rather, in that case, that we have these beings which can really appreciate + +36:52.900 --> 36:57.060 + how amazing it is. + +36:57.060 --> 37:02.260 + And in that picture, what would be the role of creativity, what a few people ask about + +37:02.260 --> 37:03.260 + creativity? + +37:03.260 --> 37:04.260 + Yeah. + +37:04.260 --> 37:08.700 + And do you think, when you think about intelligence, I mean, certainly the story you told at the + +37:08.700 --> 37:14.100 + beginning of your book involved, you know, creating movies and so on, sort of making + +37:14.100 --> 37:18.580 + money, you know, you can make a lot of money in our modern world with music and movies. + +37:18.580 --> 37:23.100 + So if you are an intelligent system, you may want to get good at that. + +37:23.100 --> 37:26.300 + But that's not necessarily what I mean by creativity. + +37:26.300 --> 37:32.620 + Is it important on that complex goals where the sea is rising for there to be something + +37:32.620 --> 37:39.940 + creative, or am I being very human centric and thinking creativity somehow special relative + +37:39.940 --> 37:41.940 + to intelligence? + +37:41.940 --> 37:50.940 + My hunch is that we should think of creativity simply as an aspect of intelligence. + +37:50.940 --> 37:57.820 + And we have to be very careful with human vanity. + +37:57.820 --> 38:01.540 + We have this tendency to very often want to say, as soon as machines can do something, + +38:01.540 --> 38:05.980 + we try to diminish it and say, oh, but that's not like real intelligence, you know, is + +38:05.980 --> 38:12.620 + it not creative or this or that, the other thing, if we ask ourselves to write down a + +38:12.620 --> 38:18.500 + definition of what we actually mean by being creative, what we mean by Andrew Wiles, what + +38:18.500 --> 38:23.660 + he did there, for example, don't we often mean that someone takes a very unexpected + +38:23.660 --> 38:26.060 + leap? + +38:26.060 --> 38:33.740 + It's not like taking 573 and multiplying by 224 by just a step of straightforward cookbook + +38:33.740 --> 38:36.500 + like rules, right? + +38:36.500 --> 38:40.660 + You can maybe make a connection between two things that people have never thought was + +38:40.660 --> 38:41.660 + connected. + +38:41.660 --> 38:42.660 + It's very surprising. + +38:42.660 --> 38:44.300 + Something like that. + +38:44.300 --> 38:50.660 + I think this is an aspect of intelligence, and this is actually one of the most important + +38:50.660 --> 38:53.260 + aspects of it. + +38:53.260 --> 38:57.940 + Maybe the reason we humans tend to be better at it than traditional computers is because + +38:57.940 --> 39:02.020 + it's something that comes more naturally if you're a neural network than if you're a + +39:02.020 --> 39:05.820 + traditional logic gates based computer machine. + +39:05.820 --> 39:11.900 + We physically have all these connections, and if you activate here, activate here, activate + +39:11.900 --> 39:20.980 + here, it ping, you know, my hunch is that if we ever build a machine where you could + +39:20.980 --> 39:31.020 + just give it the task, hey, hey, you say, hey, you know, I just realized I want to travel + +39:31.020 --> 39:32.380 + around the world instead this month. + +39:32.380 --> 39:34.700 + Can you teach my AGI course for me? + +39:34.700 --> 39:36.100 + And it's like, okay, I'll do it. + +39:36.100 --> 39:39.860 + And it does everything that you would have done and it improvises and stuff. + +39:39.860 --> 39:42.860 + That would in my mind involve a lot of creativity. + +39:42.860 --> 39:45.660 + Yeah, so it's actually a beautiful way to put it. + +39:45.660 --> 39:54.540 + I think we do try to grasp at the definition of intelligence as everything we don't understand + +39:54.540 --> 39:57.580 + how to build. + +39:57.580 --> 40:02.180 + So we as humans try to find things that we have and machines don't have, and maybe creativity + +40:02.180 --> 40:05.940 + is just one of the things, one of the words we used to describe that. + +40:05.940 --> 40:06.940 + That's a really interesting way to put it. + +40:06.940 --> 40:09.820 + I don't think we need to be that defensive. + +40:09.820 --> 40:14.700 + I don't think anything good comes out of saying, we're somehow special, you know, it's + +40:14.700 --> 40:27.540 + very wise, there are many examples in history of where trying to pretend they were somehow + +40:27.540 --> 40:36.220 + superior to all other intelligent beings has led to pretty bad results, right? + +40:36.220 --> 40:39.700 + Nazi Germany, they said that they were somehow superior to other people. + +40:39.700 --> 40:44.580 + Today, we still do a lot of cruelty to animals by saying they were so superior somehow on + +40:44.580 --> 40:50.500 + the other, they can't feel pain, slavery was justified by the same kind of really weak + +40:50.500 --> 40:52.420 + arguments. + +40:52.420 --> 40:58.700 + And I don't think if we actually go ahead and build artificial general intelligence, + +40:58.700 --> 41:01.100 + it can do things better than us. + +41:01.100 --> 41:08.980 + I don't think we should try to found our self worth on some sort of bogus claims of superiority + +41:08.980 --> 41:11.940 + in terms of our intelligence. + +41:11.940 --> 41:21.780 + I think we should instead find our calling and the meaning of life from the experiences + +41:21.780 --> 41:22.780 + that we have. + +41:22.780 --> 41:23.780 + Right. + +41:23.780 --> 41:30.260 + You know, I can have very meaningful experiences even if there are other people who are smarter + +41:30.260 --> 41:35.860 + than me, you know, when I go to faculty meeting here and I was talking about something and + +41:35.860 --> 41:39.420 + then I certainly realized, oh, he has an old prize, he has an old prize, he has an old + +41:39.420 --> 41:40.420 + prize. + +41:40.420 --> 41:41.420 + Yeah. + +41:41.420 --> 41:47.660 + You know, it doesn't make me enjoy life any less or enjoy talking to those people less. + +41:47.660 --> 41:49.780 + Of course not. + +41:49.780 --> 41:57.420 + And contrary to that, I feel very honored and privileged to get to interact with other + +41:57.420 --> 42:00.820 + very intelligent beings that are better than me and a lot of stuff. + +42:00.820 --> 42:05.420 + So I don't think there's any reason why we can't have the same approach with intelligent + +42:05.420 --> 42:06.420 + machines. + +42:06.420 --> 42:08.900 + That's a really interesting, so people don't often think about that. + +42:08.900 --> 42:14.380 + They think about if there's machines that are more intelligent, you naturally think + +42:14.380 --> 42:19.100 + that that's not going to be a beneficial type of intelligence. + +42:19.100 --> 42:24.060 + You don't realize it could be, you know, like peers with no ball prizes that would be just + +42:24.060 --> 42:25.060 + fun to talk with. + +42:25.060 --> 42:30.580 + And they might be clever about certain topics and you can have fun having a few drinks with + +42:30.580 --> 42:31.580 + them. + +42:31.580 --> 42:38.620 + Well, also, you know, another example we can all relate to why it doesn't have to be a + +42:38.620 --> 42:42.580 + terrible thing to be impressed, the presence of people who are even smarter than us all + +42:42.580 --> 42:47.980 + around is when you and I were both two years old, I mean, our parents were much more intelligent + +42:47.980 --> 42:48.980 + than us. + +42:48.980 --> 42:49.980 + Right. + +42:49.980 --> 42:50.980 + Worked out okay. + +42:50.980 --> 42:54.140 + Because their goals were aligned with our goals. + +42:54.140 --> 43:01.380 + And that I think is really the number one key issue we have to solve if we value align + +43:01.380 --> 43:07.380 + the value alignment problem exactly because people who see too many Hollywood movies with + +43:07.380 --> 43:12.260 + lousy science fiction plot lines, they worry about the wrong thing, right? + +43:12.260 --> 43:16.500 + They worry about some machine suddenly turning evil. + +43:16.500 --> 43:21.500 + It's not malice that we should that is the concern. + +43:21.500 --> 43:23.000 + It's competence. + +43:23.000 --> 43:29.580 + By definition, intelligence makes you makes you very competent if you have a more intelligent + +43:29.580 --> 43:35.300 + goal playing machine computer playing as a less intelligent one and when we define intelligence + +43:35.300 --> 43:37.740 + as the ability to accomplish go winning, right? + +43:37.740 --> 43:40.780 + It's going to be the more intelligent one that wins. + +43:40.780 --> 43:47.860 + And if you have a human and then you have an AGI that's more intelligent in all ways + +43:47.860 --> 43:50.500 + and they have different goals, guess who's going to get their way, right? + +43:50.500 --> 43:58.060 + So I was just reading about this particular rhinoceros species that was driven extinct + +43:58.060 --> 43:59.060 + just a few years ago. + +43:59.060 --> 44:05.740 + Alan Bummer is looking at this cute picture of a mommy rhinoceros with its child, you + +44:05.740 --> 44:09.140 + know, and why did we humans drive it to extinction? + +44:09.140 --> 44:12.860 + It wasn't because we were evil rhino haters as a whole. + +44:12.860 --> 44:16.380 + It was just because we our goals weren't aligned with those of the rhinoceros and it didn't + +44:16.380 --> 44:19.660 + work out so well for the rhinoceros because we were more intelligent, right? + +44:19.660 --> 44:27.220 + So I think it's just so important that if we ever do build AGI before we unleash anything, + +44:27.220 --> 44:37.380 + we have to make sure that it learns to understand our goals, that it adopts our goals and retains + +44:37.380 --> 44:38.380 + those goals. + +44:38.380 --> 44:45.740 + So the cool interesting problem there is being able, us as human beings, trying to formulate + +44:45.740 --> 44:47.240 + our values. + +44:47.240 --> 44:52.540 + So you know, you could think of the United States Constitution as a way that people sat + +44:52.540 --> 44:59.780 + down at the time a bunch of white men, which is a good example, I should say. + +44:59.780 --> 45:03.460 + They formulated the goals for this country and a lot of people agree that those goals + +45:03.460 --> 45:05.540 + actually held up pretty well. + +45:05.540 --> 45:09.600 + It's an interesting formulation of values and failed miserably in other ways. + +45:09.600 --> 45:15.500 + So for the value alignment problem and the solution to it, we have to be able to put + +45:15.500 --> 45:23.420 + on paper or in a program, human values, how difficult do you think that is? + +45:23.420 --> 45:24.420 + Very. + +45:24.420 --> 45:25.980 + But it's so important. + +45:25.980 --> 45:30.340 + We really have to give it our best and it's difficult for two separate reasons. + +45:30.340 --> 45:37.660 + There's the technical value alignment problem of figuring out just how to make machines + +45:37.660 --> 45:40.660 + understand our goals, adopt them and retain them. + +45:40.660 --> 45:46.140 + And then there's the separate part of it, the philosophical part, whose values anyway. + +45:46.140 --> 45:51.700 + And since we, it's not like we have any great consensus on this planet on values, what mechanism + +45:51.700 --> 45:56.780 + should we create then to aggregate and decide, okay, what's a good compromise? + +45:56.780 --> 46:01.260 + That second discussion can't just be left the tech nerds like myself, right? + +46:01.260 --> 46:02.260 + That's right. + +46:02.260 --> 46:06.820 + And if we refuse to talk about it and then AGI gets built, who's going to be actually + +46:06.820 --> 46:10.660 + making the decision about whose values, it's going to be a bunch of dudes in some tech + +46:10.660 --> 46:12.380 + company, right? + +46:12.380 --> 46:18.420 + And are they necessarily so representative of all of humankind that we want to just + +46:18.420 --> 46:19.580 + endorse it to them? + +46:19.580 --> 46:25.220 + Are they even uniquely qualified to speak to future human happiness just because they're + +46:25.220 --> 46:26.460 + good at programming AI? + +46:26.460 --> 46:30.380 + I'd much rather have this be a really inclusive conversation. + +46:30.380 --> 46:32.700 + But do you think it's possible? + +46:32.700 --> 46:38.820 + You create a beautiful vision that includes sort of the diversity, cultural diversity + +46:38.820 --> 46:43.900 + and various perspectives on discussing rights, freedoms, human dignity. + +46:43.900 --> 46:46.620 + But how hard is it to come to that consensus? + +46:46.620 --> 46:52.140 + Do you think it's certainly a really important thing that we should all try to do, but do + +46:52.140 --> 46:54.460 + you think it's feasible? + +46:54.460 --> 47:01.660 + I think there's no better way to guarantee failure than to refuse to talk about it or + +47:01.660 --> 47:02.980 + refuse to try. + +47:02.980 --> 47:08.060 + And I also think it's a really bad strategy to say, okay, let's first have a discussion + +47:08.060 --> 47:09.060 + for a long time. + +47:09.060 --> 47:13.540 + And then once we reach complete consensus, then we'll try to load it into some machine. + +47:13.540 --> 47:16.980 + No, we shouldn't let perfect be the enemy of good. + +47:16.980 --> 47:22.140 + Instead, we should start with the kindergarten ethics that pretty much everybody agrees on + +47:22.140 --> 47:24.580 + and put that into our machines now. + +47:24.580 --> 47:26.100 + We're not doing that even. + +47:26.100 --> 47:32.980 + Look at anyone who builds a passenger aircraft wants it to never under any circumstances + +47:32.980 --> 47:35.900 + fly into a building or mountain, right? + +47:35.900 --> 47:38.860 + Yet the September 11 hijackers were able to do that. + +47:38.860 --> 47:44.220 + And even more embarrassingly, Andreas Lubitz, this depressed German wings pilot, when he + +47:44.220 --> 47:50.220 + flew his passenger jet into the Alps, killing over 100 people, he just told the autopilot + +47:50.220 --> 47:51.220 + to do it. + +47:51.220 --> 47:55.140 + He told the freaking computer to change the altitude to 100 meters. + +47:55.140 --> 48:01.820 + And even though it had the GPS maps, everything, the computer was like, okay, no, so we should + +48:01.820 --> 48:07.300 + take those very basic values, though, where the problem is not that we don't agree. + +48:07.300 --> 48:12.460 + The problem is just we've been too lazy to try to put it into our machines and make sure + +48:12.460 --> 48:17.460 + that from now on, airplanes will just, which all have computers in them, but we'll just + +48:17.460 --> 48:19.820 + never just refuse to do something like that. + +48:19.820 --> 48:25.580 + We go into safe mode, maybe lock the cockpit door, go to the nearest airport, and there's + +48:25.580 --> 48:31.340 + so much other technology in our world as well now where it's really coming quite timely + +48:31.340 --> 48:34.300 + to put in some sort of very basic values like this. + +48:34.300 --> 48:41.460 + Even in cars, we've had enough vehicle terrorism attacks by now where people have driven trucks + +48:41.460 --> 48:47.300 + and vans into pedestrians that it's not at all a crazy idea to just have that hardwired + +48:47.300 --> 48:51.420 + into the car, because yeah, there are a lot of, there's always going to be people who + +48:51.420 --> 48:55.620 + for some reason want to harm others, but most of those people don't have the technical + +48:55.620 --> 48:58.620 + expertise to figure out how to work around something like that. + +48:58.620 --> 49:01.780 + So if the car just won't do it, it helps. + +49:01.780 --> 49:02.940 + So let's start there. + +49:02.940 --> 49:05.020 + So there's a lot of, that's a great point. + +49:05.020 --> 49:06.900 + So not chasing perfect. + +49:06.900 --> 49:10.780 + There's a lot of things that most of the world agrees on. + +49:10.780 --> 49:11.940 + Yeah, let's start there. + +49:11.940 --> 49:12.940 + Let's start there. + +49:12.940 --> 49:18.140 + And then once we start there, we'll also get into the habit of having these kind of conversations + +49:18.140 --> 49:21.940 + about, okay, what else should we put in here and have these discussions? + +49:21.940 --> 49:24.100 + This should be a gradual process then. + +49:24.100 --> 49:25.100 + Great. + +49:25.100 --> 49:31.380 + So, but that also means describing these things and describing it to a machine. + +49:31.380 --> 49:35.620 + So one thing, we had a few conversations with Steven Wolfram. + +49:35.620 --> 49:37.140 + I'm not sure if you're familiar with Steven Wolfram. + +49:37.140 --> 49:38.500 + Oh yeah, I know him quite well. + +49:38.500 --> 49:43.380 + So he has, you know, he works with a bunch of things, but you know, cellular automata, + +49:43.380 --> 49:47.660 + these simple computable things, these computation systems. + +49:47.660 --> 49:52.380 + And he kind of mentioned that, you know, we probably have already within these systems + +49:52.380 --> 49:59.580 + already something that's AGI, meaning like we just don't know it because we can't talk + +49:59.580 --> 50:00.580 + to it. + +50:00.580 --> 50:06.380 + So if you give me this chance to try it, to try to at least form a question out of this, + +50:06.380 --> 50:12.780 + because I think it's an interesting idea to think that we can have intelligent systems, + +50:12.780 --> 50:17.260 + but we don't know how to describe something to them and they can't communicate with us. + +50:17.260 --> 50:21.220 + I know you're doing a little bit of work in explainable AI, trying to get AI to explain + +50:21.220 --> 50:22.220 + itself. + +50:22.220 --> 50:28.340 + So what are your thoughts of natural language processing or some kind of other communication? + +50:28.340 --> 50:30.220 + How does the AI explain something to us? + +50:30.220 --> 50:33.740 + How do we explain something to it, to machines? + +50:33.740 --> 50:35.420 + Or you think of it differently? + +50:35.420 --> 50:40.100 + So there are two separate parts to your question there. + +50:40.100 --> 50:43.900 + One of them has to do with communication, which is super interesting and I'll get to + +50:43.900 --> 50:44.900 + that in a sec. + +50:44.900 --> 50:50.100 + The other is whether we already have AGI, we just haven't noticed it. + +50:50.100 --> 50:54.340 + There, I beg to differ. + +50:54.340 --> 50:58.420 + And don't think there's anything in any cellular automaton or anything or the internet itself + +50:58.420 --> 51:05.400 + or whatever that has artificial general intelligence in that it didn't really do exactly everything + +51:05.400 --> 51:06.980 + we humans can do better. + +51:06.980 --> 51:14.100 + I think the day that happens, when that happens, we will very soon notice and we'll probably + +51:14.100 --> 51:17.980 + notice even before because in a very, very big way. + +51:17.980 --> 51:18.980 + For the second part though. + +51:18.980 --> 51:20.700 + Can I just, sorry. + +51:20.700 --> 51:30.260 + Because you have this beautiful way to formulate in consciousness as information processing + +51:30.260 --> 51:33.740 + and you can think of intelligence and information processing and you can think of the entire + +51:33.740 --> 51:34.740 + universe. + +51:34.740 --> 51:40.220 + These particles and these systems roaming around that have this information processing + +51:40.220 --> 51:47.500 + power, you don't think there is something with the power to process information in the + +51:47.500 --> 51:55.460 + way that we human beings do that's out there that needs to be sort of connected to. + +51:55.460 --> 51:59.980 + It seems a little bit philosophical perhaps, but there's something compelling to the idea + +51:59.980 --> 52:06.100 + that the power is already there, the focus should be more on being able to communicate + +52:06.100 --> 52:07.100 + with it. + +52:07.100 --> 52:15.340 + Well, I agree that in a certain sense, the hardware processing power is already out there + +52:15.340 --> 52:21.180 + because our universe itself can think of it as being a computer already. + +52:21.180 --> 52:25.540 + It's constantly computing what water waves, how it devolved the water waves and the river + +52:25.540 --> 52:29.860 + Charles and how to move the air molecules around that Seth Lloyd has pointed out. + +52:29.860 --> 52:33.940 + My colleague here that you can even in a very rigorous way think of our entire universe + +52:33.940 --> 52:35.660 + is just being a quantum computer. + +52:35.660 --> 52:40.900 + It's pretty clear that our universe supports this amazing processing power because you + +52:40.900 --> 52:46.580 + can even within this physics computer that we live in, we can even build actual laptops + +52:46.580 --> 52:47.580 + and stuff. + +52:47.580 --> 52:49.140 + So clearly the power is there. + +52:49.140 --> 52:53.420 + It's just that most of the compute power that nature has, it's in my opinion kind of wasting + +52:53.420 --> 52:57.140 + on boring stuff like simulating yet another ocean wave somewhere where no one is even + +52:57.140 --> 52:58.140 + looking. + +52:58.140 --> 53:03.820 + So in a sense, what life does, what we are doing when we build computers is we're rechanneling + +53:03.820 --> 53:09.380 + all this compute that nature is doing anyway into doing things that are more interesting + +53:09.380 --> 53:14.220 + than just yet another ocean wave and do something cool here. + +53:14.220 --> 53:21.100 + So the raw hardware power is there for sure, and even just computing what's going to happen + +53:21.100 --> 53:25.540 + for the next five seconds in this water ball, you know, it takes a ridiculous amount of + +53:25.540 --> 53:28.060 + compute if you do it on a human computer. + +53:28.060 --> 53:30.040 + This water ball just did it. + +53:30.040 --> 53:36.020 + But that does not mean that this water ball has AGI and this because AGI means it should + +53:36.020 --> 53:40.300 + also be able to like I've written my book done this interview. + +53:40.300 --> 53:42.100 + And I don't think it's just communication problems. + +53:42.100 --> 53:47.020 + I don't think it can do it. + +53:47.020 --> 53:51.780 + So Buddhists say when they watch the water and that there is some beauty, that there's + +53:51.780 --> 53:55.380 + some depth and beauty in nature that they can communicate with. + +53:55.380 --> 54:01.180 + Communication is also very important because I mean, look, part of my job is being a teacher + +54:01.180 --> 54:09.940 + and I know some very intelligent professors even who just have a better hard time communicating. + +54:09.940 --> 54:14.620 + They come up with all these brilliant ideas, but to communicate with somebody else, you + +54:14.620 --> 54:17.140 + have to also be able to simulate their own mind. + +54:17.140 --> 54:18.140 + Yes. + +54:18.140 --> 54:22.020 + And build well enough and understand that model of their mind that you can say things + +54:22.020 --> 54:24.500 + that they will understand. + +54:24.500 --> 54:26.700 + And that's quite difficult. + +54:26.700 --> 54:31.620 + And that's why today it's so frustrating if you have a computer that makes some cancer + +54:31.620 --> 54:36.260 + diagnosis and you ask it, well, why are you saying I should have a surgery? + +54:36.260 --> 54:43.620 + And if you don't want to reply, I was trained on five terabytes of data and this is my diagnosis + +54:43.620 --> 54:49.220 + boop, boop, beep, beep, doesn't really instill a lot of confidence, right? + +54:49.220 --> 54:54.420 + So I think we have a lot of work to do on communication there. + +54:54.420 --> 54:59.380 + So what kind of, I think you're doing a little bit of work in explainable AI. + +54:59.380 --> 55:01.340 + What do you think are the most promising avenues? + +55:01.340 --> 55:07.100 + Is it mostly about sort of the Alexa problem of natural language processing of being able + +55:07.100 --> 55:13.220 + to actually use human interpretable methods of communication? + +55:13.220 --> 55:17.500 + So being able to talk to a system and talk back to you, or is there some more fundamental + +55:17.500 --> 55:18.500 + problems to be solved? + +55:18.500 --> 55:21.180 + I think it's all of the above. + +55:21.180 --> 55:27.180 + The natural language processing is obviously important, but there are also more nerdy fundamental + +55:27.180 --> 55:28.180 + problems. + +55:28.180 --> 55:39.180 + Like if you take, you play chess, Russian, I have to, when did you learn Russian? + +55:39.180 --> 55:45.700 + I speak Russian very poorly, but I bought a book, teach yourself Russian, I read a lot, + +55:45.700 --> 55:47.700 + but it was very difficult. + +55:47.700 --> 55:48.700 + Wow. + +55:48.700 --> 55:49.700 + That's why I speak so poorly. + +55:49.700 --> 55:51.700 + How many languages do you know? + +55:51.700 --> 55:52.700 + Wow. + +55:52.700 --> 55:53.700 + That's really impressive. + +55:53.700 --> 55:54.700 + I don't know. + +55:54.700 --> 55:58.740 + My wife has some calculations, but my point was, if you played chess, have you looked + +55:58.740 --> 56:00.260 + at the AlphaZero games? + +56:00.260 --> 56:01.260 + Yeah. + +56:01.260 --> 56:02.260 + Oh, the actual games now. + +56:02.260 --> 56:03.260 + Check it out. + +56:03.260 --> 56:09.900 + Some of them are just mind blowing, really beautiful. + +56:09.900 --> 56:12.460 + If you ask, how did it do that? + +56:12.460 --> 56:14.500 + You got that. + +56:14.500 --> 56:20.540 + Talk to Demis Osabis, others from DeepMind, all they'll ultimately be able to give you + +56:20.540 --> 56:26.940 + is big tables of numbers, matrices that define the neural network, and you can stare at these + +56:26.940 --> 56:32.980 + tables numbers till your face turned blue, and you're not going to understand much about + +56:32.980 --> 56:35.860 + why it made that move. + +56:35.860 --> 56:40.540 + Even if you have a natural language processing that can tell you in human language about, + +56:40.540 --> 56:44.180 + oh, five, seven, point two, eight, still not going to really help. + +56:44.180 --> 56:50.660 + I think there's a whole spectrum of fun challenges there involved in taking computation that + +56:50.660 --> 56:59.940 + does intelligent things and transforming it into something equally good, equally intelligent, + +56:59.940 --> 57:02.060 + but that's more understandable. + +57:02.060 --> 57:08.180 + I think that's really valuable because I think as we put machines in charge of ever more + +57:08.180 --> 57:13.540 + infrastructure in our world, the power grid, the trading on the stock market, weapon systems, + +57:13.540 --> 57:19.620 + and so on, it's absolutely crucial that we can trust these AIs that do all we want and + +57:19.620 --> 57:25.860 + trust really comes from understanding in a very fundamental way. + +57:25.860 --> 57:29.940 + That's why I'm working on this, because I think the more if we're going to have some + +57:29.940 --> 57:34.700 + hope of ensuring that machines have adopted our goals and that they're going to retain + +57:34.700 --> 57:41.260 + them, that kind of trust, I think, needs to be based on things you can actually understand, + +57:41.260 --> 57:47.140 + preferably even improve theorems on, even with a self driving car, right? + +57:47.140 --> 57:51.020 + If someone just tells you it's been trained on tons of data and never crashed, it's less + +57:51.020 --> 57:54.460 + reassuring than if someone actually has a proof. + +57:54.460 --> 57:58.820 + Maybe it's a computer verified proof, but still it says that under no circumstances + +57:58.820 --> 58:02.420 + is this car just going to swerve into oncoming traffic. + +58:02.420 --> 58:09.460 + And that kind of information helps build trust and helps build the alignment of goals, at + +58:09.460 --> 58:12.300 + least awareness that your goals, your values are aligned. + +58:12.300 --> 58:17.620 + And I think even in the very short term, if you look at how today, this absolutely pathetic + +58:17.620 --> 58:25.980 + state of cybersecurity that we have, where is it, 3 billion Yahoo accounts are packed + +58:25.980 --> 58:34.300 + and almost every American's credit card and so on, you know, why is this happening? + +58:34.300 --> 58:39.940 + It's ultimately happening because we have software that nobody fully understood how + +58:39.940 --> 58:41.460 + it worked. + +58:41.460 --> 58:45.100 + That's why the bugs hadn't been found, right? + +58:45.100 --> 58:50.340 + And I think AI can be used very effectively for offense for hacking, but it can also be + +58:50.340 --> 59:00.580 + used for defense, hopefully, automating verifiability and creating systems that are built in different + +59:00.580 --> 59:03.140 + ways so you can actually prove things about them. + +59:03.140 --> 59:05.460 + And it's important. + +59:05.460 --> 59:09.740 + So speaking of software that nobody understands how it works, of course, a bunch of people + +59:09.740 --> 59:14.820 + ask about your paper about your thoughts of why does deep and cheap learning work so well? + +59:14.820 --> 59:19.280 + That's the paper, but what are your thoughts on deep learning, these kind of simplified + +59:19.280 --> 59:26.620 + models of our own brains that have been able to do some successful perception work, pattern + +59:26.620 --> 59:30.940 + recognition work, and now with AlphaZero and so on, do some clever things? + +59:30.940 --> 59:35.740 + What are your thoughts about the promise limitations of this piece? + +59:35.740 --> 59:37.140 + Great. + +59:37.140 --> 59:44.300 + I think there are a number of very important insights, very important lessons we can always + +59:44.300 --> 59:47.340 + draw from these kind of successes. + +59:47.340 --> 59:50.460 + One of them is when you look at the human brain, you see it's very complicated, a tenth + +59:50.460 --> 59:54.140 + of 11 neurons, and there are all these different kinds of neurons, and yada yada, and there's + +59:54.140 --> 59:57.980 + been this long debate about whether the fact that we have dozens of different kinds is + +59:57.980 --> 1:00:01.580 + actually necessary for intelligence. + +1:00:01.580 --> 1:00:06.500 + We can now, I think, quite convincingly answer that question of no, it's enough to have just + +1:00:06.500 --> 1:00:07.500 + one kind. + +1:00:07.500 --> 1:00:11.780 + If you look under the hood of AlphaZero, there's only one kind of neuron, and it's ridiculously + +1:00:11.780 --> 1:00:15.060 + simple, a simple mathematical thing. + +1:00:15.060 --> 1:00:21.380 + So it's just like in physics, if you have a gas with waves in it, it's not the detailed + +1:00:21.380 --> 1:00:24.380 + nature of the molecules that matter. + +1:00:24.380 --> 1:00:27.060 + It's the collective behavior, somehow. + +1:00:27.060 --> 1:00:33.060 + Similarly, it's this higher level structure of the network that matters, not that you + +1:00:33.060 --> 1:00:34.060 + have 20 kinds of neurons. + +1:00:34.060 --> 1:00:41.740 + I think our brain is such a complicated mess because it wasn't evolved just to be intelligent, + +1:00:41.740 --> 1:00:51.740 + it was evolved to also be self assembling, and self repairing, and evolutionarily attainable. + +1:00:51.740 --> 1:00:53.660 + And patches and so on. + +1:00:53.660 --> 1:00:58.700 + So I think it's pretty, my hunch is that we're going to understand how to build AGI before + +1:00:58.700 --> 1:01:01.060 + we fully understand how our brains work. + +1:01:01.060 --> 1:01:06.260 + Just like we understood how to build flying machines long before we were able to build + +1:01:06.260 --> 1:01:07.260 + a mechanical bird. + +1:01:07.260 --> 1:01:08.260 + Yeah, that's right. + +1:01:08.260 --> 1:01:15.300 + You've given the example of mechanical birds and airplanes, and airplanes do a pretty good + +1:01:15.300 --> 1:01:18.620 + job of flying without really mimicking bird flight. + +1:01:18.620 --> 1:01:23.180 + And even now, after 100 years later, did you see the TED talk with this German group of + +1:01:23.180 --> 1:01:24.180 + mechanical birds? + +1:01:24.180 --> 1:01:25.180 + I did not. + +1:01:25.180 --> 1:01:26.180 + I've heard you mention it. + +1:01:26.180 --> 1:01:27.180 + Check it out. + +1:01:27.180 --> 1:01:28.180 + It's amazing. + +1:01:28.180 --> 1:01:30.180 + But even after that, we still don't fly in mechanical birds because it turned out the + +1:01:30.180 --> 1:01:34.580 + way we came up with simpler, and it's better for our purposes, and I think it might be the + +1:01:34.580 --> 1:01:35.580 + same there. + +1:01:35.580 --> 1:01:38.140 + So that's one lesson. + +1:01:38.140 --> 1:01:42.020 + Another lesson is one of what our paper was about. + +1:01:42.020 --> 1:01:47.420 + Well, first, as a physicist thought, it was fascinating how there's a very close mathematical + +1:01:47.420 --> 1:01:50.900 + relationship, actually, between our artificial neural networks. + +1:01:50.900 --> 1:01:56.580 + And a lot of things that we've studied for in physics go by nerdy names like the renormalization + +1:01:56.580 --> 1:02:01.100 + group equation and Hamiltonians and yada, yada, yada. + +1:02:01.100 --> 1:02:11.380 + And when you look a little more closely at this, you have, at first, I was like, well, + +1:02:11.380 --> 1:02:18.700 + there's something crazy here that doesn't make sense because we know that if you even + +1:02:18.700 --> 1:02:23.380 + want to build a super simple neural network to tell apart cat pictures and dog pictures, + +1:02:23.380 --> 1:02:27.260 + right, that you can do that very, very well now. + +1:02:27.260 --> 1:02:31.540 + But if you think about it a little bit, you convince yourself it must be impossible because + +1:02:31.540 --> 1:02:36.420 + if I have one megapixel, even if each pixel is just black or white, there's two to the + +1:02:36.420 --> 1:02:40.900 + power of one million possible images, which is way more than there are atoms in our universe. + +1:02:40.900 --> 1:02:47.220 + So in order to, and then for each one of those, I have to assign a number, which is the probability + +1:02:47.220 --> 1:02:49.100 + that it's a dog. + +1:02:49.100 --> 1:02:55.900 + So an arbitrary function of images is a list of more numbers than there are atoms in our + +1:02:55.900 --> 1:02:56.900 + universe. + +1:02:56.900 --> 1:03:02.220 + So clearly, I can't store that under the hood of my, my GPU or my, my computer yet somehow + +1:03:02.220 --> 1:03:03.220 + works. + +1:03:03.220 --> 1:03:04.220 + So what does that mean? + +1:03:04.220 --> 1:03:12.940 + Well, it means that out of all of the problems that you could try to solve with a neural network, + +1:03:12.940 --> 1:03:17.940 + almost all of them are impossible to solve with a reasonably sized one. + +1:03:17.940 --> 1:03:24.820 + But then what we showed in our paper was, was that the, the fraction, the kind of problems, + +1:03:24.820 --> 1:03:29.740 + the fraction of all the problems that you could possibly pose that the, that we actually + +1:03:29.740 --> 1:03:34.980 + care about given the laws of physics is also an infinitesimally tiny little part. + +1:03:34.980 --> 1:03:37.180 + And amazingly, they're basically the same part. + +1:03:37.180 --> 1:03:38.180 + Yeah. + +1:03:38.180 --> 1:03:41.180 + It's almost like our world was created for, I mean, they kind of come together. + +1:03:41.180 --> 1:03:42.180 + Yeah. + +1:03:42.180 --> 1:03:44.940 + You, but you could say maybe where the world created the world that the world was created + +1:03:44.940 --> 1:03:50.300 + for us, but I have a more modest interpretation, which is that instead evolution endowments + +1:03:50.300 --> 1:03:54.700 + with neural networks, precisely for that reason, because this particular architecture has + +1:03:54.700 --> 1:04:02.380 + opposed to the one in your laptop is very, very well adapted to solving the kind of problems + +1:04:02.380 --> 1:04:05.540 + that nature kept presenting our ancestors with, right? + +1:04:05.540 --> 1:04:09.380 + So it makes sense that why do we have a brain in the first place? + +1:04:09.380 --> 1:04:12.940 + It's to be able to make predictions about the future and so on. + +1:04:12.940 --> 1:04:17.580 + So if we had a sucky system, which could never solve it, it wouldn't have a lot. + +1:04:17.580 --> 1:04:23.420 + So, but it's, so this is, this is a, I think a very beautiful fact. + +1:04:23.420 --> 1:04:24.420 + Yeah. + +1:04:24.420 --> 1:04:28.780 + And you also realize that there's, there, that we, there've been, it's been earlier + +1:04:28.780 --> 1:04:34.140 + work on, on why deeper networks are good, but we were able to show an additional cool + +1:04:34.140 --> 1:04:40.260 + fact there, which is that even incredibly simple problems, like suppose I give you a + +1:04:40.260 --> 1:04:45.020 + thousand numbers and ask you to multiply them together and you can write a few lines of + +1:04:45.020 --> 1:04:46.820 + code, boom, done, trivial. + +1:04:46.820 --> 1:04:52.580 + If you just try to do that with a neural network that has only one single hidden layer in it, + +1:04:52.580 --> 1:04:59.940 + you can do it, but you're going to need two to the power of thousand neurons to multiply + +1:04:59.940 --> 1:05:03.260 + a thousand numbers, which is again, more neurons than their atoms in our universe. + +1:05:03.260 --> 1:05:05.740 + So that's fascinating. + +1:05:05.740 --> 1:05:11.580 + But if you allow, if you allow yourself, make it a deep network of many layers, you only + +1:05:11.580 --> 1:05:15.340 + need four thousand neurons, it's perfectly feasible. + +1:05:15.340 --> 1:05:17.500 + So that's really interesting. + +1:05:17.500 --> 1:05:18.500 + Yeah. + +1:05:18.500 --> 1:05:19.500 + Yeah. + +1:05:19.500 --> 1:05:24.460 + So architecture type, I mean, you mentioned Schrodinger's equation and what are your thoughts + +1:05:24.460 --> 1:05:32.860 + about quantum computing and the role of this kind of computational unit in creating an + +1:05:32.860 --> 1:05:34.900 + intelligent system? + +1:05:34.900 --> 1:05:41.100 + In some Hollywood movies that I don't mention my name because I don't want to spoil them. + +1:05:41.100 --> 1:05:46.820 + The way they get AGI is building a quantum computer because the word quantum sounds + +1:05:46.820 --> 1:05:47.820 + cool and so on. + +1:05:47.820 --> 1:05:48.820 + That's right. + +1:05:48.820 --> 1:05:54.940 + But first of all, I think we don't need quantum computers to build AGI. + +1:05:54.940 --> 1:06:01.740 + I suspect your brain is not quantum computer in any found sense. + +1:06:01.740 --> 1:06:03.460 + So you don't even wrote a paper about that. + +1:06:03.460 --> 1:06:09.060 + Many years ago, I calculated the so called decoherence time that how long it takes until + +1:06:09.060 --> 1:06:16.900 + the quantum computerness of what your neurons are doing gets erased by just random noise + +1:06:16.900 --> 1:06:21.420 + from the environment and it's about 10 to the minus 21 seconds. + +1:06:21.420 --> 1:06:27.420 + So as cool as it would be to have a quantum computer in my head, I don't think that fast. + +1:06:27.420 --> 1:06:35.820 + On the other hand, there are very cool things you could do with quantum computers or I think + +1:06:35.820 --> 1:06:40.780 + we'll be able to do soon when we get bigger ones that might actually help machine learning + +1:06:40.780 --> 1:06:43.180 + do even better than the brain. + +1:06:43.180 --> 1:06:58.620 + So for example, one, this is just a moonshot, but hey, learning is very much same thing + +1:06:58.620 --> 1:07:00.860 + as search. + +1:07:00.860 --> 1:07:05.460 + If you're trying to train a neural network to get really learned to do something really + +1:07:05.460 --> 1:07:10.820 + well, you have some loss function, you have a bunch of knobs you can turn represented + +1:07:10.820 --> 1:07:14.420 + by a bunch of numbers and you're trying to tweak them so that it becomes as good as possible + +1:07:14.420 --> 1:07:15.420 + at this thing. + +1:07:15.420 --> 1:07:22.580 + So if you think of a landscape with some valley, where each dimension of the landscape corresponds + +1:07:22.580 --> 1:07:25.780 + to some number you can change, you're trying to find the minimum. + +1:07:25.780 --> 1:07:29.980 + And it's well known that if you have a very high dimensional landscape, complicated things, + +1:07:29.980 --> 1:07:34.140 + it's super hard to find the minimum. + +1:07:34.140 --> 1:07:37.500 + Quantum mechanics is amazingly good at this. + +1:07:37.500 --> 1:07:42.980 + If I want to know what's the lowest energy state this water can possibly have incredibly + +1:07:42.980 --> 1:07:47.860 + hard to compute, but nature will happily figure this out for you if you just cool it down, + +1:07:47.860 --> 1:07:50.860 + make it very, very cold. + +1:07:50.860 --> 1:07:55.260 + If you put a ball somewhere, it'll roll down to its minimum and this happens metaphorically + +1:07:55.260 --> 1:07:57.620 + at the energy landscape too. + +1:07:57.620 --> 1:08:02.940 + And quantum mechanics even uses some clever tricks which today's machine learning systems + +1:08:02.940 --> 1:08:03.940 + don't. + +1:08:03.940 --> 1:08:07.940 + If you're trying to find the minimum and you get stuck in the little local minimum here + +1:08:07.940 --> 1:08:14.180 + in quantum mechanics, you can actually tunnel through the barrier and get unstuck again. + +1:08:14.180 --> 1:08:15.420 + And that's really interesting. + +1:08:15.420 --> 1:08:16.420 + Yeah. + +1:08:16.420 --> 1:08:22.940 + So maybe for example, we'll one day use quantum computers that help train neural networks + +1:08:22.940 --> 1:08:23.940 + better. + +1:08:23.940 --> 1:08:24.940 + That's really interesting. + +1:08:24.940 --> 1:08:25.940 + Okay. + +1:08:25.940 --> 1:08:32.020 + So as a component of kind of the learning process, for example, let me ask sort of wrapping + +1:08:32.020 --> 1:08:34.060 + up here a little bit. + +1:08:34.060 --> 1:08:40.540 + Let me return to the questions of our human nature and love, as I mentioned. + +1:08:40.540 --> 1:08:48.020 + So do you think you mentioned sort of a helper robot that you could think of also personal + +1:08:48.020 --> 1:08:49.020 + robots. + +1:08:49.020 --> 1:08:55.300 + Do you think the way we human beings fall in love and get connected to each other is + +1:08:55.300 --> 1:09:00.420 + possible to achieve in an AI system and human level AI intelligence system. + +1:09:00.420 --> 1:09:06.100 + Do you think we would ever see that kind of connection or, you know, in all this discussion + +1:09:06.100 --> 1:09:11.460 + about solving complex goals, as this kind of human social connection, do you think that's + +1:09:11.460 --> 1:09:16.460 + one of the goals on the peaks and valleys that were the raising sea levels that we'd + +1:09:16.460 --> 1:09:17.460 + be able to achieve? + +1:09:17.460 --> 1:09:22.180 + Or do you think that's something that's ultimately, or at least in the short term, relative to + +1:09:22.180 --> 1:09:23.620 + the other goals is not achievable? + +1:09:23.620 --> 1:09:25.220 + I think it's all possible. + +1:09:25.220 --> 1:09:31.780 + And I mean, in recent, there's a very wide range of guesses, as you know, among AI researchers + +1:09:31.780 --> 1:09:35.300 + when we're going to get AGI. + +1:09:35.300 --> 1:09:39.620 + Some people, you know, like our friend Rodney Brooks said, it's going to be hundreds of + +1:09:39.620 --> 1:09:41.140 + years at least. + +1:09:41.140 --> 1:09:44.780 + And then there are many others that think it's going to happen relatively much sooner. + +1:09:44.780 --> 1:09:52.140 + Recent polls, maybe half or so, AI researchers think we're going to get AGI within decades. + +1:09:52.140 --> 1:09:56.260 + So if that happens, of course, then I think these things are all possible. + +1:09:56.260 --> 1:10:01.860 + But in terms of whether it will happen, I think we shouldn't spend so much time asking, + +1:10:01.860 --> 1:10:04.260 + what do we think will happen in the future? + +1:10:04.260 --> 1:10:08.980 + As if we are just some sort of pathetic, passive bystanders, you know, waiting for the future + +1:10:08.980 --> 1:10:12.740 + to happen to us, hey, we're the ones creating this future, right? + +1:10:12.740 --> 1:10:18.340 + So we should be proactive about it and ask ourselves what sort of future we would like + +1:10:18.340 --> 1:10:19.340 + to have happen. + +1:10:19.340 --> 1:10:20.340 + That's right. + +1:10:20.340 --> 1:10:21.340 + Trying to make it like that. + +1:10:21.340 --> 1:10:25.660 + Well, what I prefer is some sort of incredibly boring zombie like future where there's all + +1:10:25.660 --> 1:10:30.220 + these mechanical things happening and there's no passion, no emotion, no experience, maybe + +1:10:30.220 --> 1:10:31.220 + even. + +1:10:31.220 --> 1:10:35.740 + No, I would, of course, much rather prefer it if all the things that we find that we + +1:10:35.740 --> 1:10:44.180 + value the most about humanity are a subjective experience, passion, inspiration, love, you + +1:10:44.180 --> 1:10:50.780 + know, if we can create a future where those things do exist. + +1:10:50.780 --> 1:10:56.500 + You know, I think ultimately it's not our universe giving meaning to us, it's us giving + +1:10:56.500 --> 1:10:58.500 + meaning to our universe. + +1:10:58.500 --> 1:11:03.620 + And if we build more advanced intelligence, let's make sure we build it in such a way + +1:11:03.620 --> 1:11:09.100 + that meaning is part of it. + +1:11:09.100 --> 1:11:13.900 + A lot of people that seriously study this problem and think of it from different angles have + +1:11:13.900 --> 1:11:20.140 + trouble in the majority of cases, if they think through that happen, are the ones that + +1:11:20.140 --> 1:11:22.620 + are not beneficial to humanity. + +1:11:22.620 --> 1:11:27.260 + And so, yeah, so what are your thoughts? + +1:11:27.260 --> 1:11:33.820 + What should people, you know, I really don't like people to be terrified, what's the way + +1:11:33.820 --> 1:11:38.660 + for people to think about it in a way that, in a way we can solve it and we can make it + +1:11:38.660 --> 1:11:39.660 + better. + +1:11:39.660 --> 1:11:40.660 + Yeah. + +1:11:40.660 --> 1:11:44.780 + No, I don't think panicking is going to help in any way, it's not going to increase chances + +1:11:44.780 --> 1:11:46.060 + of things going well either. + +1:11:46.060 --> 1:11:49.340 + Even if you are in a situation where there is a real threat, does it help if everybody + +1:11:49.340 --> 1:11:50.620 + just freaks out? + +1:11:50.620 --> 1:11:51.620 + Right. + +1:11:51.620 --> 1:11:53.620 + No, of course not. + +1:11:53.620 --> 1:11:59.740 + I think, yeah, there are, of course, ways in which things can go horribly wrong. + +1:11:59.740 --> 1:12:04.460 + First of all, it's important when we think about this thing, this, about the problems + +1:12:04.460 --> 1:12:08.780 + and risks, to also remember how huge the upsides can be if we get it right. + +1:12:08.780 --> 1:12:13.420 + Everything we love about society and civilization is a product of intelligence. + +1:12:13.420 --> 1:12:17.980 + So if we can amplify our intelligence with machine intelligence and not anymore lose + +1:12:17.980 --> 1:12:23.380 + our loved ones, what we're told is an uncurable disease and things like this, of course, we + +1:12:23.380 --> 1:12:24.940 + should aspire to that. + +1:12:24.940 --> 1:12:28.700 + So that can be a motivator, I think, reminding yourselves that the reason we try to solve + +1:12:28.700 --> 1:12:34.140 + problems is not just because we're trying to avoid gloom, but because we're trying to + +1:12:34.140 --> 1:12:35.900 + do something great. + +1:12:35.900 --> 1:12:43.340 + But then in terms of the risks, I think the really important question is to ask, what + +1:12:43.340 --> 1:12:47.740 + can we do today that will actually help make the outcome good, right? + +1:12:47.740 --> 1:12:52.700 + And dismissing the risk is not one of them, you know, I find it quite funny often when + +1:12:52.700 --> 1:13:01.540 + I'm in discussion panels about these things, how the people who work for companies will + +1:13:01.540 --> 1:13:05.100 + always be like, oh, nothing to worry about, nothing to worry about, nothing to worry about. + +1:13:05.100 --> 1:13:09.980 + And it's always, it's only academics sometimes express concerns. + +1:13:09.980 --> 1:13:10.980 + That's not surprising at all. + +1:13:10.980 --> 1:13:17.500 + If you think about it, often Sinclair quipped, right, that it's hard to make a man believe + +1:13:17.500 --> 1:13:20.620 + in something when his income depends on not believing in it. + +1:13:20.620 --> 1:13:25.580 + And frankly, we know a lot of these people in companies that they're just as concerned + +1:13:25.580 --> 1:13:26.580 + as anyone else. + +1:13:26.580 --> 1:13:30.300 + But if you're the CEO of a company, that's not something you want to go on record saying + +1:13:30.300 --> 1:13:34.980 + when you have silly journalists who are going to put a picture of a Terminator robot when + +1:13:34.980 --> 1:13:35.980 + they quote you. + +1:13:35.980 --> 1:13:39.380 + So, so the issues are real. + +1:13:39.380 --> 1:13:45.660 + And the way I think about what the issue is, is basically, you know, the real choice we + +1:13:45.660 --> 1:13:51.980 + have is, first of all, are we going to dismiss this, the risks and say, well, you know, let's + +1:13:51.980 --> 1:13:57.140 + just go ahead and build machines that can do everything we can do better and cheaper, + +1:13:57.140 --> 1:14:00.940 + you know, let's just make ourselves obsolete as fast as possible or what could possibly + +1:14:00.940 --> 1:14:01.940 + go wrong. + +1:14:01.940 --> 1:14:02.940 + Right. + +1:14:02.940 --> 1:14:03.940 + That's one attitude. + +1:14:03.940 --> 1:14:09.380 + The opposite attitude that I think is to say, it's incredible potential, you know, let's + +1:14:09.380 --> 1:14:14.900 + think about what kind of future we're really, really excited about. + +1:14:14.900 --> 1:14:18.700 + What are the shared goals that we can really aspire towards? + +1:14:18.700 --> 1:14:22.100 + And then let's think really hard about how we can actually get there. + +1:14:22.100 --> 1:14:23.100 + So start with it. + +1:14:23.100 --> 1:14:24.460 + Don't start thinking about the risks. + +1:14:24.460 --> 1:14:26.940 + Start thinking about the goals. + +1:14:26.940 --> 1:14:30.540 + And then when you do that, then you can think about the obstacles you want to avoid, right? + +1:14:30.540 --> 1:14:34.420 + I often get students coming in right here into my office for career advice. + +1:14:34.420 --> 1:14:38.060 + Always ask them this very question, where do you want to be in the future? + +1:14:38.060 --> 1:14:42.580 + If all she can say is, oh, maybe I'll have cancer, maybe I'll run over by a truck. + +1:14:42.580 --> 1:14:44.420 + Focus on the obstacles instead of the goal. + +1:14:44.420 --> 1:14:49.340 + She's just going to end up a hypochondriac paranoid, whereas if she comes in and fire + +1:14:49.340 --> 1:14:54.060 + in her eyes and is like, I want to be there, and then we can talk about the obstacles and + +1:14:54.060 --> 1:14:56.100 + see how we can circumvent them. + +1:14:56.100 --> 1:14:59.100 + That's I think a much, much healthier attitude. + +1:14:59.100 --> 1:15:01.540 + And that's really what we're in. + +1:15:01.540 --> 1:15:09.420 + And I feel it's very challenging to come up with a vision for the future, which we're + +1:15:09.420 --> 1:15:10.660 + unequivocally excited about. + +1:15:10.660 --> 1:15:14.300 + I'm not just talking now in the vague terms like, yeah, let's cure cancer. + +1:15:14.300 --> 1:15:18.500 + I'm talking about what kind of society do we want to create? + +1:15:18.500 --> 1:15:25.380 + What do we want it to mean to be human in the age of AI, in the age of AGI? + +1:15:25.380 --> 1:15:31.460 + So if we can have this conversation, broad, inclusive conversation, and gradually start + +1:15:31.460 --> 1:15:38.100 + converging towards some future with some direction at least that we want to steer towards, right? + +1:15:38.100 --> 1:15:42.340 + Then we'll be much more motivated to constructively take on the obstacles. + +1:15:42.340 --> 1:15:54.260 + And I think if I wrap this up in a more succinct way, I think we can all agree already now that + +1:15:54.260 --> 1:16:05.540 + we should aspire to build AGI that doesn't overpower us, but that empowers us. + +1:16:05.540 --> 1:16:10.820 + And think of the many various ways that can do that, whether that's from my side of the + +1:16:10.820 --> 1:16:12.860 + world of autonomous vehicles. + +1:16:12.860 --> 1:16:17.020 + I'm personally actually from the camp that believes this human level intelligence is + +1:16:17.020 --> 1:16:22.780 + required to achieve something like vehicles that would actually be something we would + +1:16:22.780 --> 1:16:25.380 + enjoy using and being part of. + +1:16:25.380 --> 1:16:26.380 + So that's the one example. + +1:16:26.380 --> 1:16:31.140 + And certainly there's a lot of other types of robots and medicine and so on. + +1:16:31.140 --> 1:16:35.300 + So focusing on those and then coming up with the obstacles, coming up with the ways that + +1:16:35.300 --> 1:16:38.420 + that can go wrong and solving those one at a time. + +1:16:38.420 --> 1:16:42.980 + And just because you can build an autonomous vehicle, even if you could build one that + +1:16:42.980 --> 1:16:47.500 + would drive this final AGI, maybe there are some things in life that we would actually + +1:16:47.500 --> 1:16:48.500 + want to do ourselves. + +1:16:48.500 --> 1:16:49.500 + That's right. + +1:16:49.500 --> 1:16:50.500 + Right? + +1:16:50.500 --> 1:16:54.660 + Like, for example, if you think of our society as a whole, there are some things that we + +1:16:54.660 --> 1:16:57.540 + find very meaningful to do. + +1:16:57.540 --> 1:17:02.100 + And that doesn't mean we have to stop doing them just because machines can do them better. + +1:17:02.100 --> 1:17:06.660 + I'm not going to stop playing tennis just the day someone builds a tennis robot and + +1:17:06.660 --> 1:17:07.660 + beat me. + +1:17:07.660 --> 1:17:09.900 + People are still playing chess and even go. + +1:17:09.900 --> 1:17:10.900 + Yeah. + +1:17:10.900 --> 1:17:19.100 + And in this very near term, even some people are advocating basic income, replace jobs. + +1:17:19.100 --> 1:17:22.780 + But if the government is going to be willing to just hand out cash to people for doing + +1:17:22.780 --> 1:17:27.660 + nothing, then one should also seriously consider whether the government should also just hire + +1:17:27.660 --> 1:17:33.380 + a lot more teachers and nurses and the kind of jobs which people often find great fulfillment + +1:17:33.380 --> 1:17:34.380 + in doing, right? + +1:17:34.380 --> 1:17:38.900 + We get very tired of hearing politicians saying, oh, we can't afford hiring more teachers, + +1:17:38.900 --> 1:17:41.700 + but we're going to maybe have basic income. + +1:17:41.700 --> 1:17:46.340 + If we can have more serious research and thought into what gives meaning to our lives, the + +1:17:46.340 --> 1:17:50.700 + jobs give so much more than income, right? + +1:17:50.700 --> 1:18:00.020 + And then think about, in the future, what are the roles that we want to have people + +1:18:00.020 --> 1:18:03.180 + continually feeling empowered by machines? + +1:18:03.180 --> 1:18:08.900 + And I think sort of, I come from the Russia, from the Soviet Union, and I think for a lot + +1:18:08.900 --> 1:18:14.100 + of people in the 20th century, going to the moon, going to space was an inspiring thing. + +1:18:14.100 --> 1:18:21.300 + I feel like the universe of the mind, so AI, understanding, creating intelligence is that + +1:18:21.300 --> 1:18:23.380 + for the 21st century. + +1:18:23.380 --> 1:18:26.740 + So it's really surprising, and I've heard you mention this, it's really surprising to + +1:18:26.740 --> 1:18:31.940 + me both on the research funding side that it's not funded as greatly as it could be. + +1:18:31.940 --> 1:18:36.500 + But most importantly, on the politician side, that it's not part of the public discourse + +1:18:36.500 --> 1:18:44.300 + except in killer bots, terminator kind of view, that people are not yet, I think, perhaps + +1:18:44.300 --> 1:18:48.260 + excited by the possible positive future that we can build together. + +1:18:48.260 --> 1:18:54.660 + So we should be, because politicians usually just focus on the next election cycle, right? + +1:18:54.660 --> 1:18:59.340 + The single most important thing I feel we humans have learned in the entire history of science + +1:18:59.340 --> 1:19:07.460 + is they were the masters of underestimation, underestimated the size of our cosmos, again + +1:19:07.460 --> 1:19:11.380 + and again, realizing that everything we thought existed was just a small part of something + +1:19:11.380 --> 1:19:12.380 + grander, right? + +1:19:12.380 --> 1:19:18.580 + Planet, solar system, the galaxy, clusters of galaxies, the universe. + +1:19:18.580 --> 1:19:25.700 + And we now know that we have the future has just so much more potential than our ancestors + +1:19:25.700 --> 1:19:27.820 + could ever have dreamt of. + +1:19:27.820 --> 1:19:39.820 + This cosmos, imagine if all of Earth was completely devoid of life except for Cambridge, Massachusetts. + +1:19:39.820 --> 1:19:44.220 + Wouldn't it be kind of lame if all we ever aspired to was to stay in Cambridge, Massachusetts + +1:19:44.220 --> 1:19:49.660 + forever and then go extinct in one week, even though Earth was going to continue on for + +1:19:49.660 --> 1:19:50.660 + longer? + +1:19:50.660 --> 1:19:57.300 + That sort of attitude I think we have now on the cosmic scale, we can, life can flourish + +1:19:57.300 --> 1:20:00.820 + on Earth, not for four years, but for billions of years. + +1:20:00.820 --> 1:20:06.340 + I can even tell you about how to move it out of harm's way when the sun gets too hot. + +1:20:06.340 --> 1:20:11.900 + And then we have so much more resources out here, which today, maybe there are a lot of + +1:20:11.900 --> 1:20:19.380 + other planets with bacteria or cow like life on them, but most of this, all this opportunity + +1:20:19.380 --> 1:20:25.380 + seems as far as we can tell to be largely dead, like the Sahara Desert, and yet we have the + +1:20:25.380 --> 1:20:30.380 + opportunity to help life flourish around this for billions of years. + +1:20:30.380 --> 1:20:37.420 + So like, let's quit squabbling about whether some little border should be drawn one mile + +1:20:37.420 --> 1:20:43.380 + to the left or right and look up into the skies and realize, hey, we can do such incredible + +1:20:43.380 --> 1:20:44.380 + things. + +1:20:44.380 --> 1:20:45.380 + Yeah. + +1:20:45.380 --> 1:20:49.980 + And that's I think why it's really exciting that you and others are connected with some + +1:20:49.980 --> 1:20:54.740 + of the work Elon Musk is doing because he's literally going out into that space, really + +1:20:54.740 --> 1:20:56.260 + exploring our universe. + +1:20:56.260 --> 1:20:57.260 + And it's wonderful. + +1:20:57.260 --> 1:21:02.340 + That is exactly why Elon Musk is so misunderstood, right? + +1:21:02.340 --> 1:21:05.300 + Misconstrued with some kind of pessimistic doomsayer. + +1:21:05.300 --> 1:21:10.860 + The reason he cares so much about AI safety is because he more than almost anyone else + +1:21:10.860 --> 1:21:13.340 + appreciates these amazing opportunities. + +1:21:13.340 --> 1:21:16.340 + It will squander if we wipe out here on Earth. + +1:21:16.340 --> 1:21:22.740 + We're not just going to wipe out the next generation, but all generations and this incredible + +1:21:22.740 --> 1:21:25.580 + opportunity that's out there and that would be really be a waste. + +1:21:25.580 --> 1:21:32.740 + And AI, for people who think that there would be better to do without technology, let me + +1:21:32.740 --> 1:21:37.740 + just mention that if we don't improve our technology, the question isn't whether humanity + +1:21:37.740 --> 1:21:38.740 + is going to go extinct. + +1:21:38.740 --> 1:21:43.620 + The question is just whether we're going to get taken out by the next big asteroid or + +1:21:43.620 --> 1:21:49.540 + the next super volcano or something else dumb that we could easily prevent with more tech, + +1:21:49.540 --> 1:21:50.540 + right? + +1:21:50.540 --> 1:21:56.220 + If we want life to flourish throughout the cosmos, AI is the key to it. + +1:21:56.220 --> 1:22:04.780 + As I mentioned in a lot of detail in my book, even many of the most inspired sci fi writers + +1:22:04.780 --> 1:22:11.580 + I feel have totally underestimated the opportunities for space travel, especially to other galaxies, + +1:22:11.580 --> 1:22:17.100 + because they weren't thinking about the possibility of AGI, which just makes it so much easier. + +1:22:17.100 --> 1:22:18.100 + Right. + +1:22:18.100 --> 1:22:25.900 + Yeah, so that goes to a view of AGI that enables our progress, that enables a better life. + +1:22:25.900 --> 1:22:30.060 + So that's a beautiful way to put it and something to strive for. + +1:22:30.060 --> 1:22:31.580 + So Max, thank you so much. + +1:22:31.580 --> 1:22:32.580 + Thank you for your time today. + +1:22:32.580 --> 1:22:33.580 + It's been awesome. + +1:22:33.580 --> 1:22:34.580 + Thank you so much. + +1:22:34.580 --> 1:22:35.580 + Thanks. + +1:22:35.580 --> 1:22:36.580 + Merci beaucoup. + +1:22:36.580 --> 1:22:49.100 + Thank you so much for your time today and thank you so much for your time and for your + +1:22:49.100 --> 1:22:50.100 + time. + +1:22:50.100 --> 1:22:51.100 + Thank you. + +1:22:51.100 --> 1:22:52.100 + Thank you. + +1:22:52.100 --> 1:22:53.100 + Bye. + +1:22:53.100 --> 1:22:54.100 + Bye. + +1:22:54.100 --> 1:22:55.100 + Bye. + +1:22:55.100 --> 1:22:56.100 + Bye. + +1:22:56.100 --> 1:22:57.100 + Bye. + +1:22:57.100 --> 1:22:58.100 + Bye. + +1:22:58.100 --> 1:22:59.100 + Bye. + +1:22:59.100 --> 1:23:00.100 + Bye. +