lexicap / vtt /episode_001_small.vtt
Shubham Gupta
Add readme and files
a3be5d0
raw
history blame
106 kB
WEBVTT
00:00.000 --> 00:05.060
As part of MIT course 6S 099 Artificial General Intelligence, I've gotten the chance to sit
00:05.060 --> 00:06.740
down with Max Tagmark.
00:06.740 --> 00:13.780
He is a professor here at MIT, he's a physicist, spent a large part of his career studying the
00:13.780 --> 00:20.660
mysteries of our cosmological universe, but he's also studied and delved into the beneficial
00:20.660 --> 00:25.860
possibilities and the existential risks of artificial intelligence.
00:25.860 --> 00:32.220
Amongst many other things, he's the cofounder of the Future of Life Institute, author of
00:32.220 --> 00:35.140
two books, both of which I highly recommend.
00:35.140 --> 00:40.220
First, our mathematical universe, second is Life 3.0.
00:40.220 --> 00:45.060
He's truly an out of the box thinker and a fun personality, so I really enjoy talking
00:45.060 --> 00:46.060
to him.
00:46.060 --> 00:49.500
If you'd like to see more of these videos in the future, please subscribe and also click
00:49.500 --> 00:52.980
the little bell icon to make sure you don't miss any videos.
00:52.980 --> 01:00.260
Also, Twitter, LinkedIn, AGI.MIT.IDU, if you want to watch other lectures or conversations
01:00.260 --> 01:01.260
like this one.
01:01.260 --> 01:07.980
Better yet, go read Max's book, Life 3.0, chapter 7 on goals is my favorite.
01:07.980 --> 01:12.300
It's really where philosophy and engineering come together and it opens with a quote by
01:12.300 --> 01:18.460
Dostoevsky, the mystery of human existence lies not in just staying alive, but in finding
01:18.460 --> 01:20.300
something to live for.
01:20.300 --> 01:27.100
Lastly, I believe that every failure rewards us with an opportunity to learn, in that sense
01:27.100 --> 01:33.060
I've been very fortunate to fail in so many new and exciting ways and this conversation
01:33.060 --> 01:34.060
was no different.
01:34.060 --> 01:41.260
I've learned about something called Radio Frequency Interference, RFI, look it up.
01:41.260 --> 01:45.500
Apparently music and conversations from local radio stations can bleed into the audio that
01:45.500 --> 01:49.380
you're recording in such a way that almost completely ruins that audio.
01:49.380 --> 01:52.460
It's an exceptionally difficult sound source to remove.
01:52.460 --> 01:59.620
So, I've gotten the opportunity to learn how to avoid RFI in the future during recording
01:59.620 --> 02:00.620
sessions.
02:00.620 --> 02:06.260
I've also gotten the opportunity to learn how to use Adobe Audition and iZotope RX6
02:06.260 --> 02:11.740
to do some audio repair.
02:11.740 --> 02:14.940
Of course, this is an exceptionally difficult noise to remove.
02:14.940 --> 02:20.380
I am an engineer, I'm not an audio engineer, neither is anybody else in our group, but
02:20.380 --> 02:21.780
we did our best.
02:21.780 --> 02:26.780
Nevertheless, I thank you for your patience and I hope you're still able to enjoy this
02:26.780 --> 02:27.780
conversation.
02:27.780 --> 02:31.460
Do you think there's intelligent life out there in the universe?
02:31.460 --> 02:33.420
Let's open up with an easy question.
02:33.420 --> 02:36.260
I have a minority view here actually.
02:36.260 --> 02:41.180
When I give public lectures, I often ask for show of hands who thinks there's intelligent
02:41.180 --> 02:47.060
life out there somewhere else and almost everyone puts their hands up and when I ask why, they'll
02:47.060 --> 02:52.060
be like, oh, there's so many galaxies out there, there's got to be.
02:52.060 --> 02:54.660
But I'm a number nerd, right?
02:54.660 --> 02:59.180
So when you look more carefully at it, it's not so clear at all.
02:59.180 --> 03:03.140
When we talk about our universe, first of all, we don't mean all of space.
03:03.140 --> 03:05.900
We actually mean, I don't know, you can throw me the universe if you want, it's behind you
03:05.900 --> 03:06.900
there.
03:06.900 --> 03:14.540
We simply mean the spherical region of space from which light has had time to reach us
03:14.540 --> 03:19.460
so far during the 13.8 billion years since our big bang.
03:19.460 --> 03:23.020
There's more space here, but this is what we call a universe because that's all we have
03:23.020 --> 03:24.140
access to.
03:24.140 --> 03:31.220
So is there intelligent life here that's gotten to the point of building telescopes and computers?
03:31.220 --> 03:39.500
My guess is no, actually, the probability of it happening on any given planet is some
03:39.500 --> 03:42.860
number we don't know what it is.
03:42.860 --> 03:49.340
And what we do know is that the number can't be super high because there's over a billion
03:49.340 --> 03:54.780
Earth like planets in the Milky Way galaxy alone, many of which are billions of years
03:54.780 --> 04:01.740
older than Earth, and aside from some UFO believers, you know, there isn't much evidence
04:01.740 --> 04:05.740
that any super advanced civilization has come here at all.
04:05.740 --> 04:08.700
And so that's the famous Fermi paradox, right?
04:08.700 --> 04:13.620
And then if you work the numbers, what you find is that if you have no clue what the
04:13.620 --> 04:18.500
probability is of getting life on a given planet, so it could be 10 to the minus 10,
04:18.500 --> 04:23.620
10 to the minus 20, or 10 to the minus two, or any power of 10 is sort of equally likely
04:23.620 --> 04:27.700
if you want to be really open minded, that translates into it being equally likely that
04:27.700 --> 04:34.700
our nearest neighbor is 10 to the 16 meters away, 10 to the 17 meters away, 10 to the
04:34.700 --> 04:35.700
18.
04:35.700 --> 04:42.860
Now, by the time you get much less than 10 to the 16 already, we pretty much know there
04:42.860 --> 04:46.220
is nothing else that's close.
04:46.220 --> 04:49.740
And when you get because it would have discovered us, they, yeah, they would have discovered
04:49.740 --> 04:53.540
us longer or if they're really close, we would have probably noted some engineering projects
04:53.540 --> 04:54.540
that they're doing.
04:54.540 --> 05:00.140
And if it's beyond 10 to the 26 meters, that's already outside of here.
05:00.140 --> 05:06.340
So my guess is actually that there are, we are the only life in here that's gotten the
05:06.340 --> 05:14.020
point of building advanced tech, which I think is very, puts a lot of responsibility on our
05:14.020 --> 05:18.140
shoulders, not screw up, you know, I think people who take for granted that it's okay
05:18.140 --> 05:23.300
for us to screw up, have an accidental nuclear war or go extinct somehow because there's
05:23.300 --> 05:27.460
a sort of Star Trek like situation out there where some other life forms are going to come
05:27.460 --> 05:30.380
and bail us out and it doesn't matter so much.
05:30.380 --> 05:33.380
I think they're leveling us into a false sense of security.
05:33.380 --> 05:37.540
I think it's much more prudent to say, let's be really grateful for this amazing opportunity
05:37.540 --> 05:44.180
we've had and make the best of it just in case it is down to us.
05:44.180 --> 05:50.220
So from a physics perspective, do you think intelligent life, so it's unique from a sort
05:50.220 --> 05:55.860
of statistical view of the size of the universe, but from the basic matter of the universe,
05:55.860 --> 06:00.100
how difficult is it for intelligent life to come about with the kind of advanced tech
06:00.100 --> 06:06.300
building life is implied in your statement that it's really difficult to create something
06:06.300 --> 06:07.620
like a human species?
06:07.620 --> 06:14.740
Well, I think what we know is that going from no life to having life that can do our level
06:14.740 --> 06:21.140
of tech, there's some sort of to going beyond that than actually settling our whole universe
06:21.140 --> 06:22.300
with life.
06:22.300 --> 06:30.700
There's some road major roadblock there, which is some great filter as I just sometimes called
06:30.700 --> 06:37.180
which, which tough to get through, it's either that that roadblock is either behind us or
06:37.180 --> 06:38.620
in front of us.
06:38.620 --> 06:40.980
I'm hoping very much that it's behind us.
06:40.980 --> 06:46.900
I'm super excited every time we get a new report from NASA saying they failed to find
06:46.900 --> 06:53.260
any life on Mars, because that suggests that the hard part, maybe it was getting the first
06:53.260 --> 06:59.540
ribosome or some some very low level kind of stepping stone.
06:59.540 --> 07:03.620
So they were home free because if that's true, then the future is really only limited by
07:03.620 --> 07:04.620
our own imagination.
07:04.620 --> 07:11.460
It would be much suckier if it turns out that this level of life is kind of a diamond dozen,
07:11.460 --> 07:12.780
but maybe there's some other problem.
07:12.780 --> 07:17.220
Like as soon as a civilization gets advanced technology within 100 years, they get into
07:17.220 --> 07:21.740
some stupid fight with themselves and poof, you know, that would be a bummer.
07:21.740 --> 07:22.740
Yeah.
07:22.740 --> 07:28.980
So you've explored the mysteries of the universe, the cosmological universe, the one that's
07:28.980 --> 07:36.340
between us today, I think you've also begun to explore the other universe, which is sort
07:36.340 --> 07:42.860
of the mystery, the mysterious universe of the mind of intelligence, of intelligent life.
07:42.860 --> 07:48.260
So is there a common thread between your interests or the way you think about space and intelligence?
07:48.260 --> 07:49.260
Oh, yeah.
07:49.260 --> 07:57.700
When I was a teenager, I was already very fascinated by the biggest questions and I felt that the
07:57.700 --> 08:03.660
two biggest mysteries of all in science were our universe out there and our universe in
08:03.660 --> 08:04.660
here.
08:04.660 --> 08:05.660
Yeah.
08:05.660 --> 08:11.260
So it's quite natural after having spent a quarter of a century on my career thinking
08:11.260 --> 08:12.260
a lot about this one.
08:12.260 --> 08:15.980
And now I'm indulging in the luxury of doing research on this one.
08:15.980 --> 08:17.660
It's just so cool.
08:17.660 --> 08:25.260
I feel the time is ripe now for you transparently deepening our understanding of this.
08:25.260 --> 08:26.420
Just start exploring this one.
08:26.420 --> 08:32.500
Yeah, because I think a lot of people view intelligence as something mysterious that
08:32.500 --> 08:38.340
can only exist in biological organisms like us and therefore dismiss all talk about artificial
08:38.340 --> 08:41.260
general intelligence is science fiction.
08:41.260 --> 08:47.260
But from my perspective as a physicist, I am a blob of quirks and electrons moving around
08:47.260 --> 08:50.180
in a certain pattern and processing information in certain ways.
08:50.180 --> 08:53.580
And this is also a blob of quirks and electrons.
08:53.580 --> 08:57.860
I'm not smarter than the water bottle because I'm made of different kind of quirks.
08:57.860 --> 09:02.220
I'm made of up quirks and down quirks exact same kind as this.
09:02.220 --> 09:07.020
It's a there's no secret sauce, I think in me, it's it's all about the pattern of the
09:07.020 --> 09:08.820
information processing.
09:08.820 --> 09:16.020
And this means that there's no law of physics saying that we can't create technology, which
09:16.020 --> 09:21.740
can help us by being incredibly intelligent and help us crack mysteries that we couldn't.
09:21.740 --> 09:25.580
In other words, I think we've really only seen the tip of the intelligence iceberg so
09:25.580 --> 09:26.580
far.
09:26.580 --> 09:27.580
Yeah.
09:27.580 --> 09:34.380
So the perceptronium, yeah, so you coined this amazing term, it's a hypothetical state
09:34.380 --> 09:39.420
of matter, sort of thinking from a physics perspective, what is the kind of matter that
09:39.420 --> 09:44.500
can help as you're saying, subjective experience emerge, consciousness emerge.
09:44.500 --> 09:50.140
So how do you think about consciousness from this physics perspective?
09:50.140 --> 09:51.980
Very good question.
09:51.980 --> 10:03.060
So, again, I think many people have underestimated our ability to make progress on this by convincing
10:03.060 --> 10:08.500
themselves it's hopeless because somehow we're missing some ingredient that we need.
10:08.500 --> 10:13.020
There's some new consciousness particle or whatever.
10:13.020 --> 10:19.660
I happen to think that we're not missing anything and that it's not the interesting thing about
10:19.660 --> 10:25.900
consciousness that gives us this amazing subjective experience of colors and sounds and emotions
10:25.900 --> 10:32.300
and so on is rather something at the higher level about the patterns of information processing.
10:32.300 --> 10:38.300
And that's why I like to think about this idea of perceptronium.
10:38.300 --> 10:44.220
What does it mean for an arbitrary physical system to be conscious in terms of what its
10:44.220 --> 10:47.100
particles are doing or its information is doing?
10:47.100 --> 10:52.300
I hate carbon chauvinism, this attitude, you have to be made of carbon atoms to be smart
10:52.300 --> 10:53.300
or conscious.
10:53.300 --> 10:58.180
So something about the information processing that this kind of matter performs.
10:58.180 --> 11:02.700
Yeah, and you can see I have my favorite equations here describing various fundamental
11:02.700 --> 11:04.660
aspects of the world.
11:04.660 --> 11:09.620
I think one day, maybe someone who's watching this will come up with the equations that
11:09.620 --> 11:12.140
information processing has to satisfy to be conscious.
11:12.140 --> 11:19.580
And I'm quite convinced there is big discovery to be made there because let's face it, we
11:19.580 --> 11:25.900
know that some information processing is conscious because we are conscious.
11:25.900 --> 11:28.980
But we also know that a lot of information processing is not conscious.
11:28.980 --> 11:32.980
Most of the information processing happening in your brain right now is not conscious.
11:32.980 --> 11:38.380
There are like 10 megabytes per second coming in even just through your visual system.
11:38.380 --> 11:42.940
You're not conscious about your heartbeat regulation or most things.
11:42.940 --> 11:47.300
Even if I just ask you to read what it says here, you look at it and then, oh, now you
11:47.300 --> 11:48.300
know what it said.
11:48.300 --> 11:51.820
But you're not aware of how the computation actually happened.
11:51.820 --> 11:57.020
Your consciousness is like the CEO that got an email at the end with the final answer.
11:57.020 --> 12:01.140
So what is it that makes a difference?
12:01.140 --> 12:06.620
I think that's both a great science mystery, we're actually studying it a little bit in
12:06.620 --> 12:12.260
my lab here at MIT, but I also think it's a really urgent question to answer.
12:12.260 --> 12:16.460
For starters, I mean, if you're an emergency room doctor and you have an unresponsive patient
12:16.460 --> 12:24.180
coming in, wouldn't it be great if in addition to having a CT scanner, you had a conscious
12:24.180 --> 12:30.780
scanner that could figure out whether this person is actually having locked in syndrome
12:30.780 --> 12:33.580
or is actually comatose.
12:33.580 --> 12:40.740
And in the future, imagine if we build robots or the machine that we can have really good
12:40.740 --> 12:45.100
conversations with, I think it's very likely to happen, right?
12:45.100 --> 12:50.020
Wouldn't you want to know if your home helper robot is actually experiencing anything or
12:50.020 --> 12:52.980
just like a zombie?
12:52.980 --> 12:53.980
Would you prefer it?
12:53.980 --> 12:54.980
What would you prefer?
12:54.980 --> 12:57.820
Would you prefer that it's actually unconscious so that you don't have to feel guilty about
12:57.820 --> 12:59.980
switching it off or giving boring chores?
12:59.980 --> 13:02.380
What would you prefer?
13:02.380 --> 13:09.780
Well, certainly we would prefer, I would prefer the appearance of consciousness, but the question
13:09.780 --> 13:15.300
is whether the appearance of consciousness is different than consciousness itself.
13:15.300 --> 13:21.420
And sort of ask that as a question, do you think we need to understand what consciousness
13:21.420 --> 13:28.420
is, solve the hard problem of consciousness in order to build something like an AGI system?
13:28.420 --> 13:29.420
No.
13:29.420 --> 13:31.140
I don't think that.
13:31.140 --> 13:36.220
I think we will probably be able to build things even if we don't answer that question.
13:36.220 --> 13:41.100
But if we want to make sure that what happens is a good thing, we better solve it first.
13:41.100 --> 13:47.220
So it's a wonderful controversy you're raising there, where you have basically three points
13:47.220 --> 13:50.220
of view about the hard problem.
13:50.220 --> 13:55.060
There are two different points of view that both conclude that the hard problem of consciousness
13:55.060 --> 13:56.060
is BS.
13:56.060 --> 14:01.100
On one hand, you have some people like Daniel Dennett who say that consciousness is just
14:01.100 --> 14:05.140
BS because consciousness is the same thing as intelligence.
14:05.140 --> 14:06.580
There's no difference.
14:06.580 --> 14:13.620
So anything which acts conscious is conscious, just like we are.
14:13.620 --> 14:18.820
And then there are also a lot of people, including many top AI researchers I know, who say, oh,
14:18.820 --> 14:22.820
consciousness is just bullshit because of course machines can never be conscious.
14:22.820 --> 14:28.020
They're always going to skiddy zombies, never have to feel guilty about how you treat them.
14:28.020 --> 14:35.380
And then there's a third group of people, including Giulio Tononi, for example, and another, and
14:35.380 --> 14:40.020
Gustav Koch and a number of others, I would put myself on this middle camp who say that
14:40.020 --> 14:44.260
actually some information processing is conscious and some is not.
14:44.260 --> 14:49.380
So let's find the equation which can be used to determine which it is.
14:49.380 --> 14:53.980
And I think we've just been a little bit lazy kind of running away from this problem for
14:53.980 --> 14:55.100
a long time.
14:55.100 --> 15:01.940
It's been almost taboo to even mention the C word in a lot of circles because, but we
15:01.940 --> 15:03.700
should stop making excuses.
15:03.700 --> 15:10.940
This is a science question and there are ways we can even test any theory that makes predictions
15:10.940 --> 15:12.140
for this.
15:12.140 --> 15:16.060
And coming back to this helper robot, I mean, so you said you would want your helper robot
15:16.060 --> 15:21.340
to certainly act conscious and treat you, like have conversations with you and stuff.
15:21.340 --> 15:24.860
But wouldn't you, would you feel a little bit creeped out if you realized that it was
15:24.860 --> 15:31.700
just a glossed up tape recorder, you know, that was just zombie and it's a faking emotion?
15:31.700 --> 15:37.220
Would you prefer that it actually had an experience or would you prefer that it's actually not
15:37.220 --> 15:42.300
experiencing anything so you feel, you don't have to feel guilty about what you do to it?
15:42.300 --> 15:46.580
It's such a difficult question because, you know, it's like when you're in a relationship
15:46.580 --> 15:49.860
and you say, well, I love you and the other person said I love you back.
15:49.860 --> 15:53.860
It's like asking, well, do they really love you back or are they just saying they love
15:53.860 --> 15:54.860
you back?
15:54.860 --> 15:59.620
Don't you really want them to actually love you?
15:59.620 --> 16:08.100
It's hard to, it's hard to really know the difference between everything seeming like
16:08.100 --> 16:14.820
there's consciousness present, there's intelligence present, there's affection, passion, love,
16:14.820 --> 16:16.180
and it actually being there.
16:16.180 --> 16:17.180
I'm not sure.
16:17.180 --> 16:18.180
Do you have...
16:18.180 --> 16:19.180
Can I ask you a question about this?
16:19.180 --> 16:20.180
Yes.
16:20.180 --> 16:21.180
To make it a bit more pointed.
16:21.180 --> 16:23.140
So Mass General Hospital is right across the river, right?
16:23.140 --> 16:29.180
Suppose you're going in for a medical procedure and they're like, you know, for anesthesia
16:29.180 --> 16:32.180
what we're going to do is we're going to give you muscle relaxance so you won't be able
16:32.180 --> 16:36.140
to move and you're going to feel excruciating pain during the whole surgery but you won't
16:36.140 --> 16:37.660
be able to do anything about it.
16:37.660 --> 16:42.020
But then we're going to give you this drug that erases your memory of it.
16:42.020 --> 16:45.420
Would you be cool about that?
16:45.420 --> 16:51.100
What's the difference that you're conscious about it or not if there's no behavioral change,
16:51.100 --> 16:52.100
right?
16:52.100 --> 16:53.100
Right.
16:53.100 --> 16:55.220
And that's a really clear way to put it.
16:55.220 --> 17:01.100
Yeah, it feels like in that sense, experiencing it is a valuable quality.
17:01.100 --> 17:09.140
So actually being able to have subjective experiences, at least in that case, is valuable.
17:09.140 --> 17:14.060
And I think we humans have a little bit of a bad track record also of making these self
17:14.060 --> 17:17.940
serving arguments that other entities aren't conscious.
17:17.940 --> 17:20.700
You know, people often say, oh, these animals can't feel pain.
17:20.700 --> 17:21.700
Right.
17:21.700 --> 17:25.580
It's okay to boil lobsters because we asked them if it hurt and they didn't say anything.
17:25.580 --> 17:29.180
And now there was just a paper out saying lobsters did do feel pain when you boil them
17:29.180 --> 17:31.180
and they're banning it in Switzerland.
17:31.180 --> 17:36.300
And we did this with slaves too often and said, oh, they don't mind.
17:36.300 --> 17:41.180
They don't maybe aren't conscious or women don't have souls or whatever.
17:41.180 --> 17:46.540
So I'm a little bit nervous when I hear people just take as an axiom that machines can't
17:46.540 --> 17:48.900
have experience ever.
17:48.900 --> 17:52.500
I think this is just a really fascinating science question is what it is.
17:52.500 --> 17:57.420
Let's research it and try to figure out what it is that makes the difference between unconscious
17:57.420 --> 18:01.220
intelligent behavior and conscious intelligent behavior.
18:01.220 --> 18:07.140
So in terms of, so if you think of a Boston Dynamics human or robot being sort of with
18:07.140 --> 18:13.420
a broom being pushed around, it starts pushing on a consciousness question.
18:13.420 --> 18:20.060
So let me ask, do you think an AGI system, like a few neuroscientists believe needs to
18:20.060 --> 18:25.860
have a physical embodiment, needs to have a body or something like a body?
18:25.860 --> 18:28.340
No, I don't think so.
18:28.340 --> 18:30.620
You mean to have a conscious experience?
18:30.620 --> 18:33.140
To have consciousness.
18:33.140 --> 18:37.860
I do think it helps a lot to have a physical embodiment to learn the kind of things about
18:37.860 --> 18:42.820
the world that are important to us humans for sure.
18:42.820 --> 18:47.460
But I don't think the physical embodiment is necessary after you've learned it.
18:47.460 --> 18:48.860
Just have the experience.
18:48.860 --> 18:51.500
Think about it when you're dreaming, right?
18:51.500 --> 18:55.500
Your eyes are closed, you're not getting any sensory input, you're not behaving or moving
18:55.500 --> 18:59.780
in any way, but there's still an experience there, right?
18:59.780 --> 19:03.220
And so clearly the experience that you have when you see something cool in your dreams
19:03.220 --> 19:08.660
isn't coming from your eyes, it's just the information processing itself in your brain,
19:08.660 --> 19:11.100
which is that experience, right?
19:11.100 --> 19:16.660
But if I put it another way, I'll say because it comes from neuroscience is the reason you
19:16.660 --> 19:24.620
want to have a body and a physical, something like a physical system is because you want
19:24.620 --> 19:27.100
to be able to preserve something.
19:27.100 --> 19:35.740
In order to have a self, you could argue, you'd need to have some kind of embodiment
19:35.740 --> 19:38.180
of self to want to preserve.
19:38.180 --> 19:45.940
Well, now we're getting a little bit anthropomorphic, anthropomorphizing things, maybe talking about
19:45.940 --> 19:47.820
self preservation instincts.
19:47.820 --> 19:50.700
We are evolved organisms, right?
19:50.700 --> 19:57.020
So Darwinian evolution endowed us and other evolved organisms with self preservation instinct
19:57.020 --> 20:03.100
because those that didn't have those self preservation genes got cleaned out of the gene pool.
20:03.100 --> 20:09.180
But if you build an artificial general intelligence, the mind space that you can design is much,
20:09.180 --> 20:14.500
much larger than just a specific subset of minds that can evolve that have.
20:14.500 --> 20:19.260
So an AGI mind doesn't necessarily have to have any self preservation instinct.
20:19.260 --> 20:24.100
It also doesn't necessarily have to be so individualistic as us.
20:24.100 --> 20:28.140
Like imagine if you could just, first of all, we're also very afraid of death, you know,
20:28.140 --> 20:32.180
as opposed to you could back yourself up every five minutes and then your airplane is about
20:32.180 --> 20:33.180
to crash.
20:33.180 --> 20:37.340
You're like, shucks, I'm just, I'm going to lose the last five minutes of experiences
20:37.340 --> 20:41.580
since my last cloud backup, dang, you know, it's not as big a deal.
20:41.580 --> 20:47.380
Or if we could just copy experiences between our minds easily, like which we could easily
20:47.380 --> 20:50.620
do if we were silicon based, right?
20:50.620 --> 20:55.860
Then maybe we would feel a little bit more like a hive mind, actually, that maybe it's
20:55.860 --> 21:01.220
the, so, so there's, so I don't think we should take for granted at all that AGI will have
21:01.220 --> 21:06.820
to have any of those sort of competitive as alpha male instincts.
21:06.820 --> 21:07.820
Right.
21:07.820 --> 21:12.820
On the other hand, you know, this is really interesting because I think some people go
21:12.820 --> 21:17.900
too far and say, of course, we don't have to have any concerns either that advanced
21:17.900 --> 21:22.700
AI will have those instincts because we can build anything we want.
21:22.700 --> 21:27.420
That there's, there's a very nice set of arguments going back to Steve Omohandro and
21:27.420 --> 21:32.900
Nick Bostrom and others just pointing out that when we build machines, we normally build
21:32.900 --> 21:37.700
them with some kind of goal, you know, win this chess game, drive this car safely or
21:37.700 --> 21:38.700
whatever.
21:38.700 --> 21:42.540
And as soon as you put in a goal into machine, especially if it's kind of open ended goal
21:42.540 --> 21:48.460
and the machine is very intelligent, it'll break that down into a bunch of sub goals.
21:48.460 --> 21:53.500
And one of those goals will almost always be self preservation because if it breaks
21:53.500 --> 21:56.140
or dies in the process, it's not going to accomplish the goal, right?
21:56.140 --> 21:59.540
Like, suppose you just build a little, you have a little robot and you tell it to go
21:59.540 --> 22:05.460
down the store market here and, and get you some food, make you cook your Italian dinner,
22:05.460 --> 22:09.540
you know, and then someone mugs it and tries to break it on the way.
22:09.540 --> 22:15.380
That robot has an incentive to not get destroyed and defend itself for a runaway because otherwise
22:15.380 --> 22:17.780
it's going to fail and cooking your dinner.
22:17.780 --> 22:22.940
It's not afraid of death, but it really wants to complete the dinner cooking goal.
22:22.940 --> 22:24.780
So it will have a self preservation instinct.
22:24.780 --> 22:26.820
It will continue being a functional agent.
22:26.820 --> 22:27.820
Yeah.
22:27.820 --> 22:35.860
And, and, and similarly, if you give any kind of more ambitious goal to an AGI, it's very
22:35.860 --> 22:39.940
likely they want to acquire more resources so it can do that better.
22:39.940 --> 22:44.500
And it's exactly from those sort of sub goals that we might not have intended that some
22:44.500 --> 22:50.740
of the concerns about AGI safety come, you give it some goal that seems completely harmless.
22:50.740 --> 22:55.540
And then before you realize it, it's also trying to do these other things which you
22:55.540 --> 22:59.220
didn't want it to do and it's maybe smarter than us.
22:59.220 --> 23:08.220
So, so, and let me pause just because I am in a very kind of human centric way, see fear
23:08.220 --> 23:11.900
of death as a valuable motivator.
23:11.900 --> 23:17.220
So you don't think you think that's an artifact of evolution.
23:17.220 --> 23:21.980
So that's the kind of mind space evolution created that we're sort of almost obsessed
23:21.980 --> 23:22.980
about self preservation.
23:22.980 --> 23:23.980
Yeah.
23:23.980 --> 23:29.500
Some kind of genetic well, you don't think that's necessary to be afraid of death.
23:29.500 --> 23:34.980
So not just a kind of sub goal of self preservation just so you can keep doing the thing, but
23:34.980 --> 23:42.980
more fundamentally sort of have the finite thing like this ends for you at some point.
23:42.980 --> 23:43.980
Interesting.
23:43.980 --> 23:47.500
Do I think it's necessary for what precisely?
23:47.500 --> 23:51.020
For intelligence, but also for consciousness.
23:51.020 --> 23:58.220
So for those for both, do you think really like a finite death and the fear of it is
23:58.220 --> 24:01.020
important?
24:01.020 --> 24:06.980
So before I can answer, before we can agree on whether it's necessary for intelligence
24:06.980 --> 24:10.660
or for consciousness, we should be clear on how we define those two words because a lot
24:10.660 --> 24:13.340
are really smart people define them in very different ways.
24:13.340 --> 24:18.500
I was in this on this panel with AI experts and they couldn't, they couldn't agree on
24:18.500 --> 24:20.180
how to define intelligence even.
24:20.180 --> 24:24.860
So I define intelligence simply as the ability to accomplish complex goals.
24:24.860 --> 24:30.740
I like your broad definition because again, I don't want to be a carbon chauvinist.
24:30.740 --> 24:36.580
And in that case, no, certainly it doesn't require fear of death.
24:36.580 --> 24:40.100
I would say AlphaGo AlphaZero is quite intelligent.
24:40.100 --> 24:44.260
I don't think AlphaZero has any fear of being turned off because it doesn't understand the
24:44.260 --> 24:52.180
concept of even and similarly consciousness, I mean, you can certainly imagine a very simple
24:52.180 --> 24:57.660
kind of experience if certain plants have any kind of experience, I don't think they're
24:57.660 --> 25:00.940
very afraid of dying or there's nothing they can do about it anyway much.
25:00.940 --> 25:08.420
So there wasn't that much value and but more seriously, I think if you ask not just about
25:08.420 --> 25:15.460
being conscious, but maybe having what you would, we might call an exciting life for
25:15.460 --> 25:23.300
you for your passion and really appreciate the things, maybe there, somehow, maybe there
25:23.300 --> 25:29.180
perhaps it does help having a backdrop that, hey, it's finite, you know, let's make the
25:29.180 --> 25:31.380
most of this, let's live to the fullest.
25:31.380 --> 25:36.220
So if you knew you were going to just live forever, do you think you would change your
25:36.220 --> 25:40.500
career? Yeah, I mean, in some perspective, it would
25:40.500 --> 25:44.020
be an incredibly boring life living forever.
25:44.020 --> 25:49.740
So in the sort of loose, subjective terms that you said of something exciting and something
25:49.740 --> 25:55.180
in this that other humans would understand, I think, is yeah, it seems that the finiteness
25:55.180 --> 25:56.660
of it is important.
25:56.660 --> 26:02.420
Well, the good news I have for you then is based on what we understand about cosmology,
26:02.420 --> 26:10.460
things in our universe is probably finite, although big crunch or big or big, what's
26:10.460 --> 26:11.460
the extent of the infinite?
26:11.460 --> 26:16.820
Yeah, we could have a big chill or a big crunch or a big rip or death, the big snap or death
26:16.820 --> 26:17.820
bubbles.
26:17.820 --> 26:20.140
All of them are more than a billion years away.
26:20.140 --> 26:29.500
So we should we certainly have vastly more time than our ancestors thought, but still
26:29.500 --> 26:35.580
pretty hard to squeeze in an infinite number of compute cycles, even though there are some
26:35.580 --> 26:37.820
loopholes that just might be possible.
26:37.820 --> 26:44.620
But I think, you know, some people like to say that you should live as if you're about
26:44.620 --> 26:48.100
to you're going to die in five years or so, and that's sort of optimal.
26:48.100 --> 26:54.740
Maybe it's a good as some we should build our civilization as if it's all finite to
26:54.740 --> 26:55.740
be on the safe side.
26:55.740 --> 27:02.020
Right, exactly. So you mentioned in defining intelligence as the ability to solve complex
27:02.020 --> 27:03.020
goals.
27:03.020 --> 27:04.940
So where would you draw a line?
27:04.940 --> 27:10.940
How would you try to define human level intelligence and super human level intelligence?
27:10.940 --> 27:13.380
Where is consciousness part of that definition?
27:13.380 --> 27:16.860
No, consciousness does not come into this definition.
27:16.860 --> 27:21.580
So so I think of intelligence as it's a spectrum, but there are very many different kinds of
27:21.580 --> 27:22.580
goals you can have.
27:22.580 --> 27:27.140
You have a goal to be a good chess player, a good goal player, a good car driver, a good
27:27.140 --> 27:31.260
investor, good poet, etc.
27:31.260 --> 27:35.740
So intelligence that bind by its very nature, isn't something you can measure, but it's
27:35.740 --> 27:39.900
one number, some overall goodness, no, no, there are some people who are more better
27:39.900 --> 27:42.540
at this, some people are better at that.
27:42.540 --> 27:48.380
Right now we have machines that are much better than us at some very narrow tasks like multiplying
27:48.380 --> 27:57.620
large numbers fast, memorizing large databases, playing chess, playing go, soon driving cars.
27:57.620 --> 28:03.340
But there's still no machine that can match a human child in general intelligence.
28:03.340 --> 28:08.420
But artificial general intelligence, AGI, the name of your course, of course, that
28:08.420 --> 28:16.460
is by its very definition, the quest to build a machine that can do everything as well as
28:16.460 --> 28:17.460
we can.
28:17.460 --> 28:24.060
Up to the old Holy Grail of AI from back to its inception in the 60s.
28:24.060 --> 28:27.500
If that ever happens, of course, I think it's going to be the biggest transition in the
28:27.500 --> 28:33.860
history of life on Earth, but it doesn't necessarily have to wait the big impact until machines
28:33.860 --> 28:35.780
are better than us at knitting.
28:35.780 --> 28:41.940
The really big change doesn't come exactly at the moment they're better than us at everything.
28:41.940 --> 28:45.820
The really big change comes, first, their big change is when they start becoming better
28:45.820 --> 28:51.140
at us at doing most of the jobs that we do, because that takes away much of the demand
28:51.140 --> 28:53.380
for human labor.
28:53.380 --> 29:01.300
And then the really warping change comes when they become better than us at AI research.
29:01.300 --> 29:07.900
Because right now, the time scale of AI research is limited by the human research and development
29:07.900 --> 29:14.100
cycle of years, typically, along the take from one release of some software or iPhone
29:14.100 --> 29:16.300
or whatever to the next.
29:16.300 --> 29:25.820
But once Google can replace 40,000 engineers by 40,000 equivalent pieces of software or
29:25.820 --> 29:29.660
whatever, then there's no reason that has to be years.
29:29.660 --> 29:32.020
It can be, in principle, much faster.
29:32.020 --> 29:38.900
And the time scale of future progress in AI and all of science and technology will be
29:38.900 --> 29:40.980
driven by machines, not humans.
29:40.980 --> 29:49.660
So it's this simple point, which gives right this incredibly fun controversy about whether
29:49.660 --> 29:54.540
there can be intelligence explosion, so called singularity, as Werner Winge called it.
29:54.540 --> 30:00.060
The idea, as articulated by I.J. Good, is obviously way back fifties, but you can see
30:00.060 --> 30:07.220
Alan Turing and others thought about it even earlier.
30:07.220 --> 30:12.980
You asked me what exactly what I define human level intelligence.
30:12.980 --> 30:18.540
So the glib answer is just to say something which is better than us at all cognitive tasks
30:18.540 --> 30:21.980
or better than any human at all cognitive tasks.
30:21.980 --> 30:25.900
But the really interesting bar, I think, goes a little bit lower than that, actually.
30:25.900 --> 30:33.260
It's when they're better than us at AI programming and general learning so that they can, if
30:33.260 --> 30:37.340
they want to, get better than us at anything by just starting out.
30:37.340 --> 30:43.100
So there better is a key word and better is towards this kind of spectrum of the complexity
30:43.100 --> 30:45.740
of goals it's able to accomplish.
30:45.740 --> 30:53.060
So another way to, and that's certainly a very clear definition of human love.
30:53.060 --> 30:56.300
So there's, it's almost like a sea that's rising, you can do more and more and more
30:56.300 --> 30:57.300
things.
30:57.300 --> 30:59.900
It's actually a graphic that you show, it's really nice way to put it.
30:59.900 --> 31:04.340
So there's some peaks and there's an ocean level elevating and you solve more and more
31:04.340 --> 31:05.340
problems.
31:05.340 --> 31:09.220
But, you know, just kind of to take a pause and we took a bunch of questions and a lot
31:09.220 --> 31:14.380
of social networks and a bunch of people asked a sort of a slightly different direction
31:14.380 --> 31:22.260
on creativity and on things that perhaps aren't a peak.
31:22.260 --> 31:28.620
It's, you know, human beings are flawed and perhaps better means having being having contradiction
31:28.620 --> 31:30.260
being flawed in some way.
31:30.260 --> 31:34.980
So let me sort of, yeah, start and start easy, first of all.
31:34.980 --> 31:36.620
So you have a lot of cool equations.
31:36.620 --> 31:39.660
Let me ask, what's your favorite equation, first of all?
31:39.660 --> 31:43.580
I know they're all like your children, but which one is that?
31:43.580 --> 31:49.060
This is the Shreddinger equation, it's the master key of quantum mechanics of the micro
31:49.060 --> 31:50.060
world.
31:50.060 --> 31:55.340
So this equation can take everything to do with atoms and all the fuels and all the
31:55.340 --> 32:04.020
way up to… Yeah, so, okay, so quantum mechanics is certainly a beautiful mysterious formulation
32:04.020 --> 32:05.020
of our world.
32:05.020 --> 32:10.740
So I'd like to sort of ask you, just as an example, it perhaps doesn't have the same
32:10.740 --> 32:17.420
beauty as physics does, but in mathematics abstract, the Andrew Wiles who proved the
32:17.420 --> 32:19.460
Fermat's last theory.
32:19.460 --> 32:24.180
So he just saw this recently and it kind of caught my eye a little bit.
32:24.180 --> 32:27.980
This is 358 years after it was conjectured.
32:27.980 --> 32:32.940
So this very simple formulation, everybody tried to prove it, everybody failed.
32:32.940 --> 32:38.820
And so here's this guy comes along and eventually proves it and then fails to prove it and then
32:38.820 --> 32:41.340
proves it again in 94.
32:41.340 --> 32:45.940
And he said like the moment when everything connected into place, in an interview he said
32:45.940 --> 32:47.980
it was so indescribably beautiful.
32:47.980 --> 32:53.580
That moment when you finally realize the connecting piece of two conjectures, he said it was so
32:53.580 --> 32:56.940
indescribably beautiful, it was so simple and so elegant.
32:56.940 --> 33:01.540
I couldn't understand how I'd missed it and I just stared at it in disbelief for 20
33:01.540 --> 33:02.540
minutes.
33:02.540 --> 33:08.100
Then during the day I walked around the department and I keep coming back to my desk looking
33:08.100 --> 33:09.820
to see if it was still there.
33:09.820 --> 33:10.820
It was still there.
33:10.820 --> 33:11.820
I couldn't contain myself.
33:11.820 --> 33:12.820
I was so excited.
33:12.820 --> 33:16.180
It was the most important moment of my working life.
33:16.180 --> 33:18.940
Nothing I ever do again will mean as much.
33:18.940 --> 33:24.860
So that particular moment and it kind of made me think of what would it take?
33:24.860 --> 33:28.380
And I think we have all been there at small levels.
33:28.380 --> 33:34.820
Maybe let me ask, have you had a moment like that in your life where you just had an idea
33:34.820 --> 33:40.060
it's like, wow, yes.
33:40.060 --> 33:44.700
I wouldn't mention myself in the same breath as Andrew Wiles, but I certainly had a number
33:44.700 --> 33:54.820
of aha moments when I realized something very cool about physics just completely made
33:54.820 --> 33:55.820
my head explode.
33:55.820 --> 33:59.580
In fact, some of my favorite discoveries I made later, I later realized that they had
33:59.580 --> 34:03.340
been discovered earlier by someone who's sometimes got quite famous for it.
34:03.340 --> 34:07.460
So there's too late for me to even publish it, but that doesn't diminish in any way.
34:07.460 --> 34:12.340
The emotional experience you have when you realize it like, wow.
34:12.340 --> 34:17.460
So what would it take in that moment, that wow, that was yours in that moment?
34:17.460 --> 34:23.420
So what do you think it takes for an intelligent system, an AGI system, an AI system to have
34:23.420 --> 34:24.980
a moment like that?
34:24.980 --> 34:29.420
It's a tricky question because there are actually two parts to it, right?
34:29.420 --> 34:37.260
One of them is, can it accomplish that proof, can it prove that you can never write A to
34:37.260 --> 34:46.420
the N plus B to the N equals 3 to the N for all integers, etc., etc., when N is bigger
34:46.420 --> 34:49.420
than 2.
34:49.420 --> 34:51.580
That's simply the question about intelligence.
34:51.580 --> 34:54.420
Can you build machines that are that intelligent?
34:54.420 --> 34:59.860
And I think by the time we get a machine that can independently come up with that level
34:59.860 --> 35:03.460
of proofs, probably quite close to AGI.
35:03.460 --> 35:07.860
But the second question is a question about consciousness.
35:07.860 --> 35:13.060
When will we, how likely is it that such a machine would actually have any experience
35:13.060 --> 35:16.500
at all as opposed to just being like a zombie?
35:16.500 --> 35:22.940
And would we expect it to have some sort of emotional response to this or anything at
35:22.940 --> 35:31.140
all akin to human emotion where when it accomplishes its machine goal, it views it as something
35:31.140 --> 35:39.260
very positive and sublime and deeply meaningful.
35:39.260 --> 35:45.260
I would certainly hope that if in the future we do create machines that are our peers or
35:45.260 --> 35:53.700
even our descendants, I would certainly hope that they do have this sort of sublime appreciation
35:53.700 --> 36:06.020
of life in a way, my absolutely worst nightmare would be that at some point in the future,
36:06.020 --> 36:10.620
the distant future, maybe our cosmos is teeming with all this post biological life, doing
36:10.620 --> 36:13.180
all the seemingly cool stuff.
36:13.180 --> 36:20.660
And maybe the last humans by the time our species eventually fizzles out will be like,
36:20.660 --> 36:26.140
well, that's okay, because we're so proud of our descendants here and look, my worst
36:26.140 --> 36:30.580
nightmare is that we haven't solved the consciousness problem.
36:30.580 --> 36:34.100
And we haven't realized that these are all the zombies, they're not aware of anything
36:34.100 --> 36:37.900
anymore than a tape recorder, as in any kind of experience.
36:37.900 --> 36:41.660
So the whole thing has just become a play for empty benches.
36:41.660 --> 36:44.700
That would be like the ultimate zombie apocalypse to me.
36:44.700 --> 36:52.900
So I would much rather, in that case, that we have these beings which can really appreciate
36:52.900 --> 36:57.060
how amazing it is.
36:57.060 --> 37:02.260
And in that picture, what would be the role of creativity, what a few people ask about
37:02.260 --> 37:03.260
creativity?
37:03.260 --> 37:04.260
Yeah.
37:04.260 --> 37:08.700
And do you think, when you think about intelligence, I mean, certainly the story you told at the
37:08.700 --> 37:14.100
beginning of your book involved, you know, creating movies and so on, sort of making
37:14.100 --> 37:18.580
money, you know, you can make a lot of money in our modern world with music and movies.
37:18.580 --> 37:23.100
So if you are an intelligent system, you may want to get good at that.
37:23.100 --> 37:26.300
But that's not necessarily what I mean by creativity.
37:26.300 --> 37:32.620
Is it important on that complex goals where the sea is rising for there to be something
37:32.620 --> 37:39.940
creative, or am I being very human centric and thinking creativity somehow special relative
37:39.940 --> 37:41.940
to intelligence?
37:41.940 --> 37:50.940
My hunch is that we should think of creativity simply as an aspect of intelligence.
37:50.940 --> 37:57.820
And we have to be very careful with human vanity.
37:57.820 --> 38:01.540
We have this tendency to very often want to say, as soon as machines can do something,
38:01.540 --> 38:05.980
we try to diminish it and say, oh, but that's not like real intelligence, you know, is
38:05.980 --> 38:12.620
it not creative or this or that, the other thing, if we ask ourselves to write down a
38:12.620 --> 38:18.500
definition of what we actually mean by being creative, what we mean by Andrew Wiles, what
38:18.500 --> 38:23.660
he did there, for example, don't we often mean that someone takes a very unexpected
38:23.660 --> 38:26.060
leap?
38:26.060 --> 38:33.740
It's not like taking 573 and multiplying by 224 by just a step of straightforward cookbook
38:33.740 --> 38:36.500
like rules, right?
38:36.500 --> 38:40.660
You can maybe make a connection between two things that people have never thought was
38:40.660 --> 38:41.660
connected.
38:41.660 --> 38:42.660
It's very surprising.
38:42.660 --> 38:44.300
Something like that.
38:44.300 --> 38:50.660
I think this is an aspect of intelligence, and this is actually one of the most important
38:50.660 --> 38:53.260
aspects of it.
38:53.260 --> 38:57.940
Maybe the reason we humans tend to be better at it than traditional computers is because
38:57.940 --> 39:02.020
it's something that comes more naturally if you're a neural network than if you're a
39:02.020 --> 39:05.820
traditional logic gates based computer machine.
39:05.820 --> 39:11.900
We physically have all these connections, and if you activate here, activate here, activate
39:11.900 --> 39:20.980
here, it ping, you know, my hunch is that if we ever build a machine where you could
39:20.980 --> 39:31.020
just give it the task, hey, hey, you say, hey, you know, I just realized I want to travel
39:31.020 --> 39:32.380
around the world instead this month.
39:32.380 --> 39:34.700
Can you teach my AGI course for me?
39:34.700 --> 39:36.100
And it's like, okay, I'll do it.
39:36.100 --> 39:39.860
And it does everything that you would have done and it improvises and stuff.
39:39.860 --> 39:42.860
That would in my mind involve a lot of creativity.
39:42.860 --> 39:45.660
Yeah, so it's actually a beautiful way to put it.
39:45.660 --> 39:54.540
I think we do try to grasp at the definition of intelligence as everything we don't understand
39:54.540 --> 39:57.580
how to build.
39:57.580 --> 40:02.180
So we as humans try to find things that we have and machines don't have, and maybe creativity
40:02.180 --> 40:05.940
is just one of the things, one of the words we used to describe that.
40:05.940 --> 40:06.940
That's a really interesting way to put it.
40:06.940 --> 40:09.820
I don't think we need to be that defensive.
40:09.820 --> 40:14.700
I don't think anything good comes out of saying, we're somehow special, you know, it's
40:14.700 --> 40:27.540
very wise, there are many examples in history of where trying to pretend they were somehow
40:27.540 --> 40:36.220
superior to all other intelligent beings has led to pretty bad results, right?
40:36.220 --> 40:39.700
Nazi Germany, they said that they were somehow superior to other people.
40:39.700 --> 40:44.580
Today, we still do a lot of cruelty to animals by saying they were so superior somehow on
40:44.580 --> 40:50.500
the other, they can't feel pain, slavery was justified by the same kind of really weak
40:50.500 --> 40:52.420
arguments.
40:52.420 --> 40:58.700
And I don't think if we actually go ahead and build artificial general intelligence,
40:58.700 --> 41:01.100
it can do things better than us.
41:01.100 --> 41:08.980
I don't think we should try to found our self worth on some sort of bogus claims of superiority
41:08.980 --> 41:11.940
in terms of our intelligence.
41:11.940 --> 41:21.780
I think we should instead find our calling and the meaning of life from the experiences
41:21.780 --> 41:22.780
that we have.
41:22.780 --> 41:23.780
Right.
41:23.780 --> 41:30.260
You know, I can have very meaningful experiences even if there are other people who are smarter
41:30.260 --> 41:35.860
than me, you know, when I go to faculty meeting here and I was talking about something and
41:35.860 --> 41:39.420
then I certainly realized, oh, he has an old prize, he has an old prize, he has an old
41:39.420 --> 41:40.420
prize.
41:40.420 --> 41:41.420
Yeah.
41:41.420 --> 41:47.660
You know, it doesn't make me enjoy life any less or enjoy talking to those people less.
41:47.660 --> 41:49.780
Of course not.
41:49.780 --> 41:57.420
And contrary to that, I feel very honored and privileged to get to interact with other
41:57.420 --> 42:00.820
very intelligent beings that are better than me and a lot of stuff.
42:00.820 --> 42:05.420
So I don't think there's any reason why we can't have the same approach with intelligent
42:05.420 --> 42:06.420
machines.
42:06.420 --> 42:08.900
That's a really interesting, so people don't often think about that.
42:08.900 --> 42:14.380
They think about if there's machines that are more intelligent, you naturally think
42:14.380 --> 42:19.100
that that's not going to be a beneficial type of intelligence.
42:19.100 --> 42:24.060
You don't realize it could be, you know, like peers with no ball prizes that would be just
42:24.060 --> 42:25.060
fun to talk with.
42:25.060 --> 42:30.580
And they might be clever about certain topics and you can have fun having a few drinks with
42:30.580 --> 42:31.580
them.
42:31.580 --> 42:38.620
Well, also, you know, another example we can all relate to why it doesn't have to be a
42:38.620 --> 42:42.580
terrible thing to be impressed, the presence of people who are even smarter than us all
42:42.580 --> 42:47.980
around is when you and I were both two years old, I mean, our parents were much more intelligent
42:47.980 --> 42:48.980
than us.
42:48.980 --> 42:49.980
Right.
42:49.980 --> 42:50.980
Worked out okay.
42:50.980 --> 42:54.140
Because their goals were aligned with our goals.
42:54.140 --> 43:01.380
And that I think is really the number one key issue we have to solve if we value align
43:01.380 --> 43:07.380
the value alignment problem exactly because people who see too many Hollywood movies with
43:07.380 --> 43:12.260
lousy science fiction plot lines, they worry about the wrong thing, right?
43:12.260 --> 43:16.500
They worry about some machine suddenly turning evil.
43:16.500 --> 43:21.500
It's not malice that we should that is the concern.
43:21.500 --> 43:23.000
It's competence.
43:23.000 --> 43:29.580
By definition, intelligence makes you makes you very competent if you have a more intelligent
43:29.580 --> 43:35.300
goal playing machine computer playing as a less intelligent one and when we define intelligence
43:35.300 --> 43:37.740
as the ability to accomplish go winning, right?
43:37.740 --> 43:40.780
It's going to be the more intelligent one that wins.
43:40.780 --> 43:47.860
And if you have a human and then you have an AGI that's more intelligent in all ways
43:47.860 --> 43:50.500
and they have different goals, guess who's going to get their way, right?
43:50.500 --> 43:58.060
So I was just reading about this particular rhinoceros species that was driven extinct
43:58.060 --> 43:59.060
just a few years ago.
43:59.060 --> 44:05.740
Alan Bummer is looking at this cute picture of a mommy rhinoceros with its child, you
44:05.740 --> 44:09.140
know, and why did we humans drive it to extinction?
44:09.140 --> 44:12.860
It wasn't because we were evil rhino haters as a whole.
44:12.860 --> 44:16.380
It was just because we our goals weren't aligned with those of the rhinoceros and it didn't
44:16.380 --> 44:19.660
work out so well for the rhinoceros because we were more intelligent, right?
44:19.660 --> 44:27.220
So I think it's just so important that if we ever do build AGI before we unleash anything,
44:27.220 --> 44:37.380
we have to make sure that it learns to understand our goals, that it adopts our goals and retains
44:37.380 --> 44:38.380
those goals.
44:38.380 --> 44:45.740
So the cool interesting problem there is being able, us as human beings, trying to formulate
44:45.740 --> 44:47.240
our values.
44:47.240 --> 44:52.540
So you know, you could think of the United States Constitution as a way that people sat
44:52.540 --> 44:59.780
down at the time a bunch of white men, which is a good example, I should say.
44:59.780 --> 45:03.460
They formulated the goals for this country and a lot of people agree that those goals
45:03.460 --> 45:05.540
actually held up pretty well.
45:05.540 --> 45:09.600
It's an interesting formulation of values and failed miserably in other ways.
45:09.600 --> 45:15.500
So for the value alignment problem and the solution to it, we have to be able to put
45:15.500 --> 45:23.420
on paper or in a program, human values, how difficult do you think that is?
45:23.420 --> 45:24.420
Very.
45:24.420 --> 45:25.980
But it's so important.
45:25.980 --> 45:30.340
We really have to give it our best and it's difficult for two separate reasons.
45:30.340 --> 45:37.660
There's the technical value alignment problem of figuring out just how to make machines
45:37.660 --> 45:40.660
understand our goals, adopt them and retain them.
45:40.660 --> 45:46.140
And then there's the separate part of it, the philosophical part, whose values anyway.
45:46.140 --> 45:51.700
And since we, it's not like we have any great consensus on this planet on values, what mechanism
45:51.700 --> 45:56.780
should we create then to aggregate and decide, okay, what's a good compromise?
45:56.780 --> 46:01.260
That second discussion can't just be left the tech nerds like myself, right?
46:01.260 --> 46:02.260
That's right.
46:02.260 --> 46:06.820
And if we refuse to talk about it and then AGI gets built, who's going to be actually
46:06.820 --> 46:10.660
making the decision about whose values, it's going to be a bunch of dudes in some tech
46:10.660 --> 46:12.380
company, right?
46:12.380 --> 46:18.420
And are they necessarily so representative of all of humankind that we want to just
46:18.420 --> 46:19.580
endorse it to them?
46:19.580 --> 46:25.220
Are they even uniquely qualified to speak to future human happiness just because they're
46:25.220 --> 46:26.460
good at programming AI?
46:26.460 --> 46:30.380
I'd much rather have this be a really inclusive conversation.
46:30.380 --> 46:32.700
But do you think it's possible?
46:32.700 --> 46:38.820
You create a beautiful vision that includes sort of the diversity, cultural diversity
46:38.820 --> 46:43.900
and various perspectives on discussing rights, freedoms, human dignity.
46:43.900 --> 46:46.620
But how hard is it to come to that consensus?
46:46.620 --> 46:52.140
Do you think it's certainly a really important thing that we should all try to do, but do
46:52.140 --> 46:54.460
you think it's feasible?
46:54.460 --> 47:01.660
I think there's no better way to guarantee failure than to refuse to talk about it or
47:01.660 --> 47:02.980
refuse to try.
47:02.980 --> 47:08.060
And I also think it's a really bad strategy to say, okay, let's first have a discussion
47:08.060 --> 47:09.060
for a long time.
47:09.060 --> 47:13.540
And then once we reach complete consensus, then we'll try to load it into some machine.
47:13.540 --> 47:16.980
No, we shouldn't let perfect be the enemy of good.
47:16.980 --> 47:22.140
Instead, we should start with the kindergarten ethics that pretty much everybody agrees on
47:22.140 --> 47:24.580
and put that into our machines now.
47:24.580 --> 47:26.100
We're not doing that even.
47:26.100 --> 47:32.980
Look at anyone who builds a passenger aircraft wants it to never under any circumstances
47:32.980 --> 47:35.900
fly into a building or mountain, right?
47:35.900 --> 47:38.860
Yet the September 11 hijackers were able to do that.
47:38.860 --> 47:44.220
And even more embarrassingly, Andreas Lubitz, this depressed German wings pilot, when he
47:44.220 --> 47:50.220
flew his passenger jet into the Alps, killing over 100 people, he just told the autopilot
47:50.220 --> 47:51.220
to do it.
47:51.220 --> 47:55.140
He told the freaking computer to change the altitude to 100 meters.
47:55.140 --> 48:01.820
And even though it had the GPS maps, everything, the computer was like, okay, no, so we should
48:01.820 --> 48:07.300
take those very basic values, though, where the problem is not that we don't agree.
48:07.300 --> 48:12.460
The problem is just we've been too lazy to try to put it into our machines and make sure
48:12.460 --> 48:17.460
that from now on, airplanes will just, which all have computers in them, but we'll just
48:17.460 --> 48:19.820
never just refuse to do something like that.
48:19.820 --> 48:25.580
We go into safe mode, maybe lock the cockpit door, go to the nearest airport, and there's
48:25.580 --> 48:31.340
so much other technology in our world as well now where it's really coming quite timely
48:31.340 --> 48:34.300
to put in some sort of very basic values like this.
48:34.300 --> 48:41.460
Even in cars, we've had enough vehicle terrorism attacks by now where people have driven trucks
48:41.460 --> 48:47.300
and vans into pedestrians that it's not at all a crazy idea to just have that hardwired
48:47.300 --> 48:51.420
into the car, because yeah, there are a lot of, there's always going to be people who
48:51.420 --> 48:55.620
for some reason want to harm others, but most of those people don't have the technical
48:55.620 --> 48:58.620
expertise to figure out how to work around something like that.
48:58.620 --> 49:01.780
So if the car just won't do it, it helps.
49:01.780 --> 49:02.940
So let's start there.
49:02.940 --> 49:05.020
So there's a lot of, that's a great point.
49:05.020 --> 49:06.900
So not chasing perfect.
49:06.900 --> 49:10.780
There's a lot of things that most of the world agrees on.
49:10.780 --> 49:11.940
Yeah, let's start there.
49:11.940 --> 49:12.940
Let's start there.
49:12.940 --> 49:18.140
And then once we start there, we'll also get into the habit of having these kind of conversations
49:18.140 --> 49:21.940
about, okay, what else should we put in here and have these discussions?
49:21.940 --> 49:24.100
This should be a gradual process then.
49:24.100 --> 49:25.100
Great.
49:25.100 --> 49:31.380
So, but that also means describing these things and describing it to a machine.
49:31.380 --> 49:35.620
So one thing, we had a few conversations with Steven Wolfram.
49:35.620 --> 49:37.140
I'm not sure if you're familiar with Steven Wolfram.
49:37.140 --> 49:38.500
Oh yeah, I know him quite well.
49:38.500 --> 49:43.380
So he has, you know, he works with a bunch of things, but you know, cellular automata,
49:43.380 --> 49:47.660
these simple computable things, these computation systems.
49:47.660 --> 49:52.380
And he kind of mentioned that, you know, we probably have already within these systems
49:52.380 --> 49:59.580
already something that's AGI, meaning like we just don't know it because we can't talk
49:59.580 --> 50:00.580
to it.
50:00.580 --> 50:06.380
So if you give me this chance to try it, to try to at least form a question out of this,
50:06.380 --> 50:12.780
because I think it's an interesting idea to think that we can have intelligent systems,
50:12.780 --> 50:17.260
but we don't know how to describe something to them and they can't communicate with us.
50:17.260 --> 50:21.220
I know you're doing a little bit of work in explainable AI, trying to get AI to explain
50:21.220 --> 50:22.220
itself.
50:22.220 --> 50:28.340
So what are your thoughts of natural language processing or some kind of other communication?
50:28.340 --> 50:30.220
How does the AI explain something to us?
50:30.220 --> 50:33.740
How do we explain something to it, to machines?
50:33.740 --> 50:35.420
Or you think of it differently?
50:35.420 --> 50:40.100
So there are two separate parts to your question there.
50:40.100 --> 50:43.900
One of them has to do with communication, which is super interesting and I'll get to
50:43.900 --> 50:44.900
that in a sec.
50:44.900 --> 50:50.100
The other is whether we already have AGI, we just haven't noticed it.
50:50.100 --> 50:54.340
There, I beg to differ.
50:54.340 --> 50:58.420
And don't think there's anything in any cellular automaton or anything or the internet itself
50:58.420 --> 51:05.400
or whatever that has artificial general intelligence in that it didn't really do exactly everything
51:05.400 --> 51:06.980
we humans can do better.
51:06.980 --> 51:14.100
I think the day that happens, when that happens, we will very soon notice and we'll probably
51:14.100 --> 51:17.980
notice even before because in a very, very big way.
51:17.980 --> 51:18.980
For the second part though.
51:18.980 --> 51:20.700
Can I just, sorry.
51:20.700 --> 51:30.260
Because you have this beautiful way to formulate in consciousness as information processing
51:30.260 --> 51:33.740
and you can think of intelligence and information processing and you can think of the entire
51:33.740 --> 51:34.740
universe.
51:34.740 --> 51:40.220
These particles and these systems roaming around that have this information processing
51:40.220 --> 51:47.500
power, you don't think there is something with the power to process information in the
51:47.500 --> 51:55.460
way that we human beings do that's out there that needs to be sort of connected to.
51:55.460 --> 51:59.980
It seems a little bit philosophical perhaps, but there's something compelling to the idea
51:59.980 --> 52:06.100
that the power is already there, the focus should be more on being able to communicate
52:06.100 --> 52:07.100
with it.
52:07.100 --> 52:15.340
Well, I agree that in a certain sense, the hardware processing power is already out there
52:15.340 --> 52:21.180
because our universe itself can think of it as being a computer already.
52:21.180 --> 52:25.540
It's constantly computing what water waves, how it devolved the water waves and the river
52:25.540 --> 52:29.860
Charles and how to move the air molecules around that Seth Lloyd has pointed out.
52:29.860 --> 52:33.940
My colleague here that you can even in a very rigorous way think of our entire universe
52:33.940 --> 52:35.660
is just being a quantum computer.
52:35.660 --> 52:40.900
It's pretty clear that our universe supports this amazing processing power because you
52:40.900 --> 52:46.580
can even within this physics computer that we live in, we can even build actual laptops
52:46.580 --> 52:47.580
and stuff.
52:47.580 --> 52:49.140
So clearly the power is there.
52:49.140 --> 52:53.420
It's just that most of the compute power that nature has, it's in my opinion kind of wasting
52:53.420 --> 52:57.140
on boring stuff like simulating yet another ocean wave somewhere where no one is even
52:57.140 --> 52:58.140
looking.
52:58.140 --> 53:03.820
So in a sense, what life does, what we are doing when we build computers is we're rechanneling
53:03.820 --> 53:09.380
all this compute that nature is doing anyway into doing things that are more interesting
53:09.380 --> 53:14.220
than just yet another ocean wave and do something cool here.
53:14.220 --> 53:21.100
So the raw hardware power is there for sure, and even just computing what's going to happen
53:21.100 --> 53:25.540
for the next five seconds in this water ball, you know, it takes a ridiculous amount of
53:25.540 --> 53:28.060
compute if you do it on a human computer.
53:28.060 --> 53:30.040
This water ball just did it.
53:30.040 --> 53:36.020
But that does not mean that this water ball has AGI and this because AGI means it should
53:36.020 --> 53:40.300
also be able to like I've written my book done this interview.
53:40.300 --> 53:42.100
And I don't think it's just communication problems.
53:42.100 --> 53:47.020
I don't think it can do it.
53:47.020 --> 53:51.780
So Buddhists say when they watch the water and that there is some beauty, that there's
53:51.780 --> 53:55.380
some depth and beauty in nature that they can communicate with.
53:55.380 --> 54:01.180
Communication is also very important because I mean, look, part of my job is being a teacher
54:01.180 --> 54:09.940
and I know some very intelligent professors even who just have a better hard time communicating.
54:09.940 --> 54:14.620
They come up with all these brilliant ideas, but to communicate with somebody else, you
54:14.620 --> 54:17.140
have to also be able to simulate their own mind.
54:17.140 --> 54:18.140
Yes.
54:18.140 --> 54:22.020
And build well enough and understand that model of their mind that you can say things
54:22.020 --> 54:24.500
that they will understand.
54:24.500 --> 54:26.700
And that's quite difficult.
54:26.700 --> 54:31.620
And that's why today it's so frustrating if you have a computer that makes some cancer
54:31.620 --> 54:36.260
diagnosis and you ask it, well, why are you saying I should have a surgery?
54:36.260 --> 54:43.620
And if you don't want to reply, I was trained on five terabytes of data and this is my diagnosis
54:43.620 --> 54:49.220
boop, boop, beep, beep, doesn't really instill a lot of confidence, right?
54:49.220 --> 54:54.420
So I think we have a lot of work to do on communication there.
54:54.420 --> 54:59.380
So what kind of, I think you're doing a little bit of work in explainable AI.
54:59.380 --> 55:01.340
What do you think are the most promising avenues?
55:01.340 --> 55:07.100
Is it mostly about sort of the Alexa problem of natural language processing of being able
55:07.100 --> 55:13.220
to actually use human interpretable methods of communication?
55:13.220 --> 55:17.500
So being able to talk to a system and talk back to you, or is there some more fundamental
55:17.500 --> 55:18.500
problems to be solved?
55:18.500 --> 55:21.180
I think it's all of the above.
55:21.180 --> 55:27.180
The natural language processing is obviously important, but there are also more nerdy fundamental
55:27.180 --> 55:28.180
problems.
55:28.180 --> 55:39.180
Like if you take, you play chess, Russian, I have to, when did you learn Russian?
55:39.180 --> 55:45.700
I speak Russian very poorly, but I bought a book, teach yourself Russian, I read a lot,
55:45.700 --> 55:47.700
but it was very difficult.
55:47.700 --> 55:48.700
Wow.
55:48.700 --> 55:49.700
That's why I speak so poorly.
55:49.700 --> 55:51.700
How many languages do you know?
55:51.700 --> 55:52.700
Wow.
55:52.700 --> 55:53.700
That's really impressive.
55:53.700 --> 55:54.700
I don't know.
55:54.700 --> 55:58.740
My wife has some calculations, but my point was, if you played chess, have you looked
55:58.740 --> 56:00.260
at the AlphaZero games?
56:00.260 --> 56:01.260
Yeah.
56:01.260 --> 56:02.260
Oh, the actual games now.
56:02.260 --> 56:03.260
Check it out.
56:03.260 --> 56:09.900
Some of them are just mind blowing, really beautiful.
56:09.900 --> 56:12.460
If you ask, how did it do that?
56:12.460 --> 56:14.500
You got that.
56:14.500 --> 56:20.540
Talk to Demis Osabis, others from DeepMind, all they'll ultimately be able to give you
56:20.540 --> 56:26.940
is big tables of numbers, matrices that define the neural network, and you can stare at these
56:26.940 --> 56:32.980
tables numbers till your face turned blue, and you're not going to understand much about
56:32.980 --> 56:35.860
why it made that move.
56:35.860 --> 56:40.540
Even if you have a natural language processing that can tell you in human language about,
56:40.540 --> 56:44.180
oh, five, seven, point two, eight, still not going to really help.
56:44.180 --> 56:50.660
I think there's a whole spectrum of fun challenges there involved in taking computation that
56:50.660 --> 56:59.940
does intelligent things and transforming it into something equally good, equally intelligent,
56:59.940 --> 57:02.060
but that's more understandable.
57:02.060 --> 57:08.180
I think that's really valuable because I think as we put machines in charge of ever more
57:08.180 --> 57:13.540
infrastructure in our world, the power grid, the trading on the stock market, weapon systems,
57:13.540 --> 57:19.620
and so on, it's absolutely crucial that we can trust these AIs that do all we want and
57:19.620 --> 57:25.860
trust really comes from understanding in a very fundamental way.
57:25.860 --> 57:29.940
That's why I'm working on this, because I think the more if we're going to have some
57:29.940 --> 57:34.700
hope of ensuring that machines have adopted our goals and that they're going to retain
57:34.700 --> 57:41.260
them, that kind of trust, I think, needs to be based on things you can actually understand,
57:41.260 --> 57:47.140
preferably even improve theorems on, even with a self driving car, right?
57:47.140 --> 57:51.020
If someone just tells you it's been trained on tons of data and never crashed, it's less
57:51.020 --> 57:54.460
reassuring than if someone actually has a proof.
57:54.460 --> 57:58.820
Maybe it's a computer verified proof, but still it says that under no circumstances
57:58.820 --> 58:02.420
is this car just going to swerve into oncoming traffic.
58:02.420 --> 58:09.460
And that kind of information helps build trust and helps build the alignment of goals, at
58:09.460 --> 58:12.300
least awareness that your goals, your values are aligned.
58:12.300 --> 58:17.620
And I think even in the very short term, if you look at how today, this absolutely pathetic
58:17.620 --> 58:25.980
state of cybersecurity that we have, where is it, 3 billion Yahoo accounts are packed
58:25.980 --> 58:34.300
and almost every American's credit card and so on, you know, why is this happening?
58:34.300 --> 58:39.940
It's ultimately happening because we have software that nobody fully understood how
58:39.940 --> 58:41.460
it worked.
58:41.460 --> 58:45.100
That's why the bugs hadn't been found, right?
58:45.100 --> 58:50.340
And I think AI can be used very effectively for offense for hacking, but it can also be
58:50.340 --> 59:00.580
used for defense, hopefully, automating verifiability and creating systems that are built in different
59:00.580 --> 59:03.140
ways so you can actually prove things about them.
59:03.140 --> 59:05.460
And it's important.
59:05.460 --> 59:09.740
So speaking of software that nobody understands how it works, of course, a bunch of people
59:09.740 --> 59:14.820
ask about your paper about your thoughts of why does deep and cheap learning work so well?
59:14.820 --> 59:19.280
That's the paper, but what are your thoughts on deep learning, these kind of simplified
59:19.280 --> 59:26.620
models of our own brains that have been able to do some successful perception work, pattern
59:26.620 --> 59:30.940
recognition work, and now with AlphaZero and so on, do some clever things?
59:30.940 --> 59:35.740
What are your thoughts about the promise limitations of this piece?
59:35.740 --> 59:37.140
Great.
59:37.140 --> 59:44.300
I think there are a number of very important insights, very important lessons we can always
59:44.300 --> 59:47.340
draw from these kind of successes.
59:47.340 --> 59:50.460
One of them is when you look at the human brain, you see it's very complicated, a tenth
59:50.460 --> 59:54.140
of 11 neurons, and there are all these different kinds of neurons, and yada yada, and there's
59:54.140 --> 59:57.980
been this long debate about whether the fact that we have dozens of different kinds is
59:57.980 --> 1:00:01.580
actually necessary for intelligence.
1:00:01.580 --> 1:00:06.500
We can now, I think, quite convincingly answer that question of no, it's enough to have just
1:00:06.500 --> 1:00:07.500
one kind.
1:00:07.500 --> 1:00:11.780
If you look under the hood of AlphaZero, there's only one kind of neuron, and it's ridiculously
1:00:11.780 --> 1:00:15.060
simple, a simple mathematical thing.
1:00:15.060 --> 1:00:21.380
So it's just like in physics, if you have a gas with waves in it, it's not the detailed
1:00:21.380 --> 1:00:24.380
nature of the molecules that matter.
1:00:24.380 --> 1:00:27.060
It's the collective behavior, somehow.
1:00:27.060 --> 1:00:33.060
Similarly, it's this higher level structure of the network that matters, not that you
1:00:33.060 --> 1:00:34.060
have 20 kinds of neurons.
1:00:34.060 --> 1:00:41.740
I think our brain is such a complicated mess because it wasn't evolved just to be intelligent,
1:00:41.740 --> 1:00:51.740
it was evolved to also be self assembling, and self repairing, and evolutionarily attainable.
1:00:51.740 --> 1:00:53.660
And patches and so on.
1:00:53.660 --> 1:00:58.700
So I think it's pretty, my hunch is that we're going to understand how to build AGI before
1:00:58.700 --> 1:01:01.060
we fully understand how our brains work.
1:01:01.060 --> 1:01:06.260
Just like we understood how to build flying machines long before we were able to build
1:01:06.260 --> 1:01:07.260
a mechanical bird.
1:01:07.260 --> 1:01:08.260
Yeah, that's right.
1:01:08.260 --> 1:01:15.300
You've given the example of mechanical birds and airplanes, and airplanes do a pretty good
1:01:15.300 --> 1:01:18.620
job of flying without really mimicking bird flight.
1:01:18.620 --> 1:01:23.180
And even now, after 100 years later, did you see the TED talk with this German group of
1:01:23.180 --> 1:01:24.180
mechanical birds?
1:01:24.180 --> 1:01:25.180
I did not.
1:01:25.180 --> 1:01:26.180
I've heard you mention it.
1:01:26.180 --> 1:01:27.180
Check it out.
1:01:27.180 --> 1:01:28.180
It's amazing.
1:01:28.180 --> 1:01:30.180
But even after that, we still don't fly in mechanical birds because it turned out the
1:01:30.180 --> 1:01:34.580
way we came up with simpler, and it's better for our purposes, and I think it might be the
1:01:34.580 --> 1:01:35.580
same there.
1:01:35.580 --> 1:01:38.140
So that's one lesson.
1:01:38.140 --> 1:01:42.020
Another lesson is one of what our paper was about.
1:01:42.020 --> 1:01:47.420
Well, first, as a physicist thought, it was fascinating how there's a very close mathematical
1:01:47.420 --> 1:01:50.900
relationship, actually, between our artificial neural networks.
1:01:50.900 --> 1:01:56.580
And a lot of things that we've studied for in physics go by nerdy names like the renormalization
1:01:56.580 --> 1:02:01.100
group equation and Hamiltonians and yada, yada, yada.
1:02:01.100 --> 1:02:11.380
And when you look a little more closely at this, you have, at first, I was like, well,
1:02:11.380 --> 1:02:18.700
there's something crazy here that doesn't make sense because we know that if you even
1:02:18.700 --> 1:02:23.380
want to build a super simple neural network to tell apart cat pictures and dog pictures,
1:02:23.380 --> 1:02:27.260
right, that you can do that very, very well now.
1:02:27.260 --> 1:02:31.540
But if you think about it a little bit, you convince yourself it must be impossible because
1:02:31.540 --> 1:02:36.420
if I have one megapixel, even if each pixel is just black or white, there's two to the
1:02:36.420 --> 1:02:40.900
power of one million possible images, which is way more than there are atoms in our universe.
1:02:40.900 --> 1:02:47.220
So in order to, and then for each one of those, I have to assign a number, which is the probability
1:02:47.220 --> 1:02:49.100
that it's a dog.
1:02:49.100 --> 1:02:55.900
So an arbitrary function of images is a list of more numbers than there are atoms in our
1:02:55.900 --> 1:02:56.900
universe.
1:02:56.900 --> 1:03:02.220
So clearly, I can't store that under the hood of my, my GPU or my, my computer yet somehow
1:03:02.220 --> 1:03:03.220
works.
1:03:03.220 --> 1:03:04.220
So what does that mean?
1:03:04.220 --> 1:03:12.940
Well, it means that out of all of the problems that you could try to solve with a neural network,
1:03:12.940 --> 1:03:17.940
almost all of them are impossible to solve with a reasonably sized one.
1:03:17.940 --> 1:03:24.820
But then what we showed in our paper was, was that the, the fraction, the kind of problems,
1:03:24.820 --> 1:03:29.740
the fraction of all the problems that you could possibly pose that the, that we actually
1:03:29.740 --> 1:03:34.980
care about given the laws of physics is also an infinitesimally tiny little part.
1:03:34.980 --> 1:03:37.180
And amazingly, they're basically the same part.
1:03:37.180 --> 1:03:38.180
Yeah.
1:03:38.180 --> 1:03:41.180
It's almost like our world was created for, I mean, they kind of come together.
1:03:41.180 --> 1:03:42.180
Yeah.
1:03:42.180 --> 1:03:44.940
You, but you could say maybe where the world created the world that the world was created
1:03:44.940 --> 1:03:50.300
for us, but I have a more modest interpretation, which is that instead evolution endowments
1:03:50.300 --> 1:03:54.700
with neural networks, precisely for that reason, because this particular architecture has
1:03:54.700 --> 1:04:02.380
opposed to the one in your laptop is very, very well adapted to solving the kind of problems
1:04:02.380 --> 1:04:05.540
that nature kept presenting our ancestors with, right?
1:04:05.540 --> 1:04:09.380
So it makes sense that why do we have a brain in the first place?
1:04:09.380 --> 1:04:12.940
It's to be able to make predictions about the future and so on.
1:04:12.940 --> 1:04:17.580
So if we had a sucky system, which could never solve it, it wouldn't have a lot.
1:04:17.580 --> 1:04:23.420
So, but it's, so this is, this is a, I think a very beautiful fact.
1:04:23.420 --> 1:04:24.420
Yeah.
1:04:24.420 --> 1:04:28.780
And you also realize that there's, there, that we, there've been, it's been earlier
1:04:28.780 --> 1:04:34.140
work on, on why deeper networks are good, but we were able to show an additional cool
1:04:34.140 --> 1:04:40.260
fact there, which is that even incredibly simple problems, like suppose I give you a
1:04:40.260 --> 1:04:45.020
thousand numbers and ask you to multiply them together and you can write a few lines of
1:04:45.020 --> 1:04:46.820
code, boom, done, trivial.
1:04:46.820 --> 1:04:52.580
If you just try to do that with a neural network that has only one single hidden layer in it,
1:04:52.580 --> 1:04:59.940
you can do it, but you're going to need two to the power of thousand neurons to multiply
1:04:59.940 --> 1:05:03.260
a thousand numbers, which is again, more neurons than their atoms in our universe.
1:05:03.260 --> 1:05:05.740
So that's fascinating.
1:05:05.740 --> 1:05:11.580
But if you allow, if you allow yourself, make it a deep network of many layers, you only
1:05:11.580 --> 1:05:15.340
need four thousand neurons, it's perfectly feasible.
1:05:15.340 --> 1:05:17.500
So that's really interesting.
1:05:17.500 --> 1:05:18.500
Yeah.
1:05:18.500 --> 1:05:19.500
Yeah.
1:05:19.500 --> 1:05:24.460
So architecture type, I mean, you mentioned Schrodinger's equation and what are your thoughts
1:05:24.460 --> 1:05:32.860
about quantum computing and the role of this kind of computational unit in creating an
1:05:32.860 --> 1:05:34.900
intelligent system?
1:05:34.900 --> 1:05:41.100
In some Hollywood movies that I don't mention my name because I don't want to spoil them.
1:05:41.100 --> 1:05:46.820
The way they get AGI is building a quantum computer because the word quantum sounds
1:05:46.820 --> 1:05:47.820
cool and so on.
1:05:47.820 --> 1:05:48.820
That's right.
1:05:48.820 --> 1:05:54.940
But first of all, I think we don't need quantum computers to build AGI.
1:05:54.940 --> 1:06:01.740
I suspect your brain is not quantum computer in any found sense.
1:06:01.740 --> 1:06:03.460
So you don't even wrote a paper about that.
1:06:03.460 --> 1:06:09.060
Many years ago, I calculated the so called decoherence time that how long it takes until
1:06:09.060 --> 1:06:16.900
the quantum computerness of what your neurons are doing gets erased by just random noise
1:06:16.900 --> 1:06:21.420
from the environment and it's about 10 to the minus 21 seconds.
1:06:21.420 --> 1:06:27.420
So as cool as it would be to have a quantum computer in my head, I don't think that fast.
1:06:27.420 --> 1:06:35.820
On the other hand, there are very cool things you could do with quantum computers or I think
1:06:35.820 --> 1:06:40.780
we'll be able to do soon when we get bigger ones that might actually help machine learning
1:06:40.780 --> 1:06:43.180
do even better than the brain.
1:06:43.180 --> 1:06:58.620
So for example, one, this is just a moonshot, but hey, learning is very much same thing
1:06:58.620 --> 1:07:00.860
as search.
1:07:00.860 --> 1:07:05.460
If you're trying to train a neural network to get really learned to do something really
1:07:05.460 --> 1:07:10.820
well, you have some loss function, you have a bunch of knobs you can turn represented
1:07:10.820 --> 1:07:14.420
by a bunch of numbers and you're trying to tweak them so that it becomes as good as possible
1:07:14.420 --> 1:07:15.420
at this thing.
1:07:15.420 --> 1:07:22.580
So if you think of a landscape with some valley, where each dimension of the landscape corresponds
1:07:22.580 --> 1:07:25.780
to some number you can change, you're trying to find the minimum.
1:07:25.780 --> 1:07:29.980
And it's well known that if you have a very high dimensional landscape, complicated things,
1:07:29.980 --> 1:07:34.140
it's super hard to find the minimum.
1:07:34.140 --> 1:07:37.500
Quantum mechanics is amazingly good at this.
1:07:37.500 --> 1:07:42.980
If I want to know what's the lowest energy state this water can possibly have incredibly
1:07:42.980 --> 1:07:47.860
hard to compute, but nature will happily figure this out for you if you just cool it down,
1:07:47.860 --> 1:07:50.860
make it very, very cold.
1:07:50.860 --> 1:07:55.260
If you put a ball somewhere, it'll roll down to its minimum and this happens metaphorically
1:07:55.260 --> 1:07:57.620
at the energy landscape too.
1:07:57.620 --> 1:08:02.940
And quantum mechanics even uses some clever tricks which today's machine learning systems
1:08:02.940 --> 1:08:03.940
don't.
1:08:03.940 --> 1:08:07.940
If you're trying to find the minimum and you get stuck in the little local minimum here
1:08:07.940 --> 1:08:14.180
in quantum mechanics, you can actually tunnel through the barrier and get unstuck again.
1:08:14.180 --> 1:08:15.420
And that's really interesting.
1:08:15.420 --> 1:08:16.420
Yeah.
1:08:16.420 --> 1:08:22.940
So maybe for example, we'll one day use quantum computers that help train neural networks
1:08:22.940 --> 1:08:23.940
better.
1:08:23.940 --> 1:08:24.940
That's really interesting.
1:08:24.940 --> 1:08:25.940
Okay.
1:08:25.940 --> 1:08:32.020
So as a component of kind of the learning process, for example, let me ask sort of wrapping
1:08:32.020 --> 1:08:34.060
up here a little bit.
1:08:34.060 --> 1:08:40.540
Let me return to the questions of our human nature and love, as I mentioned.
1:08:40.540 --> 1:08:48.020
So do you think you mentioned sort of a helper robot that you could think of also personal
1:08:48.020 --> 1:08:49.020
robots.
1:08:49.020 --> 1:08:55.300
Do you think the way we human beings fall in love and get connected to each other is
1:08:55.300 --> 1:09:00.420
possible to achieve in an AI system and human level AI intelligence system.
1:09:00.420 --> 1:09:06.100
Do you think we would ever see that kind of connection or, you know, in all this discussion
1:09:06.100 --> 1:09:11.460
about solving complex goals, as this kind of human social connection, do you think that's
1:09:11.460 --> 1:09:16.460
one of the goals on the peaks and valleys that were the raising sea levels that we'd
1:09:16.460 --> 1:09:17.460
be able to achieve?
1:09:17.460 --> 1:09:22.180
Or do you think that's something that's ultimately, or at least in the short term, relative to
1:09:22.180 --> 1:09:23.620
the other goals is not achievable?
1:09:23.620 --> 1:09:25.220
I think it's all possible.
1:09:25.220 --> 1:09:31.780
And I mean, in recent, there's a very wide range of guesses, as you know, among AI researchers
1:09:31.780 --> 1:09:35.300
when we're going to get AGI.
1:09:35.300 --> 1:09:39.620
Some people, you know, like our friend Rodney Brooks said, it's going to be hundreds of
1:09:39.620 --> 1:09:41.140
years at least.
1:09:41.140 --> 1:09:44.780
And then there are many others that think it's going to happen relatively much sooner.
1:09:44.780 --> 1:09:52.140
Recent polls, maybe half or so, AI researchers think we're going to get AGI within decades.
1:09:52.140 --> 1:09:56.260
So if that happens, of course, then I think these things are all possible.
1:09:56.260 --> 1:10:01.860
But in terms of whether it will happen, I think we shouldn't spend so much time asking,
1:10:01.860 --> 1:10:04.260
what do we think will happen in the future?
1:10:04.260 --> 1:10:08.980
As if we are just some sort of pathetic, passive bystanders, you know, waiting for the future
1:10:08.980 --> 1:10:12.740
to happen to us, hey, we're the ones creating this future, right?
1:10:12.740 --> 1:10:18.340
So we should be proactive about it and ask ourselves what sort of future we would like
1:10:18.340 --> 1:10:19.340
to have happen.
1:10:19.340 --> 1:10:20.340
That's right.
1:10:20.340 --> 1:10:21.340
Trying to make it like that.
1:10:21.340 --> 1:10:25.660
Well, what I prefer is some sort of incredibly boring zombie like future where there's all
1:10:25.660 --> 1:10:30.220
these mechanical things happening and there's no passion, no emotion, no experience, maybe
1:10:30.220 --> 1:10:31.220
even.
1:10:31.220 --> 1:10:35.740
No, I would, of course, much rather prefer it if all the things that we find that we
1:10:35.740 --> 1:10:44.180
value the most about humanity are a subjective experience, passion, inspiration, love, you
1:10:44.180 --> 1:10:50.780
know, if we can create a future where those things do exist.
1:10:50.780 --> 1:10:56.500
You know, I think ultimately it's not our universe giving meaning to us, it's us giving
1:10:56.500 --> 1:10:58.500
meaning to our universe.
1:10:58.500 --> 1:11:03.620
And if we build more advanced intelligence, let's make sure we build it in such a way
1:11:03.620 --> 1:11:09.100
that meaning is part of it.
1:11:09.100 --> 1:11:13.900
A lot of people that seriously study this problem and think of it from different angles have
1:11:13.900 --> 1:11:20.140
trouble in the majority of cases, if they think through that happen, are the ones that
1:11:20.140 --> 1:11:22.620
are not beneficial to humanity.
1:11:22.620 --> 1:11:27.260
And so, yeah, so what are your thoughts?
1:11:27.260 --> 1:11:33.820
What should people, you know, I really don't like people to be terrified, what's the way
1:11:33.820 --> 1:11:38.660
for people to think about it in a way that, in a way we can solve it and we can make it
1:11:38.660 --> 1:11:39.660
better.
1:11:39.660 --> 1:11:40.660
Yeah.
1:11:40.660 --> 1:11:44.780
No, I don't think panicking is going to help in any way, it's not going to increase chances
1:11:44.780 --> 1:11:46.060
of things going well either.
1:11:46.060 --> 1:11:49.340
Even if you are in a situation where there is a real threat, does it help if everybody
1:11:49.340 --> 1:11:50.620
just freaks out?
1:11:50.620 --> 1:11:51.620
Right.
1:11:51.620 --> 1:11:53.620
No, of course not.
1:11:53.620 --> 1:11:59.740
I think, yeah, there are, of course, ways in which things can go horribly wrong.
1:11:59.740 --> 1:12:04.460
First of all, it's important when we think about this thing, this, about the problems
1:12:04.460 --> 1:12:08.780
and risks, to also remember how huge the upsides can be if we get it right.
1:12:08.780 --> 1:12:13.420
Everything we love about society and civilization is a product of intelligence.
1:12:13.420 --> 1:12:17.980
So if we can amplify our intelligence with machine intelligence and not anymore lose
1:12:17.980 --> 1:12:23.380
our loved ones, what we're told is an uncurable disease and things like this, of course, we
1:12:23.380 --> 1:12:24.940
should aspire to that.
1:12:24.940 --> 1:12:28.700
So that can be a motivator, I think, reminding yourselves that the reason we try to solve
1:12:28.700 --> 1:12:34.140
problems is not just because we're trying to avoid gloom, but because we're trying to
1:12:34.140 --> 1:12:35.900
do something great.
1:12:35.900 --> 1:12:43.340
But then in terms of the risks, I think the really important question is to ask, what
1:12:43.340 --> 1:12:47.740
can we do today that will actually help make the outcome good, right?
1:12:47.740 --> 1:12:52.700
And dismissing the risk is not one of them, you know, I find it quite funny often when
1:12:52.700 --> 1:13:01.540
I'm in discussion panels about these things, how the people who work for companies will
1:13:01.540 --> 1:13:05.100
always be like, oh, nothing to worry about, nothing to worry about, nothing to worry about.
1:13:05.100 --> 1:13:09.980
And it's always, it's only academics sometimes express concerns.
1:13:09.980 --> 1:13:10.980
That's not surprising at all.
1:13:10.980 --> 1:13:17.500
If you think about it, often Sinclair quipped, right, that it's hard to make a man believe
1:13:17.500 --> 1:13:20.620
in something when his income depends on not believing in it.
1:13:20.620 --> 1:13:25.580
And frankly, we know a lot of these people in companies that they're just as concerned
1:13:25.580 --> 1:13:26.580
as anyone else.
1:13:26.580 --> 1:13:30.300
But if you're the CEO of a company, that's not something you want to go on record saying
1:13:30.300 --> 1:13:34.980
when you have silly journalists who are going to put a picture of a Terminator robot when
1:13:34.980 --> 1:13:35.980
they quote you.
1:13:35.980 --> 1:13:39.380
So, so the issues are real.
1:13:39.380 --> 1:13:45.660
And the way I think about what the issue is, is basically, you know, the real choice we
1:13:45.660 --> 1:13:51.980
have is, first of all, are we going to dismiss this, the risks and say, well, you know, let's
1:13:51.980 --> 1:13:57.140
just go ahead and build machines that can do everything we can do better and cheaper,
1:13:57.140 --> 1:14:00.940
you know, let's just make ourselves obsolete as fast as possible or what could possibly
1:14:00.940 --> 1:14:01.940
go wrong.
1:14:01.940 --> 1:14:02.940
Right.
1:14:02.940 --> 1:14:03.940
That's one attitude.
1:14:03.940 --> 1:14:09.380
The opposite attitude that I think is to say, it's incredible potential, you know, let's
1:14:09.380 --> 1:14:14.900
think about what kind of future we're really, really excited about.
1:14:14.900 --> 1:14:18.700
What are the shared goals that we can really aspire towards?
1:14:18.700 --> 1:14:22.100
And then let's think really hard about how we can actually get there.
1:14:22.100 --> 1:14:23.100
So start with it.
1:14:23.100 --> 1:14:24.460
Don't start thinking about the risks.
1:14:24.460 --> 1:14:26.940
Start thinking about the goals.
1:14:26.940 --> 1:14:30.540
And then when you do that, then you can think about the obstacles you want to avoid, right?
1:14:30.540 --> 1:14:34.420
I often get students coming in right here into my office for career advice.
1:14:34.420 --> 1:14:38.060
Always ask them this very question, where do you want to be in the future?
1:14:38.060 --> 1:14:42.580
If all she can say is, oh, maybe I'll have cancer, maybe I'll run over by a truck.
1:14:42.580 --> 1:14:44.420
Focus on the obstacles instead of the goal.
1:14:44.420 --> 1:14:49.340
She's just going to end up a hypochondriac paranoid, whereas if she comes in and fire
1:14:49.340 --> 1:14:54.060
in her eyes and is like, I want to be there, and then we can talk about the obstacles and
1:14:54.060 --> 1:14:56.100
see how we can circumvent them.
1:14:56.100 --> 1:14:59.100
That's I think a much, much healthier attitude.
1:14:59.100 --> 1:15:01.540
And that's really what we're in.
1:15:01.540 --> 1:15:09.420
And I feel it's very challenging to come up with a vision for the future, which we're
1:15:09.420 --> 1:15:10.660
unequivocally excited about.
1:15:10.660 --> 1:15:14.300
I'm not just talking now in the vague terms like, yeah, let's cure cancer.
1:15:14.300 --> 1:15:18.500
I'm talking about what kind of society do we want to create?
1:15:18.500 --> 1:15:25.380
What do we want it to mean to be human in the age of AI, in the age of AGI?
1:15:25.380 --> 1:15:31.460
So if we can have this conversation, broad, inclusive conversation, and gradually start
1:15:31.460 --> 1:15:38.100
converging towards some future with some direction at least that we want to steer towards, right?
1:15:38.100 --> 1:15:42.340
Then we'll be much more motivated to constructively take on the obstacles.
1:15:42.340 --> 1:15:54.260
And I think if I wrap this up in a more succinct way, I think we can all agree already now that
1:15:54.260 --> 1:16:05.540
we should aspire to build AGI that doesn't overpower us, but that empowers us.
1:16:05.540 --> 1:16:10.820
And think of the many various ways that can do that, whether that's from my side of the
1:16:10.820 --> 1:16:12.860
world of autonomous vehicles.
1:16:12.860 --> 1:16:17.020
I'm personally actually from the camp that believes this human level intelligence is
1:16:17.020 --> 1:16:22.780
required to achieve something like vehicles that would actually be something we would
1:16:22.780 --> 1:16:25.380
enjoy using and being part of.
1:16:25.380 --> 1:16:26.380
So that's the one example.
1:16:26.380 --> 1:16:31.140
And certainly there's a lot of other types of robots and medicine and so on.
1:16:31.140 --> 1:16:35.300
So focusing on those and then coming up with the obstacles, coming up with the ways that
1:16:35.300 --> 1:16:38.420
that can go wrong and solving those one at a time.
1:16:38.420 --> 1:16:42.980
And just because you can build an autonomous vehicle, even if you could build one that
1:16:42.980 --> 1:16:47.500
would drive this final AGI, maybe there are some things in life that we would actually
1:16:47.500 --> 1:16:48.500
want to do ourselves.
1:16:48.500 --> 1:16:49.500
That's right.
1:16:49.500 --> 1:16:50.500
Right?
1:16:50.500 --> 1:16:54.660
Like, for example, if you think of our society as a whole, there are some things that we
1:16:54.660 --> 1:16:57.540
find very meaningful to do.
1:16:57.540 --> 1:17:02.100
And that doesn't mean we have to stop doing them just because machines can do them better.
1:17:02.100 --> 1:17:06.660
I'm not going to stop playing tennis just the day someone builds a tennis robot and
1:17:06.660 --> 1:17:07.660
beat me.
1:17:07.660 --> 1:17:09.900
People are still playing chess and even go.
1:17:09.900 --> 1:17:10.900
Yeah.
1:17:10.900 --> 1:17:19.100
And in this very near term, even some people are advocating basic income, replace jobs.
1:17:19.100 --> 1:17:22.780
But if the government is going to be willing to just hand out cash to people for doing
1:17:22.780 --> 1:17:27.660
nothing, then one should also seriously consider whether the government should also just hire
1:17:27.660 --> 1:17:33.380
a lot more teachers and nurses and the kind of jobs which people often find great fulfillment
1:17:33.380 --> 1:17:34.380
in doing, right?
1:17:34.380 --> 1:17:38.900
We get very tired of hearing politicians saying, oh, we can't afford hiring more teachers,
1:17:38.900 --> 1:17:41.700
but we're going to maybe have basic income.
1:17:41.700 --> 1:17:46.340
If we can have more serious research and thought into what gives meaning to our lives, the
1:17:46.340 --> 1:17:50.700
jobs give so much more than income, right?
1:17:50.700 --> 1:18:00.020
And then think about, in the future, what are the roles that we want to have people
1:18:00.020 --> 1:18:03.180
continually feeling empowered by machines?
1:18:03.180 --> 1:18:08.900
And I think sort of, I come from the Russia, from the Soviet Union, and I think for a lot
1:18:08.900 --> 1:18:14.100
of people in the 20th century, going to the moon, going to space was an inspiring thing.
1:18:14.100 --> 1:18:21.300
I feel like the universe of the mind, so AI, understanding, creating intelligence is that
1:18:21.300 --> 1:18:23.380
for the 21st century.
1:18:23.380 --> 1:18:26.740
So it's really surprising, and I've heard you mention this, it's really surprising to
1:18:26.740 --> 1:18:31.940
me both on the research funding side that it's not funded as greatly as it could be.
1:18:31.940 --> 1:18:36.500
But most importantly, on the politician side, that it's not part of the public discourse
1:18:36.500 --> 1:18:44.300
except in killer bots, terminator kind of view, that people are not yet, I think, perhaps
1:18:44.300 --> 1:18:48.260
excited by the possible positive future that we can build together.
1:18:48.260 --> 1:18:54.660
So we should be, because politicians usually just focus on the next election cycle, right?
1:18:54.660 --> 1:18:59.340
The single most important thing I feel we humans have learned in the entire history of science
1:18:59.340 --> 1:19:07.460
is they were the masters of underestimation, underestimated the size of our cosmos, again
1:19:07.460 --> 1:19:11.380
and again, realizing that everything we thought existed was just a small part of something
1:19:11.380 --> 1:19:12.380
grander, right?
1:19:12.380 --> 1:19:18.580
Planet, solar system, the galaxy, clusters of galaxies, the universe.
1:19:18.580 --> 1:19:25.700
And we now know that we have the future has just so much more potential than our ancestors
1:19:25.700 --> 1:19:27.820
could ever have dreamt of.
1:19:27.820 --> 1:19:39.820
This cosmos, imagine if all of Earth was completely devoid of life except for Cambridge, Massachusetts.
1:19:39.820 --> 1:19:44.220
Wouldn't it be kind of lame if all we ever aspired to was to stay in Cambridge, Massachusetts
1:19:44.220 --> 1:19:49.660
forever and then go extinct in one week, even though Earth was going to continue on for
1:19:49.660 --> 1:19:50.660
longer?
1:19:50.660 --> 1:19:57.300
That sort of attitude I think we have now on the cosmic scale, we can, life can flourish
1:19:57.300 --> 1:20:00.820
on Earth, not for four years, but for billions of years.
1:20:00.820 --> 1:20:06.340
I can even tell you about how to move it out of harm's way when the sun gets too hot.
1:20:06.340 --> 1:20:11.900
And then we have so much more resources out here, which today, maybe there are a lot of
1:20:11.900 --> 1:20:19.380
other planets with bacteria or cow like life on them, but most of this, all this opportunity
1:20:19.380 --> 1:20:25.380
seems as far as we can tell to be largely dead, like the Sahara Desert, and yet we have the
1:20:25.380 --> 1:20:30.380
opportunity to help life flourish around this for billions of years.
1:20:30.380 --> 1:20:37.420
So like, let's quit squabbling about whether some little border should be drawn one mile
1:20:37.420 --> 1:20:43.380
to the left or right and look up into the skies and realize, hey, we can do such incredible
1:20:43.380 --> 1:20:44.380
things.
1:20:44.380 --> 1:20:45.380
Yeah.
1:20:45.380 --> 1:20:49.980
And that's I think why it's really exciting that you and others are connected with some
1:20:49.980 --> 1:20:54.740
of the work Elon Musk is doing because he's literally going out into that space, really
1:20:54.740 --> 1:20:56.260
exploring our universe.
1:20:56.260 --> 1:20:57.260
And it's wonderful.
1:20:57.260 --> 1:21:02.340
That is exactly why Elon Musk is so misunderstood, right?
1:21:02.340 --> 1:21:05.300
Misconstrued with some kind of pessimistic doomsayer.
1:21:05.300 --> 1:21:10.860
The reason he cares so much about AI safety is because he more than almost anyone else
1:21:10.860 --> 1:21:13.340
appreciates these amazing opportunities.
1:21:13.340 --> 1:21:16.340
It will squander if we wipe out here on Earth.
1:21:16.340 --> 1:21:22.740
We're not just going to wipe out the next generation, but all generations and this incredible
1:21:22.740 --> 1:21:25.580
opportunity that's out there and that would be really be a waste.
1:21:25.580 --> 1:21:32.740
And AI, for people who think that there would be better to do without technology, let me
1:21:32.740 --> 1:21:37.740
just mention that if we don't improve our technology, the question isn't whether humanity
1:21:37.740 --> 1:21:38.740
is going to go extinct.
1:21:38.740 --> 1:21:43.620
The question is just whether we're going to get taken out by the next big asteroid or
1:21:43.620 --> 1:21:49.540
the next super volcano or something else dumb that we could easily prevent with more tech,
1:21:49.540 --> 1:21:50.540
right?
1:21:50.540 --> 1:21:56.220
If we want life to flourish throughout the cosmos, AI is the key to it.
1:21:56.220 --> 1:22:04.780
As I mentioned in a lot of detail in my book, even many of the most inspired sci fi writers
1:22:04.780 --> 1:22:11.580
I feel have totally underestimated the opportunities for space travel, especially to other galaxies,
1:22:11.580 --> 1:22:17.100
because they weren't thinking about the possibility of AGI, which just makes it so much easier.
1:22:17.100 --> 1:22:18.100
Right.
1:22:18.100 --> 1:22:25.900
Yeah, so that goes to a view of AGI that enables our progress, that enables a better life.
1:22:25.900 --> 1:22:30.060
So that's a beautiful way to put it and something to strive for.
1:22:30.060 --> 1:22:31.580
So Max, thank you so much.
1:22:31.580 --> 1:22:32.580
Thank you for your time today.
1:22:32.580 --> 1:22:33.580
It's been awesome.
1:22:33.580 --> 1:22:34.580
Thank you so much.
1:22:34.580 --> 1:22:35.580
Thanks.
1:22:35.580 --> 1:22:36.580
Merci beaucoup.
1:22:36.580 --> 1:22:49.100
Thank you so much for your time today and thank you so much for your time and for your
1:22:49.100 --> 1:22:50.100
time.
1:22:50.100 --> 1:22:51.100
Thank you.
1:22:51.100 --> 1:22:52.100
Thank you.
1:22:52.100 --> 1:22:53.100
Bye.
1:22:53.100 --> 1:22:54.100
Bye.
1:22:54.100 --> 1:22:55.100
Bye.
1:22:55.100 --> 1:22:56.100
Bye.
1:22:56.100 --> 1:22:57.100
Bye.
1:22:57.100 --> 1:22:58.100
Bye.
1:22:58.100 --> 1:22:59.100
Bye.
1:22:59.100 --> 1:23:00.100
Bye.