|
WEBVTT |
|
|
|
00:00.000 --> 00:03.720 |
|
The following is a conversation with Francois Chollet. |
|
|
|
00:03.720 --> 00:05.760 |
|
He's the creator of Keras, |
|
|
|
00:05.760 --> 00:08.080 |
|
which is an open source deep learning library |
|
|
|
00:08.080 --> 00:11.480 |
|
that is designed to enable fast, user friendly experimentation |
|
|
|
00:11.480 --> 00:13.600 |
|
with deep neural networks. |
|
|
|
00:13.600 --> 00:16.680 |
|
It serves as an interface to several deep learning libraries, |
|
|
|
00:16.680 --> 00:19.040 |
|
most popular of which is TensorFlow, |
|
|
|
00:19.040 --> 00:22.600 |
|
and it was integrated into the TensorFlow main code base |
|
|
|
00:22.600 --> 00:24.080 |
|
a while ago. |
|
|
|
00:24.080 --> 00:27.000 |
|
Meaning, if you want to create, train, |
|
|
|
00:27.000 --> 00:28.640 |
|
and use neural networks, |
|
|
|
00:28.640 --> 00:31.040 |
|
probably the easiest and most popular option |
|
|
|
00:31.040 --> 00:33.840 |
|
is to use Keras inside TensorFlow. |
|
|
|
00:34.840 --> 00:37.240 |
|
Aside from creating an exceptionally useful |
|
|
|
00:37.240 --> 00:38.680 |
|
and popular library, |
|
|
|
00:38.680 --> 00:41.920 |
|
Francois is also a world class AI researcher |
|
|
|
00:41.920 --> 00:43.680 |
|
and software engineer at Google. |
|
|
|
00:44.560 --> 00:46.960 |
|
And he's definitely an outspoken, |
|
|
|
00:46.960 --> 00:50.560 |
|
if not controversial personality in the AI world, |
|
|
|
00:50.560 --> 00:52.920 |
|
especially in the realm of ideas |
|
|
|
00:52.920 --> 00:55.920 |
|
around the future of artificial intelligence. |
|
|
|
00:55.920 --> 00:58.600 |
|
This is the Artificial Intelligence Podcast. |
|
|
|
00:58.600 --> 01:01.000 |
|
If you enjoy it, subscribe on YouTube, |
|
|
|
01:01.000 --> 01:02.760 |
|
give it five stars on iTunes, |
|
|
|
01:02.760 --> 01:04.160 |
|
support it on Patreon, |
|
|
|
01:04.160 --> 01:06.120 |
|
or simply connect with me on Twitter |
|
|
|
01:06.120 --> 01:09.960 |
|
at Lex Friedman, spelled F R I D M A N. |
|
|
|
01:09.960 --> 01:13.840 |
|
And now, here's my conversation with Francois Chollet. |
|
|
|
01:14.880 --> 01:17.320 |
|
You're known for not sugarcoating your opinions |
|
|
|
01:17.320 --> 01:19.160 |
|
and speaking your mind about ideas in AI, |
|
|
|
01:19.160 --> 01:21.160 |
|
especially on Twitter. |
|
|
|
01:21.160 --> 01:22.760 |
|
It's one of my favorite Twitter accounts. |
|
|
|
01:22.760 --> 01:26.320 |
|
So what's one of the more controversial ideas |
|
|
|
01:26.320 --> 01:29.360 |
|
you've expressed online and gotten some heat for? |
|
|
|
01:30.440 --> 01:31.360 |
|
How do you pick? |
|
|
|
01:33.080 --> 01:33.920 |
|
How do I pick? |
|
|
|
01:33.920 --> 01:36.880 |
|
Yeah, no, I think if you go through the trouble |
|
|
|
01:36.880 --> 01:39.640 |
|
of maintaining a Twitter account, |
|
|
|
01:39.640 --> 01:41.840 |
|
you might as well speak your mind, you know? |
|
|
|
01:41.840 --> 01:44.600 |
|
Otherwise, what's even the point of having a Twitter account? |
|
|
|
01:44.600 --> 01:45.480 |
|
It's like having a nice car |
|
|
|
01:45.480 --> 01:47.560 |
|
and just leaving it in the garage. |
|
|
|
01:48.600 --> 01:50.840 |
|
Yeah, so what's one thing for which I got |
|
|
|
01:50.840 --> 01:53.600 |
|
a lot of pushback? |
|
|
|
01:53.600 --> 01:56.640 |
|
Perhaps, you know, that time I wrote something |
|
|
|
01:56.640 --> 02:00.920 |
|
about the idea of intelligence explosion, |
|
|
|
02:00.920 --> 02:04.520 |
|
and I was questioning the idea |
|
|
|
02:04.520 --> 02:06.840 |
|
and the reasoning behind this idea. |
|
|
|
02:06.840 --> 02:09.640 |
|
And I got a lot of pushback on that. |
|
|
|
02:09.640 --> 02:11.840 |
|
I got a lot of flak for it. |
|
|
|
02:11.840 --> 02:13.600 |
|
So yeah, so intelligence explosion, |
|
|
|
02:13.600 --> 02:14.960 |
|
I'm sure you're familiar with the idea, |
|
|
|
02:14.960 --> 02:18.800 |
|
but it's the idea that if you were to build |
|
|
|
02:18.800 --> 02:22.920 |
|
general AI problem solving algorithms, |
|
|
|
02:22.920 --> 02:26.000 |
|
well, the problem of building such an AI, |
|
|
|
02:27.480 --> 02:30.520 |
|
that itself is a problem that could be solved by your AI, |
|
|
|
02:30.520 --> 02:31.880 |
|
and maybe it could be solved better |
|
|
|
02:31.880 --> 02:33.760 |
|
than what humans can do. |
|
|
|
02:33.760 --> 02:36.840 |
|
So your AI could start tweaking its own algorithm, |
|
|
|
02:36.840 --> 02:39.520 |
|
could start making a better version of itself, |
|
|
|
02:39.520 --> 02:43.240 |
|
and so on iteratively in a recursive fashion. |
|
|
|
02:43.240 --> 02:47.320 |
|
And so you would end up with an AI |
|
|
|
02:47.320 --> 02:50.080 |
|
with exponentially increasing intelligence. |
|
|
|
02:50.080 --> 02:50.920 |
|
That's right. |
|
|
|
02:50.920 --> 02:55.880 |
|
And I was basically questioning this idea, |
|
|
|
02:55.880 --> 02:59.040 |
|
first of all, because the notion of intelligence explosion |
|
|
|
02:59.040 --> 03:02.200 |
|
uses an implicit definition of intelligence |
|
|
|
03:02.200 --> 03:05.360 |
|
that doesn't sound quite right to me. |
|
|
|
03:05.360 --> 03:10.360 |
|
It considers intelligence as a property of a brain |
|
|
|
03:11.200 --> 03:13.680 |
|
that you can consider in isolation, |
|
|
|
03:13.680 --> 03:16.640 |
|
like the height of a building, for instance. |
|
|
|
03:16.640 --> 03:19.040 |
|
But that's not really what intelligence is. |
|
|
|
03:19.040 --> 03:22.200 |
|
Intelligence emerges from the interaction |
|
|
|
03:22.200 --> 03:25.240 |
|
between a brain, a body, |
|
|
|
03:25.240 --> 03:28.320 |
|
like embodied intelligence, and an environment. |
|
|
|
03:28.320 --> 03:30.720 |
|
And if you're missing one of these pieces, |
|
|
|
03:30.720 --> 03:33.800 |
|
then you cannot really define intelligence anymore. |
|
|
|
03:33.800 --> 03:36.800 |
|
So just tweaking a brain to make it smaller and smaller |
|
|
|
03:36.800 --> 03:39.120 |
|
doesn't actually make any sense to me. |
|
|
|
03:39.120 --> 03:39.960 |
|
So first of all, |
|
|
|
03:39.960 --> 03:43.000 |
|
you're crushing the dreams of many people, right? |
|
|
|
03:43.000 --> 03:46.000 |
|
So there's a, let's look at like Sam Harris. |
|
|
|
03:46.000 --> 03:48.680 |
|
Actually, a lot of physicists, Max Tegmark, |
|
|
|
03:48.680 --> 03:52.120 |
|
people who think the universe |
|
|
|
03:52.120 --> 03:54.640 |
|
is an information processing system, |
|
|
|
03:54.640 --> 03:57.680 |
|
our brain is kind of an information processing system. |
|
|
|
03:57.680 --> 03:59.400 |
|
So what's the theoretical limit? |
|
|
|
03:59.400 --> 04:03.160 |
|
Like, it doesn't make sense that there should be some, |
|
|
|
04:04.800 --> 04:07.520 |
|
it seems naive to think that our own brain |
|
|
|
04:07.520 --> 04:10.000 |
|
is somehow the limit of the capabilities |
|
|
|
04:10.000 --> 04:11.600 |
|
of this information system. |
|
|
|
04:11.600 --> 04:13.600 |
|
I'm playing devil's advocate here. |
|
|
|
04:13.600 --> 04:15.600 |
|
This information processing system. |
|
|
|
04:15.600 --> 04:17.760 |
|
And then if you just scale it, |
|
|
|
04:17.760 --> 04:19.360 |
|
if you're able to build something |
|
|
|
04:19.360 --> 04:20.920 |
|
that's on par with the brain, |
|
|
|
04:20.920 --> 04:24.040 |
|
you just, the process that builds it just continues |
|
|
|
04:24.040 --> 04:26.400 |
|
and it'll improve exponentially. |
|
|
|
04:26.400 --> 04:30.160 |
|
So that's the logic that's used actually |
|
|
|
04:30.160 --> 04:32.560 |
|
by almost everybody |
|
|
|
04:32.560 --> 04:36.920 |
|
that is worried about super human intelligence. |
|
|
|
04:36.920 --> 04:39.120 |
|
So you're trying to make, |
|
|
|
04:39.120 --> 04:40.960 |
|
so most people who are skeptical of that |
|
|
|
04:40.960 --> 04:43.000 |
|
are kind of like, this doesn't, |
|
|
|
04:43.000 --> 04:46.520 |
|
their thought process, this doesn't feel right. |
|
|
|
04:46.520 --> 04:47.680 |
|
Like that's for me as well. |
|
|
|
04:47.680 --> 04:49.760 |
|
So I'm more like, it doesn't, |
|
|
|
04:51.440 --> 04:52.800 |
|
the whole thing is shrouded in mystery |
|
|
|
04:52.800 --> 04:55.840 |
|
where you can't really say anything concrete, |
|
|
|
04:55.840 --> 04:57.880 |
|
but you could say this doesn't feel right. |
|
|
|
04:57.880 --> 05:00.640 |
|
This doesn't feel like that's how the brain works. |
|
|
|
05:00.640 --> 05:02.400 |
|
And you're trying to with your blog posts |
|
|
|
05:02.400 --> 05:05.680 |
|
and now making it a little more explicit. |
|
|
|
05:05.680 --> 05:10.680 |
|
So one idea is that the brain isn't exist alone. |
|
|
|
05:10.680 --> 05:13.200 |
|
It exists within the environment. |
|
|
|
05:13.200 --> 05:15.680 |
|
So you can't exponentially, |
|
|
|
05:15.680 --> 05:18.000 |
|
you would have to somehow exponentially improve |
|
|
|
05:18.000 --> 05:20.920 |
|
the environment and the brain together almost. |
|
|
|
05:20.920 --> 05:25.920 |
|
Yeah, in order to create something that's much smarter |
|
|
|
05:25.960 --> 05:27.840 |
|
in some kind of, |
|
|
|
05:27.840 --> 05:29.960 |
|
of course we don't have a definition of intelligence. |
|
|
|
05:29.960 --> 05:31.280 |
|
That's correct, that's correct. |
|
|
|
05:31.280 --> 05:34.280 |
|
I don't think, you should look at very smart people today, |
|
|
|
05:34.280 --> 05:37.280 |
|
even humans, not even talking about AIs. |
|
|
|
05:37.280 --> 05:38.640 |
|
I don't think their brain |
|
|
|
05:38.640 --> 05:41.960 |
|
and the performance of their brain is the bottleneck |
|
|
|
05:41.960 --> 05:45.200 |
|
to their expressed intelligence, to their achievements. |
|
|
|
05:46.600 --> 05:49.960 |
|
You cannot just tweak one part of this system, |
|
|
|
05:49.960 --> 05:52.840 |
|
like of this brain, body, environment system |
|
|
|
05:52.840 --> 05:55.960 |
|
and expect that capabilities like what emerges |
|
|
|
05:55.960 --> 06:00.280 |
|
out of this system to just explode exponentially. |
|
|
|
06:00.280 --> 06:04.200 |
|
Because anytime you improve one part of a system |
|
|
|
06:04.200 --> 06:06.760 |
|
with many interdependencies like this, |
|
|
|
06:06.760 --> 06:09.520 |
|
there's a new bottleneck that arises, right? |
|
|
|
06:09.520 --> 06:12.280 |
|
And I don't think even today for very smart people, |
|
|
|
06:12.280 --> 06:15.000 |
|
their brain is not the bottleneck |
|
|
|
06:15.000 --> 06:17.560 |
|
to the sort of problems they can solve, right? |
|
|
|
06:17.560 --> 06:19.800 |
|
In fact, many very smart people today, |
|
|
|
06:20.760 --> 06:22.520 |
|
you know, they are not actually solving |
|
|
|
06:22.520 --> 06:24.800 |
|
any big scientific problems, they're not Einstein. |
|
|
|
06:24.800 --> 06:28.280 |
|
They're like Einstein, but you know, the patent clerk days. |
|
|
|
06:29.800 --> 06:31.920 |
|
Like Einstein became Einstein |
|
|
|
06:31.920 --> 06:36.080 |
|
because this was a meeting of a genius |
|
|
|
06:36.080 --> 06:39.480 |
|
with a big problem at the right time, right? |
|
|
|
06:39.480 --> 06:42.480 |
|
But maybe this meeting could have never happened |
|
|
|
06:42.480 --> 06:44.960 |
|
and then Einstein would have just been a patent clerk, right? |
|
|
|
06:44.960 --> 06:48.400 |
|
And in fact, many people today are probably like |
|
|
|
06:49.760 --> 06:52.240 |
|
genius level smart, but you wouldn't know |
|
|
|
06:52.240 --> 06:54.800 |
|
because they're not really expressing any of that. |
|
|
|
06:54.800 --> 06:55.640 |
|
Wow, that's brilliant. |
|
|
|
06:55.640 --> 06:58.520 |
|
So we can think of the world, Earth, |
|
|
|
06:58.520 --> 07:02.720 |
|
but also the universe as just as a space of problems. |
|
|
|
07:02.720 --> 07:05.160 |
|
So all these problems and tasks are roaming it |
|
|
|
07:05.160 --> 07:06.880 |
|
of various difficulty. |
|
|
|
07:06.880 --> 07:10.120 |
|
And there's agents, creatures like ourselves |
|
|
|
07:10.120 --> 07:13.360 |
|
and animals and so on that are also roaming it. |
|
|
|
07:13.360 --> 07:16.480 |
|
And then you get coupled with a problem |
|
|
|
07:16.480 --> 07:17.640 |
|
and then you solve it. |
|
|
|
07:17.640 --> 07:19.880 |
|
But without that coupling, |
|
|
|
07:19.880 --> 07:22.560 |
|
you can't demonstrate your quote unquote intelligence. |
|
|
|
07:22.560 --> 07:24.480 |
|
Exactly, intelligence is the meeting |
|
|
|
07:24.480 --> 07:27.480 |
|
of great problem solving capabilities |
|
|
|
07:27.480 --> 07:28.760 |
|
with a great problem. |
|
|
|
07:28.760 --> 07:30.560 |
|
And if you don't have the problem, |
|
|
|
07:30.560 --> 07:32.280 |
|
you don't really express any intelligence. |
|
|
|
07:32.280 --> 07:34.760 |
|
All you're left with is potential intelligence, |
|
|
|
07:34.760 --> 07:36.240 |
|
like the performance of your brain |
|
|
|
07:36.240 --> 07:38.680 |
|
or how high your IQ is, |
|
|
|
07:38.680 --> 07:42.080 |
|
which in itself is just a number, right? |
|
|
|
07:42.080 --> 07:46.520 |
|
So you mentioned problem solving capacity. |
|
|
|
07:46.520 --> 07:47.360 |
|
Yeah. |
|
|
|
07:47.360 --> 07:51.800 |
|
What do you think of as problem solving capacity? |
|
|
|
07:51.800 --> 07:55.160 |
|
Can you try to define intelligence? |
|
|
|
07:56.640 --> 08:00.000 |
|
Like what does it mean to be more or less intelligent? |
|
|
|
08:00.000 --> 08:03.000 |
|
Is it completely coupled to a particular problem |
|
|
|
08:03.000 --> 08:05.720 |
|
or is there something a little bit more universal? |
|
|
|
08:05.720 --> 08:07.440 |
|
Yeah, I do believe all intelligence |
|
|
|
08:07.440 --> 08:09.080 |
|
is specialized intelligence. |
|
|
|
08:09.080 --> 08:12.200 |
|
Even human intelligence has some degree of generality. |
|
|
|
08:12.200 --> 08:15.320 |
|
Well, all intelligent systems have some degree of generality |
|
|
|
08:15.320 --> 08:19.400 |
|
but they're always specialized in one category of problems. |
|
|
|
08:19.400 --> 08:21.880 |
|
So the human intelligence is specialized |
|
|
|
08:21.880 --> 08:23.560 |
|
in the human experience. |
|
|
|
08:23.560 --> 08:25.560 |
|
And that shows at various levels, |
|
|
|
08:25.560 --> 08:30.200 |
|
that shows in some prior knowledge that's innate |
|
|
|
08:30.200 --> 08:32.040 |
|
that we have at birth. |
|
|
|
08:32.040 --> 08:35.360 |
|
Knowledge about things like agents, |
|
|
|
08:35.360 --> 08:38.080 |
|
goal driven behavior, visual priors |
|
|
|
08:38.080 --> 08:43.080 |
|
about what makes an object, priors about time and so on. |
|
|
|
08:43.520 --> 08:45.360 |
|
That shows also in the way we learn. |
|
|
|
08:45.360 --> 08:47.160 |
|
For instance, it's very, very easy for us |
|
|
|
08:47.160 --> 08:48.600 |
|
to pick up language. |
|
|
|
08:49.560 --> 08:52.080 |
|
It's very, very easy for us to learn certain things |
|
|
|
08:52.080 --> 08:54.920 |
|
because we are basically hard coded to learn them. |
|
|
|
08:54.920 --> 08:58.280 |
|
And we are specialized in solving certain kinds of problem |
|
|
|
08:58.280 --> 08:59.720 |
|
and we are quite useless |
|
|
|
08:59.720 --> 09:01.440 |
|
when it comes to other kinds of problems. |
|
|
|
09:01.440 --> 09:06.160 |
|
For instance, we are not really designed |
|
|
|
09:06.160 --> 09:08.800 |
|
to handle very long term problems. |
|
|
|
09:08.800 --> 09:12.880 |
|
We have no capability of seeing the very long term. |
|
|
|
09:12.880 --> 09:16.880 |
|
We don't have very much working memory. |
|
|
|
09:18.000 --> 09:20.080 |
|
So how do you think about long term? |
|
|
|
09:20.080 --> 09:21.360 |
|
Do you think long term planning, |
|
|
|
09:21.360 --> 09:24.880 |
|
are we talking about scale of years, millennia? |
|
|
|
09:24.880 --> 09:26.400 |
|
What do you mean by long term? |
|
|
|
09:26.400 --> 09:28.120 |
|
We're not very good. |
|
|
|
09:28.120 --> 09:29.760 |
|
Well, human intelligence is specialized |
|
|
|
09:29.760 --> 09:30.720 |
|
in the human experience. |
|
|
|
09:30.720 --> 09:32.800 |
|
And human experience is very short. |
|
|
|
09:32.800 --> 09:34.240 |
|
One lifetime is short. |
|
|
|
09:34.240 --> 09:35.880 |
|
Even within one lifetime, |
|
|
|
09:35.880 --> 09:40.000 |
|
we have a very hard time envisioning things |
|
|
|
09:40.000 --> 09:41.360 |
|
on a scale of years. |
|
|
|
09:41.360 --> 09:43.240 |
|
It's very difficult to project yourself |
|
|
|
09:43.240 --> 09:46.960 |
|
at a scale of five years, at a scale of 10 years and so on. |
|
|
|
09:46.960 --> 09:50.000 |
|
We can solve only fairly narrowly scoped problems. |
|
|
|
09:50.000 --> 09:52.320 |
|
So when it comes to solving bigger problems, |
|
|
|
09:52.320 --> 09:53.760 |
|
larger scale problems, |
|
|
|
09:53.760 --> 09:56.360 |
|
we are not actually doing it on an individual level. |
|
|
|
09:56.360 --> 09:59.280 |
|
So it's not actually our brain doing it. |
|
|
|
09:59.280 --> 10:03.040 |
|
We have this thing called civilization, right? |
|
|
|
10:03.040 --> 10:06.600 |
|
Which is itself a sort of problem solving system, |
|
|
|
10:06.600 --> 10:10.000 |
|
a sort of artificially intelligent system, right? |
|
|
|
10:10.000 --> 10:12.120 |
|
And it's not running on one brain, |
|
|
|
10:12.120 --> 10:14.080 |
|
it's running on a network of brains. |
|
|
|
10:14.080 --> 10:15.640 |
|
In fact, it's running on much more |
|
|
|
10:15.640 --> 10:16.760 |
|
than a network of brains. |
|
|
|
10:16.760 --> 10:20.080 |
|
It's running on a lot of infrastructure, |
|
|
|
10:20.080 --> 10:23.040 |
|
like books and computers and the internet |
|
|
|
10:23.040 --> 10:25.800 |
|
and human institutions and so on. |
|
|
|
10:25.800 --> 10:30.240 |
|
And that is capable of handling problems |
|
|
|
10:30.240 --> 10:33.760 |
|
on a much greater scale than any individual human. |
|
|
|
10:33.760 --> 10:37.600 |
|
If you look at computer science, for instance, |
|
|
|
10:37.600 --> 10:39.840 |
|
that's an institution that solves problems |
|
|
|
10:39.840 --> 10:42.560 |
|
and it is superhuman, right? |
|
|
|
10:42.560 --> 10:44.200 |
|
It operates on a greater scale. |
|
|
|
10:44.200 --> 10:46.880 |
|
It can solve much bigger problems |
|
|
|
10:46.880 --> 10:49.080 |
|
than an individual human could. |
|
|
|
10:49.080 --> 10:52.160 |
|
And science itself, science as a system, as an institution, |
|
|
|
10:52.160 --> 10:57.120 |
|
is a kind of artificially intelligent problem solving |
|
|
|
10:57.120 --> 10:59.360 |
|
algorithm that is superhuman. |
|
|
|
10:59.360 --> 11:02.800 |
|
Yeah, it's, at least computer science |
|
|
|
11:02.800 --> 11:07.720 |
|
is like a theorem prover at a scale of thousands, |
|
|
|
11:07.720 --> 11:10.400 |
|
maybe hundreds of thousands of human beings. |
|
|
|
11:10.400 --> 11:14.680 |
|
At that scale, what do you think is an intelligent agent? |
|
|
|
11:14.680 --> 11:18.280 |
|
So there's us humans at the individual level, |
|
|
|
11:18.280 --> 11:22.400 |
|
there is millions, maybe billions of bacteria in our skin. |
|
|
|
11:23.880 --> 11:26.400 |
|
There is, that's at the smaller scale. |
|
|
|
11:26.400 --> 11:29.160 |
|
You can even go to the particle level |
|
|
|
11:29.160 --> 11:31.000 |
|
as systems that behave, |
|
|
|
11:31.840 --> 11:34.360 |
|
you can say intelligently in some ways. |
|
|
|
11:35.440 --> 11:37.840 |
|
And then you can look at the earth as a single organism, |
|
|
|
11:37.840 --> 11:39.200 |
|
you can look at our galaxy |
|
|
|
11:39.200 --> 11:42.160 |
|
and even the universe as a single organism. |
|
|
|
11:42.160 --> 11:44.680 |
|
Do you think, how do you think about scale |
|
|
|
11:44.680 --> 11:46.280 |
|
in defining intelligent systems? |
|
|
|
11:46.280 --> 11:50.440 |
|
And we're here at Google, there is millions of devices |
|
|
|
11:50.440 --> 11:53.360 |
|
doing computation just in a distributed way. |
|
|
|
11:53.360 --> 11:55.880 |
|
How do you think about intelligence versus scale? |
|
|
|
11:55.880 --> 11:59.400 |
|
You can always characterize anything as a system. |
|
|
|
12:00.640 --> 12:03.600 |
|
I think people who talk about things |
|
|
|
12:03.600 --> 12:05.320 |
|
like intelligence explosion, |
|
|
|
12:05.320 --> 12:08.760 |
|
tend to focus on one agent is basically one brain, |
|
|
|
12:08.760 --> 12:10.960 |
|
like one brain considered in isolation, |
|
|
|
12:10.960 --> 12:13.200 |
|
like a brain, a jaw that's controlling a body |
|
|
|
12:13.200 --> 12:16.280 |
|
in a very like top to bottom kind of fashion. |
|
|
|
12:16.280 --> 12:19.480 |
|
And that body is pursuing goals into an environment. |
|
|
|
12:19.480 --> 12:20.720 |
|
So it's a very hierarchical view. |
|
|
|
12:20.720 --> 12:22.880 |
|
You have the brain at the top of the pyramid, |
|
|
|
12:22.880 --> 12:25.960 |
|
then you have the body just plainly receiving orders. |
|
|
|
12:25.960 --> 12:27.640 |
|
And then the body is manipulating objects |
|
|
|
12:27.640 --> 12:28.920 |
|
in the environment and so on. |
|
|
|
12:28.920 --> 12:32.920 |
|
So everything is subordinate to this one thing, |
|
|
|
12:32.920 --> 12:34.720 |
|
this epicenter, which is the brain. |
|
|
|
12:34.720 --> 12:37.120 |
|
But in real life, intelligent agents |
|
|
|
12:37.120 --> 12:39.240 |
|
don't really work like this, right? |
|
|
|
12:39.240 --> 12:40.920 |
|
There is no strong delimitation |
|
|
|
12:40.920 --> 12:43.400 |
|
between the brain and the body to start with. |
|
|
|
12:43.400 --> 12:45.000 |
|
You have to look not just at the brain, |
|
|
|
12:45.000 --> 12:46.560 |
|
but at the nervous system. |
|
|
|
12:46.560 --> 12:48.840 |
|
But then the nervous system and the body |
|
|
|
12:48.840 --> 12:50.760 |
|
are naturally two separate entities. |
|
|
|
12:50.760 --> 12:53.960 |
|
So you have to look at an entire animal as one agent. |
|
|
|
12:53.960 --> 12:57.000 |
|
But then you start realizing as you observe an animal |
|
|
|
12:57.000 --> 13:00.200 |
|
over any length of time, |
|
|
|
13:00.200 --> 13:03.160 |
|
that a lot of the intelligence of an animal |
|
|
|
13:03.160 --> 13:04.600 |
|
is actually externalized. |
|
|
|
13:04.600 --> 13:06.240 |
|
That's especially true for humans. |
|
|
|
13:06.240 --> 13:08.880 |
|
A lot of our intelligence is externalized. |
|
|
|
13:08.880 --> 13:10.360 |
|
When you write down some notes, |
|
|
|
13:10.360 --> 13:11.960 |
|
that is externalized intelligence. |
|
|
|
13:11.960 --> 13:14.000 |
|
When you write a computer program, |
|
|
|
13:14.000 --> 13:16.000 |
|
you are externalizing cognition. |
|
|
|
13:16.000 --> 13:19.720 |
|
So it's externalizing books, it's externalized in computers, |
|
|
|
13:19.720 --> 13:21.520 |
|
the internet, in other humans. |
|
|
|
13:23.080 --> 13:25.400 |
|
It's externalizing language and so on. |
|
|
|
13:25.400 --> 13:30.400 |
|
So there is no hard delimitation |
|
|
|
13:30.480 --> 13:32.640 |
|
of what makes an intelligent agent. |
|
|
|
13:32.640 --> 13:33.880 |
|
It's all about context. |
|
|
|
13:34.960 --> 13:38.800 |
|
Okay, but AlphaGo is better at Go |
|
|
|
13:38.800 --> 13:40.200 |
|
than the best human player. |
|
|
|
13:42.520 --> 13:45.000 |
|
There's levels of skill here. |
|
|
|
13:45.000 --> 13:48.600 |
|
So do you think there's such a ability, |
|
|
|
13:48.600 --> 13:52.800 |
|
such a concept as intelligence explosion |
|
|
|
13:52.800 --> 13:54.760 |
|
in a specific task? |
|
|
|
13:54.760 --> 13:57.360 |
|
And then, well, yeah. |
|
|
|
13:57.360 --> 14:00.120 |
|
Do you think it's possible to have a category of tasks |
|
|
|
14:00.120 --> 14:02.080 |
|
on which you do have something |
|
|
|
14:02.080 --> 14:05.040 |
|
like an exponential growth of ability |
|
|
|
14:05.040 --> 14:07.440 |
|
to solve that particular problem? |
|
|
|
14:07.440 --> 14:10.320 |
|
I think if you consider a specific vertical, |
|
|
|
14:10.320 --> 14:13.720 |
|
it's probably possible to some extent. |
|
|
|
14:15.320 --> 14:18.320 |
|
I also don't think we have to speculate about it |
|
|
|
14:18.320 --> 14:22.280 |
|
because we have real world examples |
|
|
|
14:22.280 --> 14:26.920 |
|
of recursively self improving intelligent systems, right? |
|
|
|
14:26.920 --> 14:30.920 |
|
So for instance, science is a problem solving system, |
|
|
|
14:30.920 --> 14:32.600 |
|
a knowledge generation system, |
|
|
|
14:32.600 --> 14:36.240 |
|
like a system that experiences the world in some sense |
|
|
|
14:36.240 --> 14:40.160 |
|
and then gradually understands it and can act on it. |
|
|
|
14:40.160 --> 14:42.120 |
|
And that system is superhuman |
|
|
|
14:42.120 --> 14:45.600 |
|
and it is clearly recursively self improving |
|
|
|
14:45.600 --> 14:47.560 |
|
because science feeds into technology. |
|
|
|
14:47.560 --> 14:50.200 |
|
Technology can be used to build better tools, |
|
|
|
14:50.200 --> 14:52.880 |
|
better computers, better instrumentation and so on, |
|
|
|
14:52.880 --> 14:56.720 |
|
which in turn can make science faster, right? |
|
|
|
14:56.720 --> 15:00.560 |
|
So science is probably the closest thing we have today |
|
|
|
15:00.560 --> 15:04.760 |
|
to a recursively self improving superhuman AI. |
|
|
|
15:04.760 --> 15:08.520 |
|
And you can just observe is science, |
|
|
|
15:08.520 --> 15:10.320 |
|
is scientific progress to the exploding, |
|
|
|
15:10.320 --> 15:12.800 |
|
which itself is an interesting question. |
|
|
|
15:12.800 --> 15:15.560 |
|
You can use that as a basis to try to understand |
|
|
|
15:15.560 --> 15:17.920 |
|
what will happen with a superhuman AI |
|
|
|
15:17.920 --> 15:21.000 |
|
that has a science like behavior. |
|
|
|
15:21.000 --> 15:23.320 |
|
Let me linger on it a little bit more. |
|
|
|
15:23.320 --> 15:27.600 |
|
What is your intuition why an intelligence explosion |
|
|
|
15:27.600 --> 15:28.560 |
|
is not possible? |
|
|
|
15:28.560 --> 15:30.920 |
|
Like taking the scientific, |
|
|
|
15:30.920 --> 15:33.240 |
|
all the semi scientific revolutions, |
|
|
|
15:33.240 --> 15:38.080 |
|
why can't we slightly accelerate that process? |
|
|
|
15:38.080 --> 15:41.200 |
|
So you can absolutely accelerate |
|
|
|
15:41.200 --> 15:43.120 |
|
any problem solving process. |
|
|
|
15:43.120 --> 15:46.720 |
|
So a recursively self improvement |
|
|
|
15:46.720 --> 15:48.640 |
|
is absolutely a real thing. |
|
|
|
15:48.640 --> 15:51.880 |
|
But what happens with a recursively self improving system |
|
|
|
15:51.880 --> 15:53.680 |
|
is typically not explosion |
|
|
|
15:53.680 --> 15:56.520 |
|
because no system exists in isolation. |
|
|
|
15:56.520 --> 15:58.640 |
|
And so tweaking one part of the system |
|
|
|
15:58.640 --> 16:00.880 |
|
means that suddenly another part of the system |
|
|
|
16:00.880 --> 16:02.200 |
|
becomes a bottleneck. |
|
|
|
16:02.200 --> 16:03.800 |
|
And if you look at science, for instance, |
|
|
|
16:03.800 --> 16:06.800 |
|
which is clearly a recursively self improving, |
|
|
|
16:06.800 --> 16:09.040 |
|
clearly a problem solving system, |
|
|
|
16:09.040 --> 16:12.000 |
|
scientific progress is not actually exploding. |
|
|
|
16:12.000 --> 16:13.520 |
|
If you look at science, |
|
|
|
16:13.520 --> 16:16.480 |
|
what you see is the picture of a system |
|
|
|
16:16.480 --> 16:19.240 |
|
that is consuming an exponentially increasing |
|
|
|
16:19.240 --> 16:20.520 |
|
amount of resources, |
|
|
|
16:20.520 --> 16:23.960 |
|
but it's having a linear output |
|
|
|
16:23.960 --> 16:26.000 |
|
in terms of scientific progress. |
|
|
|
16:26.000 --> 16:28.960 |
|
And maybe that will seem like a very strong claim. |
|
|
|
16:28.960 --> 16:31.160 |
|
Many people are actually saying that, |
|
|
|
16:31.160 --> 16:34.560 |
|
scientific progress is exponential, |
|
|
|
16:34.560 --> 16:36.120 |
|
but when they're claiming this, |
|
|
|
16:36.120 --> 16:38.400 |
|
they're actually looking at indicators |
|
|
|
16:38.400 --> 16:43.080 |
|
of resource consumption by science. |
|
|
|
16:43.080 --> 16:46.680 |
|
For instance, the number of papers being published, |
|
|
|
16:47.560 --> 16:49.960 |
|
the number of patents being filed and so on, |
|
|
|
16:49.960 --> 16:53.600 |
|
which are just completely correlated |
|
|
|
16:53.600 --> 16:58.480 |
|
with how many people are working on science today. |
|
|
|
16:58.480 --> 17:00.640 |
|
So it's actually an indicator of resource consumption, |
|
|
|
17:00.640 --> 17:03.200 |
|
but what you should look at is the output, |
|
|
|
17:03.200 --> 17:06.680 |
|
is progress in terms of the knowledge |
|
|
|
17:06.680 --> 17:08.040 |
|
that science generates, |
|
|
|
17:08.040 --> 17:10.640 |
|
in terms of the scope and significance |
|
|
|
17:10.640 --> 17:12.520 |
|
of the problems that we solve. |
|
|
|
17:12.520 --> 17:16.720 |
|
And some people have actually been trying to measure that. |
|
|
|
17:16.720 --> 17:20.160 |
|
Like Michael Nielsen, for instance, |
|
|
|
17:20.160 --> 17:21.920 |
|
he had a very nice paper, |
|
|
|
17:21.920 --> 17:23.720 |
|
I think that was last year about it. |
|
|
|
17:25.200 --> 17:28.360 |
|
So his approach to measure scientific progress |
|
|
|
17:28.360 --> 17:33.360 |
|
was to look at the timeline of scientific discoveries |
|
|
|
17:33.720 --> 17:37.160 |
|
over the past, you know, 100, 150 years. |
|
|
|
17:37.160 --> 17:41.360 |
|
And for each major discovery, |
|
|
|
17:41.360 --> 17:44.360 |
|
ask a panel of experts to rate |
|
|
|
17:44.360 --> 17:46.760 |
|
the significance of the discovery. |
|
|
|
17:46.760 --> 17:49.600 |
|
And if the output of science as an institution |
|
|
|
17:49.600 --> 17:50.440 |
|
were exponential, |
|
|
|
17:50.440 --> 17:55.440 |
|
you would expect the temporal density of significance |
|
|
|
17:56.600 --> 17:58.160 |
|
to go up exponentially. |
|
|
|
17:58.160 --> 18:00.960 |
|
Maybe because there's a faster rate of discoveries, |
|
|
|
18:00.960 --> 18:02.960 |
|
maybe because the discoveries are, you know, |
|
|
|
18:02.960 --> 18:04.920 |
|
increasingly more important. |
|
|
|
18:04.920 --> 18:06.800 |
|
And what actually happens |
|
|
|
18:06.800 --> 18:10.040 |
|
if you plot this temporal density of significance |
|
|
|
18:10.040 --> 18:11.320 |
|
measured in this way, |
|
|
|
18:11.320 --> 18:14.520 |
|
is that you see very much a flat graph. |
|
|
|
18:14.520 --> 18:16.600 |
|
You see a flat graph across all disciplines, |
|
|
|
18:16.600 --> 18:19.720 |
|
across physics, biology, medicine, and so on. |
|
|
|
18:19.720 --> 18:22.480 |
|
And it actually makes a lot of sense |
|
|
|
18:22.480 --> 18:23.320 |
|
if you think about it, |
|
|
|
18:23.320 --> 18:26.000 |
|
because think about the progress of physics |
|
|
|
18:26.000 --> 18:28.000 |
|
110 years ago, right? |
|
|
|
18:28.000 --> 18:30.080 |
|
It was a time of crazy change. |
|
|
|
18:30.080 --> 18:31.960 |
|
Think about the progress of technology, |
|
|
|
18:31.960 --> 18:34.360 |
|
you know, 170 years ago, |
|
|
|
18:34.360 --> 18:35.400 |
|
when we started having, you know, |
|
|
|
18:35.400 --> 18:37.560 |
|
replacing horses with cars, |
|
|
|
18:37.560 --> 18:40.000 |
|
when we started having electricity and so on. |
|
|
|
18:40.000 --> 18:41.520 |
|
It was a time of incredible change. |
|
|
|
18:41.520 --> 18:44.600 |
|
And today is also a time of very, very fast change, |
|
|
|
18:44.600 --> 18:48.040 |
|
but it would be an unfair characterization |
|
|
|
18:48.040 --> 18:50.560 |
|
to say that today technology and science |
|
|
|
18:50.560 --> 18:52.920 |
|
are moving way faster than they did 50 years ago |
|
|
|
18:52.920 --> 18:54.360 |
|
or 100 years ago. |
|
|
|
18:54.360 --> 18:59.360 |
|
And if you do try to rigorously plot |
|
|
|
18:59.520 --> 19:04.520 |
|
the temporal density of the significance, |
|
|
|
19:04.880 --> 19:07.320 |
|
yeah, of significance, sorry, |
|
|
|
19:07.320 --> 19:09.720 |
|
you do see very flat curves. |
|
|
|
19:09.720 --> 19:12.040 |
|
And you can check out the paper |
|
|
|
19:12.040 --> 19:16.000 |
|
that Michael Nielsen had about this idea. |
|
|
|
19:16.000 --> 19:20.000 |
|
And so the way I interpret it is, |
|
|
|
19:20.000 --> 19:24.160 |
|
as you make progress in a given field, |
|
|
|
19:24.160 --> 19:26.120 |
|
or in a given subfield of science, |
|
|
|
19:26.120 --> 19:28.680 |
|
it becomes exponentially more difficult |
|
|
|
19:28.680 --> 19:30.440 |
|
to make further progress. |
|
|
|
19:30.440 --> 19:35.000 |
|
Like the very first person to work on information theory. |
|
|
|
19:35.000 --> 19:36.440 |
|
If you enter a new field, |
|
|
|
19:36.440 --> 19:37.920 |
|
and it's still the very early years, |
|
|
|
19:37.920 --> 19:41.160 |
|
there's a lot of low hanging fruit you can pick. |
|
|
|
19:41.160 --> 19:42.000 |
|
That's right, yeah. |
|
|
|
19:42.000 --> 19:43.960 |
|
But the next generation of researchers |
|
|
|
19:43.960 --> 19:48.160 |
|
is gonna have to dig much harder, actually, |
|
|
|
19:48.160 --> 19:50.640 |
|
to make smaller discoveries, |
|
|
|
19:50.640 --> 19:52.640 |
|
probably larger number of smaller discoveries, |
|
|
|
19:52.640 --> 19:54.640 |
|
and to achieve the same amount of impact, |
|
|
|
19:54.640 --> 19:57.480 |
|
you're gonna need a much greater head count. |
|
|
|
19:57.480 --> 20:00.040 |
|
And that's exactly the picture you're seeing with science, |
|
|
|
20:00.040 --> 20:03.760 |
|
that the number of scientists and engineers |
|
|
|
20:03.760 --> 20:06.520 |
|
is in fact increasing exponentially. |
|
|
|
20:06.520 --> 20:08.400 |
|
The amount of computational resources |
|
|
|
20:08.400 --> 20:10.040 |
|
that are available to science |
|
|
|
20:10.040 --> 20:11.880 |
|
is increasing exponentially and so on. |
|
|
|
20:11.880 --> 20:15.560 |
|
So the resource consumption of science is exponential, |
|
|
|
20:15.560 --> 20:18.200 |
|
but the output in terms of progress, |
|
|
|
20:18.200 --> 20:21.000 |
|
in terms of significance, is linear. |
|
|
|
20:21.000 --> 20:23.120 |
|
And the reason why is because, |
|
|
|
20:23.120 --> 20:26.000 |
|
and even though science is regressively self improving, |
|
|
|
20:26.000 --> 20:28.440 |
|
meaning that scientific progress |
|
|
|
20:28.440 --> 20:30.240 |
|
turns into technological progress, |
|
|
|
20:30.240 --> 20:32.960 |
|
which in turn helps science. |
|
|
|
20:32.960 --> 20:35.280 |
|
If you look at computers, for instance, |
|
|
|
20:35.280 --> 20:38.480 |
|
our products of science and computers |
|
|
|
20:38.480 --> 20:41.560 |
|
are tremendously useful in speeding up science. |
|
|
|
20:41.560 --> 20:43.840 |
|
The internet, same thing, the internet is a technology |
|
|
|
20:43.840 --> 20:47.480 |
|
that's made possible by very recent scientific advances. |
|
|
|
20:47.480 --> 20:52.400 |
|
And itself, because it enables scientists to network, |
|
|
|
20:52.400 --> 20:55.520 |
|
to communicate, to exchange papers and ideas much faster, |
|
|
|
20:55.520 --> 20:57.440 |
|
it is a way to speed up scientific progress. |
|
|
|
20:57.440 --> 20:58.440 |
|
So even though you're looking |
|
|
|
20:58.440 --> 21:01.440 |
|
at a regressively self improving system, |
|
|
|
21:01.440 --> 21:04.080 |
|
it is consuming exponentially more resources |
|
|
|
21:04.080 --> 21:09.080 |
|
to produce the same amount of problem solving, very much. |
|
|
|
21:09.200 --> 21:11.080 |
|
So that's a fascinating way to paint it, |
|
|
|
21:11.080 --> 21:14.960 |
|
and certainly that holds for the deep learning community. |
|
|
|
21:14.960 --> 21:18.120 |
|
If you look at the temporal, what did you call it, |
|
|
|
21:18.120 --> 21:21.240 |
|
the temporal density of significant ideas, |
|
|
|
21:21.240 --> 21:23.920 |
|
if you look at in deep learning, |
|
|
|
21:24.840 --> 21:26.960 |
|
I think, I'd have to think about that, |
|
|
|
21:26.960 --> 21:29.040 |
|
but if you really look at significant ideas |
|
|
|
21:29.040 --> 21:32.400 |
|
in deep learning, they might even be decreasing. |
|
|
|
21:32.400 --> 21:37.400 |
|
So I do believe the per paper significance is decreasing, |
|
|
|
21:39.600 --> 21:41.240 |
|
but the amount of papers |
|
|
|
21:41.240 --> 21:43.440 |
|
is still today exponentially increasing. |
|
|
|
21:43.440 --> 21:45.880 |
|
So I think if you look at an aggregate, |
|
|
|
21:45.880 --> 21:48.840 |
|
my guess is that you would see a linear progress. |
|
|
|
21:48.840 --> 21:53.840 |
|
If you were to sum the significance of all papers, |
|
|
|
21:56.120 --> 21:58.640 |
|
you would see roughly in your progress. |
|
|
|
21:58.640 --> 22:03.640 |
|
And in my opinion, it is not a coincidence |
|
|
|
22:03.880 --> 22:05.800 |
|
that you're seeing linear progress in science |
|
|
|
22:05.800 --> 22:07.720 |
|
despite exponential resource consumption. |
|
|
|
22:07.720 --> 22:10.280 |
|
I think the resource consumption |
|
|
|
22:10.280 --> 22:15.280 |
|
is dynamically adjusting itself to maintain linear progress |
|
|
|
22:15.880 --> 22:18.560 |
|
because we as a community expect linear progress, |
|
|
|
22:18.560 --> 22:21.240 |
|
meaning that if we start investing less |
|
|
|
22:21.240 --> 22:23.600 |
|
and seeing less progress, it means that suddenly |
|
|
|
22:23.600 --> 22:26.800 |
|
there are some lower hanging fruits that become available |
|
|
|
22:26.800 --> 22:31.280 |
|
and someone's gonna step up and pick them, right? |
|
|
|
22:31.280 --> 22:36.280 |
|
So it's very much like a market for discoveries and ideas. |
|
|
|
22:36.920 --> 22:38.720 |
|
But there's another fundamental part |
|
|
|
22:38.720 --> 22:41.800 |
|
which you're highlighting, which as a hypothesis |
|
|
|
22:41.800 --> 22:45.160 |
|
as science or like the space of ideas, |
|
|
|
22:45.160 --> 22:48.160 |
|
any one path you travel down, |
|
|
|
22:48.160 --> 22:51.080 |
|
it gets exponentially more difficult |
|
|
|
22:51.080 --> 22:54.720 |
|
to get a new way to develop new ideas. |
|
|
|
22:54.720 --> 22:57.640 |
|
And your sense is that's gonna hold |
|
|
|
22:57.640 --> 23:01.520 |
|
across our mysterious universe. |
|
|
|
23:01.520 --> 23:03.360 |
|
Yes, well, exponential progress |
|
|
|
23:03.360 --> 23:05.480 |
|
triggers exponential friction. |
|
|
|
23:05.480 --> 23:07.440 |
|
So that if you tweak one part of the system, |
|
|
|
23:07.440 --> 23:10.680 |
|
suddenly some other part becomes a bottleneck, right? |
|
|
|
23:10.680 --> 23:14.880 |
|
For instance, let's say you develop some device |
|
|
|
23:14.880 --> 23:17.160 |
|
that measures its own acceleration |
|
|
|
23:17.160 --> 23:18.720 |
|
and then it has some engine |
|
|
|
23:18.720 --> 23:20.800 |
|
and it outputs even more acceleration |
|
|
|
23:20.800 --> 23:22.360 |
|
in proportion of its own acceleration |
|
|
|
23:22.360 --> 23:23.320 |
|
and you drop it somewhere, |
|
|
|
23:23.320 --> 23:25.240 |
|
it's not gonna reach infinite speed |
|
|
|
23:25.240 --> 23:27.880 |
|
because it exists in a certain context. |
|
|
|
23:29.080 --> 23:31.000 |
|
So the air around it is gonna generate friction |
|
|
|
23:31.000 --> 23:34.320 |
|
and it's gonna block it at some top speed. |
|
|
|
23:34.320 --> 23:37.480 |
|
And even if you were to consider the broader context |
|
|
|
23:37.480 --> 23:39.840 |
|
and lift the bottleneck there, |
|
|
|
23:39.840 --> 23:42.240 |
|
like the bottleneck of friction, |
|
|
|
23:43.120 --> 23:45.120 |
|
then some other part of the system |
|
|
|
23:45.120 --> 23:48.120 |
|
would start stepping in and creating exponential friction, |
|
|
|
23:48.120 --> 23:49.920 |
|
maybe the speed of flight or whatever. |
|
|
|
23:49.920 --> 23:51.920 |
|
And this definitely holds true |
|
|
|
23:51.920 --> 23:54.960 |
|
when you look at the problem solving algorithm |
|
|
|
23:54.960 --> 23:58.160 |
|
that is being run by science as an institution, |
|
|
|
23:58.160 --> 23:59.720 |
|
science as a system. |
|
|
|
23:59.720 --> 24:01.720 |
|
As you make more and more progress, |
|
|
|
24:01.720 --> 24:05.800 |
|
despite having this recursive self improvement component, |
|
|
|
24:06.760 --> 24:09.840 |
|
you are encountering exponential friction. |
|
|
|
24:09.840 --> 24:13.480 |
|
The more researchers you have working on different ideas, |
|
|
|
24:13.480 --> 24:14.880 |
|
the more overhead you have |
|
|
|
24:14.880 --> 24:18.040 |
|
in terms of communication across researchers. |
|
|
|
24:18.040 --> 24:22.920 |
|
If you look at, you were mentioning quantum mechanics, right? |
|
|
|
24:22.920 --> 24:26.880 |
|
Well, if you want to start making significant discoveries |
|
|
|
24:26.880 --> 24:29.680 |
|
today, significant progress in quantum mechanics, |
|
|
|
24:29.680 --> 24:33.000 |
|
there is an amount of knowledge you have to ingest, |
|
|
|
24:33.000 --> 24:34.080 |
|
which is huge. |
|
|
|
24:34.080 --> 24:36.520 |
|
So there's a very large overhead |
|
|
|
24:36.520 --> 24:39.240 |
|
to even start to contribute. |
|
|
|
24:39.240 --> 24:40.680 |
|
There's a large amount of overhead |
|
|
|
24:40.680 --> 24:44.040 |
|
to synchronize across researchers and so on. |
|
|
|
24:44.040 --> 24:47.440 |
|
And of course, the significant practical experiments |
|
|
|
24:48.600 --> 24:52.160 |
|
are going to require exponentially expensive equipment |
|
|
|
24:52.160 --> 24:56.480 |
|
because the easier ones have already been run, right? |
|
|
|
24:56.480 --> 25:00.480 |
|
So in your senses, there's no way escaping, |
|
|
|
25:00.480 --> 25:04.480 |
|
there's no way of escaping this kind of friction |
|
|
|
25:04.480 --> 25:08.600 |
|
with artificial intelligence systems. |
|
|
|
25:08.600 --> 25:11.520 |
|
Yeah, no, I think science is a very good way |
|
|
|
25:11.520 --> 25:14.280 |
|
to model what would happen with a superhuman |
|
|
|
25:14.280 --> 25:16.440 |
|
recursive research improving AI. |
|
|
|
25:16.440 --> 25:18.240 |
|
That's your sense, I mean, the... |
|
|
|
25:18.240 --> 25:19.680 |
|
That's my intuition. |
|
|
|
25:19.680 --> 25:23.400 |
|
It's not like a mathematical proof of anything. |
|
|
|
25:23.400 --> 25:24.400 |
|
That's not my point. |
|
|
|
25:24.400 --> 25:26.600 |
|
Like, I'm not trying to prove anything. |
|
|
|
25:26.600 --> 25:27.920 |
|
I'm just trying to make an argument |
|
|
|
25:27.920 --> 25:31.160 |
|
to question the narrative of intelligence explosion, |
|
|
|
25:31.160 --> 25:32.880 |
|
which is quite a dominant narrative. |
|
|
|
25:32.880 --> 25:35.840 |
|
And you do get a lot of pushback if you go against it. |
|
|
|
25:35.840 --> 25:39.320 |
|
Because, so for many people, right, |
|
|
|
25:39.320 --> 25:42.200 |
|
AI is not just a subfield of computer science. |
|
|
|
25:42.200 --> 25:44.120 |
|
It's more like a belief system. |
|
|
|
25:44.120 --> 25:48.640 |
|
Like this belief that the world is headed towards an event, |
|
|
|
25:48.640 --> 25:55.040 |
|
the singularity, past which, you know, AI will become... |
|
|
|
25:55.040 --> 25:57.080 |
|
will go exponential very much, |
|
|
|
25:57.080 --> 25:58.600 |
|
and the world will be transformed, |
|
|
|
25:58.600 --> 26:00.840 |
|
and humans will become obsolete. |
|
|
|
26:00.840 --> 26:03.880 |
|
And if you go against this narrative, |
|
|
|
26:03.880 --> 26:06.920 |
|
because it is not really a scientific argument, |
|
|
|
26:06.920 --> 26:08.880 |
|
but more of a belief system, |
|
|
|
26:08.880 --> 26:11.240 |
|
it is part of the identity of many people. |
|
|
|
26:11.240 --> 26:12.600 |
|
If you go against this narrative, |
|
|
|
26:12.600 --> 26:14.400 |
|
it's like you're attacking the identity |
|
|
|
26:14.400 --> 26:15.560 |
|
of people who believe in it. |
|
|
|
26:15.560 --> 26:17.640 |
|
It's almost like saying God doesn't exist, |
|
|
|
26:17.640 --> 26:19.000 |
|
or something. |
|
|
|
26:19.000 --> 26:21.880 |
|
So you do get a lot of pushback |
|
|
|
26:21.880 --> 26:24.040 |
|
if you try to question these ideas. |
|
|
|
26:24.040 --> 26:26.520 |
|
First of all, I believe most people, |
|
|
|
26:26.520 --> 26:29.240 |
|
they might not be as eloquent or explicit as you're being, |
|
|
|
26:29.240 --> 26:30.920 |
|
but most people in computer science |
|
|
|
26:30.920 --> 26:33.000 |
|
are most people who actually have built |
|
|
|
26:33.000 --> 26:36.360 |
|
anything that you could call AI, quote, unquote, |
|
|
|
26:36.360 --> 26:38.080 |
|
would agree with you. |
|
|
|
26:38.080 --> 26:40.560 |
|
They might not be describing in the same kind of way. |
|
|
|
26:40.560 --> 26:43.960 |
|
It's more, so the pushback you're getting |
|
|
|
26:43.960 --> 26:48.080 |
|
is from people who get attached to the narrative |
|
|
|
26:48.080 --> 26:51.000 |
|
from, not from a place of science, |
|
|
|
26:51.000 --> 26:53.400 |
|
but from a place of imagination. |
|
|
|
26:53.400 --> 26:54.760 |
|
That's correct, that's correct. |
|
|
|
26:54.760 --> 26:56.920 |
|
So why do you think that's so appealing? |
|
|
|
26:56.920 --> 27:01.920 |
|
Because the usual dreams that people have |
|
|
|
27:02.120 --> 27:03.960 |
|
when you create a superintelligence system |
|
|
|
27:03.960 --> 27:05.120 |
|
past the singularity, |
|
|
|
27:05.120 --> 27:08.600 |
|
that what people imagine is somehow always destructive. |
|
|
|
27:09.440 --> 27:12.240 |
|
Do you have, if you were put on your psychology hat, |
|
|
|
27:12.240 --> 27:17.240 |
|
what's, why is it so appealing to imagine |
|
|
|
27:17.400 --> 27:20.760 |
|
the ways that all of human civilization will be destroyed? |
|
|
|
27:20.760 --> 27:22.080 |
|
I think it's a good story. |
|
|
|
27:22.080 --> 27:23.120 |
|
You know, it's a good story. |
|
|
|
27:23.120 --> 27:28.120 |
|
And very interestingly, it mirrors a religious stories, |
|
|
|
27:28.160 --> 27:30.560 |
|
right, religious mythology. |
|
|
|
27:30.560 --> 27:34.360 |
|
If you look at the mythology of most civilizations, |
|
|
|
27:34.360 --> 27:38.280 |
|
it's about the world being headed towards some final events |
|
|
|
27:38.280 --> 27:40.480 |
|
in which the world will be destroyed |
|
|
|
27:40.480 --> 27:42.800 |
|
and some new world order will arise |
|
|
|
27:42.800 --> 27:44.920 |
|
that will be mostly spiritual, |
|
|
|
27:44.920 --> 27:49.400 |
|
like the apocalypse followed by a paradise probably, right? |
|
|
|
27:49.400 --> 27:52.600 |
|
It's a very appealing story on a fundamental level. |
|
|
|
27:52.600 --> 27:54.560 |
|
And we all need stories. |
|
|
|
27:54.560 --> 27:58.160 |
|
We all need stories to structure the way we see the world, |
|
|
|
27:58.160 --> 27:59.960 |
|
especially at timescales |
|
|
|
27:59.960 --> 28:04.520 |
|
that are beyond our ability to make predictions, right? |
|
|
|
28:04.520 --> 28:08.840 |
|
So on a more serious non exponential explosion, |
|
|
|
28:08.840 --> 28:13.840 |
|
question, do you think there will be a time |
|
|
|
28:15.000 --> 28:19.800 |
|
when we'll create something like human level intelligence |
|
|
|
28:19.800 --> 28:23.800 |
|
or intelligent systems that will make you sit back |
|
|
|
28:23.800 --> 28:28.520 |
|
and be just surprised at damn how smart this thing is? |
|
|
|
28:28.520 --> 28:30.160 |
|
That doesn't require exponential growth |
|
|
|
28:30.160 --> 28:32.120 |
|
or an exponential improvement, |
|
|
|
28:32.120 --> 28:35.600 |
|
but what's your sense of the timeline and so on |
|
|
|
28:35.600 --> 28:40.600 |
|
that you'll be really surprised at certain capabilities? |
|
|
|
28:41.080 --> 28:42.560 |
|
And we'll talk about limitations and deep learning. |
|
|
|
28:42.560 --> 28:44.480 |
|
So do you think in your lifetime, |
|
|
|
28:44.480 --> 28:46.600 |
|
you'll be really damn surprised? |
|
|
|
28:46.600 --> 28:51.440 |
|
Around 2013, 2014, I was many times surprised |
|
|
|
28:51.440 --> 28:53.960 |
|
by the capabilities of deep learning actually. |
|
|
|
28:53.960 --> 28:55.920 |
|
That was before we had assessed exactly |
|
|
|
28:55.920 --> 28:57.880 |
|
what deep learning could do and could not do. |
|
|
|
28:57.880 --> 29:00.600 |
|
And it felt like a time of immense potential. |
|
|
|
29:00.600 --> 29:03.080 |
|
And then we started narrowing it down, |
|
|
|
29:03.080 --> 29:04.360 |
|
but I was very surprised. |
|
|
|
29:04.360 --> 29:07.120 |
|
I would say it has already happened. |
|
|
|
29:07.120 --> 29:10.800 |
|
Was there a moment, there must've been a day in there |
|
|
|
29:10.800 --> 29:14.360 |
|
where your surprise was almost bordering |
|
|
|
29:14.360 --> 29:19.360 |
|
on the belief of the narrative that we just discussed. |
|
|
|
29:19.440 --> 29:20.800 |
|
Was there a moment, |
|
|
|
29:20.800 --> 29:22.400 |
|
because you've written quite eloquently |
|
|
|
29:22.400 --> 29:23.960 |
|
about the limits of deep learning, |
|
|
|
29:23.960 --> 29:25.760 |
|
was there a moment that you thought |
|
|
|
29:25.760 --> 29:27.720 |
|
that maybe deep learning is limitless? |
|
|
|
29:30.000 --> 29:32.400 |
|
No, I don't think I've ever believed this. |
|
|
|
29:32.400 --> 29:35.560 |
|
What was really shocking is that it worked. |
|
|
|
29:35.560 --> 29:37.640 |
|
It worked at all, yeah. |
|
|
|
29:37.640 --> 29:40.520 |
|
But there's a big jump between being able |
|
|
|
29:40.520 --> 29:43.400 |
|
to do really good computer vision |
|
|
|
29:43.400 --> 29:44.920 |
|
and human level intelligence. |
|
|
|
29:44.920 --> 29:49.520 |
|
So I don't think at any point I wasn't under the impression |
|
|
|
29:49.520 --> 29:51.280 |
|
that the results we got in computer vision |
|
|
|
29:51.280 --> 29:54.080 |
|
meant that we were very close to human level intelligence. |
|
|
|
29:54.080 --> 29:56.040 |
|
I don't think we're very close to human level intelligence. |
|
|
|
29:56.040 --> 29:58.520 |
|
I do believe that there's no reason |
|
|
|
29:58.520 --> 30:01.760 |
|
why we won't achieve it at some point. |
|
|
|
30:01.760 --> 30:06.400 |
|
I also believe that it's the problem |
|
|
|
30:06.400 --> 30:08.560 |
|
with talking about human level intelligence |
|
|
|
30:08.560 --> 30:11.240 |
|
that implicitly you're considering |
|
|
|
30:11.240 --> 30:14.360 |
|
like an axis of intelligence with different levels, |
|
|
|
30:14.360 --> 30:16.720 |
|
but that's not really how intelligence works. |
|
|
|
30:16.720 --> 30:19.480 |
|
Intelligence is very multi dimensional. |
|
|
|
30:19.480 --> 30:22.480 |
|
And so there's the question of capabilities, |
|
|
|
30:22.480 --> 30:25.560 |
|
but there's also the question of being human like, |
|
|
|
30:25.560 --> 30:27.040 |
|
and it's two very different things. |
|
|
|
30:27.040 --> 30:28.280 |
|
Like you can build potentially |
|
|
|
30:28.280 --> 30:30.640 |
|
very advanced intelligent agents |
|
|
|
30:30.640 --> 30:32.640 |
|
that are not human like at all. |
|
|
|
30:32.640 --> 30:35.240 |
|
And you can also build very human like agents. |
|
|
|
30:35.240 --> 30:37.840 |
|
And these are two very different things, right? |
|
|
|
30:37.840 --> 30:38.760 |
|
Right. |
|
|
|
30:38.760 --> 30:42.240 |
|
Let's go from the philosophical to the practical. |
|
|
|
30:42.240 --> 30:44.240 |
|
Can you give me a history of Keras |
|
|
|
30:44.240 --> 30:46.440 |
|
and all the major deep learning frameworks |
|
|
|
30:46.440 --> 30:48.480 |
|
that you kind of remember in relation to Keras |
|
|
|
30:48.480 --> 30:52.040 |
|
and in general, TensorFlow, Theano, the old days. |
|
|
|
30:52.040 --> 30:55.400 |
|
Can you give a brief overview Wikipedia style history |
|
|
|
30:55.400 --> 30:59.120 |
|
and your role in it before we return to AGI discussions? |
|
|
|
30:59.120 --> 31:00.720 |
|
Yeah, that's a broad topic. |
|
|
|
31:00.720 --> 31:04.040 |
|
So I started working on Keras. |
|
|
|
31:04.920 --> 31:06.240 |
|
It was the name Keras at the time. |
|
|
|
31:06.240 --> 31:08.320 |
|
I actually picked the name like |
|
|
|
31:08.320 --> 31:10.200 |
|
just the day I was going to release it. |
|
|
|
31:10.200 --> 31:14.800 |
|
So I started working on it in February, 2015. |
|
|
|
31:14.800 --> 31:17.240 |
|
And so at the time there weren't too many people |
|
|
|
31:17.240 --> 31:20.320 |
|
working on deep learning, maybe like fewer than 10,000. |
|
|
|
31:20.320 --> 31:22.840 |
|
The software tooling was not really developed. |
|
|
|
31:25.320 --> 31:28.800 |
|
So the main deep learning library was Cafe, |
|
|
|
31:28.800 --> 31:30.840 |
|
which was mostly C++. |
|
|
|
31:30.840 --> 31:32.760 |
|
Why do you say Cafe was the main one? |
|
|
|
31:32.760 --> 31:36.000 |
|
Cafe was vastly more popular than Theano |
|
|
|
31:36.000 --> 31:38.920 |
|
in late 2014, early 2015. |
|
|
|
31:38.920 --> 31:42.400 |
|
Cafe was the one library that everyone was using |
|
|
|
31:42.400 --> 31:43.400 |
|
for computer vision. |
|
|
|
31:43.400 --> 31:46.120 |
|
And computer vision was the most popular problem |
|
|
|
31:46.120 --> 31:46.960 |
|
in deep learning at the time. |
|
|
|
31:46.960 --> 31:47.800 |
|
Absolutely. |
|
|
|
31:47.800 --> 31:50.440 |
|
Like ConvNets was like the subfield of deep learning |
|
|
|
31:50.440 --> 31:53.160 |
|
that everyone was working on. |
|
|
|
31:53.160 --> 31:57.680 |
|
So myself, so in late 2014, |
|
|
|
31:57.680 --> 32:00.600 |
|
I was actually interested in RNNs, |
|
|
|
32:00.600 --> 32:01.760 |
|
in recurrent neural networks, |
|
|
|
32:01.760 --> 32:05.800 |
|
which was a very niche topic at the time, right? |
|
|
|
32:05.800 --> 32:08.640 |
|
It really took off around 2016. |
|
|
|
32:08.640 --> 32:11.080 |
|
And so I was looking for good tools. |
|
|
|
32:11.080 --> 32:14.800 |
|
I had used Torch 7, I had used Theano, |
|
|
|
32:14.800 --> 32:17.640 |
|
used Theano a lot in Kaggle competitions. |
|
|
|
32:19.320 --> 32:20.840 |
|
I had used Cafe. |
|
|
|
32:20.840 --> 32:25.840 |
|
And there was no like good solution for RNNs at the time. |
|
|
|
32:25.840 --> 32:28.640 |
|
Like there was no reusable open source implementation |
|
|
|
32:28.640 --> 32:30.000 |
|
of an LSTM, for instance. |
|
|
|
32:30.000 --> 32:32.920 |
|
So I decided to build my own. |
|
|
|
32:32.920 --> 32:35.440 |
|
And at first, the pitch for that was, |
|
|
|
32:35.440 --> 32:39.960 |
|
it was gonna be mostly around LSTM recurrent neural networks. |
|
|
|
32:39.960 --> 32:41.360 |
|
It was gonna be in Python. |
|
|
|
32:42.280 --> 32:44.280 |
|
An important decision at the time |
|
|
|
32:44.280 --> 32:45.440 |
|
that was kind of not obvious |
|
|
|
32:45.440 --> 32:50.360 |
|
is that the models would be defined via Python code, |
|
|
|
32:50.360 --> 32:54.400 |
|
which was kind of like going against the mainstream |
|
|
|
32:54.400 --> 32:58.000 |
|
at the time because Cafe, Pylon 2, and so on, |
|
|
|
32:58.000 --> 33:00.600 |
|
like all the big libraries were actually going |
|
|
|
33:00.600 --> 33:03.520 |
|
with the approach of setting configuration files |
|
|
|
33:03.520 --> 33:05.560 |
|
in YAML to define models. |
|
|
|
33:05.560 --> 33:08.840 |
|
So some libraries were using code to define models, |
|
|
|
33:08.840 --> 33:12.280 |
|
like Torch 7, obviously, but that was not Python. |
|
|
|
33:12.280 --> 33:16.680 |
|
Lasagne was like a Theano based very early library |
|
|
|
33:16.680 --> 33:18.640 |
|
that was, I think, developed, I don't remember exactly, |
|
|
|
33:18.640 --> 33:20.240 |
|
probably late 2014. |
|
|
|
33:20.240 --> 33:21.200 |
|
It's Python as well. |
|
|
|
33:21.200 --> 33:22.040 |
|
It's Python as well. |
|
|
|
33:22.040 --> 33:24.320 |
|
It was like on top of Theano. |
|
|
|
33:24.320 --> 33:28.320 |
|
And so I started working on something |
|
|
|
33:29.480 --> 33:32.520 |
|
and the value proposition at the time was that |
|
|
|
33:32.520 --> 33:36.240 |
|
not only what I think was the first |
|
|
|
33:36.240 --> 33:38.800 |
|
reusable open source implementation of LSTM, |
|
|
|
33:40.400 --> 33:44.440 |
|
you could combine RNNs and covenants |
|
|
|
33:44.440 --> 33:45.440 |
|
with the same library, |
|
|
|
33:45.440 --> 33:46.920 |
|
which is not really possible before, |
|
|
|
33:46.920 --> 33:49.080 |
|
like Cafe was only doing covenants. |
|
|
|
33:50.440 --> 33:52.560 |
|
And it was kind of easy to use |
|
|
|
33:52.560 --> 33:54.440 |
|
because, so before I was using Theano, |
|
|
|
33:54.440 --> 33:55.680 |
|
I was actually using scikitlin |
|
|
|
33:55.680 --> 33:58.320 |
|
and I loved scikitlin for its usability. |
|
|
|
33:58.320 --> 34:01.560 |
|
So I drew a lot of inspiration from scikitlin |
|
|
|
34:01.560 --> 34:02.400 |
|
when I made Keras. |
|
|
|
34:02.400 --> 34:05.600 |
|
It's almost like scikitlin for neural networks. |
|
|
|
34:05.600 --> 34:06.680 |
|
The fit function. |
|
|
|
34:06.680 --> 34:07.960 |
|
Exactly, the fit function, |
|
|
|
34:07.960 --> 34:10.800 |
|
like reducing a complex string loop |
|
|
|
34:10.800 --> 34:12.880 |
|
to a single function call, right? |
|
|
|
34:12.880 --> 34:14.880 |
|
And of course, some people will say, |
|
|
|
34:14.880 --> 34:16.320 |
|
this is hiding a lot of details, |
|
|
|
34:16.320 --> 34:18.680 |
|
but that's exactly the point, right? |
|
|
|
34:18.680 --> 34:20.280 |
|
The magic is the point. |
|
|
|
34:20.280 --> 34:22.680 |
|
So it's magical, but in a good way. |
|
|
|
34:22.680 --> 34:24.960 |
|
It's magical in the sense that it's delightful. |
|
|
|
34:24.960 --> 34:26.160 |
|
Yeah, yeah. |
|
|
|
34:26.160 --> 34:27.640 |
|
I'm actually quite surprised. |
|
|
|
34:27.640 --> 34:29.600 |
|
I didn't know that it was born out of desire |
|
|
|
34:29.600 --> 34:32.480 |
|
to implement RNNs and LSTMs. |
|
|
|
34:32.480 --> 34:33.320 |
|
It was. |
|
|
|
34:33.320 --> 34:34.160 |
|
That's fascinating. |
|
|
|
34:34.160 --> 34:36.040 |
|
So you were actually one of the first people |
|
|
|
34:36.040 --> 34:37.960 |
|
to really try to attempt |
|
|
|
34:37.960 --> 34:41.000 |
|
to get the major architectures together. |
|
|
|
34:41.000 --> 34:42.760 |
|
And it's also interesting. |
|
|
|
34:42.760 --> 34:45.160 |
|
You made me realize that that was a design decision at all |
|
|
|
34:45.160 --> 34:47.360 |
|
is defining the model and code. |
|
|
|
34:47.360 --> 34:49.920 |
|
Just, I'm putting myself in your shoes, |
|
|
|
34:49.920 --> 34:53.200 |
|
whether the YAML, especially if cafe was the most popular. |
|
|
|
34:53.200 --> 34:54.720 |
|
It was the most popular by far. |
|
|
|
34:54.720 --> 34:58.480 |
|
If I was, if I were, yeah, I don't, |
|
|
|
34:58.480 --> 34:59.560 |
|
I didn't like the YAML thing, |
|
|
|
34:59.560 --> 35:02.840 |
|
but it makes more sense that you will put |
|
|
|
35:02.840 --> 35:05.720 |
|
in a configuration file, the definition of a model. |
|
|
|
35:05.720 --> 35:07.200 |
|
That's an interesting gutsy move |
|
|
|
35:07.200 --> 35:10.040 |
|
to stick with defining it in code. |
|
|
|
35:10.040 --> 35:11.600 |
|
Just if you look back. |
|
|
|
35:11.600 --> 35:13.480 |
|
Other libraries were doing it as well, |
|
|
|
35:13.480 --> 35:16.320 |
|
but it was definitely the more niche option. |
|
|
|
35:16.320 --> 35:17.160 |
|
Yeah. |
|
|
|
35:17.160 --> 35:18.360 |
|
Okay, Keras and then. |
|
|
|
35:18.360 --> 35:21.520 |
|
So I released Keras in March, 2015, |
|
|
|
35:21.520 --> 35:24.160 |
|
and it got users pretty much from the start. |
|
|
|
35:24.160 --> 35:25.800 |
|
So the deep learning community was very, very small |
|
|
|
35:25.800 --> 35:27.240 |
|
at the time. |
|
|
|
35:27.240 --> 35:30.600 |
|
Lots of people were starting to be interested in LSTM. |
|
|
|
35:30.600 --> 35:32.440 |
|
So it was gonna release it at the right time |
|
|
|
35:32.440 --> 35:35.560 |
|
because it was offering an easy to use LSTM implementation. |
|
|
|
35:35.560 --> 35:37.680 |
|
Exactly at the time where lots of people started |
|
|
|
35:37.680 --> 35:42.280 |
|
to be intrigued by the capabilities of RNN, RNNs for NLP. |
|
|
|
35:42.280 --> 35:43.920 |
|
So it grew from there. |
|
|
|
35:43.920 --> 35:48.920 |
|
Then I joined Google about six months later, |
|
|
|
35:51.480 --> 35:54.920 |
|
and that was actually completely unrelated to Keras. |
|
|
|
35:54.920 --> 35:57.080 |
|
So I actually joined a research team |
|
|
|
35:57.080 --> 35:59.520 |
|
working on image classification, |
|
|
|
35:59.520 --> 36:00.680 |
|
mostly like computer vision. |
|
|
|
36:00.680 --> 36:02.320 |
|
So I was doing computer vision research |
|
|
|
36:02.320 --> 36:03.640 |
|
at Google initially. |
|
|
|
36:03.640 --> 36:05.520 |
|
And immediately when I joined Google, |
|
|
|
36:05.520 --> 36:10.520 |
|
I was exposed to the early internal version of TensorFlow. |
|
|
|
36:10.520 --> 36:13.920 |
|
And the way it appeared to me at the time, |
|
|
|
36:13.920 --> 36:15.720 |
|
and it was definitely the way it was at the time |
|
|
|
36:15.720 --> 36:20.760 |
|
is that this was an improved version of Theano. |
|
|
|
36:20.760 --> 36:24.720 |
|
So I immediately knew I had to port Keras |
|
|
|
36:24.720 --> 36:26.800 |
|
to this new TensorFlow thing. |
|
|
|
36:26.800 --> 36:29.800 |
|
And I was actually very busy as a noobler, |
|
|
|
36:29.800 --> 36:30.720 |
|
as a new Googler. |
|
|
|
36:31.600 --> 36:34.520 |
|
So I had not time to work on that. |
|
|
|
36:34.520 --> 36:38.680 |
|
But then in November, I think it was November, 2015, |
|
|
|
36:38.680 --> 36:41.240 |
|
TensorFlow got released. |
|
|
|
36:41.240 --> 36:44.560 |
|
And it was kind of like my wake up call |
|
|
|
36:44.560 --> 36:47.320 |
|
that, hey, I had to actually go and make it happen. |
|
|
|
36:47.320 --> 36:52.200 |
|
So in December, I ported Keras to run on top of TensorFlow, |
|
|
|
36:52.200 --> 36:53.320 |
|
but it was not exactly a port. |
|
|
|
36:53.320 --> 36:55.280 |
|
It was more like a refactoring |
|
|
|
36:55.280 --> 36:57.920 |
|
where I was abstracting away |
|
|
|
36:57.920 --> 37:00.480 |
|
all the backend functionality into one module |
|
|
|
37:00.480 --> 37:02.320 |
|
so that the same code base |
|
|
|
37:02.320 --> 37:05.080 |
|
could run on top of multiple backends. |
|
|
|
37:05.080 --> 37:07.440 |
|
So on top of TensorFlow or Theano. |
|
|
|
37:07.440 --> 37:09.760 |
|
And for the next year, |
|
|
|
37:09.760 --> 37:14.760 |
|
Theano stayed as the default option. |
|
|
|
37:15.400 --> 37:20.400 |
|
It was easier to use, somewhat less buggy. |
|
|
|
37:20.640 --> 37:23.360 |
|
It was much faster, especially when it came to audience. |
|
|
|
37:23.360 --> 37:26.360 |
|
But eventually, TensorFlow overtook it. |
|
|
|
37:27.480 --> 37:30.200 |
|
And TensorFlow, the early TensorFlow, |
|
|
|
37:30.200 --> 37:33.960 |
|
has similar architectural decisions as Theano, right? |
|
|
|
37:33.960 --> 37:37.440 |
|
So it was a natural transition. |
|
|
|
37:37.440 --> 37:38.320 |
|
Yeah, absolutely. |
|
|
|
37:38.320 --> 37:42.960 |
|
So what, I mean, that still Keras is a side, |
|
|
|
37:42.960 --> 37:45.280 |
|
almost fun project, right? |
|
|
|
37:45.280 --> 37:49.040 |
|
Yeah, so it was not my job assignment. |
|
|
|
37:49.040 --> 37:50.360 |
|
It was not. |
|
|
|
37:50.360 --> 37:52.240 |
|
I was doing it on the side. |
|
|
|
37:52.240 --> 37:55.840 |
|
And even though it grew to have a lot of users |
|
|
|
37:55.840 --> 37:59.600 |
|
for a deep learning library at the time, like Stroud 2016, |
|
|
|
37:59.600 --> 38:02.480 |
|
but I wasn't doing it as my main job. |
|
|
|
38:02.480 --> 38:04.760 |
|
So things started changing in, |
|
|
|
38:04.760 --> 38:09.760 |
|
I think it must have been maybe October, 2016. |
|
|
|
38:10.200 --> 38:11.320 |
|
So one year later. |
|
|
|
38:12.360 --> 38:15.240 |
|
So Rajat, who was the lead on TensorFlow, |
|
|
|
38:15.240 --> 38:19.240 |
|
basically showed up one day in our building |
|
|
|
38:19.240 --> 38:20.080 |
|
where I was doing like, |
|
|
|
38:20.080 --> 38:21.640 |
|
so I was doing research and things like, |
|
|
|
38:21.640 --> 38:24.640 |
|
so I did a lot of computer vision research, |
|
|
|
38:24.640 --> 38:27.560 |
|
also collaborations with Christian Zighetti |
|
|
|
38:27.560 --> 38:29.640 |
|
and deep learning for theorem proving. |
|
|
|
38:29.640 --> 38:32.920 |
|
It was a really interesting research topic. |
|
|
|
38:34.520 --> 38:37.640 |
|
And so Rajat was saying, |
|
|
|
38:37.640 --> 38:41.040 |
|
hey, we saw Keras, we like it. |
|
|
|
38:41.040 --> 38:42.440 |
|
We saw that you're at Google. |
|
|
|
38:42.440 --> 38:45.280 |
|
Why don't you come over for like a quarter |
|
|
|
38:45.280 --> 38:47.280 |
|
and work with us? |
|
|
|
38:47.280 --> 38:49.240 |
|
And I was like, yeah, that sounds like a great opportunity. |
|
|
|
38:49.240 --> 38:50.400 |
|
Let's do it. |
|
|
|
38:50.400 --> 38:55.400 |
|
And so I started working on integrating the Keras API |
|
|
|
38:55.720 --> 38:57.320 |
|
into TensorFlow more tightly. |
|
|
|
38:57.320 --> 39:02.320 |
|
So what followed up is a sort of like temporary |
|
|
|
39:02.640 --> 39:05.480 |
|
TensorFlow only version of Keras |
|
|
|
39:05.480 --> 39:09.320 |
|
that was in TensorFlow.com Trib for a while. |
|
|
|
39:09.320 --> 39:12.200 |
|
And finally moved to TensorFlow Core. |
|
|
|
39:12.200 --> 39:15.360 |
|
And I've never actually gotten back |
|
|
|
39:15.360 --> 39:17.600 |
|
to my old team doing research. |
|
|
|
39:17.600 --> 39:22.320 |
|
Well, it's kind of funny that somebody like you |
|
|
|
39:22.320 --> 39:27.320 |
|
who dreams of, or at least sees the power of AI systems |
|
|
|
39:28.960 --> 39:31.680 |
|
that reason and theorem proving we'll talk about |
|
|
|
39:31.680 --> 39:36.520 |
|
has also created a system that makes the most basic |
|
|
|
39:36.520 --> 39:40.400 |
|
kind of Lego building that is deep learning |
|
|
|
39:40.400 --> 39:42.640 |
|
super accessible, super easy. |
|
|
|
39:42.640 --> 39:43.800 |
|
So beautifully so. |
|
|
|
39:43.800 --> 39:47.720 |
|
It's a funny irony that you're both, |
|
|
|
39:47.720 --> 39:49.120 |
|
you're responsible for both things, |
|
|
|
39:49.120 --> 39:54.000 |
|
but so TensorFlow 2.0 is kind of, there's a sprint. |
|
|
|
39:54.000 --> 39:55.080 |
|
I don't know how long it'll take, |
|
|
|
39:55.080 --> 39:56.960 |
|
but there's a sprint towards the finish. |
|
|
|
39:56.960 --> 40:01.040 |
|
What do you look, what are you working on these days? |
|
|
|
40:01.040 --> 40:02.160 |
|
What are you excited about? |
|
|
|
40:02.160 --> 40:04.280 |
|
What are you excited about in 2.0? |
|
|
|
40:04.280 --> 40:05.760 |
|
I mean, eager execution. |
|
|
|
40:05.760 --> 40:08.440 |
|
There's so many things that just make it a lot easier |
|
|
|
40:08.440 --> 40:09.760 |
|
to work. |
|
|
|
40:09.760 --> 40:13.640 |
|
What are you excited about and what's also really hard? |
|
|
|
40:13.640 --> 40:15.800 |
|
What are the problems you have to kind of solve? |
|
|
|
40:15.800 --> 40:19.080 |
|
So I've spent the past year and a half working on |
|
|
|
40:19.080 --> 40:22.920 |
|
TensorFlow 2.0 and it's been a long journey. |
|
|
|
40:22.920 --> 40:25.080 |
|
I'm actually extremely excited about it. |
|
|
|
40:25.080 --> 40:26.440 |
|
I think it's a great product. |
|
|
|
40:26.440 --> 40:29.360 |
|
It's a delightful product compared to TensorFlow 1.0. |
|
|
|
40:29.360 --> 40:31.440 |
|
We've made huge progress. |
|
|
|
40:32.640 --> 40:37.400 |
|
So on the Keras side, what I'm really excited about is that, |
|
|
|
40:37.400 --> 40:42.400 |
|
so previously Keras has been this very easy to use |
|
|
|
40:42.400 --> 40:45.840 |
|
high level interface to do deep learning. |
|
|
|
40:45.840 --> 40:47.280 |
|
But if you wanted to, |
|
|
|
40:50.520 --> 40:53.040 |
|
if you wanted a lot of flexibility, |
|
|
|
40:53.040 --> 40:57.520 |
|
the Keras framework was probably not the optimal way |
|
|
|
40:57.520 --> 40:59.760 |
|
to do things compared to just writing everything |
|
|
|
40:59.760 --> 41:00.600 |
|
from scratch. |
|
|
|
41:01.800 --> 41:04.680 |
|
So in some way, the framework was getting in the way. |
|
|
|
41:04.680 --> 41:07.960 |
|
And in TensorFlow 2.0, you don't have this at all, actually. |
|
|
|
41:07.960 --> 41:11.040 |
|
You have the usability of the high level interface, |
|
|
|
41:11.040 --> 41:14.480 |
|
but you have the flexibility of this lower level interface. |
|
|
|
41:14.480 --> 41:16.800 |
|
And you have this spectrum of workflows |
|
|
|
41:16.800 --> 41:21.560 |
|
where you can get more or less usability |
|
|
|
41:21.560 --> 41:26.560 |
|
and flexibility trade offs depending on your needs, right? |
|
|
|
41:26.640 --> 41:29.680 |
|
You can write everything from scratch |
|
|
|
41:29.680 --> 41:32.320 |
|
and you get a lot of help doing so |
|
|
|
41:32.320 --> 41:36.400 |
|
by subclassing models and writing some train loops |
|
|
|
41:36.400 --> 41:38.200 |
|
using ego execution. |
|
|
|
41:38.200 --> 41:40.160 |
|
It's very flexible, it's very easy to debug, |
|
|
|
41:40.160 --> 41:41.400 |
|
it's very powerful. |
|
|
|
41:42.280 --> 41:45.000 |
|
But all of this integrates seamlessly |
|
|
|
41:45.000 --> 41:49.440 |
|
with higher level features up to the classic Keras workflows, |
|
|
|
41:49.440 --> 41:51.560 |
|
which are very scikit learn like |
|
|
|
41:51.560 --> 41:56.040 |
|
and are ideal for a data scientist, |
|
|
|
41:56.040 --> 41:58.240 |
|
machine learning engineer type of profile. |
|
|
|
41:58.240 --> 42:00.840 |
|
So now you can have the same framework |
|
|
|
42:00.840 --> 42:02.880 |
|
offering the same set of APIs |
|
|
|
42:02.880 --> 42:05.000 |
|
that enable a spectrum of workflows |
|
|
|
42:05.000 --> 42:08.560 |
|
that are more or less low level, more or less high level |
|
|
|
42:08.560 --> 42:13.520 |
|
that are suitable for profiles ranging from researchers |
|
|
|
42:13.520 --> 42:15.560 |
|
to data scientists and everything in between. |
|
|
|
42:15.560 --> 42:16.960 |
|
Yeah, so that's super exciting. |
|
|
|
42:16.960 --> 42:18.400 |
|
I mean, it's not just that, |
|
|
|
42:18.400 --> 42:21.680 |
|
it's connected to all kinds of tooling. |
|
|
|
42:21.680 --> 42:24.520 |
|
You can go on mobile, you can go with TensorFlow Lite, |
|
|
|
42:24.520 --> 42:27.240 |
|
you can go in the cloud or serving and so on. |
|
|
|
42:27.240 --> 42:28.960 |
|
It all is connected together. |
|
|
|
42:28.960 --> 42:31.880 |
|
Now some of the best software written ever |
|
|
|
42:31.880 --> 42:36.880 |
|
is often done by one person, sometimes two. |
|
|
|
42:36.880 --> 42:40.800 |
|
So with a Google, you're now seeing sort of Keras |
|
|
|
42:40.800 --> 42:42.840 |
|
having to be integrated in TensorFlow, |
|
|
|
42:42.840 --> 42:46.800 |
|
I'm sure has a ton of engineers working on. |
|
|
|
42:46.800 --> 42:51.040 |
|
And there's, I'm sure a lot of tricky design decisions |
|
|
|
42:51.040 --> 42:52.200 |
|
to be made. |
|
|
|
42:52.200 --> 42:54.440 |
|
How does that process usually happen |
|
|
|
42:54.440 --> 42:56.800 |
|
from at least your perspective? |
|
|
|
42:56.800 --> 42:59.800 |
|
What are the debates like? |
|
|
|
43:00.720 --> 43:04.200 |
|
Is there a lot of thinking, |
|
|
|
43:04.200 --> 43:06.880 |
|
considering different options and so on? |
|
|
|
43:06.880 --> 43:08.160 |
|
Yes. |
|
|
|
43:08.160 --> 43:12.640 |
|
So a lot of the time I spend at Google |
|
|
|
43:12.640 --> 43:17.280 |
|
is actually discussing design discussions, right? |
|
|
|
43:17.280 --> 43:20.480 |
|
Writing design docs, participating in design review meetings |
|
|
|
43:20.480 --> 43:22.080 |
|
and so on. |
|
|
|
43:22.080 --> 43:25.240 |
|
This is as important as actually writing a code. |
|
|
|
43:25.240 --> 43:26.080 |
|
Right. |
|
|
|
43:26.080 --> 43:28.120 |
|
So there's a lot of thought, there's a lot of thought |
|
|
|
43:28.120 --> 43:32.280 |
|
and a lot of care that is taken |
|
|
|
43:32.280 --> 43:34.160 |
|
in coming up with these decisions |
|
|
|
43:34.160 --> 43:37.160 |
|
and taking into account all of our users |
|
|
|
43:37.160 --> 43:40.680 |
|
because TensorFlow has this extremely diverse user base, |
|
|
|
43:40.680 --> 43:41.520 |
|
right? |
|
|
|
43:41.520 --> 43:43.120 |
|
It's not like just one user segment |
|
|
|
43:43.120 --> 43:45.480 |
|
where everyone has the same needs. |
|
|
|
43:45.480 --> 43:47.640 |
|
We have small scale production users, |
|
|
|
43:47.640 --> 43:49.520 |
|
large scale production users. |
|
|
|
43:49.520 --> 43:52.800 |
|
We have startups, we have researchers, |
|
|
|
43:53.720 --> 43:55.080 |
|
you know, it's all over the place. |
|
|
|
43:55.080 --> 43:57.560 |
|
And we have to cater to all of their needs. |
|
|
|
43:57.560 --> 44:00.040 |
|
If I just look at the standard debates |
|
|
|
44:00.040 --> 44:04.000 |
|
of C++ or Python, there's some heated debates. |
|
|
|
44:04.000 --> 44:06.000 |
|
Do you have those at Google? |
|
|
|
44:06.000 --> 44:08.080 |
|
I mean, they're not heated in terms of emotionally, |
|
|
|
44:08.080 --> 44:10.800 |
|
but there's probably multiple ways to do it, right? |
|
|
|
44:10.800 --> 44:14.040 |
|
So how do you arrive through those design meetings |
|
|
|
44:14.040 --> 44:15.440 |
|
at the best way to do it? |
|
|
|
44:15.440 --> 44:19.280 |
|
Especially in deep learning where the field is evolving |
|
|
|
44:19.280 --> 44:20.880 |
|
as you're doing it. |
|
|
|
44:21.880 --> 44:23.600 |
|
Is there some magic to it? |
|
|
|
44:23.600 --> 44:26.240 |
|
Is there some magic to the process? |
|
|
|
44:26.240 --> 44:28.280 |
|
I don't know if there's magic to the process, |
|
|
|
44:28.280 --> 44:30.640 |
|
but there definitely is a process. |
|
|
|
44:30.640 --> 44:33.760 |
|
So making design decisions |
|
|
|
44:33.760 --> 44:36.080 |
|
is about satisfying a set of constraints, |
|
|
|
44:36.080 --> 44:39.920 |
|
but also trying to do so in the simplest way possible, |
|
|
|
44:39.920 --> 44:42.240 |
|
because this is what can be maintained, |
|
|
|
44:42.240 --> 44:44.920 |
|
this is what can be expanded in the future. |
|
|
|
44:44.920 --> 44:49.120 |
|
So you don't want to naively satisfy the constraints |
|
|
|
44:49.120 --> 44:51.880 |
|
by just, you know, for each capability you need available, |
|
|
|
44:51.880 --> 44:53.960 |
|
you're gonna come up with one argument in your API |
|
|
|
44:53.960 --> 44:54.800 |
|
and so on. |
|
|
|
44:54.800 --> 44:59.800 |
|
You want to design APIs that are modular and hierarchical |
|
|
|
45:00.640 --> 45:04.080 |
|
so that they have an API surface |
|
|
|
45:04.080 --> 45:07.040 |
|
that is as small as possible, right? |
|
|
|
45:07.040 --> 45:11.640 |
|
And you want this modular hierarchical architecture |
|
|
|
45:11.640 --> 45:14.560 |
|
to reflect the way that domain experts |
|
|
|
45:14.560 --> 45:16.400 |
|
think about the problem. |
|
|
|
45:16.400 --> 45:17.880 |
|
Because as a domain expert, |
|
|
|
45:17.880 --> 45:19.840 |
|
when you are reading about a new API, |
|
|
|
45:19.840 --> 45:24.760 |
|
you're reading a tutorial or some docs pages, |
|
|
|
45:24.760 --> 45:28.200 |
|
you already have a way that you're thinking about the problem. |
|
|
|
45:28.200 --> 45:32.320 |
|
You already have like certain concepts in mind |
|
|
|
45:32.320 --> 45:35.680 |
|
and you're thinking about how they relate together. |
|
|
|
45:35.680 --> 45:37.200 |
|
And when you're reading docs, |
|
|
|
45:37.200 --> 45:40.280 |
|
you're trying to build as quickly as possible |
|
|
|
45:40.280 --> 45:45.280 |
|
a mapping between the concepts featured in your API |
|
|
|
45:45.280 --> 45:46.800 |
|
and the concepts in your mind. |
|
|
|
45:46.800 --> 45:48.880 |
|
So you're trying to map your mental model |
|
|
|
45:48.880 --> 45:53.600 |
|
as a domain expert to the way things work in the API. |
|
|
|
45:53.600 --> 45:57.040 |
|
So you need an API and an underlying implementation |
|
|
|
45:57.040 --> 46:00.120 |
|
that are reflecting the way people think about these things. |
|
|
|
46:00.120 --> 46:02.880 |
|
So in minimizing the time it takes to do the mapping. |
|
|
|
46:02.880 --> 46:04.680 |
|
Yes, minimizing the time, |
|
|
|
46:04.680 --> 46:06.560 |
|
the cognitive load there is |
|
|
|
46:06.560 --> 46:10.920 |
|
in ingesting this new knowledge about your API. |
|
|
|
46:10.920 --> 46:13.160 |
|
An API should not be self referential |
|
|
|
46:13.160 --> 46:15.520 |
|
or referring to implementation details. |
|
|
|
46:15.520 --> 46:19.160 |
|
It should only be referring to domain specific concepts |
|
|
|
46:19.160 --> 46:21.360 |
|
that people already understand. |
|
|
|
46:23.240 --> 46:24.480 |
|
Brilliant. |
|
|
|
46:24.480 --> 46:27.560 |
|
So what's the future of Keras and TensorFlow look like? |
|
|
|
46:27.560 --> 46:29.640 |
|
What does TensorFlow 3.0 look like? |
|
|
|
46:30.600 --> 46:33.720 |
|
So that's kind of too far in the future for me to answer, |
|
|
|
46:33.720 --> 46:37.800 |
|
especially since I'm not even the one making these decisions. |
|
|
|
46:37.800 --> 46:39.080 |
|
Okay. |
|
|
|
46:39.080 --> 46:41.240 |
|
But so from my perspective, |
|
|
|
46:41.240 --> 46:43.200 |
|
which is just one perspective |
|
|
|
46:43.200 --> 46:46.040 |
|
among many different perspectives on the TensorFlow team, |
|
|
|
46:47.200 --> 46:52.200 |
|
I'm really excited by developing even higher level APIs, |
|
|
|
46:52.360 --> 46:53.560 |
|
higher level than Keras. |
|
|
|
46:53.560 --> 46:56.480 |
|
I'm really excited by hyperparameter tuning, |
|
|
|
46:56.480 --> 46:59.240 |
|
by automated machine learning, AutoML. |
|
|
|
47:01.120 --> 47:03.200 |
|
I think the future is not just, you know, |
|
|
|
47:03.200 --> 47:07.600 |
|
defining a model like you were assembling Lego blocks |
|
|
|
47:07.600 --> 47:09.200 |
|
and then collect fit on it. |
|
|
|
47:09.200 --> 47:13.680 |
|
It's more like an automagical model |
|
|
|
47:13.680 --> 47:16.080 |
|
that would just look at your data |
|
|
|
47:16.080 --> 47:19.040 |
|
and optimize the objective you're after, right? |
|
|
|
47:19.040 --> 47:23.040 |
|
So that's what I'm looking into. |
|
|
|
47:23.040 --> 47:26.480 |
|
Yeah, so you put the baby into a room with the problem |
|
|
|
47:26.480 --> 47:28.760 |
|
and come back a few hours later |
|
|
|
47:28.760 --> 47:30.960 |
|
with a fully solved problem. |
|
|
|
47:30.960 --> 47:33.560 |
|
Exactly, it's not like a box of Legos. |
|
|
|
47:33.560 --> 47:35.920 |
|
It's more like the combination of a kid |
|
|
|
47:35.920 --> 47:38.800 |
|
that's really good at Legos and a box of Legos. |
|
|
|
47:38.800 --> 47:41.520 |
|
It's just building the thing on its own. |
|
|
|
47:41.520 --> 47:42.680 |
|
Very nice. |
|
|
|
47:42.680 --> 47:44.160 |
|
So that's an exciting future. |
|
|
|
47:44.160 --> 47:46.080 |
|
I think there's a huge amount of applications |
|
|
|
47:46.080 --> 47:48.560 |
|
and revolutions to be had |
|
|
|
47:49.920 --> 47:52.640 |
|
under the constraints of the discussion we previously had. |
|
|
|
47:52.640 --> 47:57.480 |
|
But what do you think of the current limits of deep learning? |
|
|
|
47:57.480 --> 48:02.480 |
|
If we look specifically at these function approximators |
|
|
|
48:03.840 --> 48:06.160 |
|
that tries to generalize from data. |
|
|
|
48:06.160 --> 48:10.160 |
|
You've talked about local versus extreme generalization. |
|
|
|
48:11.120 --> 48:13.280 |
|
You mentioned that neural networks don't generalize well |
|
|
|
48:13.280 --> 48:14.560 |
|
and humans do. |
|
|
|
48:14.560 --> 48:15.760 |
|
So there's this gap. |
|
|
|
48:17.640 --> 48:20.840 |
|
And you've also mentioned that extreme generalization |
|
|
|
48:20.840 --> 48:23.960 |
|
requires something like reasoning to fill those gaps. |
|
|
|
48:23.960 --> 48:27.560 |
|
So how can we start trying to build systems like that? |
|
|
|
48:27.560 --> 48:30.600 |
|
Right, yeah, so this is by design, right? |
|
|
|
48:30.600 --> 48:37.080 |
|
Deep learning models are like huge parametric models, |
|
|
|
48:37.080 --> 48:39.280 |
|
differentiable, so continuous, |
|
|
|
48:39.280 --> 48:42.680 |
|
that go from an input space to an output space. |
|
|
|
48:42.680 --> 48:44.120 |
|
And they're trained with gradient descent. |
|
|
|
48:44.120 --> 48:47.160 |
|
So they're trained pretty much point by point. |
|
|
|
48:47.160 --> 48:50.520 |
|
They are learning a continuous geometric morphing |
|
|
|
48:50.520 --> 48:55.320 |
|
from an input vector space to an output vector space. |
|
|
|
48:55.320 --> 48:58.960 |
|
And because this is done point by point, |
|
|
|
48:58.960 --> 49:02.200 |
|
a deep neural network can only make sense |
|
|
|
49:02.200 --> 49:05.880 |
|
of points in experience space that are very close |
|
|
|
49:05.880 --> 49:08.520 |
|
to things that it has already seen in string data. |
|
|
|
49:08.520 --> 49:12.520 |
|
At best, it can do interpolation across points. |
|
|
|
49:13.840 --> 49:17.360 |
|
But that means in order to train your network, |
|
|
|
49:17.360 --> 49:21.680 |
|
you need a dense sampling of the input cross output space, |
|
|
|
49:22.880 --> 49:25.240 |
|
almost a point by point sampling, |
|
|
|
49:25.240 --> 49:27.160 |
|
which can be very expensive if you're dealing |
|
|
|
49:27.160 --> 49:29.320 |
|
with complex real world problems, |
|
|
|
49:29.320 --> 49:33.240 |
|
like autonomous driving, for instance, or robotics. |
|
|
|
49:33.240 --> 49:36.000 |
|
It's doable if you're looking at the subset |
|
|
|
49:36.000 --> 49:37.120 |
|
of the visual space. |
|
|
|
49:37.120 --> 49:38.800 |
|
But even then, it's still fairly expensive. |
|
|
|
49:38.800 --> 49:40.920 |
|
You still need millions of examples. |
|
|
|
49:40.920 --> 49:44.240 |
|
And it's only going to be able to make sense of things |
|
|
|
49:44.240 --> 49:46.880 |
|
that are very close to what it has seen before. |
|
|
|
49:46.880 --> 49:49.160 |
|
And in contrast to that, well, of course, |
|
|
|
49:49.160 --> 49:50.160 |
|
you have human intelligence. |
|
|
|
49:50.160 --> 49:53.240 |
|
But even if you're not looking at human intelligence, |
|
|
|
49:53.240 --> 49:56.800 |
|
you can look at very simple rules, algorithms. |
|
|
|
49:56.800 --> 49:58.080 |
|
If you have a symbolic rule, |
|
|
|
49:58.080 --> 50:03.080 |
|
it can actually apply to a very, very large set of inputs |
|
|
|
50:03.120 --> 50:04.880 |
|
because it is abstract. |
|
|
|
50:04.880 --> 50:09.560 |
|
It is not obtained by doing a point by point mapping. |
|
|
|
50:10.720 --> 50:14.000 |
|
For instance, if you try to learn a sorting algorithm |
|
|
|
50:14.000 --> 50:15.520 |
|
using a deep neural network, |
|
|
|
50:15.520 --> 50:18.520 |
|
well, you're very much limited to learning point by point |
|
|
|
50:20.080 --> 50:24.360 |
|
what the sorted representation of this specific list is like. |
|
|
|
50:24.360 --> 50:29.360 |
|
But instead, you could have a very, very simple |
|
|
|
50:29.400 --> 50:31.920 |
|
sorting algorithm written in a few lines. |
|
|
|
50:31.920 --> 50:34.520 |
|
Maybe it's just two nested loops. |
|
|
|
50:35.560 --> 50:40.560 |
|
And it can process any list at all because it is abstract, |
|
|
|
50:41.040 --> 50:42.240 |
|
because it is a set of rules. |
|
|
|
50:42.240 --> 50:45.160 |
|
So deep learning is really like point by point |
|
|
|
50:45.160 --> 50:48.640 |
|
geometric morphings, train with good and decent. |
|
|
|
50:48.640 --> 50:53.640 |
|
And meanwhile, abstract rules can generalize much better. |
|
|
|
50:53.640 --> 50:56.160 |
|
And I think the future is we need to combine the two. |
|
|
|
50:56.160 --> 50:59.160 |
|
So how do we, do you think, combine the two? |
|
|
|
50:59.160 --> 51:03.040 |
|
How do we combine good point by point functions |
|
|
|
51:03.040 --> 51:08.040 |
|
with programs, which is what the symbolic AI type systems? |
|
|
|
51:08.920 --> 51:11.600 |
|
At which levels the combination happen? |
|
|
|
51:11.600 --> 51:14.680 |
|
I mean, obviously we're jumping into the realm |
|
|
|
51:14.680 --> 51:16.880 |
|
of where there's no good answers. |
|
|
|
51:16.880 --> 51:20.280 |
|
It's just kind of ideas and intuitions and so on. |
|
|
|
51:20.280 --> 51:23.080 |
|
Well, if you look at the really successful AI systems |
|
|
|
51:23.080 --> 51:26.320 |
|
today, I think they are already hybrid systems |
|
|
|
51:26.320 --> 51:29.520 |
|
that are combining symbolic AI with deep learning. |
|
|
|
51:29.520 --> 51:32.520 |
|
For instance, successful robotics systems |
|
|
|
51:32.520 --> 51:36.400 |
|
are already mostly model based, rule based, |
|
|
|
51:37.400 --> 51:39.400 |
|
things like planning algorithms and so on. |
|
|
|
51:39.400 --> 51:42.200 |
|
At the same time, they're using deep learning |
|
|
|
51:42.200 --> 51:43.840 |
|
as perception modules. |
|
|
|
51:43.840 --> 51:46.000 |
|
Sometimes they're using deep learning as a way |
|
|
|
51:46.000 --> 51:50.920 |
|
to inject fuzzy intuition into a rule based process. |
|
|
|
51:50.920 --> 51:54.560 |
|
If you look at the system like in a self driving car, |
|
|
|
51:54.560 --> 51:57.240 |
|
it's not just one big end to end neural network. |
|
|
|
51:57.240 --> 51:59.000 |
|
You know, that wouldn't work at all. |
|
|
|
51:59.000 --> 52:00.760 |
|
Precisely because in order to train that, |
|
|
|
52:00.760 --> 52:05.160 |
|
you would need a dense sampling of experience base |
|
|
|
52:05.160 --> 52:06.200 |
|
when it comes to driving, |
|
|
|
52:06.200 --> 52:08.880 |
|
which is completely unrealistic, obviously. |
|
|
|
52:08.880 --> 52:12.440 |
|
Instead, the self driving car is mostly |
|
|
|
52:13.920 --> 52:18.360 |
|
symbolic, you know, it's software, it's programmed by hand. |
|
|
|
52:18.360 --> 52:21.640 |
|
So it's mostly based on explicit models. |
|
|
|
52:21.640 --> 52:25.840 |
|
In this case, mostly 3D models of the environment |
|
|
|
52:25.840 --> 52:29.520 |
|
around the car, but it's interfacing with the real world |
|
|
|
52:29.520 --> 52:31.440 |
|
using deep learning modules, right? |
|
|
|
52:31.440 --> 52:33.440 |
|
So the deep learning there serves as a way |
|
|
|
52:33.440 --> 52:36.080 |
|
to convert the raw sensory information |
|
|
|
52:36.080 --> 52:38.320 |
|
to something usable by symbolic systems. |
|
|
|
52:39.760 --> 52:42.400 |
|
Okay, well, let's linger on that a little more. |
|
|
|
52:42.400 --> 52:45.440 |
|
So dense sampling from input to output. |
|
|
|
52:45.440 --> 52:48.240 |
|
You said it's obviously very difficult. |
|
|
|
52:48.240 --> 52:50.120 |
|
Is it possible? |
|
|
|
52:50.120 --> 52:51.800 |
|
In the case of self driving, you mean? |
|
|
|
52:51.800 --> 52:53.040 |
|
Let's say self driving, right? |
|
|
|
52:53.040 --> 52:55.760 |
|
Self driving for many people, |
|
|
|
52:57.560 --> 52:59.520 |
|
let's not even talk about self driving, |
|
|
|
52:59.520 --> 53:03.880 |
|
let's talk about steering, so staying inside the lane. |
|
|
|
53:05.040 --> 53:07.080 |
|
Lane following, yeah, it's definitely a problem |
|
|
|
53:07.080 --> 53:08.880 |
|
you can solve with an end to end deep learning model, |
|
|
|
53:08.880 --> 53:10.600 |
|
but that's like one small subset. |
|
|
|
53:10.600 --> 53:11.440 |
|
Hold on a second. |
|
|
|
53:11.440 --> 53:12.760 |
|
Yeah, I don't know why you're jumping |
|
|
|
53:12.760 --> 53:14.480 |
|
from the extreme so easily, |
|
|
|
53:14.480 --> 53:16.280 |
|
because I disagree with you on that. |
|
|
|
53:16.280 --> 53:21.000 |
|
I think, well, it's not obvious to me |
|
|
|
53:21.000 --> 53:23.400 |
|
that you can solve lane following. |
|
|
|
53:23.400 --> 53:25.840 |
|
No, it's not obvious, I think it's doable. |
|
|
|
53:25.840 --> 53:30.840 |
|
I think in general, there is no hard limitations |
|
|
|
53:31.200 --> 53:33.680 |
|
to what you can learn with a deep neural network, |
|
|
|
53:33.680 --> 53:38.680 |
|
as long as the search space is rich enough, |
|
|
|
53:40.320 --> 53:42.240 |
|
is flexible enough, and as long as you have |
|
|
|
53:42.240 --> 53:45.360 |
|
this dense sampling of the input cross output space. |
|
|
|
53:45.360 --> 53:47.720 |
|
The problem is that this dense sampling |
|
|
|
53:47.720 --> 53:51.120 |
|
could mean anything from 10,000 examples |
|
|
|
53:51.120 --> 53:52.840 |
|
to like trillions and trillions. |
|
|
|
53:52.840 --> 53:54.360 |
|
So that's my question. |
|
|
|
53:54.360 --> 53:56.200 |
|
So what's your intuition? |
|
|
|
53:56.200 --> 53:58.720 |
|
And if you could just give it a chance |
|
|
|
53:58.720 --> 54:01.880 |
|
and think what kind of problems can be solved |
|
|
|
54:01.880 --> 54:04.240 |
|
by getting a huge amounts of data |
|
|
|
54:04.240 --> 54:08.000 |
|
and thereby creating a dense mapping. |
|
|
|
54:08.000 --> 54:12.480 |
|
So let's think about natural language dialogue, |
|
|
|
54:12.480 --> 54:14.000 |
|
the Turing test. |
|
|
|
54:14.000 --> 54:17.000 |
|
Do you think the Turing test can be solved |
|
|
|
54:17.000 --> 54:21.120 |
|
with a neural network alone? |
|
|
|
54:21.120 --> 54:24.440 |
|
Well, the Turing test is all about tricking people |
|
|
|
54:24.440 --> 54:26.880 |
|
into believing they're talking to a human. |
|
|
|
54:26.880 --> 54:29.040 |
|
And I don't think that's actually very difficult |
|
|
|
54:29.040 --> 54:34.040 |
|
because it's more about exploiting human perception |
|
|
|
54:35.600 --> 54:37.520 |
|
and not so much about intelligence. |
|
|
|
54:37.520 --> 54:39.680 |
|
There's a big difference between mimicking |
|
|
|
54:39.680 --> 54:42.080 |
|
intelligent behavior and actual intelligent behavior. |
|
|
|
54:42.080 --> 54:45.360 |
|
So, okay, let's look at maybe the Alexa prize and so on. |
|
|
|
54:45.360 --> 54:47.480 |
|
The different formulations of the natural language |
|
|
|
54:47.480 --> 54:50.520 |
|
conversation that are less about mimicking |
|
|
|
54:50.520 --> 54:52.800 |
|
and more about maintaining a fun conversation |
|
|
|
54:52.800 --> 54:54.720 |
|
that lasts for 20 minutes. |
|
|
|
54:54.720 --> 54:56.200 |
|
That's a little less about mimicking |
|
|
|
54:56.200 --> 54:59.080 |
|
and that's more about, I mean, it's still mimicking, |
|
|
|
54:59.080 --> 55:01.440 |
|
but it's more about being able to carry forward |
|
|
|
55:01.440 --> 55:03.640 |
|
a conversation with all the tangents that happen |
|
|
|
55:03.640 --> 55:05.080 |
|
in dialogue and so on. |
|
|
|
55:05.080 --> 55:08.320 |
|
Do you think that problem is learnable |
|
|
|
55:08.320 --> 55:13.320 |
|
with a neural network that does the point to point mapping? |
|
|
|
55:14.520 --> 55:16.280 |
|
So I think it would be very, very challenging |
|
|
|
55:16.280 --> 55:17.800 |
|
to do this with deep learning. |
|
|
|
55:17.800 --> 55:21.480 |
|
I don't think it's out of the question either. |
|
|
|
55:21.480 --> 55:23.240 |
|
I wouldn't rule it out. |
|
|
|
55:23.240 --> 55:25.400 |
|
The space of problems that can be solved |
|
|
|
55:25.400 --> 55:26.920 |
|
with a large neural network. |
|
|
|
55:26.920 --> 55:30.080 |
|
What's your sense about the space of those problems? |
|
|
|
55:30.080 --> 55:32.560 |
|
So useful problems for us. |
|
|
|
55:32.560 --> 55:34.800 |
|
In theory, it's infinite, right? |
|
|
|
55:34.800 --> 55:36.200 |
|
You can solve any problem. |
|
|
|
55:36.200 --> 55:39.800 |
|
In practice, well, deep learning is a great fit |
|
|
|
55:39.800 --> 55:41.800 |
|
for perception problems. |
|
|
|
55:41.800 --> 55:46.800 |
|
In general, any problem which is naturally amenable |
|
|
|
55:47.640 --> 55:52.200 |
|
to explicit handcrafted rules or rules that you can generate |
|
|
|
55:52.200 --> 55:54.960 |
|
by exhaustive search over some program space. |
|
|
|
55:56.080 --> 55:59.320 |
|
So perception, artificial intuition, |
|
|
|
55:59.320 --> 56:03.240 |
|
as long as you have a sufficient training dataset. |
|
|
|
56:03.240 --> 56:05.360 |
|
And that's the question, I mean, perception, |
|
|
|
56:05.360 --> 56:08.400 |
|
there's interpretation and understanding of the scene, |
|
|
|
56:08.400 --> 56:10.280 |
|
which seems to be outside the reach |
|
|
|
56:10.280 --> 56:12.960 |
|
of current perception systems. |
|
|
|
56:12.960 --> 56:15.920 |
|
So do you think larger networks will be able |
|
|
|
56:15.920 --> 56:18.280 |
|
to start to understand the physics |
|
|
|
56:18.280 --> 56:21.080 |
|
and the physics of the scene, |
|
|
|
56:21.080 --> 56:23.400 |
|
the three dimensional structure and relationships |
|
|
|
56:23.400 --> 56:25.560 |
|
of objects in the scene and so on? |
|
|
|
56:25.560 --> 56:28.320 |
|
Or really that's where symbolic AI has to step in? |
|
|
|
56:28.320 --> 56:34.480 |
|
Well, it's always possible to solve these problems |
|
|
|
56:34.480 --> 56:36.800 |
|
with deep learning. |
|
|
|
56:36.800 --> 56:38.560 |
|
It's just extremely inefficient. |
|
|
|
56:38.560 --> 56:42.000 |
|
A model would be an explicit rule based abstract model |
|
|
|
56:42.000 --> 56:45.240 |
|
would be a far better, more compressed |
|
|
|
56:45.240 --> 56:46.840 |
|
representation of physics. |
|
|
|
56:46.840 --> 56:49.080 |
|
Then learning just this mapping between |
|
|
|
56:49.080 --> 56:50.960 |
|
in this situation, this thing happens. |
|
|
|
56:50.960 --> 56:52.720 |
|
If you change the situation slightly, |
|
|
|
56:52.720 --> 56:54.760 |
|
then this other thing happens and so on. |
|
|
|
56:54.760 --> 56:57.440 |
|
Do you think it's possible to automatically generate |
|
|
|
56:57.440 --> 57:02.200 |
|
the programs that would require that kind of reasoning? |
|
|
|
57:02.200 --> 57:05.360 |
|
Or does it have to, so the way the expert systems fail, |
|
|
|
57:05.360 --> 57:07.120 |
|
there's so many facts about the world |
|
|
|
57:07.120 --> 57:08.960 |
|
had to be hand coded in. |
|
|
|
57:08.960 --> 57:14.600 |
|
Do you think it's possible to learn those logical statements |
|
|
|
57:14.600 --> 57:18.200 |
|
that are true about the world and their relationships? |
|
|
|
57:18.200 --> 57:20.360 |
|
Do you think, I mean, that's kind of what theorem proving |
|
|
|
57:20.360 --> 57:22.680 |
|
at a basic level is trying to do, right? |
|
|
|
57:22.680 --> 57:26.160 |
|
Yeah, except it's much harder to formulate statements |
|
|
|
57:26.160 --> 57:28.480 |
|
about the world compared to formulating |
|
|
|
57:28.480 --> 57:30.320 |
|
mathematical statements. |
|
|
|
57:30.320 --> 57:34.200 |
|
Statements about the world tend to be subjective. |
|
|
|
57:34.200 --> 57:39.600 |
|
So can you learn rule based models? |
|
|
|
57:39.600 --> 57:40.920 |
|
Yes, definitely. |
|
|
|
57:40.920 --> 57:43.640 |
|
That's the field of program synthesis. |
|
|
|
57:43.640 --> 57:48.040 |
|
However, today we just don't really know how to do it. |
|
|
|
57:48.040 --> 57:52.400 |
|
So it's very much a grass search or tree search problem. |
|
|
|
57:52.400 --> 57:56.800 |
|
And so we are limited to the sort of tree session grass |
|
|
|
57:56.800 --> 57:58.560 |
|
search algorithms that we have today. |
|
|
|
57:58.560 --> 58:02.760 |
|
Personally, I think genetic algorithms are very promising. |
|
|
|
58:02.760 --> 58:04.360 |
|
So almost like genetic programming. |
|
|
|
58:04.360 --> 58:05.560 |
|
Genetic programming, exactly. |
|
|
|
58:05.560 --> 58:08.840 |
|
Can you discuss the field of program synthesis? |
|
|
|
58:08.840 --> 58:14.560 |
|
Like how many people are working and thinking about it? |
|
|
|
58:14.560 --> 58:17.960 |
|
Where we are in the history of program synthesis |
|
|
|
58:17.960 --> 58:20.720 |
|
and what are your hopes for it? |
|
|
|
58:20.720 --> 58:24.600 |
|
Well, if it were deep learning, this is like the 90s. |
|
|
|
58:24.600 --> 58:29.120 |
|
So meaning that we already have existing solutions. |
|
|
|
58:29.120 --> 58:34.280 |
|
We are starting to have some basic understanding |
|
|
|
58:34.280 --> 58:35.480 |
|
of what this is about. |
|
|
|
58:35.480 --> 58:38.000 |
|
But it's still a field that is in its infancy. |
|
|
|
58:38.000 --> 58:40.440 |
|
There are very few people working on it. |
|
|
|
58:40.440 --> 58:44.480 |
|
There are very few real world applications. |
|
|
|
58:44.480 --> 58:47.640 |
|
So the one real world application I'm aware of |
|
|
|
58:47.640 --> 58:51.680 |
|
is Flash Fill in Excel. |
|
|
|
58:51.680 --> 58:55.080 |
|
It's a way to automatically learn very simple programs |
|
|
|
58:55.080 --> 58:58.200 |
|
to format cells in an Excel spreadsheet |
|
|
|
58:58.200 --> 59:00.240 |
|
from a few examples. |
|
|
|
59:00.240 --> 59:02.800 |
|
For instance, learning a way to format a date, things like that. |
|
|
|
59:02.800 --> 59:03.680 |
|
Oh, that's fascinating. |
|
|
|
59:03.680 --> 59:04.560 |
|
Yeah. |
|
|
|
59:04.560 --> 59:06.280 |
|
You know, OK, that's a fascinating topic. |
|
|
|
59:06.280 --> 59:10.480 |
|
I always wonder when I provide a few samples to Excel, |
|
|
|
59:10.480 --> 59:12.600 |
|
what it's able to figure out. |
|
|
|
59:12.600 --> 59:15.960 |
|
Like just giving it a few dates, what |
|
|
|
59:15.960 --> 59:18.480 |
|
are you able to figure out from the pattern I just gave you? |
|
|
|
59:18.480 --> 59:19.760 |
|
That's a fascinating question. |
|
|
|
59:19.760 --> 59:23.320 |
|
And it's fascinating whether that's learnable patterns. |
|
|
|
59:23.320 --> 59:25.520 |
|
And you're saying they're working on that. |
|
|
|
59:25.520 --> 59:28.200 |
|
How big is the toolbox currently? |
|
|
|
59:28.200 --> 59:29.520 |
|
Are we completely in the dark? |
|
|
|
59:29.520 --> 59:30.440 |
|
So if you said the 90s. |
|
|
|
59:30.440 --> 59:31.720 |
|
In terms of program synthesis? |
|
|
|
59:31.720 --> 59:32.360 |
|
No. |
|
|
|
59:32.360 --> 59:37.720 |
|
So I would say, so maybe 90s is even too optimistic. |
|
|
|
59:37.720 --> 59:41.080 |
|
Because by the 90s, we already understood back prop. |
|
|
|
59:41.080 --> 59:43.960 |
|
We already understood the engine of deep learning, |
|
|
|
59:43.960 --> 59:47.280 |
|
even though we couldn't really see its potential quite. |
|
|
|
59:47.280 --> 59:48.520 |
|
Today, I don't think we have found |
|
|
|
59:48.520 --> 59:50.400 |
|
the engine of program synthesis. |
|
|
|
59:50.400 --> 59:52.880 |
|
So we're in the winter before back prop. |
|
|
|
59:52.880 --> 59:54.160 |
|
Yeah. |
|
|
|
59:54.160 --> 59:55.720 |
|
In a way, yes. |
|
|
|
59:55.720 --> 1:00:00.120 |
|
So I do believe program synthesis and general discrete search |
|
|
|
1:00:00.120 --> 1:00:02.760 |
|
over rule based models is going to be |
|
|
|
1:00:02.760 --> 1:00:06.640 |
|
a cornerstone of AI research in the next century. |
|
|
|
1:00:06.640 --> 1:00:10.200 |
|
And that doesn't mean we are going to drop deep learning. |
|
|
|
1:00:10.200 --> 1:00:11.880 |
|
Deep learning is immensely useful. |
|
|
|
1:00:11.880 --> 1:00:17.200 |
|
Like, being able to learn is a very flexible, adaptable, |
|
|
|
1:00:17.200 --> 1:00:18.120 |
|
parametric model. |
|
|
|
1:00:18.120 --> 1:00:20.720 |
|
So it's got to understand that's actually immensely useful. |
|
|
|
1:00:20.720 --> 1:00:23.040 |
|
All it's doing is pattern cognition. |
|
|
|
1:00:23.040 --> 1:00:25.640 |
|
But being good at pattern cognition, given lots of data, |
|
|
|
1:00:25.640 --> 1:00:27.920 |
|
is just extremely powerful. |
|
|
|
1:00:27.920 --> 1:00:30.320 |
|
So we are still going to be working on deep learning. |
|
|
|
1:00:30.320 --> 1:00:31.840 |
|
We are going to be working on program synthesis. |
|
|
|
1:00:31.840 --> 1:00:34.680 |
|
We are going to be combining the two in increasingly automated |
|
|
|
1:00:34.680 --> 1:00:36.400 |
|
ways. |
|
|
|
1:00:36.400 --> 1:00:38.520 |
|
So let's talk a little bit about data. |
|
|
|
1:00:38.520 --> 1:00:44.600 |
|
You've tweeted, about 10,000 deep learning papers |
|
|
|
1:00:44.600 --> 1:00:47.080 |
|
have been written about hard coding priors |
|
|
|
1:00:47.080 --> 1:00:49.600 |
|
about a specific task in a neural network architecture |
|
|
|
1:00:49.600 --> 1:00:52.440 |
|
works better than a lack of a prior. |
|
|
|
1:00:52.440 --> 1:00:55.120 |
|
Basically, summarizing all these efforts, |
|
|
|
1:00:55.120 --> 1:00:56.920 |
|
they put a name to an architecture. |
|
|
|
1:00:56.920 --> 1:00:59.280 |
|
But really, what they're doing is hard coding some priors |
|
|
|
1:00:59.280 --> 1:01:01.560 |
|
that improve the performance of the system. |
|
|
|
1:01:01.560 --> 1:01:06.880 |
|
But which gets straight to the point is probably true. |
|
|
|
1:01:06.880 --> 1:01:09.800 |
|
So you say that you can always buy performance by, |
|
|
|
1:01:09.800 --> 1:01:12.920 |
|
in quotes, performance by either training on more data, |
|
|
|
1:01:12.920 --> 1:01:15.480 |
|
better data, or by injecting task information |
|
|
|
1:01:15.480 --> 1:01:18.400 |
|
to the architecture of the preprocessing. |
|
|
|
1:01:18.400 --> 1:01:21.280 |
|
However, this isn't informative about the generalization power |
|
|
|
1:01:21.280 --> 1:01:23.080 |
|
the techniques use, the fundamental ability |
|
|
|
1:01:23.080 --> 1:01:24.200 |
|
to generalize. |
|
|
|
1:01:24.200 --> 1:01:26.800 |
|
Do you think we can go far by coming up |
|
|
|
1:01:26.800 --> 1:01:29.920 |
|
with better methods for this kind of cheating, |
|
|
|
1:01:29.920 --> 1:01:33.520 |
|
for better methods of large scale annotation of data? |
|
|
|
1:01:33.520 --> 1:01:34.960 |
|
So building better priors. |
|
|
|
1:01:34.960 --> 1:01:37.280 |
|
If you automate it, it's not cheating anymore. |
|
|
|
1:01:37.280 --> 1:01:38.360 |
|
Right. |
|
|
|
1:01:38.360 --> 1:01:41.600 |
|
I'm joking about the cheating, but large scale. |
|
|
|
1:01:41.600 --> 1:01:46.560 |
|
So basically, I'm asking about something |
|
|
|
1:01:46.560 --> 1:01:48.280 |
|
that hasn't, from my perspective, |
|
|
|
1:01:48.280 --> 1:01:53.360 |
|
been researched too much is exponential improvement |
|
|
|
1:01:53.360 --> 1:01:55.960 |
|
in annotation of data. |
|
|
|
1:01:55.960 --> 1:01:58.120 |
|
Do you often think about? |
|
|
|
1:01:58.120 --> 1:02:00.840 |
|
I think it's actually been researched quite a bit. |
|
|
|
1:02:00.840 --> 1:02:02.720 |
|
You just don't see publications about it. |
|
|
|
1:02:02.720 --> 1:02:05.840 |
|
Because people who publish papers |
|
|
|
1:02:05.840 --> 1:02:07.920 |
|
are going to publish about known benchmarks. |
|
|
|
1:02:07.920 --> 1:02:09.800 |
|
Sometimes they're going to read a new benchmark. |
|
|
|
1:02:09.800 --> 1:02:12.200 |
|
People who actually have real world large scale |
|
|
|
1:02:12.200 --> 1:02:13.880 |
|
depending on problems, they're going |
|
|
|
1:02:13.880 --> 1:02:16.960 |
|
to spend a lot of resources into data annotation |
|
|
|
1:02:16.960 --> 1:02:18.400 |
|
and good data annotation pipelines, |
|
|
|
1:02:18.400 --> 1:02:19.640 |
|
but you don't see any papers about it. |
|
|
|
1:02:19.640 --> 1:02:20.400 |
|
That's interesting. |
|
|
|
1:02:20.400 --> 1:02:22.720 |
|
So do you think, certainly resources, |
|
|
|
1:02:22.720 --> 1:02:24.840 |
|
but do you think there's innovation happening? |
|
|
|
1:02:24.840 --> 1:02:25.880 |
|
Oh, yeah. |
|
|
|
1:02:25.880 --> 1:02:28.880 |
|
To clarify the point in the tweet. |
|
|
|
1:02:28.880 --> 1:02:31.160 |
|
So machine learning in general is |
|
|
|
1:02:31.160 --> 1:02:33.840 |
|
the science of generalization. |
|
|
|
1:02:33.840 --> 1:02:37.800 |
|
You want to generate knowledge that |
|
|
|
1:02:37.800 --> 1:02:40.440 |
|
can be reused across different data sets, |
|
|
|
1:02:40.440 --> 1:02:42.000 |
|
across different tasks. |
|
|
|
1:02:42.000 --> 1:02:45.280 |
|
And if instead you're looking at one data set |
|
|
|
1:02:45.280 --> 1:02:50.000 |
|
and then you are hard coding knowledge about this task |
|
|
|
1:02:50.000 --> 1:02:54.040 |
|
into your architecture, this is no more useful |
|
|
|
1:02:54.040 --> 1:02:56.760 |
|
than training a network and then saying, oh, I |
|
|
|
1:02:56.760 --> 1:03:01.920 |
|
found these weight values perform well. |
|
|
|
1:03:01.920 --> 1:03:05.680 |
|
So David Ha, I don't know if you know David, |
|
|
|
1:03:05.680 --> 1:03:08.760 |
|
he had a paper the other day about weight |
|
|
|
1:03:08.760 --> 1:03:10.400 |
|
agnostic neural networks. |
|
|
|
1:03:10.400 --> 1:03:12.120 |
|
And this is a very interesting paper |
|
|
|
1:03:12.120 --> 1:03:14.400 |
|
because it really illustrates the fact |
|
|
|
1:03:14.400 --> 1:03:17.400 |
|
that an architecture, even without weights, |
|
|
|
1:03:17.400 --> 1:03:21.360 |
|
an architecture is knowledge about a task. |
|
|
|
1:03:21.360 --> 1:03:23.640 |
|
It encodes knowledge. |
|
|
|
1:03:23.640 --> 1:03:25.840 |
|
And when it comes to architectures |
|
|
|
1:03:25.840 --> 1:03:30.440 |
|
that are uncrafted by researchers, in some cases, |
|
|
|
1:03:30.440 --> 1:03:34.160 |
|
it is very, very clear that all they are doing |
|
|
|
1:03:34.160 --> 1:03:38.880 |
|
is artificially reencoding the template that |
|
|
|
1:03:38.880 --> 1:03:44.400 |
|
corresponds to the proper way to solve the task encoding |
|
|
|
1:03:44.400 --> 1:03:45.200 |
|
a given data set. |
|
|
|
1:03:45.200 --> 1:03:48.120 |
|
For instance, I know if you looked |
|
|
|
1:03:48.120 --> 1:03:52.280 |
|
at the baby data set, which is about natural language |
|
|
|
1:03:52.280 --> 1:03:55.520 |
|
question answering, it is generated by an algorithm. |
|
|
|
1:03:55.520 --> 1:03:57.680 |
|
So this is a question answer pairs |
|
|
|
1:03:57.680 --> 1:03:59.280 |
|
that are generated by an algorithm. |
|
|
|
1:03:59.280 --> 1:04:01.520 |
|
The algorithm is solving a certain template. |
|
|
|
1:04:01.520 --> 1:04:04.400 |
|
Turns out, if you craft a network that |
|
|
|
1:04:04.400 --> 1:04:06.360 |
|
literally encodes this template, you |
|
|
|
1:04:06.360 --> 1:04:09.640 |
|
can solve this data set with nearly 100% accuracy. |
|
|
|
1:04:09.640 --> 1:04:11.160 |
|
But that doesn't actually tell you |
|
|
|
1:04:11.160 --> 1:04:14.640 |
|
anything about how to solve question answering |
|
|
|
1:04:14.640 --> 1:04:17.680 |
|
in general, which is the point. |
|
|
|
1:04:17.680 --> 1:04:19.400 |
|
The question is just to linger on it, |
|
|
|
1:04:19.400 --> 1:04:21.560 |
|
whether it's from the data side or from the size |
|
|
|
1:04:21.560 --> 1:04:23.280 |
|
of the network. |
|
|
|
1:04:23.280 --> 1:04:25.920 |
|
I don't know if you've read the blog post by Rich Sutton, |
|
|
|
1:04:25.920 --> 1:04:28.400 |
|
The Bitter Lesson, where he says, |
|
|
|
1:04:28.400 --> 1:04:31.480 |
|
the biggest lesson that we can read from 70 years of AI |
|
|
|
1:04:31.480 --> 1:04:34.720 |
|
research is that general methods that leverage computation |
|
|
|
1:04:34.720 --> 1:04:37.160 |
|
are ultimately the most effective. |
|
|
|
1:04:37.160 --> 1:04:39.720 |
|
So as opposed to figuring out methods |
|
|
|
1:04:39.720 --> 1:04:41.840 |
|
that can generalize effectively, do you |
|
|
|
1:04:41.840 --> 1:04:47.720 |
|
think we can get pretty far by just having something |
|
|
|
1:04:47.720 --> 1:04:51.520 |
|
that leverages computation and the improvement of computation? |
|
|
|
1:04:51.520 --> 1:04:54.960 |
|
Yeah, so I think Rich is making a very good point, which |
|
|
|
1:04:54.960 --> 1:04:57.560 |
|
is that a lot of these papers, which are actually |
|
|
|
1:04:57.560 --> 1:05:02.800 |
|
all about manually hardcoding prior knowledge about a task |
|
|
|
1:05:02.800 --> 1:05:04.720 |
|
into some system, it doesn't have |
|
|
|
1:05:04.720 --> 1:05:08.600 |
|
to be deep learning architecture, but into some system. |
|
|
|
1:05:08.600 --> 1:05:11.920 |
|
These papers are not actually making any impact. |
|
|
|
1:05:11.920 --> 1:05:14.800 |
|
Instead, what's making really long term impact |
|
|
|
1:05:14.800 --> 1:05:18.520 |
|
is very simple, very general systems |
|
|
|
1:05:18.520 --> 1:05:21.280 |
|
that are really agnostic to all these tricks. |
|
|
|
1:05:21.280 --> 1:05:23.320 |
|
Because these tricks do not generalize. |
|
|
|
1:05:23.320 --> 1:05:27.480 |
|
And of course, the one general and simple thing |
|
|
|
1:05:27.480 --> 1:05:33.160 |
|
that you should focus on is that which leverages computation. |
|
|
|
1:05:33.160 --> 1:05:36.200 |
|
Because computation, the availability |
|
|
|
1:05:36.200 --> 1:05:39.400 |
|
of large scale computation has been increasing exponentially |
|
|
|
1:05:39.400 --> 1:05:40.560 |
|
following Moore's law. |
|
|
|
1:05:40.560 --> 1:05:44.080 |
|
So if your algorithm is all about exploiting this, |
|
|
|
1:05:44.080 --> 1:05:47.440 |
|
then your algorithm is suddenly exponentially improving. |
|
|
|
1:05:47.440 --> 1:05:52.400 |
|
So I think Rich is definitely right. |
|
|
|
1:05:52.400 --> 1:05:57.120 |
|
However, he's right about the past 70 years. |
|
|
|
1:05:57.120 --> 1:05:59.440 |
|
He's like assessing the past 70 years. |
|
|
|
1:05:59.440 --> 1:06:02.360 |
|
I am not sure that this assessment will still |
|
|
|
1:06:02.360 --> 1:06:04.880 |
|
hold true for the next 70 years. |
|
|
|
1:06:04.880 --> 1:06:07.160 |
|
It might to some extent. |
|
|
|
1:06:07.160 --> 1:06:08.560 |
|
I suspect it will not. |
|
|
|
1:06:08.560 --> 1:06:11.560 |
|
Because the truth of his assessment |
|
|
|
1:06:11.560 --> 1:06:16.800 |
|
is a function of the context in which this research took place. |
|
|
|
1:06:16.800 --> 1:06:18.600 |
|
And the context is changing. |
|
|
|
1:06:18.600 --> 1:06:21.440 |
|
Moore's law might not be applicable anymore, |
|
|
|
1:06:21.440 --> 1:06:23.760 |
|
for instance, in the future. |
|
|
|
1:06:23.760 --> 1:06:31.200 |
|
And I do believe that when you tweak one aspect of a system, |
|
|
|
1:06:31.200 --> 1:06:32.920 |
|
when you exploit one aspect of a system, |
|
|
|
1:06:32.920 --> 1:06:36.480 |
|
some other aspect starts becoming the bottleneck. |
|
|
|
1:06:36.480 --> 1:06:38.800 |
|
Let's say you have unlimited computation. |
|
|
|
1:06:38.800 --> 1:06:41.440 |
|
Well, then data is the bottleneck. |
|
|
|
1:06:41.440 --> 1:06:43.560 |
|
And I think we are already starting |
|
|
|
1:06:43.560 --> 1:06:45.720 |
|
to be in a regime where our systems are |
|
|
|
1:06:45.720 --> 1:06:48.120 |
|
so large in scale and so data ingrained |
|
|
|
1:06:48.120 --> 1:06:50.360 |
|
that data today and the quality of data |
|
|
|
1:06:50.360 --> 1:06:53.040 |
|
and the scale of data is the bottleneck. |
|
|
|
1:06:53.040 --> 1:06:58.160 |
|
And in this environment, the bitter lesson from Rich |
|
|
|
1:06:58.160 --> 1:07:00.800 |
|
is not going to be true anymore. |
|
|
|
1:07:00.800 --> 1:07:03.960 |
|
So I think we are going to move from a focus |
|
|
|
1:07:03.960 --> 1:07:09.840 |
|
on a computation scale to focus on data efficiency. |
|
|
|
1:07:09.840 --> 1:07:10.720 |
|
Data efficiency. |
|
|
|
1:07:10.720 --> 1:07:13.120 |
|
So that's getting to the question of symbolic AI. |
|
|
|
1:07:13.120 --> 1:07:16.280 |
|
But to linger on the deep learning approaches, |
|
|
|
1:07:16.280 --> 1:07:19.240 |
|
do you have hope for either unsupervised learning |
|
|
|
1:07:19.240 --> 1:07:23.280 |
|
or reinforcement learning, which are |
|
|
|
1:07:23.280 --> 1:07:28.120 |
|
ways of being more data efficient in terms |
|
|
|
1:07:28.120 --> 1:07:31.560 |
|
of the amount of data they need that required human annotation? |
|
|
|
1:07:31.560 --> 1:07:34.280 |
|
So unsupervised learning and reinforcement learning |
|
|
|
1:07:34.280 --> 1:07:36.640 |
|
are frameworks for learning, but they are not |
|
|
|
1:07:36.640 --> 1:07:39.000 |
|
like any specific technique. |
|
|
|
1:07:39.000 --> 1:07:41.200 |
|
So usually when people say reinforcement learning, |
|
|
|
1:07:41.200 --> 1:07:43.320 |
|
what they really mean is deep reinforcement learning, |
|
|
|
1:07:43.320 --> 1:07:47.440 |
|
which is like one approach which is actually very questionable. |
|
|
|
1:07:47.440 --> 1:07:50.920 |
|
The question I was asking was unsupervised learning |
|
|
|
1:07:50.920 --> 1:07:54.680 |
|
with deep neural networks and deep reinforcement learning. |
|
|
|
1:07:54.680 --> 1:07:56.840 |
|
Well, these are not really data efficient |
|
|
|
1:07:56.840 --> 1:08:00.520 |
|
because you're still leveraging these huge parametric models |
|
|
|
1:08:00.520 --> 1:08:03.720 |
|
point by point with gradient descent. |
|
|
|
1:08:03.720 --> 1:08:08.000 |
|
It is more efficient in terms of the number of annotations, |
|
|
|
1:08:08.000 --> 1:08:09.520 |
|
the density of annotations you need. |
|
|
|
1:08:09.520 --> 1:08:13.840 |
|
So the idea being to learn the latent space around which |
|
|
|
1:08:13.840 --> 1:08:17.960 |
|
the data is organized and then map the sparse annotations |
|
|
|
1:08:17.960 --> 1:08:18.760 |
|
into it. |
|
|
|
1:08:18.760 --> 1:08:23.560 |
|
And sure, I mean, that's clearly a very good idea. |
|
|
|
1:08:23.560 --> 1:08:26.080 |
|
It's not really a topic I would be working on, |
|
|
|
1:08:26.080 --> 1:08:28.040 |
|
but it's clearly a good idea. |
|
|
|
1:08:28.040 --> 1:08:31.760 |
|
So it would get us to solve some problems that? |
|
|
|
1:08:31.760 --> 1:08:34.880 |
|
It will get us to incremental improvements |
|
|
|
1:08:34.880 --> 1:08:38.240 |
|
in labeled data efficiency. |
|
|
|
1:08:38.240 --> 1:08:43.520 |
|
Do you have concerns about short term or long term threats |
|
|
|
1:08:43.520 --> 1:08:47.800 |
|
from AI, from artificial intelligence? |
|
|
|
1:08:47.800 --> 1:08:50.640 |
|
Yes, definitely to some extent. |
|
|
|
1:08:50.640 --> 1:08:52.800 |
|
And what's the shape of those concerns? |
|
|
|
1:08:52.800 --> 1:08:56.880 |
|
This is actually something I've briefly written about. |
|
|
|
1:08:56.880 --> 1:09:02.680 |
|
But the capabilities of deep learning technology |
|
|
|
1:09:02.680 --> 1:09:05.200 |
|
can be used in many ways that are |
|
|
|
1:09:05.200 --> 1:09:09.760 |
|
concerning from mass surveillance with things |
|
|
|
1:09:09.760 --> 1:09:11.880 |
|
like facial recognition. |
|
|
|
1:09:11.880 --> 1:09:15.440 |
|
In general, tracking lots of data about everyone |
|
|
|
1:09:15.440 --> 1:09:18.920 |
|
and then being able to making sense of this data |
|
|
|
1:09:18.920 --> 1:09:22.240 |
|
to do identification, to do prediction. |
|
|
|
1:09:22.240 --> 1:09:23.160 |
|
That's concerning. |
|
|
|
1:09:23.160 --> 1:09:26.560 |
|
That's something that's being very aggressively pursued |
|
|
|
1:09:26.560 --> 1:09:31.440 |
|
by totalitarian states like China. |
|
|
|
1:09:31.440 --> 1:09:34.000 |
|
One thing I am very much concerned about |
|
|
|
1:09:34.000 --> 1:09:40.640 |
|
is that our lives are increasingly online, |
|
|
|
1:09:40.640 --> 1:09:43.280 |
|
are increasingly digital, made of information, |
|
|
|
1:09:43.280 --> 1:09:48.080 |
|
made of information consumption and information production, |
|
|
|
1:09:48.080 --> 1:09:51.800 |
|
our digital footprint, I would say. |
|
|
|
1:09:51.800 --> 1:09:56.280 |
|
And if you absorb all of this data |
|
|
|
1:09:56.280 --> 1:10:01.440 |
|
and you are in control of where you consume information, |
|
|
|
1:10:01.440 --> 1:10:06.960 |
|
social networks and so on, recommendation engines, |
|
|
|
1:10:06.960 --> 1:10:10.200 |
|
then you can build a sort of reinforcement |
|
|
|
1:10:10.200 --> 1:10:13.760 |
|
loop for human behavior. |
|
|
|
1:10:13.760 --> 1:10:18.360 |
|
You can observe the state of your mind at time t. |
|
|
|
1:10:18.360 --> 1:10:21.080 |
|
You can predict how you would react |
|
|
|
1:10:21.080 --> 1:10:23.800 |
|
to different pieces of content, how |
|
|
|
1:10:23.800 --> 1:10:27.000 |
|
to get you to move your mind in a certain direction. |
|
|
|
1:10:27.000 --> 1:10:33.160 |
|
And then you can feed you the specific piece of content |
|
|
|
1:10:33.160 --> 1:10:35.680 |
|
that would move you in a specific direction. |
|
|
|
1:10:35.680 --> 1:10:41.800 |
|
And you can do this at scale in terms |
|
|
|
1:10:41.800 --> 1:10:44.960 |
|
of doing it continuously in real time. |
|
|
|
1:10:44.960 --> 1:10:46.440 |
|
You can also do it at scale in terms |
|
|
|
1:10:46.440 --> 1:10:50.480 |
|
of scaling this to many, many people, to entire populations. |
|
|
|
1:10:50.480 --> 1:10:53.840 |
|
So potentially, artificial intelligence, |
|
|
|
1:10:53.840 --> 1:10:57.440 |
|
even in its current state, if you combine it |
|
|
|
1:10:57.440 --> 1:11:01.760 |
|
with the internet, with the fact that all of our lives |
|
|
|
1:11:01.760 --> 1:11:05.120 |
|
are moving to digital devices and digital information |
|
|
|
1:11:05.120 --> 1:11:08.720 |
|
consumption and creation, what you get |
|
|
|
1:11:08.720 --> 1:11:14.480 |
|
is the possibility to achieve mass manipulation of behavior |
|
|
|
1:11:14.480 --> 1:11:16.840 |
|
and mass psychological control. |
|
|
|
1:11:16.840 --> 1:11:18.520 |
|
And this is a very real possibility. |
|
|
|
1:11:18.520 --> 1:11:22.080 |
|
Yeah, so you're talking about any kind of recommender system. |
|
|
|
1:11:22.080 --> 1:11:26.160 |
|
Let's look at the YouTube algorithm, Facebook, |
|
|
|
1:11:26.160 --> 1:11:29.720 |
|
anything that recommends content you should watch next. |
|
|
|
1:11:29.720 --> 1:11:32.960 |
|
And it's fascinating to think that there's |
|
|
|
1:11:32.960 --> 1:11:41.120 |
|
some aspects of human behavior that you can say a problem of, |
|
|
|
1:11:41.120 --> 1:11:45.400 |
|
is this person hold Republican beliefs or Democratic beliefs? |
|
|
|
1:11:45.400 --> 1:11:50.240 |
|
And this is a trivial, that's an objective function. |
|
|
|
1:11:50.240 --> 1:11:52.600 |
|
And you can optimize, and you can measure, |
|
|
|
1:11:52.600 --> 1:11:54.360 |
|
and you can turn everybody into a Republican |
|
|
|
1:11:54.360 --> 1:11:56.080 |
|
or everybody into a Democrat. |
|
|
|
1:11:56.080 --> 1:11:57.840 |
|
I do believe it's true. |
|
|
|
1:11:57.840 --> 1:12:03.680 |
|
So the human mind is very, if you look at the human mind |
|
|
|
1:12:03.680 --> 1:12:05.320 |
|
as a kind of computer program, it |
|
|
|
1:12:05.320 --> 1:12:07.560 |
|
has a very large exploit surface. |
|
|
|
1:12:07.560 --> 1:12:09.360 |
|
It has many, many vulnerabilities. |
|
|
|
1:12:09.360 --> 1:12:10.840 |
|
Exploit surfaces, yeah. |
|
|
|
1:12:10.840 --> 1:12:13.520 |
|
Ways you can control it. |
|
|
|
1:12:13.520 --> 1:12:16.680 |
|
For instance, when it comes to your political beliefs, |
|
|
|
1:12:16.680 --> 1:12:19.400 |
|
this is very much tied to your identity. |
|
|
|
1:12:19.400 --> 1:12:23.040 |
|
So for instance, if I'm in control of your news feed |
|
|
|
1:12:23.040 --> 1:12:26.000 |
|
on your favorite social media platforms, |
|
|
|
1:12:26.000 --> 1:12:29.360 |
|
this is actually where you're getting your news from. |
|
|
|
1:12:29.360 --> 1:12:32.960 |
|
And of course, I can choose to only show you |
|
|
|
1:12:32.960 --> 1:12:37.120 |
|
news that will make you see the world in a specific way. |
|
|
|
1:12:37.120 --> 1:12:41.920 |
|
But I can also create incentives for you |
|
|
|
1:12:41.920 --> 1:12:44.720 |
|
to post about some political beliefs. |
|
|
|
1:12:44.720 --> 1:12:47.960 |
|
And then when I get you to express a statement, |
|
|
|
1:12:47.960 --> 1:12:51.840 |
|
if it's a statement that me as the controller, |
|
|
|
1:12:51.840 --> 1:12:53.800 |
|
I want to reinforce. |
|
|
|
1:12:53.800 --> 1:12:55.560 |
|
I can just show it to people who will agree, |
|
|
|
1:12:55.560 --> 1:12:56.880 |
|
and they will like it. |
|
|
|
1:12:56.880 --> 1:12:59.280 |
|
And that will reinforce the statement in your mind. |
|
|
|
1:12:59.280 --> 1:13:02.760 |
|
If this is a statement I want you to, |
|
|
|
1:13:02.760 --> 1:13:05.320 |
|
this is a belief I want you to abandon, |
|
|
|
1:13:05.320 --> 1:13:09.600 |
|
I can, on the other hand, show it to opponents. |
|
|
|
1:13:09.600 --> 1:13:10.640 |
|
We'll attack you. |
|
|
|
1:13:10.640 --> 1:13:12.840 |
|
And because they attack you, at the very least, |
|
|
|
1:13:12.840 --> 1:13:16.840 |
|
next time you will think twice about posting it. |
|
|
|
1:13:16.840 --> 1:13:20.280 |
|
But maybe you will even start believing this |
|
|
|
1:13:20.280 --> 1:13:22.840 |
|
because you got pushback. |
|
|
|
1:13:22.840 --> 1:13:28.440 |
|
So there are many ways in which social media platforms |
|
|
|
1:13:28.440 --> 1:13:30.520 |
|
can potentially control your opinions. |
|
|
|
1:13:30.520 --> 1:13:35.040 |
|
And today, so all of these things |
|
|
|
1:13:35.040 --> 1:13:38.240 |
|
are already being controlled by AI algorithms. |
|
|
|
1:13:38.240 --> 1:13:41.880 |
|
These algorithms do not have any explicit political goal |
|
|
|
1:13:41.880 --> 1:13:42.880 |
|
today. |
|
|
|
1:13:42.880 --> 1:13:48.680 |
|
Well, potentially they could, like if some totalitarian |
|
|
|
1:13:48.680 --> 1:13:52.720 |
|
government takes over social media platforms |
|
|
|
1:13:52.720 --> 1:13:55.360 |
|
and decides that now we are going to use this not just |
|
|
|
1:13:55.360 --> 1:13:58.040 |
|
for mass surveillance, but also for mass opinion control |
|
|
|
1:13:58.040 --> 1:13:59.360 |
|
and behavior control. |
|
|
|
1:13:59.360 --> 1:14:01.840 |
|
Very bad things could happen. |
|
|
|
1:14:01.840 --> 1:14:06.480 |
|
But what's really fascinating and actually quite concerning |
|
|
|
1:14:06.480 --> 1:14:11.280 |
|
is that even without an explicit intent to manipulate, |
|
|
|
1:14:11.280 --> 1:14:14.760 |
|
you're already seeing very dangerous dynamics |
|
|
|
1:14:14.760 --> 1:14:18.160 |
|
in terms of how these content recommendation |
|
|
|
1:14:18.160 --> 1:14:19.800 |
|
algorithms behave. |
|
|
|
1:14:19.800 --> 1:14:24.920 |
|
Because right now, the goal, the objective function |
|
|
|
1:14:24.920 --> 1:14:28.640 |
|
of these algorithms is to maximize engagement, |
|
|
|
1:14:28.640 --> 1:14:32.520 |
|
which seems fairly innocuous at first. |
|
|
|
1:14:32.520 --> 1:14:36.480 |
|
However, it is not because content |
|
|
|
1:14:36.480 --> 1:14:42.000 |
|
that will maximally engage people, get people to react |
|
|
|
1:14:42.000 --> 1:14:44.720 |
|
in an emotional way, get people to click on something. |
|
|
|
1:14:44.720 --> 1:14:52.200 |
|
It is very often content that is not |
|
|
|
1:14:52.200 --> 1:14:54.400 |
|
healthy to the public discourse. |
|
|
|
1:14:54.400 --> 1:14:58.200 |
|
For instance, fake news are far more |
|
|
|
1:14:58.200 --> 1:15:01.320 |
|
likely to get you to click on them than real news |
|
|
|
1:15:01.320 --> 1:15:06.960 |
|
simply because they are not constrained to reality. |
|
|
|
1:15:06.960 --> 1:15:11.360 |
|
So they can be as outrageous, as surprising, |
|
|
|
1:15:11.360 --> 1:15:15.880 |
|
as good stories as you want because they're artificial. |
|
|
|
1:15:15.880 --> 1:15:18.880 |
|
To me, that's an exciting world because so much good |
|
|
|
1:15:18.880 --> 1:15:19.560 |
|
can come. |
|
|
|
1:15:19.560 --> 1:15:24.520 |
|
So there's an opportunity to educate people. |
|
|
|
1:15:24.520 --> 1:15:31.200 |
|
You can balance people's worldview with other ideas. |
|
|
|
1:15:31.200 --> 1:15:33.800 |
|
So there's so many objective functions. |
|
|
|
1:15:33.800 --> 1:15:35.840 |
|
The space of objective functions that |
|
|
|
1:15:35.840 --> 1:15:40.720 |
|
create better civilizations is large, arguably infinite. |
|
|
|
1:15:40.720 --> 1:15:43.720 |
|
But there's also a large space that |
|
|
|
1:15:43.720 --> 1:15:51.480 |
|
creates division and destruction, civil war, |
|
|
|
1:15:51.480 --> 1:15:53.160 |
|
a lot of bad stuff. |
|
|
|
1:15:53.160 --> 1:15:56.920 |
|
And the worry is, naturally, probably that space |
|
|
|
1:15:56.920 --> 1:15:59.160 |
|
is bigger, first of all. |
|
|
|
1:15:59.160 --> 1:16:04.920 |
|
And if we don't explicitly think about what kind of effects |
|
|
|
1:16:04.920 --> 1:16:08.320 |
|
are going to be observed from different objective functions, |
|
|
|
1:16:08.320 --> 1:16:10.160 |
|
then we're going to get into trouble. |
|
|
|
1:16:10.160 --> 1:16:14.480 |
|
But the question is, how do we get into rooms |
|
|
|
1:16:14.480 --> 1:16:18.560 |
|
and have discussions, so inside Google, inside Facebook, |
|
|
|
1:16:18.560 --> 1:16:21.840 |
|
inside Twitter, and think about, OK, |
|
|
|
1:16:21.840 --> 1:16:24.840 |
|
how can we drive up engagement and, at the same time, |
|
|
|
1:16:24.840 --> 1:16:28.200 |
|
create a good society? |
|
|
|
1:16:28.200 --> 1:16:29.560 |
|
Is it even possible to have that kind |
|
|
|
1:16:29.560 --> 1:16:31.720 |
|
of philosophical discussion? |
|
|
|
1:16:31.720 --> 1:16:33.080 |
|
I think you can definitely try. |
|
|
|
1:16:33.080 --> 1:16:37.280 |
|
So from my perspective, I would feel rather uncomfortable |
|
|
|
1:16:37.280 --> 1:16:41.560 |
|
with companies that are uncomfortable with these new |
|
|
|
1:16:41.560 --> 1:16:47.120 |
|
student algorithms, with them making explicit decisions |
|
|
|
1:16:47.120 --> 1:16:50.440 |
|
to manipulate people's opinions or behaviors, |
|
|
|
1:16:50.440 --> 1:16:53.480 |
|
even if the intent is good, because that's |
|
|
|
1:16:53.480 --> 1:16:55.200 |
|
a very totalitarian mindset. |
|
|
|
1:16:55.200 --> 1:16:57.440 |
|
So instead, what I would like to see |
|
|
|
1:16:57.440 --> 1:16:58.880 |
|
is probably never going to happen, |
|
|
|
1:16:58.880 --> 1:17:00.360 |
|
because it's not super realistic, |
|
|
|
1:17:00.360 --> 1:17:02.520 |
|
but that's actually something I really care about. |
|
|
|
1:17:02.520 --> 1:17:06.280 |
|
I would like all these algorithms |
|
|
|
1:17:06.280 --> 1:17:10.560 |
|
to present configuration settings to their users, |
|
|
|
1:17:10.560 --> 1:17:14.600 |
|
so that the users can actually make the decision about how |
|
|
|
1:17:14.600 --> 1:17:19.000 |
|
they want to be impacted by these information |
|
|
|
1:17:19.000 --> 1:17:21.960 |
|
recommendation, content recommendation algorithms. |
|
|
|
1:17:21.960 --> 1:17:24.240 |
|
For instance, as a user of something |
|
|
|
1:17:24.240 --> 1:17:26.520 |
|
like YouTube or Twitter, maybe I want |
|
|
|
1:17:26.520 --> 1:17:30.280 |
|
to maximize learning about a specific topic. |
|
|
|
1:17:30.280 --> 1:17:36.800 |
|
So I want the algorithm to feed my curiosity, |
|
|
|
1:17:36.800 --> 1:17:38.760 |
|
which is in itself a very interesting problem. |
|
|
|
1:17:38.760 --> 1:17:41.200 |
|
So instead of maximizing my engagement, |
|
|
|
1:17:41.200 --> 1:17:44.600 |
|
it will maximize how fast and how much I'm learning. |
|
|
|
1:17:44.600 --> 1:17:47.360 |
|
And it will also take into account the accuracy, |
|
|
|
1:17:47.360 --> 1:17:50.680 |
|
hopefully, of the information I'm learning. |
|
|
|
1:17:50.680 --> 1:17:55.680 |
|
So yeah, the user should be able to determine exactly |
|
|
|
1:17:55.680 --> 1:17:58.560 |
|
how these algorithms are affecting their lives. |
|
|
|
1:17:58.560 --> 1:18:03.520 |
|
I don't want actually any entity making decisions |
|
|
|
1:18:03.520 --> 1:18:09.480 |
|
about in which direction they're going to try to manipulate me. |
|
|
|
1:18:09.480 --> 1:18:11.680 |
|
I want technology. |
|
|
|
1:18:11.680 --> 1:18:14.280 |
|
So AI, these algorithms are increasingly |
|
|
|
1:18:14.280 --> 1:18:18.560 |
|
going to be our interface to a world that is increasingly |
|
|
|
1:18:18.560 --> 1:18:19.960 |
|
made of information. |
|
|
|
1:18:19.960 --> 1:18:25.840 |
|
And I want everyone to be in control of this interface, |
|
|
|
1:18:25.840 --> 1:18:29.160 |
|
to interface with the world on their own terms. |
|
|
|
1:18:29.160 --> 1:18:32.840 |
|
So if someone wants these algorithms |
|
|
|
1:18:32.840 --> 1:18:37.640 |
|
to serve their own personal growth goals, |
|
|
|
1:18:37.640 --> 1:18:40.640 |
|
they should be able to configure these algorithms |
|
|
|
1:18:40.640 --> 1:18:41.800 |
|
in such a way. |
|
|
|
1:18:41.800 --> 1:18:46.680 |
|
Yeah, but so I know it's painful to have explicit decisions. |
|
|
|
1:18:46.680 --> 1:18:51.080 |
|
But there is underlying explicit decisions, |
|
|
|
1:18:51.080 --> 1:18:53.360 |
|
which is some of the most beautiful fundamental |
|
|
|
1:18:53.360 --> 1:18:57.400 |
|
philosophy that we have before us, |
|
|
|
1:18:57.400 --> 1:19:01.120 |
|
which is personal growth. |
|
|
|
1:19:01.120 --> 1:19:05.680 |
|
If I want to watch videos from which I can learn, |
|
|
|
1:19:05.680 --> 1:19:08.080 |
|
what does that mean? |
|
|
|
1:19:08.080 --> 1:19:11.800 |
|
So if I have a checkbox that wants to emphasize learning, |
|
|
|
1:19:11.800 --> 1:19:15.480 |
|
there's still an algorithm with explicit decisions in it |
|
|
|
1:19:15.480 --> 1:19:17.800 |
|
that would promote learning. |
|
|
|
1:19:17.800 --> 1:19:19.200 |
|
What does that mean for me? |
|
|
|
1:19:19.200 --> 1:19:22.800 |
|
For example, I've watched a documentary on flat Earth |
|
|
|
1:19:22.800 --> 1:19:23.640 |
|
theory, I guess. |
|
|
|
1:19:27.280 --> 1:19:28.240 |
|
I learned a lot. |
|
|
|
1:19:28.240 --> 1:19:29.800 |
|
I'm really glad I watched it. |
|
|
|
1:19:29.800 --> 1:19:32.560 |
|
It was a friend recommended it to me. |
|
|
|
1:19:32.560 --> 1:19:35.800 |
|
Because I don't have such an allergic reaction to crazy |
|
|
|
1:19:35.800 --> 1:19:37.640 |
|
people, as my fellow colleagues do. |
|
|
|
1:19:37.640 --> 1:19:40.360 |
|
But it was very eye opening. |
|
|
|
1:19:40.360 --> 1:19:42.120 |
|
And for others, it might not be. |
|
|
|
1:19:42.120 --> 1:19:45.560 |
|
From others, they might just get turned off from that, same |
|
|
|
1:19:45.560 --> 1:19:47.160 |
|
with Republican and Democrat. |
|
|
|
1:19:47.160 --> 1:19:50.200 |
|
And it's a non trivial problem. |
|
|
|
1:19:50.200 --> 1:19:52.880 |
|
And first of all, if it's done well, |
|
|
|
1:19:52.880 --> 1:19:56.560 |
|
I don't think it's something that wouldn't happen, |
|
|
|
1:19:56.560 --> 1:19:59.280 |
|
that YouTube wouldn't be promoting, |
|
|
|
1:19:59.280 --> 1:20:00.200 |
|
or Twitter wouldn't be. |
|
|
|
1:20:00.200 --> 1:20:02.280 |
|
It's just a really difficult problem, |
|
|
|
1:20:02.280 --> 1:20:05.520 |
|
how to give people control. |
|
|
|
1:20:05.520 --> 1:20:08.960 |
|
Well, it's mostly an interface design problem. |
|
|
|
1:20:08.960 --> 1:20:11.080 |
|
The way I see it, you want to create technology |
|
|
|
1:20:11.080 --> 1:20:16.400 |
|
that's like a mentor, or a coach, or an assistant, |
|
|
|
1:20:16.400 --> 1:20:20.520 |
|
so that it's not your boss. |
|
|
|
1:20:20.520 --> 1:20:22.560 |
|
You are in control of it. |
|
|
|
1:20:22.560 --> 1:20:25.760 |
|
You are telling it what to do for you. |
|
|
|
1:20:25.760 --> 1:20:27.840 |
|
And if you feel like it's manipulating you, |
|
|
|
1:20:27.840 --> 1:20:31.760 |
|
it's not actually doing what you want. |
|
|
|
1:20:31.760 --> 1:20:34.920 |
|
You should be able to switch to a different algorithm. |
|
|
|
1:20:34.920 --> 1:20:36.440 |
|
So that's fine tune control. |
|
|
|
1:20:36.440 --> 1:20:38.840 |
|
You kind of learn that you're trusting |
|
|
|
1:20:38.840 --> 1:20:40.080 |
|
the human collaboration. |
|
|
|
1:20:40.080 --> 1:20:41.920 |
|
I mean, that's how I see autonomous vehicles too, |
|
|
|
1:20:41.920 --> 1:20:44.480 |
|
is giving as much information as possible, |
|
|
|
1:20:44.480 --> 1:20:47.240 |
|
and you learn that dance yourself. |
|
|
|
1:20:47.240 --> 1:20:50.280 |
|
Yeah, Adobe, I don't know if you use Adobe product |
|
|
|
1:20:50.280 --> 1:20:52.280 |
|
for like Photoshop. |
|
|
|
1:20:52.280 --> 1:20:55.040 |
|
They're trying to see if they can inject YouTube |
|
|
|
1:20:55.040 --> 1:20:57.120 |
|
into their interface, but basically allow you |
|
|
|
1:20:57.120 --> 1:20:59.840 |
|
to show you all these videos, |
|
|
|
1:20:59.840 --> 1:21:03.320 |
|
that everybody's confused about what to do with features. |
|
|
|
1:21:03.320 --> 1:21:07.120 |
|
So basically teach people by linking to, |
|
|
|
1:21:07.120 --> 1:21:10.280 |
|
in that way, it's an assistant that uses videos |
|
|
|
1:21:10.280 --> 1:21:13.440 |
|
as a basic element of information. |
|
|
|
1:21:13.440 --> 1:21:18.240 |
|
Okay, so what practically should people do |
|
|
|
1:21:18.240 --> 1:21:24.000 |
|
to try to fight against abuses of these algorithms, |
|
|
|
1:21:24.000 --> 1:21:27.400 |
|
or algorithms that manipulate us? |
|
|
|
1:21:27.400 --> 1:21:29.280 |
|
Honestly, it's a very, very difficult problem, |
|
|
|
1:21:29.280 --> 1:21:32.800 |
|
because to start with, there is very little public awareness |
|
|
|
1:21:32.800 --> 1:21:35.040 |
|
of these issues. |
|
|
|
1:21:35.040 --> 1:21:38.520 |
|
Very few people would think there's anything wrong |
|
|
|
1:21:38.520 --> 1:21:39.720 |
|
with the unused algorithm, |
|
|
|
1:21:39.720 --> 1:21:42.040 |
|
even though there is actually something wrong already, |
|
|
|
1:21:42.040 --> 1:21:44.480 |
|
which is that it's trying to maximize engagement |
|
|
|
1:21:44.480 --> 1:21:49.880 |
|
most of the time, which has very negative side effects. |
|
|
|
1:21:49.880 --> 1:21:56.160 |
|
So ideally, so the very first thing is to stop |
|
|
|
1:21:56.160 --> 1:21:59.560 |
|
trying to purely maximize engagement, |
|
|
|
1:21:59.560 --> 1:22:06.560 |
|
try to propagate content based on popularity, right? |
|
|
|
1:22:06.560 --> 1:22:11.040 |
|
Instead, take into account the goals |
|
|
|
1:22:11.040 --> 1:22:13.560 |
|
and the profiles of each user. |
|
|
|
1:22:13.560 --> 1:22:16.920 |
|
So you will be, one example is, for instance, |
|
|
|
1:22:16.920 --> 1:22:20.800 |
|
when I look at topic recommendations on Twitter, |
|
|
|
1:22:20.800 --> 1:22:24.480 |
|
it's like, you know, they have this news tab |
|
|
|
1:22:24.480 --> 1:22:25.480 |
|
with switch recommendations. |
|
|
|
1:22:25.480 --> 1:22:28.480 |
|
It's always the worst coverage, |
|
|
|
1:22:28.480 --> 1:22:30.360 |
|
because it's content that appeals |
|
|
|
1:22:30.360 --> 1:22:34.080 |
|
to the smallest common denominator |
|
|
|
1:22:34.080 --> 1:22:37.080 |
|
to all Twitter users, because they're trying to optimize. |
|
|
|
1:22:37.080 --> 1:22:39.040 |
|
They're purely trying to optimize popularity. |
|
|
|
1:22:39.040 --> 1:22:41.320 |
|
They're purely trying to optimize engagement. |
|
|
|
1:22:41.320 --> 1:22:42.960 |
|
But that's not what I want. |
|
|
|
1:22:42.960 --> 1:22:46.080 |
|
So they should put me in control of some setting |
|
|
|
1:22:46.080 --> 1:22:50.360 |
|
so that I define what's the objective function |
|
|
|
1:22:50.360 --> 1:22:52.200 |
|
that Twitter is going to be following |
|
|
|
1:22:52.200 --> 1:22:54.120 |
|
to show me this content. |
|
|
|
1:22:54.120 --> 1:22:57.360 |
|
And honestly, so this is all about interface design. |
|
|
|
1:22:57.360 --> 1:22:59.440 |
|
And we are not, it's not realistic |
|
|
|
1:22:59.440 --> 1:23:01.760 |
|
to give users control of a bunch of knobs |
|
|
|
1:23:01.760 --> 1:23:03.400 |
|
that define algorithm. |
|
|
|
1:23:03.400 --> 1:23:06.760 |
|
Instead, we should purely put them in charge |
|
|
|
1:23:06.760 --> 1:23:09.400 |
|
of defining the objective function. |
|
|
|
1:23:09.400 --> 1:23:13.240 |
|
Like, let the user tell us what they want to achieve, |
|
|
|
1:23:13.240 --> 1:23:15.280 |
|
how they want this algorithm to impact their lives. |
|
|
|
1:23:15.280 --> 1:23:16.680 |
|
So do you think it is that, |
|
|
|
1:23:16.680 --> 1:23:19.360 |
|
or do they provide individual article by article |
|
|
|
1:23:19.360 --> 1:23:21.600 |
|
reward structure where you give a signal, |
|
|
|
1:23:21.600 --> 1:23:24.720 |
|
I'm glad I saw this, or I'm glad I didn't? |
|
|
|
1:23:24.720 --> 1:23:28.480 |
|
So like a Spotify type feedback mechanism, |
|
|
|
1:23:28.480 --> 1:23:30.680 |
|
it works to some extent. |
|
|
|
1:23:30.680 --> 1:23:32.000 |
|
I'm kind of skeptical about it |
|
|
|
1:23:32.000 --> 1:23:34.880 |
|
because the only way the algorithm, |
|
|
|
1:23:34.880 --> 1:23:39.120 |
|
the algorithm will attempt to relate your choices |
|
|
|
1:23:39.120 --> 1:23:41.040 |
|
with the choices of everyone else, |
|
|
|
1:23:41.040 --> 1:23:45.000 |
|
which might, you know, if you have an average profile |
|
|
|
1:23:45.000 --> 1:23:47.880 |
|
that works fine, I'm sure Spotify accommodations work fine |
|
|
|
1:23:47.880 --> 1:23:49.560 |
|
if you just like mainstream stuff. |
|
|
|
1:23:49.560 --> 1:23:53.960 |
|
If you don't, it can be, it's not optimal at all actually. |
|
|
|
1:23:53.960 --> 1:23:56.040 |
|
It'll be in an efficient search |
|
|
|
1:23:56.040 --> 1:24:00.800 |
|
for the part of the Spotify world that represents you. |
|
|
|
1:24:00.800 --> 1:24:02.960 |
|
So it's a tough problem, |
|
|
|
1:24:02.960 --> 1:24:07.960 |
|
but do note that even a feedback system |
|
|
|
1:24:07.960 --> 1:24:10.880 |
|
like what Spotify has does not give me control |
|
|
|
1:24:10.880 --> 1:24:15.000 |
|
over what the algorithm is trying to optimize for. |
|
|
|
1:24:16.320 --> 1:24:19.360 |
|
Well, public awareness, which is what we're doing now, |
|
|
|
1:24:19.360 --> 1:24:21.360 |
|
is a good place to start. |
|
|
|
1:24:21.360 --> 1:24:25.960 |
|
Do you have concerns about longterm existential threats |
|
|
|
1:24:25.960 --> 1:24:27.360 |
|
of artificial intelligence? |
|
|
|
1:24:28.280 --> 1:24:31.040 |
|
Well, as I was saying, |
|
|
|
1:24:31.040 --> 1:24:33.360 |
|
our world is increasingly made of information. |
|
|
|
1:24:33.360 --> 1:24:36.240 |
|
AI algorithms are increasingly going to be our interface |
|
|
|
1:24:36.240 --> 1:24:37.880 |
|
to this world of information, |
|
|
|
1:24:37.880 --> 1:24:41.480 |
|
and somebody will be in control of these algorithms. |
|
|
|
1:24:41.480 --> 1:24:45.920 |
|
And that puts us in any kind of a bad situation, right? |
|
|
|
1:24:45.920 --> 1:24:46.880 |
|
It has risks. |
|
|
|
1:24:46.880 --> 1:24:50.840 |
|
It has risks coming from potentially large companies |
|
|
|
1:24:50.840 --> 1:24:53.760 |
|
wanting to optimize their own goals, |
|
|
|
1:24:53.760 --> 1:24:55.960 |
|
maybe profit, maybe something else. |
|
|
|
1:24:55.960 --> 1:25:00.720 |
|
Also from governments who might want to use these algorithms |
|
|
|
1:25:00.720 --> 1:25:03.520 |
|
as a means of control of the population. |
|
|
|
1:25:03.520 --> 1:25:05.000 |
|
Do you think there's existential threat |
|
|
|
1:25:05.000 --> 1:25:06.320 |
|
that could arise from that? |
|
|
|
1:25:06.320 --> 1:25:09.120 |
|
So existential threat. |
|
|
|
1:25:09.120 --> 1:25:13.240 |
|
So maybe you're referring to the singularity narrative |
|
|
|
1:25:13.240 --> 1:25:15.560 |
|
where robots just take over. |
|
|
|
1:25:15.560 --> 1:25:18.320 |
|
Well, I don't, I'm not terminating robots, |
|
|
|
1:25:18.320 --> 1:25:21.000 |
|
and I don't believe it has to be a singularity. |
|
|
|
1:25:21.000 --> 1:25:24.800 |
|
We're just talking to, just like you said, |
|
|
|
1:25:24.800 --> 1:25:27.920 |
|
the algorithm controlling masses of populations. |
|
|
|
1:25:28.920 --> 1:25:31.120 |
|
The existential threat being, |
|
|
|
1:25:32.640 --> 1:25:36.760 |
|
hurt ourselves much like a nuclear war would hurt ourselves. |
|
|
|
1:25:36.760 --> 1:25:37.600 |
|
That kind of thing. |
|
|
|
1:25:37.600 --> 1:25:39.480 |
|
I don't think that requires a singularity. |
|
|
|
1:25:39.480 --> 1:25:42.560 |
|
That requires a loss of control over AI algorithm. |
|
|
|
1:25:42.560 --> 1:25:43.560 |
|
Yes. |
|
|
|
1:25:43.560 --> 1:25:47.000 |
|
So I do agree there are concerning trends. |
|
|
|
1:25:47.000 --> 1:25:52.000 |
|
Honestly, I wouldn't want to make any longterm predictions. |
|
|
|
1:25:52.960 --> 1:25:56.000 |
|
I don't think today we really have the capability |
|
|
|
1:25:56.000 --> 1:25:58.560 |
|
to see what the dangers of AI |
|
|
|
1:25:58.560 --> 1:26:01.360 |
|
are going to be in 50 years, in 100 years. |
|
|
|
1:26:01.360 --> 1:26:04.800 |
|
I do see that we are already faced |
|
|
|
1:26:04.800 --> 1:26:08.840 |
|
with concrete and present dangers |
|
|
|
1:26:08.840 --> 1:26:11.560 |
|
surrounding the negative side effects |
|
|
|
1:26:11.560 --> 1:26:14.960 |
|
of content recombination systems, of newsfeed algorithms |
|
|
|
1:26:14.960 --> 1:26:17.640 |
|
concerning algorithmic bias as well. |
|
|
|
1:26:18.640 --> 1:26:21.200 |
|
So we are delegating more and more |
|
|
|
1:26:22.240 --> 1:26:25.080 |
|
decision processes to algorithms. |
|
|
|
1:26:25.080 --> 1:26:26.760 |
|
Some of these algorithms are uncrafted, |
|
|
|
1:26:26.760 --> 1:26:29.360 |
|
some are learned from data, |
|
|
|
1:26:29.360 --> 1:26:31.920 |
|
but we are delegating control. |
|
|
|
1:26:32.920 --> 1:26:36.280 |
|
Sometimes it's a good thing, sometimes not so much. |
|
|
|
1:26:36.280 --> 1:26:39.480 |
|
And there is in general very little supervision |
|
|
|
1:26:39.480 --> 1:26:41.000 |
|
of this process, right? |
|
|
|
1:26:41.000 --> 1:26:45.400 |
|
So we are still in this period of very fast change, |
|
|
|
1:26:45.400 --> 1:26:50.400 |
|
even chaos, where society is restructuring itself, |
|
|
|
1:26:50.920 --> 1:26:53.160 |
|
turning into an information society, |
|
|
|
1:26:53.160 --> 1:26:54.520 |
|
which itself is turning into |
|
|
|
1:26:54.520 --> 1:26:58.360 |
|
an increasingly automated information passing society. |
|
|
|
1:26:58.360 --> 1:27:02.520 |
|
And well, yeah, I think the best we can do today |
|
|
|
1:27:02.520 --> 1:27:06.040 |
|
is try to raise awareness around some of these issues. |
|
|
|
1:27:06.040 --> 1:27:07.680 |
|
And I think we're actually making good progress. |
|
|
|
1:27:07.680 --> 1:27:11.720 |
|
If you look at algorithmic bias, for instance, |
|
|
|
1:27:12.760 --> 1:27:14.760 |
|
three years ago, even two years ago, |
|
|
|
1:27:14.760 --> 1:27:17.040 |
|
very, very few people were talking about it. |
|
|
|
1:27:17.040 --> 1:27:20.320 |
|
And now all the big companies are talking about it. |
|
|
|
1:27:20.320 --> 1:27:22.360 |
|
They are often not in a very serious way, |
|
|
|
1:27:22.360 --> 1:27:24.560 |
|
but at least it is part of the public discourse. |
|
|
|
1:27:24.560 --> 1:27:27.080 |
|
You see people in Congress talking about it. |
|
|
|
1:27:27.080 --> 1:27:31.960 |
|
And it all started from raising awareness. |
|
|
|
1:27:31.960 --> 1:27:32.800 |
|
Right. |
|
|
|
1:27:32.800 --> 1:27:36.080 |
|
So in terms of alignment problem, |
|
|
|
1:27:36.080 --> 1:27:39.400 |
|
trying to teach as we allow algorithms, |
|
|
|
1:27:39.400 --> 1:27:41.520 |
|
just even recommender systems on Twitter, |
|
|
|
1:27:43.640 --> 1:27:47.080 |
|
encoding human values and morals, |
|
|
|
1:27:48.280 --> 1:27:50.200 |
|
decisions that touch on ethics, |
|
|
|
1:27:50.200 --> 1:27:52.600 |
|
how hard do you think that problem is? |
|
|
|
1:27:52.600 --> 1:27:57.240 |
|
How do we have lost functions in neural networks |
|
|
|
1:27:57.240 --> 1:27:58.640 |
|
that have some component, |
|
|
|
1:27:58.640 --> 1:28:01.080 |
|
some fuzzy components of human morals? |
|
|
|
1:28:01.080 --> 1:28:06.080 |
|
Well, I think this is really all about objective function engineering, |
|
|
|
1:28:06.080 --> 1:28:10.520 |
|
which is probably going to be increasingly a topic of concern in the future. |
|
|
|
1:28:10.520 --> 1:28:14.640 |
|
Like for now, we're just using very naive loss functions |
|
|
|
1:28:14.640 --> 1:28:17.760 |
|
because the hard part is not actually what you're trying to minimize. |
|
|
|
1:28:17.760 --> 1:28:19.040 |
|
It's everything else. |
|
|
|
1:28:19.040 --> 1:28:22.840 |
|
But as the everything else is going to be increasingly automated, |
|
|
|
1:28:22.840 --> 1:28:27.040 |
|
we're going to be focusing our human attention |
|
|
|
1:28:27.040 --> 1:28:30.240 |
|
on increasingly high level components, |
|
|
|
1:28:30.240 --> 1:28:32.680 |
|
like what's actually driving the whole learning system, |
|
|
|
1:28:32.680 --> 1:28:33.960 |
|
like the objective function. |
|
|
|
1:28:33.960 --> 1:28:36.920 |
|
So loss function engineering is going to be, |
|
|
|
1:28:36.920 --> 1:28:40.640 |
|
loss function engineer is probably going to be a job title in the future. |
|
|
|
1:28:40.640 --> 1:28:44.520 |
|
And then the tooling you're creating with Keras essentially |
|
|
|
1:28:44.520 --> 1:28:47.040 |
|
takes care of all the details underneath. |
|
|
|
1:28:47.040 --> 1:28:52.720 |
|
And basically the human expert is needed for exactly that. |
|
|
|
1:28:52.720 --> 1:28:53.920 |
|
That's the idea. |
|
|
|
1:28:53.920 --> 1:28:57.640 |
|
Keras is the interface between the data you're collecting |
|
|
|
1:28:57.640 --> 1:28:59.080 |
|
and the business goals. |
|
|
|
1:28:59.080 --> 1:29:03.480 |
|
And your job as an engineer is going to be to express your business goals |
|
|
|
1:29:03.480 --> 1:29:06.720 |
|
and your understanding of your business or your product, |
|
|
|
1:29:06.720 --> 1:29:11.840 |
|
your system as a kind of loss function or a kind of set of constraints. |
|
|
|
1:29:11.840 --> 1:29:19.480 |
|
Does the possibility of creating an AGI system excite you or scare you or bore you? |
|
|
|
1:29:19.480 --> 1:29:22.080 |
|
So intelligence can never really be general. |
|
|
|
1:29:22.080 --> 1:29:26.400 |
|
You know, at best it can have some degree of generality like human intelligence. |
|
|
|
1:29:26.400 --> 1:29:30.640 |
|
It also always has some specialization in the same way that human intelligence |
|
|
|
1:29:30.640 --> 1:29:33.440 |
|
is specialized in a certain category of problems, |
|
|
|
1:29:33.440 --> 1:29:35.440 |
|
is specialized in the human experience. |
|
|
|
1:29:35.440 --> 1:29:37.280 |
|
And when people talk about AGI, |
|
|
|
1:29:37.280 --> 1:29:42.520 |
|
I'm never quite sure if they're talking about very, very smart AI, |
|
|
|
1:29:42.520 --> 1:29:45.080 |
|
so smart that it's even smarter than humans, |
|
|
|
1:29:45.080 --> 1:29:48.000 |
|
or they're talking about human like intelligence, |
|
|
|
1:29:48.000 --> 1:29:49.680 |
|
because these are different things. |
|
|
|
1:29:49.680 --> 1:29:54.760 |
|
Let's say, presumably I'm oppressing you today with my humanness. |
|
|
|
1:29:54.760 --> 1:29:59.240 |
|
So imagine that I was in fact a robot. |
|
|
|
1:29:59.240 --> 1:30:01.920 |
|
So what does that mean? |
|
|
|
1:30:01.920 --> 1:30:04.920 |
|
That I'm impressing you with natural language processing. |
|
|
|
1:30:04.920 --> 1:30:07.840 |
|
Maybe if you weren't able to see me, maybe this is a phone call. |
|
|
|
1:30:07.840 --> 1:30:10.000 |
|
So that kind of system. |
|
|
|
1:30:10.000 --> 1:30:11.120 |
|
Companion. |
|
|
|
1:30:11.120 --> 1:30:15.040 |
|
So that's very much about building human like AI. |
|
|
|
1:30:15.040 --> 1:30:18.200 |
|
And you're asking me, you know, is this an exciting perspective? |
|
|
|
1:30:18.200 --> 1:30:19.440 |
|
Yes. |
|
|
|
1:30:19.440 --> 1:30:21.760 |
|
I think so, yes. |
|
|
|
1:30:21.760 --> 1:30:28.000 |
|
Not so much because of what artificial human like intelligence could do, |
|
|
|
1:30:28.000 --> 1:30:30.880 |
|
but, you know, from an intellectual perspective, |
|
|
|
1:30:30.880 --> 1:30:34.120 |
|
I think if you could build truly human like intelligence, |
|
|
|
1:30:34.120 --> 1:30:37.240 |
|
that means you could actually understand human intelligence, |
|
|
|
1:30:37.240 --> 1:30:39.880 |
|
which is fascinating, right? |
|
|
|
1:30:39.880 --> 1:30:42.680 |
|
Human like intelligence is going to require emotions. |
|
|
|
1:30:42.680 --> 1:30:44.400 |
|
It's going to require consciousness, |
|
|
|
1:30:44.400 --> 1:30:49.720 |
|
which is not things that would normally be required by an intelligent system. |
|
|
|
1:30:49.720 --> 1:30:53.160 |
|
If you look at, you know, we were mentioning earlier like science |
|
|
|
1:30:53.160 --> 1:30:59.600 |
|
as a superhuman problem solving agent or system, |
|
|
|
1:30:59.600 --> 1:31:02.120 |
|
it does not have consciousness, it doesn't have emotions. |
|
|
|
1:31:02.120 --> 1:31:04.320 |
|
In general, so emotions, |
|
|
|
1:31:04.320 --> 1:31:07.640 |
|
I see consciousness as being on the same spectrum as emotions. |
|
|
|
1:31:07.640 --> 1:31:12.280 |
|
It is a component of the subjective experience |
|
|
|
1:31:12.280 --> 1:31:18.800 |
|
that is meant very much to guide behavior generation, right? |
|
|
|
1:31:18.800 --> 1:31:20.800 |
|
It's meant to guide your behavior. |
|
|
|
1:31:20.800 --> 1:31:24.520 |
|
In general, human intelligence and animal intelligence |
|
|
|
1:31:24.520 --> 1:31:29.280 |
|
has evolved for the purpose of behavior generation, right? |
|
|
|
1:31:29.280 --> 1:31:30.680 |
|
Including in a social context. |
|
|
|
1:31:30.680 --> 1:31:32.480 |
|
So that's why we actually need emotions. |
|
|
|
1:31:32.480 --> 1:31:34.920 |
|
That's why we need consciousness. |
|
|
|
1:31:34.920 --> 1:31:38.360 |
|
An artificial intelligence system developed in a different context |
|
|
|
1:31:38.360 --> 1:31:42.800 |
|
may well never need them, may well never be conscious like science. |
|
|
|
1:31:42.800 --> 1:31:47.960 |
|
Well, on that point, I would argue it's possible to imagine |
|
|
|
1:31:47.960 --> 1:31:51.480 |
|
that there's echoes of consciousness in science |
|
|
|
1:31:51.480 --> 1:31:55.480 |
|
when viewed as an organism, that science is consciousness. |
|
|
|
1:31:55.480 --> 1:31:59.160 |
|
So, I mean, how would you go about testing this hypothesis? |
|
|
|
1:31:59.160 --> 1:32:07.000 |
|
How do you probe the subjective experience of an abstract system like science? |
|
|
|
1:32:07.000 --> 1:32:10.400 |
|
Well, the point of probing any subjective experience is impossible |
|
|
|
1:32:10.400 --> 1:32:13.200 |
|
because I'm not science, I'm Lex. |
|
|
|
1:32:13.200 --> 1:32:20.520 |
|
So I can't probe another entity, it's no more than bacteria on my skin. |
|
|
|
1:32:20.520 --> 1:32:24.160 |
|
You're Lex, I can ask you questions about your subjective experience |
|
|
|
1:32:24.160 --> 1:32:28.440 |
|
and you can answer me, and that's how I know you're conscious. |
|
|
|
1:32:28.440 --> 1:32:31.840 |
|
Yes, but that's because we speak the same language. |
|
|
|
1:32:31.840 --> 1:32:35.520 |
|
You perhaps, we have to speak the language of science in order to ask it. |
|
|
|
1:32:35.520 --> 1:32:40.320 |
|
Honestly, I don't think consciousness, just like emotions of pain and pleasure, |
|
|
|
1:32:40.320 --> 1:32:44.160 |
|
is not something that inevitably arises |
|
|
|
1:32:44.160 --> 1:32:47.920 |
|
from any sort of sufficiently intelligent information processing. |
|
|
|
1:32:47.920 --> 1:32:53.920 |
|
It is a feature of the mind, and if you've not implemented it explicitly, it is not there. |
|
|
|
1:32:53.920 --> 1:32:58.960 |
|
So you think it's an emergent feature of a particular architecture. |
|
|
|
1:32:58.960 --> 1:33:00.320 |
|
So do you think... |
|
|
|
1:33:00.320 --> 1:33:02.000 |
|
It's a feature in the same sense. |
|
|
|
1:33:02.000 --> 1:33:08.240 |
|
So, again, the subjective experience is all about guiding behavior. |
|
|
|
1:33:08.240 --> 1:33:15.120 |
|
If the problems you're trying to solve don't really involve an embodied agent, |
|
|
|
1:33:15.120 --> 1:33:19.520 |
|
maybe in a social context, generating behavior and pursuing goals like this. |
|
|
|
1:33:19.520 --> 1:33:22.160 |
|
And if you look at science, that's not really what's happening. |
|
|
|
1:33:22.160 --> 1:33:27.920 |
|
Even though it is, it is a form of artificial AI, artificial intelligence, |
|
|
|
1:33:27.920 --> 1:33:31.920 |
|
in the sense that it is solving problems, it is accumulating knowledge, |
|
|
|
1:33:31.920 --> 1:33:35.040 |
|
accumulating solutions and so on. |
|
|
|
1:33:35.040 --> 1:33:39.440 |
|
So if you're not explicitly implementing a subjective experience, |
|
|
|
1:33:39.440 --> 1:33:44.000 |
|
implementing certain emotions and implementing consciousness, |
|
|
|
1:33:44.000 --> 1:33:47.360 |
|
it's not going to just spontaneously emerge. |
|
|
|
1:33:47.360 --> 1:33:48.080 |
|
Yeah. |
|
|
|
1:33:48.080 --> 1:33:53.200 |
|
But so for a system like, human like intelligence system that has consciousness, |
|
|
|
1:33:53.200 --> 1:33:55.840 |
|
do you think it needs to have a body? |
|
|
|
1:33:55.840 --> 1:33:56.720 |
|
Yes, definitely. |
|
|
|
1:33:56.720 --> 1:33:59.600 |
|
I mean, it doesn't have to be a physical body, right? |
|
|
|
1:33:59.600 --> 1:34:03.440 |
|
And there's not that much difference between a realistic simulation in the real world. |
|
|
|
1:34:03.440 --> 1:34:06.400 |
|
So there has to be something you have to preserve kind of thing. |
|
|
|
1:34:06.400 --> 1:34:11.840 |
|
Yes, but human like intelligence can only arise in a human like context. |
|
|
|
1:34:11.840 --> 1:34:16.800 |
|
Intelligence needs other humans in order for you to demonstrate |
|
|
|
1:34:16.800 --> 1:34:19.040 |
|
that you have human like intelligence, essentially. |
|
|
|
1:34:19.040 --> 1:34:19.540 |
|
Yes. |
|
|
|
1:34:20.320 --> 1:34:28.080 |
|
So what kind of tests and demonstration would be sufficient for you |
|
|
|
1:34:28.080 --> 1:34:30.960 |
|
to demonstrate human like intelligence? |
|
|
|
1:34:30.960 --> 1:34:31.360 |
|
Yeah. |
|
|
|
1:34:31.360 --> 1:34:35.600 |
|
Just out of curiosity, you've talked about in terms of theorem proving |
|
|
|
1:34:35.600 --> 1:34:38.000 |
|
and program synthesis, I think you've written about |
|
|
|
1:34:38.000 --> 1:34:40.480 |
|
that there's no good benchmarks for this. |
|
|
|
1:34:40.480 --> 1:34:40.720 |
|
Yeah. |
|
|
|
1:34:40.720 --> 1:34:42.000 |
|
That's one of the problems. |
|
|
|
1:34:42.000 --> 1:34:46.320 |
|
So let's talk program synthesis. |
|
|
|
1:34:46.320 --> 1:34:47.760 |
|
So what do you imagine is a good... |
|
|
|
1:34:48.800 --> 1:34:51.360 |
|
I think it's related questions for human like intelligence |
|
|
|
1:34:51.360 --> 1:34:52.560 |
|
and for program synthesis. |
|
|
|
1:34:53.360 --> 1:34:56.080 |
|
What's a good benchmark for either or both? |
|
|
|
1:34:56.080 --> 1:34:56.480 |
|
Right. |
|
|
|
1:34:56.480 --> 1:34:59.200 |
|
So I mean, you're actually asking two questions, |
|
|
|
1:34:59.200 --> 1:35:02.480 |
|
which is one is about quantifying intelligence |
|
|
|
1:35:02.480 --> 1:35:06.880 |
|
and comparing the intelligence of an artificial system |
|
|
|
1:35:06.880 --> 1:35:08.480 |
|
to the intelligence for human. |
|
|
|
1:35:08.480 --> 1:35:13.440 |
|
And the other is about the degree to which this intelligence is human like. |
|
|
|
1:35:13.440 --> 1:35:15.120 |
|
It's actually two different questions. |
|
|
|
1:35:16.560 --> 1:35:18.960 |
|
So you mentioned earlier the Turing test. |
|
|
|
1:35:19.680 --> 1:35:23.200 |
|
Well, I actually don't like the Turing test because it's very lazy. |
|
|
|
1:35:23.200 --> 1:35:28.720 |
|
It's all about completely bypassing the problem of defining and measuring intelligence |
|
|
|
1:35:28.720 --> 1:35:34.160 |
|
and instead delegating to a human judge or a panel of human judges. |
|
|
|
1:35:34.160 --> 1:35:37.120 |
|
So it's a total copout, right? |
|
|
|
1:35:38.160 --> 1:35:43.200 |
|
If you want to measure how human like an agent is, |
|
|
|
1:35:43.760 --> 1:35:46.640 |
|
I think you have to make it interact with other humans. |
|
|
|
1:35:47.600 --> 1:35:53.760 |
|
Maybe it's not necessarily a good idea to have these other humans be the judges. |
|
|
|
1:35:53.760 --> 1:35:59.280 |
|
Maybe you should just observe behavior and compare it to what a human would actually have done. |
|
|
|
1:36:00.560 --> 1:36:05.120 |
|
When it comes to measuring how smart, how clever an agent is |
|
|
|
1:36:05.120 --> 1:36:11.120 |
|
and comparing that to the degree of human intelligence. |
|
|
|
1:36:11.120 --> 1:36:12.960 |
|
So we're already talking about two things, right? |
|
|
|
1:36:13.520 --> 1:36:20.320 |
|
The degree, kind of like the magnitude of an intelligence and its direction, right? |
|
|
|
1:36:20.320 --> 1:36:23.280 |
|
Like the norm of a vector and its direction. |
|
|
|
1:36:23.280 --> 1:36:32.000 |
|
And the direction is like human likeness and the magnitude, the norm is intelligence. |
|
|
|
1:36:32.720 --> 1:36:34.080 |
|
You could call it intelligence, right? |
|
|
|
1:36:34.080 --> 1:36:41.040 |
|
So the direction, your sense, the space of directions that are human like is very narrow. |
|
|
|
1:36:41.040 --> 1:36:41.200 |
|
Yeah. |
|
|
|
1:36:42.240 --> 1:36:48.880 |
|
So the way you would measure the magnitude of intelligence in a system |
|
|
|
1:36:48.880 --> 1:36:54.640 |
|
in a way that also enables you to compare it to that of a human. |
|
|
|
1:36:54.640 --> 1:36:59.200 |
|
Well, if you look at different benchmarks for intelligence today, |
|
|
|
1:36:59.200 --> 1:37:04.160 |
|
they're all too focused on skill at a given task. |
|
|
|
1:37:04.160 --> 1:37:08.720 |
|
Like skill at playing chess, skill at playing Go, skill at playing Dota. |
|
|
|
1:37:10.720 --> 1:37:15.600 |
|
And I think that's not the right way to go about it because you can always |
|
|
|
1:37:15.600 --> 1:37:18.240 |
|
beat a human at one specific task. |
|
|
|
1:37:19.200 --> 1:37:23.920 |
|
The reason why our skill at playing Go or juggling or anything is impressive |
|
|
|
1:37:23.920 --> 1:37:28.400 |
|
is because we are expressing this skill within a certain set of constraints. |
|
|
|
1:37:28.400 --> 1:37:32.320 |
|
If you remove the constraints, the constraints that we have one lifetime, |
|
|
|
1:37:32.320 --> 1:37:36.080 |
|
that we have this body and so on, if you remove the context, |
|
|
|
1:37:36.080 --> 1:37:40.480 |
|
if you have unlimited string data, if you can have access to, you know, |
|
|
|
1:37:40.480 --> 1:37:44.640 |
|
for instance, if you look at juggling, if you have no restriction on the hardware, |
|
|
|
1:37:44.640 --> 1:37:48.400 |
|
then achieving arbitrary levels of skill is not very interesting |
|
|
|
1:37:48.400 --> 1:37:52.400 |
|
and says nothing about the amount of intelligence you've achieved. |
|
|
|
1:37:52.400 --> 1:37:57.440 |
|
So if you want to measure intelligence, you need to rigorously define what |
|
|
|
1:37:57.440 --> 1:38:02.960 |
|
intelligence is, which in itself, you know, it's a very challenging problem. |
|
|
|
1:38:02.960 --> 1:38:04.320 |
|
And do you think that's possible? |
|
|
|
1:38:04.320 --> 1:38:06.000 |
|
To define intelligence? Yes, absolutely. |
|
|
|
1:38:06.000 --> 1:38:09.760 |
|
I mean, you can provide, many people have provided, you know, some definition. |
|
|
|
1:38:10.560 --> 1:38:12.000 |
|
I have my own definition. |
|
|
|
1:38:12.000 --> 1:38:13.440 |
|
Where does your definition begin? |
|
|
|
1:38:13.440 --> 1:38:16.240 |
|
Where does your definition begin if it doesn't end? |
|
|
|
1:38:16.240 --> 1:38:21.680 |
|
Well, I think intelligence is essentially the efficiency |
|
|
|
1:38:22.320 --> 1:38:29.760 |
|
with which you turn experience into generalizable programs. |
|
|
|
1:38:29.760 --> 1:38:32.800 |
|
So what that means is it's the efficiency with which |
|
|
|
1:38:32.800 --> 1:38:36.720 |
|
you turn a sampling of experience space into |
|
|
|
1:38:36.720 --> 1:38:46.000 |
|
the ability to process a larger chunk of experience space. |
|
|
|
1:38:46.000 --> 1:38:52.560 |
|
So measuring skill can be one proxy across many different tasks, |
|
|
|
1:38:52.560 --> 1:38:54.480 |
|
can be one proxy for measuring intelligence. |
|
|
|
1:38:54.480 --> 1:38:58.720 |
|
But if you want to only measure skill, you should control for two things. |
|
|
|
1:38:58.720 --> 1:39:04.960 |
|
You should control for the amount of experience that your system has |
|
|
|
1:39:04.960 --> 1:39:08.080 |
|
and the priors that your system has. |
|
|
|
1:39:08.080 --> 1:39:13.120 |
|
But if you look at two agents and you give them the same priors |
|
|
|
1:39:13.120 --> 1:39:16.160 |
|
and you give them the same amount of experience, |
|
|
|
1:39:16.160 --> 1:39:21.360 |
|
there is one of the agents that is going to learn programs, |
|
|
|
1:39:21.360 --> 1:39:25.440 |
|
representations, something, a model that will perform well |
|
|
|
1:39:25.440 --> 1:39:28.720 |
|
on the larger chunk of experience space than the other. |
|
|
|
1:39:28.720 --> 1:39:30.960 |
|
And that is the smaller agent. |
|
|
|
1:39:30.960 --> 1:39:36.960 |
|
Yeah. So if you fix the experience, which generate better programs, |
|
|
|
1:39:37.680 --> 1:39:39.520 |
|
better meaning more generalizable. |
|
|
|
1:39:39.520 --> 1:39:40.560 |
|
That's really interesting. |
|
|
|
1:39:40.560 --> 1:39:42.400 |
|
That's a very nice, clean definition of... |
|
|
|
1:39:42.400 --> 1:39:47.280 |
|
Oh, by the way, in this definition, it is already very obvious |
|
|
|
1:39:47.280 --> 1:39:49.440 |
|
that intelligence has to be specialized |
|
|
|
1:39:49.440 --> 1:39:51.680 |
|
because you're talking about experience space |
|
|
|
1:39:51.680 --> 1:39:54.080 |
|
and you're talking about segments of experience space. |
|
|
|
1:39:54.080 --> 1:39:57.200 |
|
You're talking about priors and you're talking about experience. |
|
|
|
1:39:57.200 --> 1:40:02.480 |
|
All of these things define the context in which intelligence emerges. |
|
|
|
1:40:04.480 --> 1:40:08.640 |
|
And you can never look at the totality of experience space, right? |
|
|
|
1:40:09.760 --> 1:40:12.160 |
|
So intelligence has to be specialized. |
|
|
|
1:40:12.160 --> 1:40:14.960 |
|
But it can be sufficiently large, the experience space, |
|
|
|
1:40:14.960 --> 1:40:16.080 |
|
even though it's specialized. |
|
|
|
1:40:16.080 --> 1:40:19.120 |
|
There's a certain point when the experience space is large enough |
|
|
|
1:40:19.120 --> 1:40:21.440 |
|
to where it might as well be general. |
|
|
|
1:40:22.000 --> 1:40:23.920 |
|
It feels general. It looks general. |
|
|
|
1:40:23.920 --> 1:40:25.680 |
|
Sure. I mean, it's very relative. |
|
|
|
1:40:25.680 --> 1:40:29.360 |
|
Like, for instance, many people would say human intelligence is general. |
|
|
|
1:40:29.360 --> 1:40:31.200 |
|
In fact, it is quite specialized. |
|
|
|
1:40:32.800 --> 1:40:37.120 |
|
We can definitely build systems that start from the same innate priors |
|
|
|
1:40:37.120 --> 1:40:39.120 |
|
as what humans have at birth. |
|
|
|
1:40:39.120 --> 1:40:42.320 |
|
Because we already understand fairly well |
|
|
|
1:40:42.320 --> 1:40:44.480 |
|
what sort of priors we have as humans. |
|
|
|
1:40:44.480 --> 1:40:46.080 |
|
Like many people have worked on this problem. |
|
|
|
1:40:46.800 --> 1:40:51.040 |
|
Most notably, Elisabeth Spelke from Harvard. |
|
|
|
1:40:51.040 --> 1:40:52.240 |
|
I don't know if you know her. |
|
|
|
1:40:52.240 --> 1:40:56.000 |
|
She's worked a lot on what she calls core knowledge. |
|
|
|
1:40:56.000 --> 1:41:00.640 |
|
And it is very much about trying to determine and describe |
|
|
|
1:41:00.640 --> 1:41:02.320 |
|
what priors we are born with. |
|
|
|
1:41:02.320 --> 1:41:04.720 |
|
Like language skills and so on, all that kind of stuff. |
|
|
|
1:41:04.720 --> 1:41:05.220 |
|
Exactly. |
|
|
|
1:41:06.880 --> 1:41:11.440 |
|
So we have some pretty good understanding of what priors we are born with. |
|
|
|
1:41:11.440 --> 1:41:12.560 |
|
So we could... |
|
|
|
1:41:13.760 --> 1:41:17.760 |
|
So I've actually been working on a benchmark for the past couple years, |
|
|
|
1:41:17.760 --> 1:41:18.640 |
|
you know, on and off. |
|
|
|
1:41:18.640 --> 1:41:20.480 |
|
I hope to be able to release it at some point. |
|
|
|
1:41:20.480 --> 1:41:21.760 |
|
That's exciting. |
|
|
|
1:41:21.760 --> 1:41:25.680 |
|
The idea is to measure the intelligence of systems |
|
|
|
1:41:26.800 --> 1:41:28.640 |
|
by countering for priors, |
|
|
|
1:41:28.640 --> 1:41:30.480 |
|
countering for amount of experience, |
|
|
|
1:41:30.480 --> 1:41:34.800 |
|
and by assuming the same priors as what humans are born with. |
|
|
|
1:41:34.800 --> 1:41:39.520 |
|
So that you can actually compare these scores to human intelligence. |
|
|
|
1:41:39.520 --> 1:41:43.280 |
|
You can actually have humans pass the same test in a way that's fair. |
|
|
|
1:41:43.280 --> 1:41:52.320 |
|
Yeah. And so importantly, such a benchmark should be such that any amount |
|
|
|
1:41:52.960 --> 1:41:55.920 |
|
of practicing does not increase your score. |
|
|
|
1:41:56.480 --> 1:42:00.560 |
|
So try to picture a game where no matter how much you play this game, |
|
|
|
1:42:01.600 --> 1:42:05.040 |
|
that does not change your skill at the game. |
|
|
|
1:42:05.040 --> 1:42:05.920 |
|
Can you picture that? |
|
|
|
1:42:05.920 --> 1:42:11.040 |
|
As a person who deeply appreciates practice, I cannot actually. |
|
|
|
1:42:11.040 --> 1:42:16.560 |
|
There's actually a very simple trick. |
|
|
|
1:42:16.560 --> 1:42:19.440 |
|
So in order to come up with a task, |
|
|
|
1:42:19.440 --> 1:42:21.760 |
|
so the only thing you can measure is skill at the task. |
|
|
|
1:42:21.760 --> 1:42:22.320 |
|
Yes. |
|
|
|
1:42:22.320 --> 1:42:24.800 |
|
All tasks are going to involve priors. |
|
|
|
1:42:24.800 --> 1:42:25.600 |
|
Yes. |
|
|
|
1:42:25.600 --> 1:42:29.920 |
|
The trick is to know what they are and to describe that. |
|
|
|
1:42:29.920 --> 1:42:33.760 |
|
And then you make sure that this is the same set of priors as what humans start with. |
|
|
|
1:42:33.760 --> 1:42:38.560 |
|
So you create a task that assumes these priors, that exactly documents these priors, |
|
|
|
1:42:38.560 --> 1:42:42.240 |
|
so that the priors are made explicit and there are no other priors involved. |
|
|
|
1:42:42.240 --> 1:42:48.960 |
|
And then you generate a certain number of samples in experience space for this task, right? |
|
|
|
1:42:49.840 --> 1:42:56.320 |
|
And this, for one task, assuming that the task is new for the agent passing it, |
|
|
|
1:42:56.320 --> 1:43:04.320 |
|
that's one test of this definition of intelligence that we set up. |
|
|
|
1:43:04.320 --> 1:43:06.880 |
|
And now you can scale that to many different tasks, |
|
|
|
1:43:06.880 --> 1:43:10.480 |
|
that each task should be new to the agent passing it, right? |
|
|
|
1:43:11.360 --> 1:43:14.480 |
|
And also it should be human interpretable and understandable |
|
|
|
1:43:14.480 --> 1:43:16.880 |
|
so that you can actually have a human pass the same test. |
|
|
|
1:43:16.880 --> 1:43:19.760 |
|
And then you can compare the score of your machine and the score of your human. |
|
|
|
1:43:19.760 --> 1:43:20.720 |
|
Which could be a lot of stuff. |
|
|
|
1:43:20.720 --> 1:43:23.040 |
|
You could even start a task like MNIST. |
|
|
|
1:43:23.040 --> 1:43:28.800 |
|
Just as long as you start with the same set of priors. |
|
|
|
1:43:28.800 --> 1:43:34.080 |
|
So the problem with MNIST, humans are already trying to recognize digits, right? |
|
|
|
1:43:35.600 --> 1:43:40.960 |
|
But let's say we're considering objects that are not digits, |
|
|
|
1:43:42.400 --> 1:43:43.920 |
|
some completely arbitrary patterns. |
|
|
|
1:43:44.480 --> 1:43:48.880 |
|
Well, humans already come with visual priors about how to process that. |
|
|
|
1:43:48.880 --> 1:43:54.080 |
|
So in order to make the game fair, you would have to isolate these priors |
|
|
|
1:43:54.080 --> 1:43:57.280 |
|
and describe them and then express them as computational rules. |
|
|
|
1:43:57.280 --> 1:44:01.680 |
|
Having worked a lot with vision science people, that's exceptionally difficult. |
|
|
|
1:44:01.680 --> 1:44:03.120 |
|
A lot of progress has been made. |
|
|
|
1:44:03.120 --> 1:44:08.080 |
|
There's been a lot of good tests and basically reducing all of human vision into some good priors. |
|
|
|
1:44:08.640 --> 1:44:10.960 |
|
We're still probably far away from that perfectly, |
|
|
|
1:44:10.960 --> 1:44:14.640 |
|
but as a start for a benchmark, that's an exciting possibility. |
|
|
|
1:44:14.640 --> 1:44:24.240 |
|
Yeah, so Elisabeth Spelke actually lists objectness as one of the core knowledge priors. |
|
|
|
1:44:24.800 --> 1:44:25.920 |
|
Objectness, cool. |
|
|
|
1:44:25.920 --> 1:44:26.880 |
|
Objectness, yeah. |
|
|
|
1:44:27.440 --> 1:44:31.520 |
|
So we have priors about objectness, like about the visual space, about time, |
|
|
|
1:44:31.520 --> 1:44:34.240 |
|
about agents, about goal oriented behavior. |
|
|
|
1:44:35.280 --> 1:44:39.280 |
|
We have many different priors, but what's interesting is that, |
|
|
|
1:44:39.280 --> 1:44:43.920 |
|
sure, we have this pretty diverse and rich set of priors, |
|
|
|
1:44:43.920 --> 1:44:46.880 |
|
but it's also not that diverse, right? |
|
|
|
1:44:46.880 --> 1:44:50.800 |
|
We are not born into this world with a ton of knowledge about the world, |
|
|
|
1:44:50.800 --> 1:44:57.840 |
|
with only a small set of core knowledge. |
|
|
|
1:44:58.640 --> 1:45:05.040 |
|
Yeah, sorry, do you have a sense of how it feels to us humans that that set is not that large? |
|
|
|
1:45:05.040 --> 1:45:09.600 |
|
But just even the nature of time that we kind of integrate pretty effectively |
|
|
|
1:45:09.600 --> 1:45:11.600 |
|
through all of our perception, all of our reasoning, |
|
|
|
1:45:12.640 --> 1:45:17.680 |
|
maybe how, you know, do you have a sense of how easy it is to encode those priors? |
|
|
|
1:45:17.680 --> 1:45:24.560 |
|
Maybe it requires building a universe and then the human brain in order to encode those priors. |
|
|
|
1:45:25.440 --> 1:45:28.640 |
|
Or do you have a hope that it can be listed like an axiomatic? |
|
|
|
1:45:28.640 --> 1:45:29.280 |
|
I don't think so. |
|
|
|
1:45:29.280 --> 1:45:33.040 |
|
So you have to keep in mind that any knowledge about the world that we are |
|
|
|
1:45:33.040 --> 1:45:41.120 |
|
born with is something that has to have been encoded into our DNA by evolution at some point. |
|
|
|
1:45:41.120 --> 1:45:41.440 |
|
Right. |
|
|
|
1:45:41.440 --> 1:45:45.440 |
|
And DNA is a very, very low bandwidth medium. |
|
|
|
1:45:46.000 --> 1:45:51.200 |
|
Like it's extremely long and expensive to encode anything into DNA because first of all, |
|
|
|
1:45:52.560 --> 1:45:57.440 |
|
you need some sort of evolutionary pressure to guide this writing process. |
|
|
|
1:45:57.440 --> 1:46:03.440 |
|
And then, you know, the higher level of information you're trying to write, the longer it's going to take. |
|
|
|
1:46:04.480 --> 1:46:13.520 |
|
And the thing in the environment that you're trying to encode knowledge about has to be stable |
|
|
|
1:46:13.520 --> 1:46:15.280 |
|
over this duration. |
|
|
|
1:46:15.280 --> 1:46:20.960 |
|
So you can only encode into DNA things that constitute an evolutionary advantage. |
|
|
|
1:46:20.960 --> 1:46:25.280 |
|
So this is actually a very small subset of all possible knowledge about the world. |
|
|
|
1:46:25.280 --> 1:46:32.080 |
|
You can only encode things that are stable, that are true, over very, very long periods of time, |
|
|
|
1:46:32.080 --> 1:46:33.680 |
|
typically millions of years. |
|
|
|
1:46:33.680 --> 1:46:38.720 |
|
For instance, we might have some visual prior about the shape of snakes, right? |
|
|
|
1:46:38.720 --> 1:46:43.920 |
|
But what makes a face, what's the difference between a face and an art face? |
|
|
|
1:46:44.560 --> 1:46:48.080 |
|
But consider this interesting question. |
|
|
|
1:46:48.080 --> 1:46:56.640 |
|
Do we have any innate sense of the visual difference between a male face and a female face? |
|
|
|
1:46:56.640 --> 1:46:57.600 |
|
What do you think? |
|
|
|
1:46:58.640 --> 1:46:59.840 |
|
For a human, I mean. |
|
|
|
1:46:59.840 --> 1:47:04.000 |
|
I would have to look back into evolutionary history when the genders emerged. |
|
|
|
1:47:04.000 --> 1:47:06.240 |
|
But yeah, most... |
|
|
|
1:47:06.240 --> 1:47:09.840 |
|
I mean, the faces of humans are quite different from the faces of great apes. |
|
|
|
1:47:10.640 --> 1:47:11.600 |
|
Great apes, right? |
|
|
|
1:47:12.880 --> 1:47:13.600 |
|
Yeah. |
|
|
|
1:47:13.600 --> 1:47:14.800 |
|
That's interesting. |
|
|
|
1:47:14.800 --> 1:47:22.800 |
|
Yeah, you couldn't tell the face of a female chimpanzee from the face of a male chimpanzee, |
|
|
|
1:47:22.800 --> 1:47:23.440 |
|
probably. |
|
|
|
1:47:23.440 --> 1:47:26.160 |
|
Yeah, and I don't think most humans have all that ability. |
|
|
|
1:47:26.160 --> 1:47:33.280 |
|
So we do have innate knowledge of what makes a face, but it's actually impossible for us to |
|
|
|
1:47:33.280 --> 1:47:40.320 |
|
have any DNA encoded knowledge of the difference between a female human face and a male human face |
|
|
|
1:47:40.320 --> 1:47:50.560 |
|
because that knowledge, that information came up into the world actually very recently. |
|
|
|
1:47:50.560 --> 1:47:56.400 |
|
If you look at the slowness of the process of encoding knowledge into DNA. |
|
|
|
1:47:56.400 --> 1:47:57.360 |
|
Yeah, so that's interesting. |
|
|
|
1:47:57.360 --> 1:48:02.080 |
|
That's a really powerful argument that DNA is a low bandwidth and it takes a long time to encode. |
|
|
|
1:48:02.800 --> 1:48:05.200 |
|
That naturally creates a very efficient encoding. |
|
|
|
1:48:05.200 --> 1:48:12.800 |
|
But one important consequence of this is that, so yes, we are born into this world with a bunch of |
|
|
|
1:48:12.800 --> 1:48:17.600 |
|
knowledge, sometimes high level knowledge about the world, like the shape, the rough shape of a |
|
|
|
1:48:17.600 --> 1:48:19.520 |
|
snake, of the rough shape of a face. |
|
|
|
1:48:20.480 --> 1:48:26.960 |
|
But importantly, because this knowledge takes so long to write, almost all of this innate |
|
|
|
1:48:26.960 --> 1:48:32.080 |
|
knowledge is shared with our cousins, with great apes, right? |
|
|
|
1:48:32.080 --> 1:48:35.600 |
|
So it is not actually this innate knowledge that makes us special. |
|
|
|
1:48:36.320 --> 1:48:42.000 |
|
But to throw it right back at you from the earlier on in our discussion, it's that encoding |
|
|
|
1:48:42.960 --> 1:48:48.320 |
|
might also include the entirety of the environment of Earth. |
|
|
|
1:48:49.360 --> 1:48:49.920 |
|
To some extent. |
|
|
|
1:48:49.920 --> 1:48:56.480 |
|
So it can include things that are important to survival and production, so for which there is |
|
|
|
1:48:56.480 --> 1:49:02.880 |
|
some evolutionary pressure, and things that are stable, constant over very, very, very long time |
|
|
|
1:49:02.880 --> 1:49:03.380 |
|
periods. |
|
|
|
1:49:04.160 --> 1:49:06.320 |
|
And honestly, it's not that much information. |
|
|
|
1:49:06.320 --> 1:49:14.400 |
|
There's also, besides the bandwidths constraint and the constraints of the writing process, |
|
|
|
1:49:14.400 --> 1:49:21.440 |
|
there's also memory constraints, like DNA, the part of DNA that deals with the human brain, |
|
|
|
1:49:21.440 --> 1:49:22.640 |
|
it's actually fairly small. |
|
|
|
1:49:22.640 --> 1:49:25.520 |
|
It's like, you know, on the order of megabytes, right? |
|
|
|
1:49:25.520 --> 1:49:29.600 |
|
There's not that much high level knowledge about the world you can encode. |
|
|
|
1:49:31.600 --> 1:49:38.880 |
|
That's quite brilliant and hopeful for a benchmark that you're referring to of encoding |
|
|
|
1:49:38.880 --> 1:49:39.360 |
|
priors. |
|
|
|
1:49:39.360 --> 1:49:43.120 |
|
I actually look forward to, I'm skeptical whether you can do it in the next couple of |
|
|
|
1:49:43.120 --> 1:49:44.320 |
|
years, but hopefully. |
|
|
|
1:49:45.040 --> 1:49:45.760 |
|
I've been working. |
|
|
|
1:49:45.760 --> 1:49:49.920 |
|
So honestly, it's a very simple benchmark, and it's not like a big breakthrough or anything. |
|
|
|
1:49:49.920 --> 1:49:53.200 |
|
It's more like a fun side project, right? |
|
|
|
1:49:53.200 --> 1:49:55.680 |
|
But these fun, so is ImageNet. |
|
|
|
1:49:56.480 --> 1:50:04.080 |
|
These fun side projects could launch entire groups of efforts towards creating reasoning |
|
|
|
1:50:04.080 --> 1:50:04.960 |
|
systems and so on. |
|
|
|
1:50:04.960 --> 1:50:05.440 |
|
And I think... |
|
|
|
1:50:05.440 --> 1:50:06.160 |
|
Yeah, that's the goal. |
|
|
|
1:50:06.160 --> 1:50:12.080 |
|
It's trying to measure strong generalization, to measure the strength of abstraction in |
|
|
|
1:50:12.080 --> 1:50:16.960 |
|
our minds, well, in our minds and in artificial intelligence agencies. |
|
|
|
1:50:16.960 --> 1:50:24.800 |
|
And if there's anything true about this science organism is its individual cells love competition. |
|
|
|
1:50:24.800 --> 1:50:26.800 |
|
So and benchmarks encourage competition. |
|
|
|
1:50:26.800 --> 1:50:29.520 |
|
So that's an exciting possibility. |
|
|
|
1:50:29.520 --> 1:50:32.640 |
|
If you, do you think an AI winter is coming? |
|
|
|
1:50:33.520 --> 1:50:34.640 |
|
And how do we prevent it? |
|
|
|
1:50:35.440 --> 1:50:36.080 |
|
Not really. |
|
|
|
1:50:36.080 --> 1:50:42.160 |
|
So an AI winter is something that would occur when there's a big mismatch between how we |
|
|
|
1:50:42.160 --> 1:50:47.280 |
|
are selling the capabilities of AI and the actual capabilities of AI. |
|
|
|
1:50:47.280 --> 1:50:50.560 |
|
And today, some deep learning is creating a lot of value. |
|
|
|
1:50:50.560 --> 1:50:56.240 |
|
And it will keep creating a lot of value in the sense that these models are applicable |
|
|
|
1:50:56.240 --> 1:51:00.000 |
|
to a very wide range of problems that are relevant today. |
|
|
|
1:51:00.000 --> 1:51:05.120 |
|
And we are only just getting started with applying these algorithms to every problem |
|
|
|
1:51:05.120 --> 1:51:06.320 |
|
they could be solving. |
|
|
|
1:51:06.320 --> 1:51:10.160 |
|
So deep learning will keep creating a lot of value for the time being. |
|
|
|
1:51:10.160 --> 1:51:15.920 |
|
What's concerning, however, is that there's a lot of hype around deep learning and around |
|
|
|
1:51:15.920 --> 1:51:16.240 |
|
AI. |
|
|
|
1:51:16.240 --> 1:51:22.000 |
|
There are lots of people are overselling the capabilities of these systems, not just |
|
|
|
1:51:22.000 --> 1:51:27.760 |
|
the capabilities, but also overselling the fact that they might be more or less, you |
|
|
|
1:51:27.760 --> 1:51:36.640 |
|
know, brain like, like given the kind of a mystical aspect, these technologies and also |
|
|
|
1:51:36.640 --> 1:51:43.840 |
|
overselling the pace of progress, which, you know, it might look fast in the sense that |
|
|
|
1:51:43.840 --> 1:51:46.480 |
|
we have this exponentially increasing number of papers. |
|
|
|
1:51:47.760 --> 1:51:52.960 |
|
But again, that's just a simple consequence of the fact that we have ever more people |
|
|
|
1:51:52.960 --> 1:51:53.840 |
|
coming into the field. |
|
|
|
1:51:54.400 --> 1:51:57.440 |
|
It doesn't mean the progress is actually exponentially fast. |
|
|
|
1:51:58.640 --> 1:52:02.720 |
|
Let's say you're trying to raise money for your startup or your research lab. |
|
|
|
1:52:02.720 --> 1:52:09.120 |
|
You might want to tell, you know, a grandiose story to investors about how deep learning |
|
|
|
1:52:09.120 --> 1:52:14.240 |
|
is just like the brain and how it can solve all these incredible problems like self driving |
|
|
|
1:52:14.240 --> 1:52:15.760 |
|
and robotics and so on. |
|
|
|
1:52:15.760 --> 1:52:19.440 |
|
And maybe you can tell them that the field is progressing so fast and we are going to |
|
|
|
1:52:19.440 --> 1:52:23.040 |
|
have AGI within 15 years or even 10 years. |
|
|
|
1:52:23.040 --> 1:52:25.920 |
|
And none of this is true. |
|
|
|
1:52:25.920 --> 1:52:32.800 |
|
And every time you're like saying these things and an investor or, you know, a decision maker |
|
|
|
1:52:32.800 --> 1:52:41.680 |
|
believes them, well, this is like the equivalent of taking on credit card debt, but for trust, |
|
|
|
1:52:41.680 --> 1:52:42.480 |
|
right? |
|
|
|
1:52:42.480 --> 1:52:50.160 |
|
And maybe this will, you know, this will be what enables you to raise a lot of money, |
|
|
|
1:52:50.160 --> 1:52:54.320 |
|
but ultimately you are creating damage, you are damaging the field. |
|
|
|
1:52:54.320 --> 1:53:00.160 |
|
So that's the concern is that that debt, that's what happens with the other AI winters is |
|
|
|
1:53:00.160 --> 1:53:04.160 |
|
the concern is you actually tweeted about this with autonomous vehicles, right? |
|
|
|
1:53:04.160 --> 1:53:08.960 |
|
There's almost every single company now have promised that they will have full autonomous |
|
|
|
1:53:08.960 --> 1:53:11.760 |
|
vehicles by 2021, 2022. |
|
|
|
1:53:11.760 --> 1:53:18.080 |
|
That's a good example of the consequences of over hyping the capabilities of AI and |
|
|
|
1:53:18.080 --> 1:53:19.280 |
|
the pace of progress. |
|
|
|
1:53:19.280 --> 1:53:25.200 |
|
So because I work especially a lot recently in this area, I have a deep concern of what |
|
|
|
1:53:25.200 --> 1:53:30.400 |
|
happens when all of these companies after I've invested billions have a meeting and |
|
|
|
1:53:30.400 --> 1:53:33.600 |
|
say, how much do we actually, first of all, do we have an autonomous vehicle? |
|
|
|
1:53:33.600 --> 1:53:35.280 |
|
The answer will definitely be no. |
|
|
|
1:53:35.840 --> 1:53:40.560 |
|
And second will be, wait a minute, we've invested one, two, three, four billion dollars |
|
|
|
1:53:40.560 --> 1:53:43.120 |
|
into this and we made no profit. |
|
|
|
1:53:43.120 --> 1:53:49.200 |
|
And the reaction to that may be going very hard in other directions that might impact |
|
|
|
1:53:49.200 --> 1:53:50.400 |
|
even other industries. |
|
|
|
1:53:50.400 --> 1:53:55.520 |
|
And that's what we call an AI winter is when there is backlash where no one believes any |
|
|
|
1:53:55.520 --> 1:53:59.360 |
|
of these promises anymore because they've turned that to be big lies the first time |
|
|
|
1:53:59.360 --> 1:54:00.240 |
|
around. |
|
|
|
1:54:00.240 --> 1:54:06.000 |
|
And this will definitely happen to some extent for autonomous vehicles because the public |
|
|
|
1:54:06.000 --> 1:54:13.360 |
|
and decision makers have been convinced that around 2015, they've been convinced by these |
|
|
|
1:54:13.360 --> 1:54:19.600 |
|
people who are trying to raise money for their startups and so on, that L5 driving was coming |
|
|
|
1:54:19.600 --> 1:54:22.880 |
|
in maybe 2016, maybe 2017, maybe 2018. |
|
|
|
1:54:22.880 --> 1:54:26.080 |
|
Now we're in 2019, we're still waiting for it. |
|
|
|
1:54:27.600 --> 1:54:32.800 |
|
And so I don't believe we are going to have a full on AI winter because we have these |
|
|
|
1:54:32.800 --> 1:54:36.640 |
|
technologies that are producing a tremendous amount of real value. |
|
|
|
1:54:37.680 --> 1:54:39.920 |
|
But there is also too much hype. |
|
|
|
1:54:39.920 --> 1:54:43.520 |
|
So there will be some backlash, especially there will be backlash. |
|
|
|
1:54:44.960 --> 1:54:53.040 |
|
So some startups are trying to sell the dream of AGI and the fact that AGI is going to create |
|
|
|
1:54:53.040 --> 1:54:53.760 |
|
infinite value. |
|
|
|
1:54:53.760 --> 1:54:55.680 |
|
Like AGI is like a free lunch. |
|
|
|
1:54:55.680 --> 1:55:02.800 |
|
Like if you can develop an AI system that passes a certain threshold of IQ or something, |
|
|
|
1:55:02.800 --> 1:55:04.400 |
|
then suddenly you have infinite value. |
|
|
|
1:55:04.400 --> 1:55:14.160 |
|
And well, there are actually lots of investors buying into this idea and they will wait maybe |
|
|
|
1:55:14.160 --> 1:55:17.760 |
|
10, 15 years and nothing will happen. |
|
|
|
1:55:17.760 --> 1:55:22.560 |
|
And the next time around, well, maybe there will be a new generation of investors. |
|
|
|
1:55:22.560 --> 1:55:23.360 |
|
No one will care. |
|
|
|
1:55:24.800 --> 1:55:27.280 |
|
Human memory is fairly short after all. |
|
|
|
1:55:27.280 --> 1:55:34.320 |
|
I don't know about you, but because I've spoken about AGI sometimes poetically, I get a lot |
|
|
|
1:55:34.320 --> 1:55:42.000 |
|
of emails from people giving me, they're usually like a large manifestos of they've, they say |
|
|
|
1:55:42.000 --> 1:55:47.200 |
|
to me that they have created an AGI system or they know how to do it. |
|
|
|
1:55:47.200 --> 1:55:48.880 |
|
And there's a long write up of how to do it. |
|
|
|
1:55:48.880 --> 1:55:50.560 |
|
I get a lot of these emails, yeah. |
|
|
|
1:55:50.560 --> 1:55:57.760 |
|
They're a little bit feel like it's generated by an AI system actually, but there's usually |
|
|
|
1:55:57.760 --> 1:56:06.640 |
|
no diagram, you have a transformer generating crank papers about AGI. |
|
|
|
1:56:06.640 --> 1:56:12.160 |
|
So the question is about, because you've been such a good, you have a good radar for crank |
|
|
|
1:56:12.160 --> 1:56:16.720 |
|
papers, how do we know they're not onto something? |
|
|
|
1:56:16.720 --> 1:56:24.240 |
|
How do I, so when you start to talk about AGI or anything like the reasoning benchmarks |
|
|
|
1:56:24.240 --> 1:56:28.160 |
|
and so on, so something that doesn't have a benchmark, it's really difficult to know. |
|
|
|
1:56:29.120 --> 1:56:34.560 |
|
I mean, I talked to Jeff Hawkins, who's really looking at neuroscience approaches to how, |
|
|
|
1:56:35.200 --> 1:56:41.520 |
|
and there's some, there's echoes of really interesting ideas in at least Jeff's case, |
|
|
|
1:56:41.520 --> 1:56:42.320 |
|
which he's showing. |
|
|
|
1:56:43.280 --> 1:56:45.040 |
|
How do you usually think about this? |
|
|
|
1:56:46.640 --> 1:56:52.880 |
|
Like preventing yourself from being too narrow minded and elitist about deep learning, it |
|
|
|
1:56:52.880 --> 1:56:56.720 |
|
has to work on these particular benchmarks, otherwise it's trash. |
|
|
|
1:56:56.720 --> 1:57:05.280 |
|
Well, you know, the thing is, intelligence does not exist in the abstract. |
|
|
|
1:57:05.280 --> 1:57:07.200 |
|
Intelligence has to be applied. |
|
|
|
1:57:07.200 --> 1:57:11.040 |
|
So if you don't have a benchmark, if you have an improvement in some benchmark, maybe it's |
|
|
|
1:57:11.040 --> 1:57:12.400 |
|
a new benchmark, right? |
|
|
|
1:57:12.400 --> 1:57:16.640 |
|
Maybe it's not something we've been looking at before, but you do need a problem that |
|
|
|
1:57:16.640 --> 1:57:17.360 |
|
you're trying to solve. |
|
|
|
1:57:17.360 --> 1:57:20.000 |
|
You're not going to come up with a solution without a problem. |
|
|
|
1:57:20.000 --> 1:57:25.520 |
|
So you, general intelligence, I mean, you've clearly highlighted generalization. |
|
|
|
1:57:26.320 --> 1:57:31.200 |
|
If you want to claim that you have an intelligence system, it should come with a benchmark. |
|
|
|
1:57:31.200 --> 1:57:35.760 |
|
It should, yes, it should display capabilities of some kind. |
|
|
|
1:57:35.760 --> 1:57:41.840 |
|
It should show that it can create some form of value, even if it's a very artificial form |
|
|
|
1:57:41.840 --> 1:57:42.800 |
|
of value. |
|
|
|
1:57:42.800 --> 1:57:48.800 |
|
And that's also the reason why you don't actually need to care about telling which papers have |
|
|
|
1:57:48.800 --> 1:57:52.000 |
|
actually some hidden potential and which do not. |
|
|
|
1:57:53.120 --> 1:57:59.200 |
|
Because if there is a new technique that's actually creating value, this is going to |
|
|
|
1:57:59.200 --> 1:58:02.480 |
|
be brought to light very quickly because it's actually making a difference. |
|
|
|
1:58:02.480 --> 1:58:08.160 |
|
So it's the difference between something that is ineffectual and something that is actually |
|
|
|
1:58:08.160 --> 1:58:08.800 |
|
useful. |
|
|
|
1:58:08.800 --> 1:58:14.080 |
|
And ultimately usefulness is our guide, not just in this field, but if you look at science |
|
|
|
1:58:14.080 --> 1:58:18.720 |
|
in general, maybe there are many, many people over the years that have had some really interesting |
|
|
|
1:58:19.440 --> 1:58:22.800 |
|
theories of everything, but they were just completely useless. |
|
|
|
1:58:22.800 --> 1:58:27.280 |
|
And you don't actually need to tell the interesting theories from the useless theories. |
|
|
|
1:58:28.000 --> 1:58:34.080 |
|
All you need is to see, is this actually having an effect on something else? |
|
|
|
1:58:34.080 --> 1:58:35.360 |
|
Is this actually useful? |
|
|
|
1:58:35.360 --> 1:58:36.800 |
|
Is this making an impact or not? |
|
|
|
1:58:37.600 --> 1:58:38.640 |
|
That's beautifully put. |
|
|
|
1:58:38.640 --> 1:58:43.680 |
|
I mean, the same applies to quantum mechanics, to string theory, to the holographic principle. |
|
|
|
1:58:43.680 --> 1:58:45.280 |
|
We are doing deep learning because it works. |
|
|
|
1:58:46.960 --> 1:58:52.720 |
|
Before it started working, people considered people working on neural networks as cranks |
|
|
|
1:58:52.720 --> 1:58:53.120 |
|
very much. |
|
|
|
1:58:54.560 --> 1:58:56.320 |
|
No one was working on this anymore. |
|
|
|
1:58:56.320 --> 1:58:59.120 |
|
And now it's working, which is what makes it valuable. |
|
|
|
1:58:59.120 --> 1:59:00.320 |
|
It's not about being right. |
|
|
|
1:59:01.120 --> 1:59:02.560 |
|
It's about being effective. |
|
|
|
1:59:02.560 --> 1:59:08.080 |
|
And nevertheless, the individual entities of this scientific mechanism, just like Yoshua |
|
|
|
1:59:08.080 --> 1:59:12.480 |
|
Banjo or Jan Lekun, they, while being called cranks, stuck with it. |
|
|
|
1:59:12.480 --> 1:59:12.880 |
|
Right? |
|
|
|
1:59:12.880 --> 1:59:13.280 |
|
Yeah. |
|
|
|
1:59:13.280 --> 1:59:17.840 |
|
And so us individual agents, even if everyone's laughing at us, just stick with it. |
|
|
|
1:59:18.880 --> 1:59:21.840 |
|
If you believe you have something, you should stick with it and see it through. |
|
|
|
1:59:23.520 --> 1:59:25.920 |
|
That's a beautiful inspirational message to end on. |
|
|
|
1:59:25.920 --> 1:59:27.600 |
|
Francois, thank you so much for talking today. |
|
|
|
1:59:27.600 --> 1:59:28.640 |
|
That was amazing. |
|
|
|
1:59:28.640 --> 1:59:44.000 |
|
Thank you. |
|
|
|
|