|
WEBVTT |
|
|
|
00:00.000 --> 00:03.760 |
|
The following is a conversation with Ian Goodfellow. |
|
|
|
00:03.760 --> 00:06.360 |
|
He's the author of the popular textbook on deep learning |
|
|
|
00:06.360 --> 00:08.880 |
|
simply titled Deep Learning. |
|
|
|
00:08.880 --> 00:12.320 |
|
He coined the term of generative adversarial networks, |
|
|
|
00:12.320 --> 00:14.560 |
|
otherwise known as GANs. |
|
|
|
00:14.560 --> 00:18.160 |
|
And with his 2014 paper is responsible |
|
|
|
00:18.160 --> 00:20.440 |
|
for launching the incredible growth |
|
|
|
00:20.440 --> 00:22.120 |
|
of research and innovation |
|
|
|
00:22.120 --> 00:24.720 |
|
in this subfield of deep learning. |
|
|
|
00:24.720 --> 00:27.520 |
|
He got his BS and MS at Stanford, |
|
|
|
00:27.520 --> 00:30.120 |
|
his PhD at University of Montreal |
|
|
|
00:30.120 --> 00:33.200 |
|
with Yoshua Benjo and Aaron Kervel. |
|
|
|
00:33.200 --> 00:35.240 |
|
He held several research positions, |
|
|
|
00:35.240 --> 00:37.560 |
|
including at OpenAI, Google Brain, |
|
|
|
00:37.560 --> 00:41.560 |
|
and now at Apple as the director of machine learning. |
|
|
|
00:41.560 --> 00:45.400 |
|
This recording happened while Ian was still a Google Brain, |
|
|
|
00:45.400 --> 00:48.520 |
|
but we don't talk about anything specific to Google |
|
|
|
00:48.520 --> 00:50.760 |
|
or any other organization. |
|
|
|
00:50.760 --> 00:52.480 |
|
This conversation is part |
|
|
|
00:52.480 --> 00:54.520 |
|
of the artificial intelligence podcast. |
|
|
|
00:54.520 --> 00:56.680 |
|
If you enjoy it, subscribe on YouTube, |
|
|
|
00:56.680 --> 00:59.600 |
|
iTunes, or simply connect with me on Twitter |
|
|
|
00:59.600 --> 01:03.000 |
|
at Lex Freedman, spelled F R I D. |
|
|
|
01:03.000 --> 01:07.080 |
|
And now here's my conversation with Ian Goodfellow. |
|
|
|
01:08.240 --> 01:11.000 |
|
You open your popular deep learning book |
|
|
|
01:11.000 --> 01:13.600 |
|
with a Russian doll type diagram |
|
|
|
01:13.600 --> 01:15.880 |
|
that shows deep learning is a subset |
|
|
|
01:15.880 --> 01:17.160 |
|
of representation learning, |
|
|
|
01:17.160 --> 01:19.960 |
|
which in turn is a subset of machine learning |
|
|
|
01:19.960 --> 01:22.520 |
|
and finally a subset of AI. |
|
|
|
01:22.520 --> 01:25.280 |
|
So this kind of implies that there may be limits |
|
|
|
01:25.280 --> 01:27.720 |
|
to deep learning in the context of AI. |
|
|
|
01:27.720 --> 01:31.560 |
|
So what do you think is the current limits of deep learning |
|
|
|
01:31.560 --> 01:33.120 |
|
and are those limits something |
|
|
|
01:33.120 --> 01:35.760 |
|
that we can overcome with time? |
|
|
|
01:35.760 --> 01:37.720 |
|
Yeah, I think one of the biggest limitations |
|
|
|
01:37.720 --> 01:39.320 |
|
of deep learning is that right now |
|
|
|
01:39.320 --> 01:42.920 |
|
it requires really a lot of data, especially labeled data. |
|
|
|
01:43.960 --> 01:45.480 |
|
There are some unsupervised |
|
|
|
01:45.480 --> 01:47.160 |
|
and semi supervised learning algorithms |
|
|
|
01:47.160 --> 01:49.480 |
|
that can reduce the amount of labeled data you need, |
|
|
|
01:49.480 --> 01:52.200 |
|
but they still require a lot of unlabeled data. |
|
|
|
01:52.200 --> 01:54.200 |
|
Reinforcement learning algorithms, they don't need labels, |
|
|
|
01:54.200 --> 01:56.280 |
|
but they need really a lot of experiences. |
|
|
|
01:57.280 --> 01:58.920 |
|
As human beings, we don't learn to play a pong |
|
|
|
01:58.920 --> 02:01.520 |
|
by failing at pong two million times. |
|
|
|
02:02.720 --> 02:05.880 |
|
So just getting the generalization ability better |
|
|
|
02:05.880 --> 02:08.040 |
|
is one of the most important bottlenecks |
|
|
|
02:08.040 --> 02:10.520 |
|
in the capability of the technology today. |
|
|
|
02:10.520 --> 02:12.360 |
|
And then I guess I'd also say deep learning |
|
|
|
02:12.360 --> 02:15.620 |
|
is like a component of a bigger system. |
|
|
|
02:16.600 --> 02:19.040 |
|
So far, nobody is really proposing to have |
|
|
|
02:20.600 --> 02:22.000 |
|
only what you'd call deep learning |
|
|
|
02:22.000 --> 02:25.520 |
|
as the entire ingredient of intelligence. |
|
|
|
02:25.520 --> 02:29.860 |
|
You use deep learning as sub modules of other systems, |
|
|
|
02:29.860 --> 02:32.320 |
|
like AlphaGo has a deep learning model |
|
|
|
02:32.320 --> 02:34.160 |
|
that estimates the value function. |
|
|
|
02:35.200 --> 02:36.600 |
|
Most reinforcement learning algorithms |
|
|
|
02:36.600 --> 02:37.880 |
|
have a deep learning module |
|
|
|
02:37.880 --> 02:40.320 |
|
that estimates which action to take next, |
|
|
|
02:40.320 --> 02:42.480 |
|
but you might have other components. |
|
|
|
02:42.480 --> 02:46.120 |
|
So you're basically building a function estimator. |
|
|
|
02:46.120 --> 02:48.600 |
|
Do you think it's possible? |
|
|
|
02:48.600 --> 02:51.000 |
|
You said nobody's kind of been thinking about this so far, |
|
|
|
02:51.000 --> 02:54.320 |
|
but do you think neural networks could be made to reason |
|
|
|
02:54.320 --> 02:57.720 |
|
in the way symbolic systems did in the 80s and 90s |
|
|
|
02:57.720 --> 03:00.160 |
|
to do more, create more like programs |
|
|
|
03:00.160 --> 03:01.440 |
|
as opposed to functions? |
|
|
|
03:01.440 --> 03:03.920 |
|
Yeah, I think we already see that a little bit. |
|
|
|
03:04.880 --> 03:08.860 |
|
I already kind of think of neural nets as a kind of program. |
|
|
|
03:08.860 --> 03:12.920 |
|
I think of deep learning as basically learning programs |
|
|
|
03:12.920 --> 03:15.280 |
|
that have more than one step. |
|
|
|
03:15.280 --> 03:16.960 |
|
So if you draw a flow chart |
|
|
|
03:16.960 --> 03:19.540 |
|
or if you draw a TensorFlow graph |
|
|
|
03:19.540 --> 03:21.880 |
|
describing your machine learning model, |
|
|
|
03:21.880 --> 03:23.520 |
|
I think of the depth of that graph |
|
|
|
03:23.520 --> 03:25.880 |
|
as describing the number of steps that run in sequence |
|
|
|
03:25.880 --> 03:27.640 |
|
and then the width of that graph |
|
|
|
03:27.640 --> 03:30.120 |
|
as the number of steps that run in parallel. |
|
|
|
03:30.120 --> 03:31.680 |
|
Now it's been long enough |
|
|
|
03:31.680 --> 03:32.880 |
|
that we've had deep learning working |
|
|
|
03:32.880 --> 03:33.880 |
|
that it's a little bit silly |
|
|
|
03:33.880 --> 03:35.740 |
|
to even discuss shallow learning anymore, |
|
|
|
03:35.740 --> 03:38.880 |
|
but back when I first got involved in AI, |
|
|
|
03:38.880 --> 03:40.080 |
|
when we used machine learning, |
|
|
|
03:40.080 --> 03:41.280 |
|
we were usually learning things |
|
|
|
03:41.280 --> 03:43.680 |
|
like support vector machines. |
|
|
|
03:43.680 --> 03:45.640 |
|
You could have a lot of input features to the model |
|
|
|
03:45.640 --> 03:48.120 |
|
and you could multiply each feature by a different weight. |
|
|
|
03:48.120 --> 03:51.240 |
|
All those multiplications were done in parallel to each other |
|
|
|
03:51.240 --> 03:52.720 |
|
and there wasn't a lot done in series. |
|
|
|
03:52.720 --> 03:54.360 |
|
I think what we got with deep learning |
|
|
|
03:54.360 --> 03:58.400 |
|
was really the ability to have steps of a program |
|
|
|
03:58.400 --> 04:00.320 |
|
that run in sequence. |
|
|
|
04:00.320 --> 04:03.200 |
|
And I think that we've actually started to see |
|
|
|
04:03.200 --> 04:05.040 |
|
that what's important with deep learning |
|
|
|
04:05.040 --> 04:08.000 |
|
is more the fact that we have a multi step program |
|
|
|
04:08.000 --> 04:10.800 |
|
rather than the fact that we've learned a representation. |
|
|
|
04:10.800 --> 04:15.120 |
|
If you look at things like Resnuts, for example, |
|
|
|
04:15.120 --> 04:18.660 |
|
they take one particular kind of representation |
|
|
|
04:18.660 --> 04:21.040 |
|
and they update it several times. |
|
|
|
04:21.040 --> 04:23.560 |
|
Back when deep learning first really took off |
|
|
|
04:23.560 --> 04:25.760 |
|
in the academic world in 2006, |
|
|
|
04:25.760 --> 04:28.400 |
|
when Jeff Hinton showed that you could train |
|
|
|
04:28.400 --> 04:30.160 |
|
deep belief networks, |
|
|
|
04:30.160 --> 04:31.960 |
|
everybody who was interested in the idea |
|
|
|
04:31.960 --> 04:33.560 |
|
thought of it as each layer |
|
|
|
04:33.560 --> 04:35.960 |
|
learns a different level of abstraction, |
|
|
|
04:35.960 --> 04:37.840 |
|
that the first layer trained on images |
|
|
|
04:37.840 --> 04:38.960 |
|
learns something like edges |
|
|
|
04:38.960 --> 04:40.420 |
|
and the second layer learns corners |
|
|
|
04:40.420 --> 04:43.320 |
|
and eventually you get these kind of grandmother cell units |
|
|
|
04:43.320 --> 04:45.920 |
|
that recognize specific objects. |
|
|
|
04:45.920 --> 04:48.560 |
|
Today, I think most people think of it more |
|
|
|
04:48.560 --> 04:52.000 |
|
as a computer program where as you add more layers, |
|
|
|
04:52.000 --> 04:55.120 |
|
you can do more updates before you output your final number. |
|
|
|
04:55.120 --> 04:57.160 |
|
But I don't think anybody believes that |
|
|
|
04:57.160 --> 05:02.040 |
|
layer 150 of the Resnet is a grandmother cell |
|
|
|
05:02.040 --> 05:05.080 |
|
and layer 100 is contours or something like that. |
|
|
|
05:06.040 --> 05:08.160 |
|
Okay, so you're not thinking of it |
|
|
|
05:08.160 --> 05:11.520 |
|
as a singular representation that keeps building. |
|
|
|
05:11.520 --> 05:15.960 |
|
You think of it as a program sort of almost like a state. |
|
|
|
05:15.960 --> 05:18.600 |
|
The representation is a state of understanding. |
|
|
|
05:18.600 --> 05:21.520 |
|
Yeah, I think of it as a program that makes several updates |
|
|
|
05:21.520 --> 05:23.840 |
|
and arrives at better and better understandings, |
|
|
|
05:23.840 --> 05:27.500 |
|
but it's not replacing the representation at each step. |
|
|
|
05:27.500 --> 05:29.160 |
|
It's refining it. |
|
|
|
05:29.160 --> 05:31.660 |
|
And in some sense, that's a little bit like reasoning. |
|
|
|
05:31.660 --> 05:33.560 |
|
It's not reasoning in the form of deduction, |
|
|
|
05:33.560 --> 05:36.960 |
|
but it's reasoning in the form of taking a thought |
|
|
|
05:36.960 --> 05:39.440 |
|
and refining it and refining it carefully |
|
|
|
05:39.440 --> 05:41.240 |
|
until it's good enough to use. |
|
|
|
05:41.240 --> 05:43.560 |
|
So do you think, and I hope you don't mind, |
|
|
|
05:43.560 --> 05:46.040 |
|
we'll jump philosophical every once in a while. |
|
|
|
05:46.040 --> 05:50.480 |
|
Do you think of, you know, cognition, human cognition, |
|
|
|
05:50.480 --> 05:53.520 |
|
or even consciousness as simply a result |
|
|
|
05:53.520 --> 05:58.120 |
|
of this kind of sequential representation learning? |
|
|
|
05:58.120 --> 06:00.440 |
|
Do you think that can emerge? |
|
|
|
06:00.440 --> 06:02.440 |
|
Cognition, yes, I think so. |
|
|
|
06:02.440 --> 06:05.160 |
|
Consciousness, it's really hard to even define |
|
|
|
06:05.160 --> 06:06.400 |
|
what we mean by that. |
|
|
|
06:07.400 --> 06:09.840 |
|
I guess there's, consciousness is often defined |
|
|
|
06:09.840 --> 06:12.120 |
|
as things like having self awareness, |
|
|
|
06:12.120 --> 06:15.200 |
|
and that's relatively easy to turn it |
|
|
|
06:15.200 --> 06:17.200 |
|
to something actionable for a computer scientist |
|
|
|
06:17.200 --> 06:18.400 |
|
to reason about. |
|
|
|
06:18.400 --> 06:20.080 |
|
People also define consciousness in terms |
|
|
|
06:20.080 --> 06:24.000 |
|
of having qualitative states of experience, like qualia. |
|
|
|
06:24.000 --> 06:25.280 |
|
There's all these philosophical problems, |
|
|
|
06:25.280 --> 06:27.880 |
|
like could you imagine a zombie |
|
|
|
06:27.880 --> 06:30.760 |
|
who does all the same information processing as a human, |
|
|
|
06:30.760 --> 06:33.500 |
|
but doesn't really have the qualitative experiences |
|
|
|
06:33.500 --> 06:34.720 |
|
that we have? |
|
|
|
06:34.720 --> 06:37.580 |
|
That sort of thing, I have no idea how to formalize |
|
|
|
06:37.580 --> 06:39.960 |
|
or turn it into a scientific question. |
|
|
|
06:39.960 --> 06:41.600 |
|
I don't know how you could run an experiment |
|
|
|
06:41.600 --> 06:44.880 |
|
to tell whether a person is a zombie or not. |
|
|
|
06:44.880 --> 06:46.680 |
|
And similarly, I don't know how you could run |
|
|
|
06:46.680 --> 06:49.680 |
|
an experiment to tell whether an advanced AI system |
|
|
|
06:49.680 --> 06:53.080 |
|
had become conscious in the sense of qualia or not. |
|
|
|
06:53.080 --> 06:54.600 |
|
But in the more practical sense, |
|
|
|
06:54.600 --> 06:56.320 |
|
like almost like self attention, |
|
|
|
06:56.320 --> 06:58.920 |
|
you think consciousness and cognition can, |
|
|
|
06:58.920 --> 07:03.240 |
|
in an impressive way, emerge from current types |
|
|
|
07:03.240 --> 07:05.600 |
|
of architectures that we think of as determining. |
|
|
|
07:05.600 --> 07:07.920 |
|
Or if you think of consciousness |
|
|
|
07:07.920 --> 07:12.160 |
|
in terms of self awareness and just making plans |
|
|
|
07:12.160 --> 07:15.120 |
|
based on the fact that the agent itself |
|
|
|
07:15.120 --> 07:18.000 |
|
exists in the world, reinforcement learning algorithms |
|
|
|
07:18.000 --> 07:20.840 |
|
are already more or less forced to model |
|
|
|
07:20.840 --> 07:23.040 |
|
the agent's effect on the environment. |
|
|
|
07:23.040 --> 07:26.340 |
|
So that more limited version of consciousness |
|
|
|
07:26.340 --> 07:30.560 |
|
is already something that we get limited versions |
|
|
|
07:30.560 --> 07:32.960 |
|
of with reinforcement learning algorithms |
|
|
|
07:32.960 --> 07:34.640 |
|
if they're trained well. |
|
|
|
07:34.640 --> 07:37.440 |
|
But you say limited. |
|
|
|
07:37.440 --> 07:39.920 |
|
So the big question really is how you jump |
|
|
|
07:39.920 --> 07:42.120 |
|
from limited to human level, right? |
|
|
|
07:42.120 --> 07:44.640 |
|
And whether it's possible, |
|
|
|
07:46.840 --> 07:49.000 |
|
even just building common sense reasoning |
|
|
|
07:49.000 --> 07:50.520 |
|
seems to be exceptionally difficult. |
|
|
|
07:50.520 --> 07:52.480 |
|
So if we scale things up, |
|
|
|
07:52.480 --> 07:55.000 |
|
if we get much better on supervised learning, |
|
|
|
07:55.000 --> 07:56.600 |
|
if we get better at labeling, |
|
|
|
07:56.600 --> 08:00.640 |
|
if we get bigger datasets, more compute, |
|
|
|
08:00.640 --> 08:03.880 |
|
do you think we'll start to see really impressive things |
|
|
|
08:03.880 --> 08:08.760 |
|
that go from limited to something echoes |
|
|
|
08:08.760 --> 08:10.320 |
|
of human level cognition? |
|
|
|
08:10.320 --> 08:11.200 |
|
I think so, yeah. |
|
|
|
08:11.200 --> 08:13.360 |
|
I'm optimistic about what can happen |
|
|
|
08:13.360 --> 08:16.440 |
|
just with more computation and more data. |
|
|
|
08:16.440 --> 08:20.120 |
|
I do think it'll be important to get the right kind of data. |
|
|
|
08:20.120 --> 08:23.160 |
|
Today, most of the machine learning systems we train |
|
|
|
08:23.160 --> 08:27.560 |
|
are mostly trained on one type of data for each model. |
|
|
|
08:27.560 --> 08:31.380 |
|
But the human brain, we get all of our different senses |
|
|
|
08:31.380 --> 08:33.880 |
|
and we have many different experiences |
|
|
|
08:33.880 --> 08:36.320 |
|
like riding a bike, driving a car, |
|
|
|
08:36.320 --> 08:37.940 |
|
talking to people, reading. |
|
|
|
08:39.160 --> 08:42.440 |
|
I think when we get that kind of integrated dataset |
|
|
|
08:42.440 --> 08:44.400 |
|
working with a machine learning model |
|
|
|
08:44.400 --> 08:47.640 |
|
that can actually close the loop and interact, |
|
|
|
08:47.640 --> 08:50.480 |
|
we may find that algorithms not so different |
|
|
|
08:50.480 --> 08:51.840 |
|
from what we have today, |
|
|
|
08:51.840 --> 08:53.240 |
|
learn really interesting things |
|
|
|
08:53.240 --> 08:54.400 |
|
when you scale them up a lot |
|
|
|
08:54.400 --> 08:58.240 |
|
and train them on a large amount of multimodal data. |
|
|
|
08:58.240 --> 08:59.640 |
|
So multimodal is really interesting, |
|
|
|
08:59.640 --> 09:04.000 |
|
but within, like you're working adversarial examples. |
|
|
|
09:04.000 --> 09:09.000 |
|
So selecting within model, within one mode of data, |
|
|
|
09:11.120 --> 09:13.800 |
|
selecting better at what are the difficult cases |
|
|
|
09:13.800 --> 09:16.120 |
|
from which you're most useful to learn from. |
|
|
|
09:16.120 --> 09:18.880 |
|
Oh, yeah, like could we get a whole lot of mileage |
|
|
|
09:18.880 --> 09:22.280 |
|
out of designing a model that's resistant |
|
|
|
09:22.280 --> 09:24.080 |
|
to adversarial examples or something like that? |
|
|
|
09:24.080 --> 09:26.280 |
|
Right, that's the question. |
|
|
|
09:26.280 --> 09:27.760 |
|
My thinking on that has evolved a lot |
|
|
|
09:27.760 --> 09:28.920 |
|
over the last few years. |
|
|
|
09:28.920 --> 09:31.280 |
|
When I first started to really invest |
|
|
|
09:31.280 --> 09:32.760 |
|
in studying adversarial examples, |
|
|
|
09:32.760 --> 09:36.320 |
|
I was thinking of it mostly as adversarial examples |
|
|
|
09:36.320 --> 09:39.000 |
|
reveal a big problem with machine learning. |
|
|
|
09:39.000 --> 09:41.160 |
|
And we would like to close the gap |
|
|
|
09:41.160 --> 09:44.160 |
|
between how machine learning models respond |
|
|
|
09:44.160 --> 09:46.560 |
|
to adversarial examples and how humans respond. |
|
|
|
09:47.640 --> 09:49.160 |
|
After studying the problem more, |
|
|
|
09:49.160 --> 09:51.960 |
|
I still think that adversarial examples are important. |
|
|
|
09:51.960 --> 09:55.440 |
|
I think of them now more of as a security liability |
|
|
|
09:55.440 --> 09:57.800 |
|
than as an issue that necessarily shows |
|
|
|
09:57.800 --> 09:59.880 |
|
there's something uniquely wrong |
|
|
|
09:59.880 --> 10:02.800 |
|
with machine learning as opposed to humans. |
|
|
|
10:02.800 --> 10:04.600 |
|
Also, do you see them as a tool |
|
|
|
10:04.600 --> 10:06.480 |
|
to improve the performance of the system? |
|
|
|
10:06.480 --> 10:10.760 |
|
Not on the security side, but literally just accuracy. |
|
|
|
10:10.760 --> 10:13.480 |
|
I do see them as a kind of tool on that side, |
|
|
|
10:13.480 --> 10:16.640 |
|
but maybe not quite as much as I used to think. |
|
|
|
10:16.640 --> 10:18.520 |
|
We've started to find that there's a trade off |
|
|
|
10:18.520 --> 10:21.680 |
|
between accuracy on adversarial examples |
|
|
|
10:21.680 --> 10:24.360 |
|
and accuracy on clean examples. |
|
|
|
10:24.360 --> 10:28.320 |
|
Back in 2014, when I did the first adversarily trained |
|
|
|
10:28.320 --> 10:30.840 |
|
classifier that showed resistance |
|
|
|
10:30.840 --> 10:33.040 |
|
to some kinds of adversarial examples, |
|
|
|
10:33.040 --> 10:36.040 |
|
it also got better at the clean data on MNIST. |
|
|
|
10:36.040 --> 10:37.720 |
|
And that's something we've replicated several times |
|
|
|
10:37.720 --> 10:39.640 |
|
on MNIST, that when we train |
|
|
|
10:39.640 --> 10:41.480 |
|
against weak adversarial examples, |
|
|
|
10:41.480 --> 10:43.880 |
|
MNIST classifiers get more accurate. |
|
|
|
10:43.880 --> 10:47.080 |
|
So far that hasn't really held up on other data sets |
|
|
|
10:47.080 --> 10:48.880 |
|
and hasn't held up when we train |
|
|
|
10:48.880 --> 10:50.720 |
|
against stronger adversaries. |
|
|
|
10:50.720 --> 10:53.160 |
|
It seems like when you confront |
|
|
|
10:53.160 --> 10:55.720 |
|
a really strong adversary, |
|
|
|
10:55.720 --> 10:58.080 |
|
you tend to have to give something up. |
|
|
|
10:58.080 --> 11:00.520 |
|
Interesting, but it's such a compelling idea |
|
|
|
11:00.520 --> 11:04.800 |
|
because it feels like that's how us humans learn |
|
|
|
11:04.800 --> 11:06.320 |
|
to do the difficult cases. |
|
|
|
11:06.320 --> 11:08.760 |
|
We try to think of what would we screw up |
|
|
|
11:08.760 --> 11:11.000 |
|
and then we make sure we fix that. |
|
|
|
11:11.000 --> 11:13.680 |
|
It's also in a lot of branches of engineering, |
|
|
|
11:13.680 --> 11:15.800 |
|
you do a worst case analysis |
|
|
|
11:15.800 --> 11:18.720 |
|
and make sure that your system will work in the worst case. |
|
|
|
11:18.720 --> 11:20.400 |
|
And then that guarantees that it'll work |
|
|
|
11:20.400 --> 11:24.360 |
|
in all of the messy average cases that happen |
|
|
|
11:24.360 --> 11:27.440 |
|
when you go out into a really randomized world. |
|
|
|
11:27.440 --> 11:29.560 |
|
Yeah, with driving with autonomous vehicles, |
|
|
|
11:29.560 --> 11:31.840 |
|
there seems to be a desire to just look |
|
|
|
11:31.840 --> 11:34.880 |
|
for think adversarially, |
|
|
|
11:34.880 --> 11:36.920 |
|
try to figure out how to mess up the system. |
|
|
|
11:36.920 --> 11:40.640 |
|
And if you can be robust to all those difficult cases, |
|
|
|
11:40.640 --> 11:43.600 |
|
then you can, it's a hand wavy empirical way |
|
|
|
11:43.600 --> 11:45.800 |
|
to show your system is safe. |
|
|
|
11:45.800 --> 11:47.000 |
|
Yeah, yeah. |
|
|
|
11:47.000 --> 11:49.120 |
|
Today, most adversarial example research |
|
|
|
11:49.120 --> 11:51.640 |
|
isn't really focused on a particular use case, |
|
|
|
11:51.640 --> 11:54.000 |
|
but there are a lot of different use cases |
|
|
|
11:54.000 --> 11:55.080 |
|
where you'd like to make sure |
|
|
|
11:55.080 --> 11:57.720 |
|
that the adversary can't interfere |
|
|
|
11:57.720 --> 12:00.200 |
|
with the operation of your system. |
|
|
|
12:00.200 --> 12:01.040 |
|
Like in finance, |
|
|
|
12:01.040 --> 12:03.320 |
|
if you have an algorithm making trades for you, |
|
|
|
12:03.320 --> 12:04.640 |
|
people go to a lot of an effort |
|
|
|
12:04.640 --> 12:06.680 |
|
to obfuscate their algorithm. |
|
|
|
12:06.680 --> 12:08.080 |
|
That's both to protect their IP |
|
|
|
12:08.080 --> 12:10.880 |
|
because you don't want to research |
|
|
|
12:10.880 --> 12:13.560 |
|
and develop a profitable trading algorithm |
|
|
|
12:13.560 --> 12:16.120 |
|
then have somebody else capture the gains. |
|
|
|
12:16.120 --> 12:17.160 |
|
But it's at least partly |
|
|
|
12:17.160 --> 12:19.000 |
|
because you don't want people to make adversarial |
|
|
|
12:19.000 --> 12:21.240 |
|
examples that fool your algorithm |
|
|
|
12:21.240 --> 12:22.560 |
|
into making bad trades. |
|
|
|
12:24.360 --> 12:26.560 |
|
Or I guess one area that's been popular |
|
|
|
12:26.560 --> 12:30.160 |
|
in the academic literature is speech recognition. |
|
|
|
12:30.160 --> 12:34.400 |
|
If you use speech recognition to hear an audio waveform |
|
|
|
12:34.400 --> 12:37.680 |
|
and then turn that into a command |
|
|
|
12:37.680 --> 12:39.640 |
|
that a phone executes for you, |
|
|
|
12:39.640 --> 12:41.840 |
|
you don't want a malicious adversary |
|
|
|
12:41.840 --> 12:43.600 |
|
to be able to produce audio |
|
|
|
12:43.600 --> 12:46.280 |
|
that gets interpreted as malicious commands, |
|
|
|
12:46.280 --> 12:47.800 |
|
especially if a human in the room |
|
|
|
12:47.800 --> 12:50.320 |
|
doesn't realize that something like that is happening. |
|
|
|
12:50.320 --> 12:52.000 |
|
In speech recognition, |
|
|
|
12:52.000 --> 12:53.920 |
|
has there been much success |
|
|
|
12:53.920 --> 12:58.440 |
|
in being able to create adversarial examples |
|
|
|
12:58.440 --> 12:59.760 |
|
that fool the system? |
|
|
|
12:59.760 --> 13:00.880 |
|
Yeah, actually. |
|
|
|
13:00.880 --> 13:02.440 |
|
I guess the first work that I'm aware of |
|
|
|
13:02.440 --> 13:05.120 |
|
is a paper called Hidden Voice Commands |
|
|
|
13:05.120 --> 13:08.480 |
|
that came out in 2016, I believe. |
|
|
|
13:08.480 --> 13:09.560 |
|
And they were able to show |
|
|
|
13:09.560 --> 13:11.920 |
|
that they could make sounds |
|
|
|
13:11.920 --> 13:14.960 |
|
that are not understandable by a human |
|
|
|
13:14.960 --> 13:18.400 |
|
but are recognized as the target phrase |
|
|
|
13:18.400 --> 13:21.360 |
|
that the attacker wants the phone to recognize it as. |
|
|
|
13:21.360 --> 13:24.040 |
|
Since then, things have gotten a little bit better |
|
|
|
13:24.040 --> 13:27.600 |
|
on the attacker side when worse on the defender side. |
|
|
|
13:28.680 --> 13:33.360 |
|
It's become possible to make sounds |
|
|
|
13:33.360 --> 13:35.600 |
|
that sound like normal speech |
|
|
|
13:35.600 --> 13:39.000 |
|
but are actually interpreted as a different sentence |
|
|
|
13:39.000 --> 13:40.720 |
|
than the human hears. |
|
|
|
13:40.720 --> 13:42.720 |
|
The level of perceptibility |
|
|
|
13:42.720 --> 13:45.360 |
|
of the adversarial perturbation is still kind of high. |
|
|
|
13:46.600 --> 13:48.160 |
|
When you listen to the recording, |
|
|
|
13:48.160 --> 13:51.040 |
|
it sounds like there's some noise in the background, |
|
|
|
13:51.040 --> 13:52.960 |
|
just like rustling sounds. |
|
|
|
13:52.960 --> 13:54.360 |
|
But those rustling sounds are actually |
|
|
|
13:54.360 --> 13:55.560 |
|
the adversarial perturbation |
|
|
|
13:55.560 --> 13:58.040 |
|
that makes the phone hear a completely different sentence. |
|
|
|
13:58.040 --> 14:00.120 |
|
Yeah, that's so fascinating. |
|
|
|
14:00.120 --> 14:01.640 |
|
Peter Norvig mentioned that you're writing |
|
|
|
14:01.640 --> 14:04.280 |
|
the deep learning chapter for the fourth edition |
|
|
|
14:04.280 --> 14:05.840 |
|
of the Artificial Intelligence, |
|
|
|
14:05.840 --> 14:07.320 |
|
the Modern Approach Book. |
|
|
|
14:07.320 --> 14:10.680 |
|
So how do you even begin summarizing |
|
|
|
14:10.680 --> 14:12.960 |
|
the field of deep learning in a chapter? |
|
|
|
14:12.960 --> 14:16.840 |
|
Well, in my case, I waited like a year |
|
|
|
14:16.840 --> 14:19.080 |
|
before I actually wrote anything. |
|
|
|
14:19.080 --> 14:20.280 |
|
Is it? |
|
|
|
14:20.280 --> 14:22.600 |
|
Even having written a full length textbook before, |
|
|
|
14:22.600 --> 14:25.560 |
|
it's still pretty intimidating |
|
|
|
14:25.560 --> 14:27.800 |
|
to try to start writing just one chapter |
|
|
|
14:27.800 --> 14:29.040 |
|
that covers everything. |
|
|
|
14:31.080 --> 14:33.160 |
|
One thing that helped me make that plan |
|
|
|
14:33.160 --> 14:34.280 |
|
was actually the experience |
|
|
|
14:34.280 --> 14:36.680 |
|
of having written the full book before |
|
|
|
14:36.680 --> 14:39.080 |
|
and then watching how the field changed |
|
|
|
14:39.080 --> 14:40.920 |
|
after the book came out. |
|
|
|
14:40.920 --> 14:42.280 |
|
I realized there's a lot of topics |
|
|
|
14:42.280 --> 14:44.960 |
|
that were maybe extraneous in the first book |
|
|
|
14:44.960 --> 14:47.560 |
|
and just seeing what stood the test |
|
|
|
14:47.560 --> 14:49.360 |
|
of a few years of being published |
|
|
|
14:49.360 --> 14:52.160 |
|
and what seems a little bit less important |
|
|
|
14:52.160 --> 14:53.760 |
|
to have included now helped me pare down |
|
|
|
14:53.760 --> 14:55.920 |
|
the topics I wanted to cover for the book. |
|
|
|
14:56.840 --> 14:59.560 |
|
It's also really nice now that the field |
|
|
|
14:59.560 --> 15:00.920 |
|
is kind of stabilized to the point |
|
|
|
15:00.920 --> 15:04.720 |
|
where some core ideas from the 1980s are still used today. |
|
|
|
15:04.720 --> 15:06.640 |
|
When I first started studying machine learning, |
|
|
|
15:06.640 --> 15:09.520 |
|
almost everything from the 1980s had been rejected |
|
|
|
15:09.520 --> 15:11.320 |
|
and now some of it has come back. |
|
|
|
15:11.320 --> 15:13.440 |
|
So that stuff that's really stood the test of time |
|
|
|
15:13.440 --> 15:15.880 |
|
is what I focused on putting into the book. |
|
|
|
15:16.880 --> 15:21.240 |
|
There's also, I guess, two different philosophies |
|
|
|
15:21.240 --> 15:23.120 |
|
about how you might write a book. |
|
|
|
15:23.120 --> 15:24.760 |
|
One philosophy is you try to write a reference |
|
|
|
15:24.760 --> 15:26.160 |
|
that covers everything. |
|
|
|
15:26.160 --> 15:27.960 |
|
The other philosophy is you try to provide |
|
|
|
15:27.960 --> 15:30.320 |
|
a high level summary that gives people |
|
|
|
15:30.320 --> 15:32.360 |
|
the language to understand a field |
|
|
|
15:32.360 --> 15:34.920 |
|
and tells them what the most important concepts are. |
|
|
|
15:34.920 --> 15:37.080 |
|
The first deep learning book that I wrote |
|
|
|
15:37.080 --> 15:39.240 |
|
with Joshua and Aaron was somewhere |
|
|
|
15:39.240 --> 15:41.240 |
|
between the two philosophies, |
|
|
|
15:41.240 --> 15:43.640 |
|
that it's trying to be both a reference |
|
|
|
15:43.640 --> 15:45.760 |
|
and an introductory guide. |
|
|
|
15:45.760 --> 15:48.920 |
|
Writing this chapter for Russell and Norvig's book, |
|
|
|
15:48.920 --> 15:52.800 |
|
I was able to focus more on just a concise introduction |
|
|
|
15:52.800 --> 15:54.240 |
|
of the key concepts and the language |
|
|
|
15:54.240 --> 15:56.000 |
|
you need to read about them more. |
|
|
|
15:56.000 --> 15:57.560 |
|
In a lot of cases, I actually just wrote paragraphs |
|
|
|
15:57.560 --> 16:00.080 |
|
that said, here's a rapidly evolving area |
|
|
|
16:00.080 --> 16:02.400 |
|
that you should pay attention to. |
|
|
|
16:02.400 --> 16:04.760 |
|
It's pointless to try to tell you what the latest |
|
|
|
16:04.760 --> 16:09.760 |
|
and best version of a learn to learn model is. |
|
|
|
16:11.440 --> 16:13.640 |
|
I can point you to a paper that's recent right now, |
|
|
|
16:13.640 --> 16:16.880 |
|
but there isn't a whole lot of a reason to delve |
|
|
|
16:16.880 --> 16:20.440 |
|
into exactly what's going on with the latest |
|
|
|
16:20.440 --> 16:22.960 |
|
learning to learn approach or the latest module |
|
|
|
16:22.960 --> 16:24.960 |
|
produced by a learning to learn algorithm. |
|
|
|
16:24.960 --> 16:26.760 |
|
You should know that learning to learn is a thing |
|
|
|
16:26.760 --> 16:29.480 |
|
and that it may very well be the source |
|
|
|
16:29.480 --> 16:32.200 |
|
of the latest and greatest convolutional net |
|
|
|
16:32.200 --> 16:34.520 |
|
or recurrent net module that you would want to use |
|
|
|
16:34.520 --> 16:36.040 |
|
in your latest project. |
|
|
|
16:36.040 --> 16:38.200 |
|
But there isn't a lot of point in trying to summarize |
|
|
|
16:38.200 --> 16:42.280 |
|
exactly which architecture and which learning approach |
|
|
|
16:42.280 --> 16:44.040 |
|
got to which level of performance. |
|
|
|
16:44.040 --> 16:49.040 |
|
So you maybe focus more on the basics of the methodology. |
|
|
|
16:49.240 --> 16:52.480 |
|
So from back propagation to feed forward |
|
|
|
16:52.480 --> 16:55.160 |
|
to recurrent networks, convolutional, that kind of thing. |
|
|
|
16:55.160 --> 16:56.480 |
|
Yeah, yeah. |
|
|
|
16:56.480 --> 17:00.320 |
|
So if I were to ask you, I remember I took algorithms |
|
|
|
17:00.320 --> 17:03.720 |
|
and data structures algorithms, of course. |
|
|
|
17:03.720 --> 17:08.120 |
|
I remember the professor asked, what is an algorithm? |
|
|
|
17:09.200 --> 17:12.200 |
|
And he yelled at everybody in a good way |
|
|
|
17:12.200 --> 17:14.040 |
|
that nobody was answering it correctly. |
|
|
|
17:14.040 --> 17:16.360 |
|
Everybody knew what the algorithm, it was graduate course. |
|
|
|
17:16.360 --> 17:18.120 |
|
Everybody knew what an algorithm was, |
|
|
|
17:18.120 --> 17:19.800 |
|
but they weren't able to answer it well. |
|
|
|
17:19.800 --> 17:22.360 |
|
So let me ask you, in that same spirit, |
|
|
|
17:22.360 --> 17:23.580 |
|
what is deep learning? |
|
|
|
17:24.520 --> 17:29.520 |
|
I would say deep learning is any kind of machine learning |
|
|
|
17:29.520 --> 17:34.520 |
|
that involves learning parameters of more than one |
|
|
|
17:34.720 --> 17:36.020 |
|
consecutive step. |
|
|
|
17:37.280 --> 17:40.760 |
|
So that, I mean, shallow learning is things where |
|
|
|
17:40.760 --> 17:43.760 |
|
you learn a lot of operations that happen in parallel. |
|
|
|
17:43.760 --> 17:46.720 |
|
You might have a system that makes multiple steps, |
|
|
|
17:46.720 --> 17:51.000 |
|
like you might have hand designed feature extractors, |
|
|
|
17:51.000 --> 17:52.600 |
|
but really only one step is learned. |
|
|
|
17:52.600 --> 17:55.440 |
|
Deep learning is anything where you have multiple |
|
|
|
17:55.440 --> 17:56.880 |
|
operations in sequence. |
|
|
|
17:56.880 --> 17:59.400 |
|
And that includes the things that are really popular |
|
|
|
17:59.400 --> 18:01.280 |
|
today, like convolutional networks |
|
|
|
18:01.280 --> 18:04.640 |
|
and recurrent networks, but it also includes some |
|
|
|
18:04.640 --> 18:08.280 |
|
of the things that have died out, like Bolton machines, |
|
|
|
18:08.280 --> 18:10.880 |
|
where we weren't using back propagation. |
|
|
|
18:11.960 --> 18:14.240 |
|
Today, I hear a lot of people define deep learning |
|
|
|
18:14.240 --> 18:19.240 |
|
as gradient descent applied to these differentiable |
|
|
|
18:20.400 --> 18:24.240 |
|
functions, and I think that's a legitimate usage |
|
|
|
18:24.240 --> 18:25.920 |
|
of the term, it's just different from the way |
|
|
|
18:25.920 --> 18:27.800 |
|
that I use the term myself. |
|
|
|
18:27.800 --> 18:32.360 |
|
So what's an example of deep learning that is not |
|
|
|
18:32.360 --> 18:34.720 |
|
gradient descent and differentiable functions? |
|
|
|
18:34.720 --> 18:37.400 |
|
In your, I mean, not specifically perhaps, |
|
|
|
18:37.400 --> 18:39.760 |
|
but more even looking into the future. |
|
|
|
18:39.760 --> 18:44.300 |
|
What's your thought about that space of approaches? |
|
|
|
18:44.300 --> 18:46.340 |
|
Yeah, so I tend to think of machine learning algorithms |
|
|
|
18:46.340 --> 18:50.200 |
|
as decomposed into really three different pieces. |
|
|
|
18:50.200 --> 18:53.000 |
|
There's the model, which can be something like a neural net |
|
|
|
18:53.000 --> 18:56.600 |
|
or a Bolton machine or a recurrent model. |
|
|
|
18:56.600 --> 18:59.520 |
|
And that basically just describes how do you take data |
|
|
|
18:59.520 --> 19:03.480 |
|
and how do you take parameters and what function do you use |
|
|
|
19:03.480 --> 19:07.320 |
|
to make a prediction given the data and the parameters? |
|
|
|
19:07.320 --> 19:09.920 |
|
Another piece of the learning algorithm is |
|
|
|
19:09.920 --> 19:13.880 |
|
the optimization algorithm, or not every algorithm |
|
|
|
19:13.880 --> 19:15.920 |
|
can be really described in terms of optimization, |
|
|
|
19:15.920 --> 19:18.880 |
|
but what's the algorithm for updating the parameters |
|
|
|
19:18.880 --> 19:21.680 |
|
or updating whatever the state of the network is? |
|
|
|
19:22.600 --> 19:26.280 |
|
And then the last part is the data set, |
|
|
|
19:26.280 --> 19:29.200 |
|
like how do you actually represent the world |
|
|
|
19:29.200 --> 19:32.120 |
|
as it comes into your machine learning system? |
|
|
|
19:33.160 --> 19:35.800 |
|
So I think of deep learning as telling us something |
|
|
|
19:35.800 --> 19:39.040 |
|
about what does the model look like? |
|
|
|
19:39.040 --> 19:41.240 |
|
And basically to qualify as deep, |
|
|
|
19:41.240 --> 19:44.560 |
|
I say that it just has to have multiple layers. |
|
|
|
19:44.560 --> 19:47.360 |
|
That can be multiple steps in a feed forward |
|
|
|
19:47.360 --> 19:49.240 |
|
differentiable computation. |
|
|
|
19:49.240 --> 19:52.040 |
|
That can be multiple layers in a graphical model. |
|
|
|
19:52.040 --> 19:53.560 |
|
There's a lot of ways that you could satisfy me |
|
|
|
19:53.560 --> 19:56.160 |
|
that something has multiple steps |
|
|
|
19:56.160 --> 19:58.920 |
|
that are each parameterized separately. |
|
|
|
19:58.920 --> 20:00.640 |
|
I think of gradient descent as being all about |
|
|
|
20:00.640 --> 20:01.560 |
|
that other piece, |
|
|
|
20:01.560 --> 20:04.240 |
|
the how do you actually update the parameters piece? |
|
|
|
20:04.240 --> 20:05.960 |
|
So you could imagine having a deep model |
|
|
|
20:05.960 --> 20:08.680 |
|
like a convolutional net and training it with something |
|
|
|
20:08.680 --> 20:11.280 |
|
like evolution or a genetic algorithm. |
|
|
|
20:11.280 --> 20:14.640 |
|
And I would say that still qualifies as deep learning. |
|
|
|
20:14.640 --> 20:16.040 |
|
And then in terms of models |
|
|
|
20:16.040 --> 20:18.760 |
|
that aren't necessarily differentiable, |
|
|
|
20:18.760 --> 20:22.480 |
|
I guess Bolton machines are probably the main example |
|
|
|
20:22.480 --> 20:25.560 |
|
of something where you can't really take a derivative |
|
|
|
20:25.560 --> 20:28.000 |
|
and use that for the learning process. |
|
|
|
20:28.000 --> 20:32.320 |
|
But you can still argue that the model has many steps |
|
|
|
20:32.320 --> 20:33.760 |
|
of processing that it applies |
|
|
|
20:33.760 --> 20:35.800 |
|
when you run inference in the model. |
|
|
|
20:35.800 --> 20:38.960 |
|
So it's the steps of processing that's key. |
|
|
|
20:38.960 --> 20:41.320 |
|
So Jeff Hinton suggests that we need to throw away |
|
|
|
20:41.320 --> 20:44.960 |
|
back propagation and start all over. |
|
|
|
20:44.960 --> 20:46.520 |
|
What do you think about that? |
|
|
|
20:46.520 --> 20:48.600 |
|
What could an alternative direction |
|
|
|
20:48.600 --> 20:51.000 |
|
of training neural networks look like? |
|
|
|
20:51.000 --> 20:52.880 |
|
I don't know that back propagation |
|
|
|
20:52.880 --> 20:54.680 |
|
is going to go away entirely. |
|
|
|
20:54.680 --> 20:57.120 |
|
Most of the time when we decide |
|
|
|
20:57.120 --> 20:59.200 |
|
that a machine learning algorithm |
|
|
|
20:59.200 --> 21:03.440 |
|
isn't on the critical path to research for improving AI, |
|
|
|
21:03.440 --> 21:04.640 |
|
the algorithm doesn't die, |
|
|
|
21:04.640 --> 21:07.760 |
|
it just becomes used for some specialized set of things. |
|
|
|
21:08.760 --> 21:11.160 |
|
A lot of algorithms like logistic regression |
|
|
|
21:11.160 --> 21:14.000 |
|
don't seem that exciting to AI researchers |
|
|
|
21:14.000 --> 21:16.760 |
|
who are working on things like speech recognition |
|
|
|
21:16.760 --> 21:18.400 |
|
or autonomous cars today, |
|
|
|
21:18.400 --> 21:21.080 |
|
but there's still a lot of use for logistic regression |
|
|
|
21:21.080 --> 21:23.960 |
|
and things like analyzing really noisy data |
|
|
|
21:23.960 --> 21:25.640 |
|
in medicine and finance |
|
|
|
21:25.640 --> 21:28.720 |
|
or making really rapid predictions |
|
|
|
21:28.720 --> 21:30.680 |
|
in really time limited contexts. |
|
|
|
21:30.680 --> 21:33.440 |
|
So I think back propagation and gradient descent |
|
|
|
21:33.440 --> 21:34.520 |
|
are around to stay, |
|
|
|
21:34.520 --> 21:38.760 |
|
but they may not end up being everything |
|
|
|
21:38.760 --> 21:40.840 |
|
that we need to get to real human level |
|
|
|
21:40.840 --> 21:42.360 |
|
or super human AI. |
|
|
|
21:42.360 --> 21:44.680 |
|
Are you optimistic about us discovering? |
|
|
|
21:44.680 --> 21:49.680 |
|
You know, back propagation has been around for a few decades. |
|
|
|
21:50.240 --> 21:54.080 |
|
So are you optimistic about us as a community |
|
|
|
21:54.080 --> 21:56.800 |
|
being able to discover something better? |
|
|
|
21:56.800 --> 21:57.640 |
|
Yeah, I am. |
|
|
|
21:57.640 --> 22:01.840 |
|
I think we likely will find something that works better. |
|
|
|
22:01.840 --> 22:05.520 |
|
You could imagine things like having stacks of models |
|
|
|
22:05.520 --> 22:08.720 |
|
where some of the lower level models predict parameters |
|
|
|
22:08.720 --> 22:10.200 |
|
of the higher level models. |
|
|
|
22:10.200 --> 22:12.160 |
|
And so at the top level, |
|
|
|
22:12.160 --> 22:13.480 |
|
you're not learning in terms of literally |
|
|
|
22:13.480 --> 22:15.800 |
|
calculating gradients, but just predicting |
|
|
|
22:15.800 --> 22:17.680 |
|
how different values will perform. |
|
|
|
22:17.680 --> 22:19.560 |
|
You can kind of see that already in some areas |
|
|
|
22:19.560 --> 22:21.400 |
|
like Bayesian optimization, |
|
|
|
22:21.400 --> 22:22.960 |
|
where you have a Gaussian process |
|
|
|
22:22.960 --> 22:24.800 |
|
that predicts how well different parameter values |
|
|
|
22:24.800 --> 22:25.880 |
|
will perform. |
|
|
|
22:25.880 --> 22:27.680 |
|
We already use those kinds of algorithms |
|
|
|
22:27.680 --> 22:30.240 |
|
for things like hyper parameter optimization. |
|
|
|
22:30.240 --> 22:31.640 |
|
And in general, we know a lot of things |
|
|
|
22:31.640 --> 22:33.240 |
|
other than back prop that work really well |
|
|
|
22:33.240 --> 22:35.000 |
|
for specific problems. |
|
|
|
22:35.000 --> 22:38.240 |
|
The main thing we haven't found is a way of taking one |
|
|
|
22:38.240 --> 22:41.160 |
|
of these other non back prop based algorithms |
|
|
|
22:41.160 --> 22:43.520 |
|
and having it really advance the state of the art |
|
|
|
22:43.520 --> 22:46.160 |
|
on an AI level problem. |
|
|
|
22:46.160 --> 22:47.120 |
|
Right. |
|
|
|
22:47.120 --> 22:49.600 |
|
But I wouldn't be surprised if eventually we find |
|
|
|
22:49.600 --> 22:51.560 |
|
that some of these algorithms that, |
|
|
|
22:51.560 --> 22:52.760 |
|
even the ones that already exist, |
|
|
|
22:52.760 --> 22:54.200 |
|
not even necessarily a new one, |
|
|
|
22:54.200 --> 22:59.200 |
|
we might find some way of customizing one of these algorithms |
|
|
|
22:59.200 --> 23:00.560 |
|
to do something really interesting |
|
|
|
23:00.560 --> 23:05.240 |
|
at the level of cognition or the level of, |
|
|
|
23:06.400 --> 23:08.680 |
|
I think one system that we really don't have working |
|
|
|
23:08.680 --> 23:12.920 |
|
quite right yet is like short term memory. |
|
|
|
23:12.920 --> 23:14.480 |
|
We have things like LSTMs, |
|
|
|
23:14.480 --> 23:17.000 |
|
they're called long short term memory. |
|
|
|
23:17.000 --> 23:20.000 |
|
They still don't do quite what a human does |
|
|
|
23:20.000 --> 23:21.720 |
|
with short term memory. |
|
|
|
23:22.840 --> 23:26.920 |
|
Like gradient descent to learn a specific fact |
|
|
|
23:26.920 --> 23:29.360 |
|
has to do multiple steps on that fact. |
|
|
|
23:29.360 --> 23:34.120 |
|
Like if I tell you, the meeting today is at 3pm, |
|
|
|
23:34.120 --> 23:35.440 |
|
I don't need to say over and over again. |
|
|
|
23:35.440 --> 23:38.640 |
|
It's at 3pm, it's at 3pm, it's at 3pm, it's at 3pm. |
|
|
|
23:38.640 --> 23:40.400 |
|
For you to do a gradient step on each one, |
|
|
|
23:40.400 --> 23:43.160 |
|
you just hear it once and you remember it. |
|
|
|
23:43.160 --> 23:46.920 |
|
There's been some work on things like self attention |
|
|
|
23:46.920 --> 23:50.400 |
|
and attention like mechanisms like the neural Turing machine |
|
|
|
23:50.400 --> 23:53.160 |
|
that can write to memory cells and update themselves |
|
|
|
23:53.160 --> 23:54.880 |
|
with facts like that right away. |
|
|
|
23:54.880 --> 23:56.880 |
|
But I don't think we've really nailed it yet. |
|
|
|
23:56.880 --> 24:01.880 |
|
And that's one area where I'd imagine that new optimization |
|
|
|
24:02.080 --> 24:04.240 |
|
algorithms or different ways of applying existing |
|
|
|
24:04.240 --> 24:07.280 |
|
optimization algorithms could give us a way |
|
|
|
24:07.280 --> 24:10.120 |
|
of just lightning fast updating the state |
|
|
|
24:10.120 --> 24:12.400 |
|
of a machine learning system to contain |
|
|
|
24:12.400 --> 24:14.920 |
|
a specific fact like that without needing to have it |
|
|
|
24:14.920 --> 24:17.000 |
|
presented over and over and over again. |
|
|
|
24:17.000 --> 24:21.440 |
|
So some of the success of symbolic systems in the 80s |
|
|
|
24:21.440 --> 24:26.200 |
|
is they were able to assemble these kinds of facts better. |
|
|
|
24:26.200 --> 24:29.080 |
|
But there's a lot of expert input required |
|
|
|
24:29.080 --> 24:31.120 |
|
and it's very limited in that sense. |
|
|
|
24:31.120 --> 24:34.720 |
|
Do you ever look back to that as something |
|
|
|
24:34.720 --> 24:36.560 |
|
that we'll have to return to eventually |
|
|
|
24:36.560 --> 24:38.440 |
|
sort of dust off the book from the shelf |
|
|
|
24:38.440 --> 24:42.400 |
|
and think about how we build knowledge, representation, |
|
|
|
24:42.400 --> 24:43.240 |
|
knowledge. |
|
|
|
24:43.240 --> 24:44.840 |
|
Like will we have to use graph searches? |
|
|
|
24:44.840 --> 24:45.800 |
|
Graph searches, right. |
|
|
|
24:45.800 --> 24:47.720 |
|
And like first order logic and entailment |
|
|
|
24:47.720 --> 24:48.560 |
|
and things like that. |
|
|
|
24:48.560 --> 24:49.560 |
|
That kind of thing, yeah, exactly. |
|
|
|
24:49.560 --> 24:51.200 |
|
In my particular line of work, |
|
|
|
24:51.200 --> 24:54.560 |
|
which has mostly been machine learning security |
|
|
|
24:54.560 --> 24:56.720 |
|
and also generative modeling, |
|
|
|
24:56.720 --> 25:00.560 |
|
I haven't usually found myself moving in that direction. |
|
|
|
25:00.560 --> 25:03.520 |
|
For generative models, I could see a little bit of, |
|
|
|
25:03.520 --> 25:06.520 |
|
it could be useful if you had something like a, |
|
|
|
25:06.520 --> 25:09.680 |
|
a differentiable knowledge base |
|
|
|
25:09.680 --> 25:11.000 |
|
or some other kind of knowledge base |
|
|
|
25:11.000 --> 25:13.840 |
|
where it's possible for some of our fuzzier |
|
|
|
25:13.840 --> 25:16.880 |
|
machine learning algorithms to interact with a knowledge base. |
|
|
|
25:16.880 --> 25:19.040 |
|
I mean, your network is kind of like that. |
|
|
|
25:19.040 --> 25:21.440 |
|
It's a differentiable knowledge base of sorts. |
|
|
|
25:21.440 --> 25:22.280 |
|
Yeah. |
|
|
|
25:22.280 --> 25:27.280 |
|
But if we had a really easy way of giving feedback |
|
|
|
25:27.600 --> 25:29.240 |
|
to machine learning models, |
|
|
|
25:29.240 --> 25:32.400 |
|
that would clearly help a lot with, with generative models. |
|
|
|
25:32.400 --> 25:34.680 |
|
And so you could imagine one way of getting there would be, |
|
|
|
25:34.680 --> 25:36.720 |
|
get a lot better at natural language processing. |
|
|
|
25:36.720 --> 25:38.920 |
|
But another way of getting there would be, |
|
|
|
25:38.920 --> 25:40.280 |
|
take some kind of knowledge base |
|
|
|
25:40.280 --> 25:42.800 |
|
and figure out a way for it to actually interact |
|
|
|
25:42.800 --> 25:44.080 |
|
with a neural network. |
|
|
|
25:44.080 --> 25:46.080 |
|
Being able to have a chat with a neural network. |
|
|
|
25:46.080 --> 25:47.920 |
|
Yeah. |
|
|
|
25:47.920 --> 25:50.920 |
|
So like one thing in generative models we see a lot today is, |
|
|
|
25:50.920 --> 25:54.480 |
|
you'll get things like faces that are not symmetrical. |
|
|
|
25:54.480 --> 25:56.800 |
|
Like, like people that have two eyes |
|
|
|
25:56.800 --> 25:58.200 |
|
that are different colors. |
|
|
|
25:58.200 --> 25:59.560 |
|
And I mean, there are people with eyes |
|
|
|
25:59.560 --> 26:00.840 |
|
that are different colors in real life, |
|
|
|
26:00.840 --> 26:03.480 |
|
but not nearly as many of them as you tend to see |
|
|
|
26:03.480 --> 26:06.120 |
|
in the machine learning generated data. |
|
|
|
26:06.120 --> 26:08.120 |
|
So if you had either a knowledge base |
|
|
|
26:08.120 --> 26:10.200 |
|
that could contain the fact, |
|
|
|
26:10.200 --> 26:13.360 |
|
people's faces are generally approximately symmetric |
|
|
|
26:13.360 --> 26:15.920 |
|
and eye color is especially likely |
|
|
|
26:15.920 --> 26:17.920 |
|
to be the same on both sides. |
|
|
|
26:17.920 --> 26:20.160 |
|
Being able to just inject that hint |
|
|
|
26:20.160 --> 26:22.000 |
|
into the machine learning model |
|
|
|
26:22.000 --> 26:23.800 |
|
without having to discover that itself |
|
|
|
26:23.800 --> 26:25.760 |
|
after studying a lot of data |
|
|
|
26:25.760 --> 26:28.360 |
|
would be a really useful feature. |
|
|
|
26:28.360 --> 26:30.120 |
|
I could see a lot of ways of getting there |
|
|
|
26:30.120 --> 26:32.200 |
|
without bringing back some of the 1980s technology, |
|
|
|
26:32.200 --> 26:35.160 |
|
but I also see some ways that you could imagine |
|
|
|
26:35.160 --> 26:38.240 |
|
extending the 1980s technology to play nice with neural nets |
|
|
|
26:38.240 --> 26:40.040 |
|
and have it help get there. |
|
|
|
26:40.040 --> 26:40.880 |
|
Awesome. |
|
|
|
26:40.880 --> 26:44.360 |
|
So you talked about the story of you coming up |
|
|
|
26:44.360 --> 26:47.040 |
|
with the idea of GANs at a bar with some friends. |
|
|
|
26:47.040 --> 26:50.400 |
|
You were arguing that this, you know, |
|
|
|
26:50.400 --> 26:53.080 |
|
GANs would work generative adversarial networks |
|
|
|
26:53.080 --> 26:54.680 |
|
and the others didn't think so. |
|
|
|
26:54.680 --> 26:58.400 |
|
Then you went home at midnight, coded up and it worked. |
|
|
|
26:58.400 --> 27:01.320 |
|
So if I was a friend of yours at the bar, |
|
|
|
27:01.320 --> 27:02.720 |
|
I would also have doubts. |
|
|
|
27:02.720 --> 27:03.880 |
|
It's a really nice idea, |
|
|
|
27:03.880 --> 27:06.800 |
|
but I'm very skeptical that it would work. |
|
|
|
27:06.800 --> 27:09.280 |
|
What was the basis of their skepticism? |
|
|
|
27:09.280 --> 27:13.200 |
|
What was the basis of your intuition why it should work? |
|
|
|
27:14.360 --> 27:16.840 |
|
I don't wanna be someone who goes around promoting alcohol |
|
|
|
27:16.840 --> 27:18.280 |
|
for the purposes of science, |
|
|
|
27:18.280 --> 27:21.040 |
|
but in this case, I do actually think |
|
|
|
27:21.040 --> 27:23.080 |
|
that drinking helped a little bit. |
|
|
|
27:23.080 --> 27:25.360 |
|
When your inhibitions are lowered, |
|
|
|
27:25.360 --> 27:27.400 |
|
you're more willing to try out things |
|
|
|
27:27.400 --> 27:29.640 |
|
that you wouldn't try out otherwise. |
|
|
|
27:29.640 --> 27:32.480 |
|
So I have noticed in general |
|
|
|
27:32.480 --> 27:34.560 |
|
that I'm less prone to shooting down some of my own ideas |
|
|
|
27:34.560 --> 27:37.960 |
|
when I have had a little bit to drink. |
|
|
|
27:37.960 --> 27:40.800 |
|
I think if I had had that idea at lunchtime, |
|
|
|
27:40.800 --> 27:42.280 |
|
I probably would have thought it. |
|
|
|
27:42.280 --> 27:43.720 |
|
It's hard enough to train one neural net. |
|
|
|
27:43.720 --> 27:44.880 |
|
You can't train a second neural net |
|
|
|
27:44.880 --> 27:48.080 |
|
in the inner loop of the outer neural net. |
|
|
|
27:48.080 --> 27:49.800 |
|
That was basically my friend's objection |
|
|
|
27:49.800 --> 27:52.720 |
|
was that trying to train two neural nets at the same time |
|
|
|
27:52.720 --> 27:54.280 |
|
would be too hard. |
|
|
|
27:54.280 --> 27:56.120 |
|
So it was more about the training process |
|
|
|
27:56.120 --> 28:01.120 |
|
unless, so my skepticism would be, I'm sure you could train it |
|
|
|
28:01.160 --> 28:03.200 |
|
but the thing would converge to |
|
|
|
28:03.200 --> 28:05.840 |
|
would not be able to generate anything reasonable |
|
|
|
28:05.840 --> 28:08.240 |
|
and any kind of reasonable realism. |
|
|
|
28:08.240 --> 28:11.360 |
|
Yeah, so part of what all of us were thinking about |
|
|
|
28:11.360 --> 28:15.280 |
|
when we had this conversation was deep Bolton machines, |
|
|
|
28:15.280 --> 28:17.000 |
|
which a lot of us in the lab, including me, |
|
|
|
28:17.000 --> 28:19.480 |
|
were a big fan of deep Bolton machines at the time. |
|
|
|
28:20.640 --> 28:24.240 |
|
They involved two separate processes running at the same time. |
|
|
|
28:24.240 --> 28:27.400 |
|
One of them is called the positive phase |
|
|
|
28:27.400 --> 28:30.440 |
|
where you load data into the model |
|
|
|
28:30.440 --> 28:32.920 |
|
and tell the model to make the data more likely. |
|
|
|
28:32.920 --> 28:34.480 |
|
The other one is called the negative phase |
|
|
|
28:34.480 --> 28:36.280 |
|
where you draw samples from the model |
|
|
|
28:36.280 --> 28:38.880 |
|
and tell the model to make those samples less likely. |
|
|
|
28:40.480 --> 28:42.400 |
|
In a deep Bolton machine, it's not trivial |
|
|
|
28:42.400 --> 28:43.320 |
|
to generate a sample. |
|
|
|
28:43.320 --> 28:46.280 |
|
You have to actually run an iterative process |
|
|
|
28:46.280 --> 28:48.520 |
|
that gets better and better samples |
|
|
|
28:48.520 --> 28:50.720 |
|
coming closer and closer to the distribution |
|
|
|
28:50.720 --> 28:52.120 |
|
the model represents. |
|
|
|
28:52.120 --> 28:53.240 |
|
So during the training process, |
|
|
|
28:53.240 --> 28:56.560 |
|
you're always running these two systems at the same time. |
|
|
|
28:56.560 --> 28:58.360 |
|
One that's updating the parameters of the model |
|
|
|
28:58.360 --> 28:59.880 |
|
and another one that's trying to generate samples |
|
|
|
28:59.880 --> 29:01.120 |
|
from the model. |
|
|
|
29:01.120 --> 29:03.720 |
|
And they worked really well on things like MNIST, |
|
|
|
29:03.720 --> 29:05.200 |
|
but a lot of us in the lab, including me, |
|
|
|
29:05.200 --> 29:08.840 |
|
had tried to get deep Bolton machines to scale past MNIST |
|
|
|
29:08.840 --> 29:11.320 |
|
to things like generating color photos, |
|
|
|
29:11.320 --> 29:13.480 |
|
and we just couldn't get the two processes |
|
|
|
29:13.480 --> 29:15.360 |
|
to stay synchronized. |
|
|
|
29:16.720 --> 29:18.120 |
|
So when I had the idea for GANs, |
|
|
|
29:18.120 --> 29:19.720 |
|
a lot of people thought that the discriminator |
|
|
|
29:19.720 --> 29:21.960 |
|
would have more or less the same problem |
|
|
|
29:21.960 --> 29:25.360 |
|
as the negative phase in the Bolton machine, |
|
|
|
29:25.360 --> 29:27.840 |
|
that trying to train the discriminator in the inner loop, |
|
|
|
29:27.840 --> 29:29.960 |
|
you just couldn't get it to keep up |
|
|
|
29:29.960 --> 29:31.560 |
|
with the generator in the outer loop. |
|
|
|
29:31.560 --> 29:33.360 |
|
And that would prevent it from |
|
|
|
29:33.360 --> 29:35.240 |
|
converging to anything useful. |
|
|
|
29:35.240 --> 29:36.880 |
|
Yeah, I share that intuition. |
|
|
|
29:36.880 --> 29:37.720 |
|
Yeah. |
|
|
|
29:39.560 --> 29:42.000 |
|
But turns out to not be the case. |
|
|
|
29:42.000 --> 29:43.800 |
|
A lot of the time with machine learning algorithms, |
|
|
|
29:43.800 --> 29:45.200 |
|
it's really hard to predict ahead of time |
|
|
|
29:45.200 --> 29:46.960 |
|
how well they'll actually perform. |
|
|
|
29:46.960 --> 29:48.160 |
|
You have to just run the experiment |
|
|
|
29:48.160 --> 29:49.200 |
|
and see what happens. |
|
|
|
29:49.200 --> 29:53.480 |
|
And I would say I still today don't have like one factor |
|
|
|
29:53.480 --> 29:54.840 |
|
I can put my finger on and say, |
|
|
|
29:54.840 --> 29:58.360 |
|
this is why GANs worked for photo generation |
|
|
|
29:58.360 --> 30:00.240 |
|
and deep Bolton machines don't. |
|
|
|
30:02.000 --> 30:04.560 |
|
There are a lot of theory papers showing that |
|
|
|
30:04.560 --> 30:06.400 |
|
under some theoretical settings, |
|
|
|
30:06.400 --> 30:09.640 |
|
the GAN algorithm does actually converge. |
|
|
|
30:10.720 --> 30:14.200 |
|
But those settings are restricted enough |
|
|
|
30:14.200 --> 30:17.560 |
|
that they don't necessarily explain the whole picture |
|
|
|
30:17.560 --> 30:20.760 |
|
in terms of all the results that we see in practice. |
|
|
|
30:20.760 --> 30:22.360 |
|
So taking a step back, |
|
|
|
30:22.360 --> 30:24.880 |
|
can you, in the same way as we talked about deep learning, |
|
|
|
30:24.880 --> 30:28.440 |
|
can you tell me what generative adversarial networks are? |
|
|
|
30:29.480 --> 30:31.400 |
|
Yeah, so generative adversarial networks |
|
|
|
30:31.400 --> 30:34.000 |
|
are a particular kind of generative model. |
|
|
|
30:34.000 --> 30:36.320 |
|
A generative model is a machine learning model |
|
|
|
30:36.320 --> 30:38.880 |
|
that can train on some set of data. |
|
|
|
30:38.880 --> 30:41.280 |
|
Like say you have a collection of photos of cats |
|
|
|
30:41.280 --> 30:44.040 |
|
and you want to generate more photos of cats, |
|
|
|
30:44.040 --> 30:47.120 |
|
or you want to estimate a probability distribution |
|
|
|
30:47.120 --> 30:49.840 |
|
over cats so you can ask how likely it is |
|
|
|
30:49.840 --> 30:51.840 |
|
that some new image is a photo of a cat. |
|
|
|
30:52.920 --> 30:55.840 |
|
GANs are one way of doing this. |
|
|
|
30:55.840 --> 30:59.200 |
|
Some generative models are good at creating new data. |
|
|
|
30:59.200 --> 31:00.840 |
|
Other generative models are good |
|
|
|
31:00.840 --> 31:02.600 |
|
at estimating that density function |
|
|
|
31:02.600 --> 31:06.600 |
|
and telling you how likely particular pieces of data are |
|
|
|
31:06.600 --> 31:09.760 |
|
to come from the same distribution as the training data. |
|
|
|
31:09.760 --> 31:12.440 |
|
GANs are more focused on generating samples |
|
|
|
31:12.440 --> 31:15.640 |
|
rather than estimating the density function. |
|
|
|
31:15.640 --> 31:17.720 |
|
There are some kinds of GANs, like flow GAN, |
|
|
|
31:17.720 --> 31:18.560 |
|
that can do both, |
|
|
|
31:18.560 --> 31:21.680 |
|
but mostly GANs are about generating samples, |
|
|
|
31:21.680 --> 31:24.240 |
|
generating new photos of cats that look realistic. |
|
|
|
31:25.240 --> 31:29.360 |
|
And they do that completely from scratch. |
|
|
|
31:29.360 --> 31:32.240 |
|
It's analogous to human imagination |
|
|
|
31:32.240 --> 31:34.760 |
|
when a GAN creates a new image of a cat. |
|
|
|
31:34.760 --> 31:39.320 |
|
It's using a neural network to produce a cat |
|
|
|
31:39.320 --> 31:41.040 |
|
that has not existed before. |
|
|
|
31:41.040 --> 31:44.560 |
|
It isn't doing something like compositing photos together. |
|
|
|
31:44.560 --> 31:47.080 |
|
You're not literally taking the eye off of one cat |
|
|
|
31:47.080 --> 31:49.000 |
|
and the ear off of another cat. |
|
|
|
31:49.000 --> 31:51.320 |
|
It's more of this digestive process |
|
|
|
31:51.320 --> 31:53.920 |
|
where the neural net trains in a lot of data |
|
|
|
31:53.920 --> 31:55.560 |
|
and comes up with some representation |
|
|
|
31:55.560 --> 31:57.360 |
|
of the probability distribution |
|
|
|
31:57.360 --> 31:59.760 |
|
and generates entirely new cats. |
|
|
|
31:59.760 --> 32:00.880 |
|
There are a lot of different ways |
|
|
|
32:00.880 --> 32:01.960 |
|
of building a generative model. |
|
|
|
32:01.960 --> 32:05.640 |
|
What's specific to GANs is that we have a two player game |
|
|
|
32:05.640 --> 32:08.080 |
|
in the game theoretic sense. |
|
|
|
32:08.080 --> 32:10.280 |
|
And as the players in this game compete, |
|
|
|
32:10.280 --> 32:13.920 |
|
one of them becomes able to generate realistic data. |
|
|
|
32:13.920 --> 32:16.120 |
|
The first player is called the generator. |
|
|
|
32:16.120 --> 32:20.640 |
|
It produces output data, such as just images, for example. |
|
|
|
32:20.640 --> 32:22.400 |
|
And at the start of the learning process, |
|
|
|
32:22.400 --> 32:25.120 |
|
it'll just produce completely random images. |
|
|
|
32:25.120 --> 32:27.360 |
|
The other player is called the discriminator. |
|
|
|
32:27.360 --> 32:29.680 |
|
The discriminator takes images as input |
|
|
|
32:29.680 --> 32:31.560 |
|
and guesses whether they're real or fake. |
|
|
|
32:32.480 --> 32:34.200 |
|
You train it both on real data, |
|
|
|
32:34.200 --> 32:36.120 |
|
so photos that come from your training set, |
|
|
|
32:36.120 --> 32:37.840 |
|
actual photos of cats. |
|
|
|
32:37.840 --> 32:39.880 |
|
And you try to say that those are real. |
|
|
|
32:39.880 --> 32:41.920 |
|
You also train it on images |
|
|
|
32:41.920 --> 32:43.840 |
|
that come from the generator network. |
|
|
|
32:43.840 --> 32:46.720 |
|
And you train it to say that those are fake. |
|
|
|
32:46.720 --> 32:49.200 |
|
As the two players compete in this game, |
|
|
|
32:49.200 --> 32:50.920 |
|
the discriminator tries to become better |
|
|
|
32:50.920 --> 32:53.280 |
|
at recognizing whether images are real or fake. |
|
|
|
32:53.280 --> 32:54.760 |
|
And the generator becomes better |
|
|
|
32:54.760 --> 32:56.960 |
|
at fooling the discriminator into thinking |
|
|
|
32:56.960 --> 32:59.560 |
|
that its outputs are real. |
|
|
|
33:00.760 --> 33:03.560 |
|
And you can analyze this through the language of game theory |
|
|
|
33:03.560 --> 33:06.920 |
|
and find that there's a Nash equilibrium |
|
|
|
33:06.920 --> 33:08.600 |
|
where the generator has captured |
|
|
|
33:08.600 --> 33:10.800 |
|
the correct probability distribution. |
|
|
|
33:10.800 --> 33:12.160 |
|
So in the cat example, |
|
|
|
33:12.160 --> 33:14.560 |
|
it makes perfectly realistic cat photos. |
|
|
|
33:14.560 --> 33:17.160 |
|
And the discriminator is unable to do better |
|
|
|
33:17.160 --> 33:18.720 |
|
than random guessing, |
|
|
|
33:18.720 --> 33:21.800 |
|
because all the samples coming from both the data |
|
|
|
33:21.800 --> 33:24.000 |
|
and the generator look equally likely |
|
|
|
33:24.000 --> 33:25.840 |
|
to have come from either source. |
|
|
|
33:25.840 --> 33:28.320 |
|
So do you ever sit back |
|
|
|
33:28.320 --> 33:31.280 |
|
and does it just blow your mind that this thing works? |
|
|
|
33:31.280 --> 33:35.840 |
|
So from very, so it's able to estimate the density function |
|
|
|
33:35.840 --> 33:38.640 |
|
enough to generate realistic images. |
|
|
|
33:38.640 --> 33:43.640 |
|
I mean, yeah, do you ever sit back and think, |
|
|
|
33:43.640 --> 33:46.760 |
|
how does this even, this is quite incredible, |
|
|
|
33:46.760 --> 33:49.280 |
|
especially where against have gone in terms of realism. |
|
|
|
33:49.280 --> 33:51.600 |
|
Yeah, and not just to flatter my own work, |
|
|
|
33:51.600 --> 33:53.840 |
|
but generative models, |
|
|
|
33:53.840 --> 33:55.400 |
|
all of them have this property |
|
|
|
33:55.400 --> 33:58.800 |
|
that if they really did what we asked them to do, |
|
|
|
33:58.800 --> 34:01.040 |
|
they would do nothing but memorize the training data. |
|
|
|
34:01.040 --> 34:02.920 |
|
Right, exactly. |
|
|
|
34:02.920 --> 34:05.720 |
|
Models that are based on maximizing the likelihood, |
|
|
|
34:05.720 --> 34:08.200 |
|
the way that you obtain the maximum likelihood |
|
|
|
34:08.200 --> 34:09.720 |
|
for a specific training set |
|
|
|
34:09.720 --> 34:12.440 |
|
is you assign all of your probability mass |
|
|
|
34:12.440 --> 34:15.120 |
|
to the training examples and nowhere else. |
|
|
|
34:15.120 --> 34:18.440 |
|
For GANs, the game is played using a training set. |
|
|
|
34:18.440 --> 34:21.160 |
|
So the way that you become unbeatable in the game |
|
|
|
34:21.160 --> 34:23.440 |
|
is you literally memorize training examples. |
|
|
|
34:25.360 --> 34:28.880 |
|
One of my former interns wrote a paper, |
|
|
|
34:28.880 --> 34:31.040 |
|
his name is Vaishnav Nagarajan, |
|
|
|
34:31.040 --> 34:33.080 |
|
and he showed that it's actually hard |
|
|
|
34:33.080 --> 34:36.120 |
|
for the generator to memorize the training data, |
|
|
|
34:36.120 --> 34:39.160 |
|
hard in a statistical learning theory sense, |
|
|
|
34:39.160 --> 34:42.200 |
|
that you can actually create reasons |
|
|
|
34:42.200 --> 34:47.200 |
|
for why it would require quite a lot of learning steps |
|
|
|
34:48.400 --> 34:52.200 |
|
and a lot of observations of different latent variables |
|
|
|
34:52.200 --> 34:54.360 |
|
before you could memorize the training data. |
|
|
|
34:54.360 --> 34:55.680 |
|
That still doesn't really explain |
|
|
|
34:55.680 --> 34:58.280 |
|
why when you produce samples that are new, |
|
|
|
34:58.280 --> 34:59.880 |
|
why do you get compelling images |
|
|
|
34:59.880 --> 35:02.400 |
|
rather than just garbage that's different |
|
|
|
35:02.400 --> 35:03.800 |
|
from the training set. |
|
|
|
35:03.800 --> 35:06.960 |
|
And I don't think we really have a good answer for that, |
|
|
|
35:06.960 --> 35:07.920 |
|
especially if you think about |
|
|
|
35:07.920 --> 35:10.240 |
|
how many possible images are out there |
|
|
|
35:10.240 --> 35:15.240 |
|
and how few images the generative model sees during training. |
|
|
|
35:15.440 --> 35:16.920 |
|
It seems just unreasonable |
|
|
|
35:16.920 --> 35:19.200 |
|
that generative models create new images |
|
|
|
35:19.200 --> 35:22.080 |
|
as well as they do, especially considering |
|
|
|
35:22.080 --> 35:23.760 |
|
that we're basically training them to memorize |
|
|
|
35:23.760 --> 35:25.000 |
|
rather than generalize. |
|
|
|
35:26.240 --> 35:28.920 |
|
I think part of the answer is there's a paper |
|
|
|
35:28.920 --> 35:31.480 |
|
called Deep Image Prior where they show |
|
|
|
35:31.480 --> 35:33.080 |
|
that you can take a convolutional net |
|
|
|
35:33.080 --> 35:35.000 |
|
and you don't even need to learn the parameters of it at all. |
|
|
|
35:35.000 --> 35:37.640 |
|
You just use the model architecture. |
|
|
|
35:37.640 --> 35:41.080 |
|
And it's already useful for things like in painting images. |
|
|
|
35:41.080 --> 35:43.760 |
|
I think that shows us that the convolutional network |
|
|
|
35:43.760 --> 35:45.880 |
|
architecture captures something really important |
|
|
|
35:45.880 --> 35:47.960 |
|
about the structure of images. |
|
|
|
35:47.960 --> 35:50.960 |
|
And we don't need to actually use learning |
|
|
|
35:50.960 --> 35:52.200 |
|
to capture all the information |
|
|
|
35:52.200 --> 35:54.000 |
|
coming out of the convolutional net. |
|
|
|
35:55.240 --> 35:58.400 |
|
That would imply that it would be much harder |
|
|
|
35:58.400 --> 36:01.240 |
|
to make generative models in other domains. |
|
|
|
36:01.240 --> 36:03.600 |
|
So far, we're able to make reasonable speech models |
|
|
|
36:03.600 --> 36:04.880 |
|
and things like that. |
|
|
|
36:04.880 --> 36:07.440 |
|
But to be honest, we haven't actually explored |
|
|
|
36:07.440 --> 36:09.800 |
|
a whole lot of different data sets all that much. |
|
|
|
36:09.800 --> 36:13.920 |
|
We don't, for example, see a lot of deep learning models |
|
|
|
36:13.920 --> 36:18.440 |
|
of like biology data sets |
|
|
|
36:18.440 --> 36:19.880 |
|
where you have lots of microarrays |
|
|
|
36:19.880 --> 36:22.240 |
|
measuring the amount of different enzymes |
|
|
|
36:22.240 --> 36:23.080 |
|
and things like that. |
|
|
|
36:23.080 --> 36:25.240 |
|
So we may find that some of the progress |
|
|
|
36:25.240 --> 36:27.360 |
|
that we've seen for images and speech turns out |
|
|
|
36:27.360 --> 36:30.120 |
|
to really rely heavily on the model architecture. |
|
|
|
36:30.120 --> 36:32.960 |
|
And we were able to do what we did for vision |
|
|
|
36:32.960 --> 36:36.080 |
|
by trying to reverse engineer the human visual system. |
|
|
|
36:37.040 --> 36:39.800 |
|
And maybe it'll turn out that we can't just |
|
|
|
36:39.800 --> 36:42.560 |
|
use that same trick for arbitrary kinds of data. |
|
|
|
36:43.480 --> 36:45.920 |
|
Right, so there's aspect of the human vision system, |
|
|
|
36:45.920 --> 36:49.280 |
|
the hardware of it that makes it, |
|
|
|
36:49.280 --> 36:51.120 |
|
without learning, without cognition, |
|
|
|
36:51.120 --> 36:53.640 |
|
just makes it really effective at detecting the patterns |
|
|
|
36:53.640 --> 36:54.960 |
|
we see in the visual world. |
|
|
|
36:54.960 --> 36:57.280 |
|
Yeah, that's really interesting. |
|
|
|
36:57.280 --> 37:02.280 |
|
What, in a big quick overview in your view, |
|
|
|
37:04.640 --> 37:06.280 |
|
what types of GANs are there |
|
|
|
37:06.280 --> 37:10.080 |
|
and what other generative models besides GANs are there? |
|
|
|
37:10.080 --> 37:13.360 |
|
Yeah, so it's maybe a little bit easier to start |
|
|
|
37:13.360 --> 37:14.640 |
|
with what kinds of generative models |
|
|
|
37:14.640 --> 37:15.920 |
|
are there other than GANs. |
|
|
|
37:16.840 --> 37:20.840 |
|
So most generative models are likelihood based |
|
|
|
37:20.840 --> 37:23.920 |
|
where to train them, you have a model |
|
|
|
37:23.920 --> 37:27.320 |
|
that tells you how much probability it assigns |
|
|
|
37:27.320 --> 37:29.080 |
|
to a particular example, |
|
|
|
37:29.080 --> 37:31.480 |
|
and you just maximize the probability assigned |
|
|
|
37:31.480 --> 37:33.680 |
|
to all the training examples. |
|
|
|
37:33.680 --> 37:36.200 |
|
It turns out that it's hard to design a model |
|
|
|
37:36.200 --> 37:39.200 |
|
that can create really complicated images |
|
|
|
37:39.200 --> 37:42.280 |
|
or really complicated audio waveforms |
|
|
|
37:42.280 --> 37:46.200 |
|
and still have it be possible to estimate |
|
|
|
37:46.200 --> 37:51.200 |
|
the likelihood function from a computational point of view. |
|
|
|
37:51.200 --> 37:53.200 |
|
Most interesting models that you would just write |
|
|
|
37:53.200 --> 37:56.200 |
|
down intuitively, it turns out that it's almost impossible |
|
|
|
37:56.200 --> 37:58.200 |
|
to calculate the amount of probability |
|
|
|
37:58.200 --> 38:00.200 |
|
they assign to a particular point. |
|
|
|
38:00.200 --> 38:04.200 |
|
So there's a few different schools of generative models |
|
|
|
38:04.200 --> 38:06.200 |
|
in the likelihood family. |
|
|
|
38:06.200 --> 38:09.200 |
|
One approach is to very carefully design the model |
|
|
|
38:09.200 --> 38:12.200 |
|
so that it is computationally tractable |
|
|
|
38:12.200 --> 38:15.200 |
|
to measure the density it assigns to a particular point. |
|
|
|
38:15.200 --> 38:18.200 |
|
So there are things like auto regressive models, |
|
|
|
38:18.200 --> 38:23.200 |
|
like pixel CNN, those basically break down |
|
|
|
38:23.200 --> 38:26.200 |
|
the probability distribution into a product |
|
|
|
38:26.200 --> 38:28.200 |
|
over every single feature. |
|
|
|
38:28.200 --> 38:32.200 |
|
So for an image, you estimate the probability of each pixel |
|
|
|
38:32.200 --> 38:35.200 |
|
given all of the pixels that came before it. |
|
|
|
38:35.200 --> 38:37.200 |
|
There's tricks where if you want to measure |
|
|
|
38:37.200 --> 38:40.200 |
|
the density function, you can actually calculate |
|
|
|
38:40.200 --> 38:43.200 |
|
the density for all these pixels more or less in parallel. |
|
|
|
38:44.200 --> 38:46.200 |
|
Generating the image still tends to require you |
|
|
|
38:46.200 --> 38:50.200 |
|
to go one pixel at a time, and that can be very slow. |
|
|
|
38:50.200 --> 38:52.200 |
|
But there are, again, tricks for doing this |
|
|
|
38:52.200 --> 38:54.200 |
|
in a hierarchical pattern where you can keep |
|
|
|
38:54.200 --> 38:56.200 |
|
the runtime under control. |
|
|
|
38:56.200 --> 38:59.200 |
|
Are the quality of the images it generates |
|
|
|
38:59.200 --> 39:02.200 |
|
putting runtime aside pretty good? |
|
|
|
39:02.200 --> 39:04.200 |
|
They're reasonable, yeah. |
|
|
|
39:04.200 --> 39:07.200 |
|
I would say a lot of the best results |
|
|
|
39:07.200 --> 39:10.200 |
|
are from GANs these days, but it can be hard to tell |
|
|
|
39:10.200 --> 39:14.200 |
|
how much of that is based on who's studying |
|
|
|
39:14.200 --> 39:17.200 |
|
which type of algorithm, if that makes sense. |
|
|
|
39:17.200 --> 39:19.200 |
|
The amount of effort invested in it. |
|
|
|
39:19.200 --> 39:21.200 |
|
Yeah, or the kind of expertise. |
|
|
|
39:21.200 --> 39:23.200 |
|
So a lot of people who've traditionally been excited |
|
|
|
39:23.200 --> 39:25.200 |
|
about graphics or art and things like that |
|
|
|
39:25.200 --> 39:27.200 |
|
have gotten interested in GANs. |
|
|
|
39:27.200 --> 39:29.200 |
|
And to some extent, it's hard to tell, |
|
|
|
39:29.200 --> 39:32.200 |
|
are GANs doing better because they have a lot of |
|
|
|
39:32.200 --> 39:34.200 |
|
graphics and art experts behind them? |
|
|
|
39:34.200 --> 39:36.200 |
|
Or are GANs doing better because |
|
|
|
39:36.200 --> 39:38.200 |
|
they're more computationally efficient? |
|
|
|
39:38.200 --> 39:40.200 |
|
Or are GANs doing better because |
|
|
|
39:40.200 --> 39:43.200 |
|
they prioritize the realism of samples |
|
|
|
39:43.200 --> 39:45.200 |
|
over the accuracy of the density function? |
|
|
|
39:45.200 --> 39:47.200 |
|
I think all of those are potentially |
|
|
|
39:47.200 --> 39:51.200 |
|
valid explanations, and it's hard to tell. |
|
|
|
39:51.200 --> 39:53.200 |
|
So can you give a brief history of GANs |
|
|
|
39:53.200 --> 39:59.200 |
|
from 2014 with Paper 13? |
|
|
|
39:59.200 --> 40:01.200 |
|
Yeah, so a few highlights. |
|
|
|
40:01.200 --> 40:03.200 |
|
In the first paper, we just showed that |
|
|
|
40:03.200 --> 40:05.200 |
|
GANs basically work. |
|
|
|
40:05.200 --> 40:07.200 |
|
If you look back at the samples we had now, |
|
|
|
40:07.200 --> 40:09.200 |
|
they look terrible. |
|
|
|
40:09.200 --> 40:11.200 |
|
On the CFAR 10 data set, you can't even |
|
|
|
40:11.200 --> 40:13.200 |
|
see the effects in them. |
|
|
|
40:13.200 --> 40:15.200 |
|
Your paper, sorry, you used CFAR 10? |
|
|
|
40:15.200 --> 40:17.200 |
|
We used MNIST, which is Little Handwritten Digits. |
|
|
|
40:17.200 --> 40:19.200 |
|
We used the Toronto Face Database, |
|
|
|
40:19.200 --> 40:22.200 |
|
which is small grayscale photos of faces. |
|
|
|
40:22.200 --> 40:24.200 |
|
We did have recognizable faces. |
|
|
|
40:24.200 --> 40:26.200 |
|
My colleague Bing Xu put together |
|
|
|
40:26.200 --> 40:29.200 |
|
the first GAN face model for that paper. |
|
|
|
40:29.200 --> 40:32.200 |
|
We also had the CFAR 10 data set, |
|
|
|
40:32.200 --> 40:35.200 |
|
which is things like very small 32x32 pixels |
|
|
|
40:35.200 --> 40:40.200 |
|
of cars and cats and dogs. |
|
|
|
40:40.200 --> 40:43.200 |
|
For that, we didn't get recognizable objects, |
|
|
|
40:43.200 --> 40:46.200 |
|
but all the deep learning people back then |
|
|
|
40:46.200 --> 40:48.200 |
|
were really used to looking at these failed samples |
|
|
|
40:48.200 --> 40:50.200 |
|
and kind of reading them like tea leaves. |
|
|
|
40:50.200 --> 40:53.200 |
|
And people who are used to reading the tea leaves |
|
|
|
40:53.200 --> 40:56.200 |
|
recognize that our tea leaves at least look different. |
|
|
|
40:56.200 --> 40:58.200 |
|
Maybe not necessarily better, |
|
|
|
40:58.200 --> 41:01.200 |
|
but there was something unusual about them. |
|
|
|
41:01.200 --> 41:03.200 |
|
And that got a lot of us excited. |
|
|
|
41:03.200 --> 41:06.200 |
|
One of the next really big steps was LAPGAN |
|
|
|
41:06.200 --> 41:10.200 |
|
by Emily Denton and Sumith Chintala at Facebook AI Research, |
|
|
|
41:10.200 --> 41:14.200 |
|
where they actually got really good high resolution photos |
|
|
|
41:14.200 --> 41:16.200 |
|
working with GANs for the first time. |
|
|
|
41:16.200 --> 41:18.200 |
|
They had a complicated system |
|
|
|
41:18.200 --> 41:20.200 |
|
where they generated the image starting at low res |
|
|
|
41:20.200 --> 41:22.200 |
|
and then scaling up to high res, |
|
|
|
41:22.200 --> 41:24.200 |
|
but they were able to get it to work. |
|
|
|
41:24.200 --> 41:30.200 |
|
And then in 2015, I believe later that same year, |
|
|
|
41:30.200 --> 41:35.200 |
|
Alec Radford and Sumith Chintala and Luke Metz |
|
|
|
41:35.200 --> 41:38.200 |
|
published the DC GAN paper, |
|
|
|
41:38.200 --> 41:41.200 |
|
which it stands for Deep Convolutional GAN. |
|
|
|
41:41.200 --> 41:43.200 |
|
It's kind of a nonunique name |
|
|
|
41:43.200 --> 41:46.200 |
|
because these days basically all GANs |
|
|
|
41:46.200 --> 41:48.200 |
|
and even some before that were deep and convolutional, |
|
|
|
41:48.200 --> 41:52.200 |
|
but they just kind of picked a name for a really great recipe |
|
|
|
41:52.200 --> 41:55.200 |
|
where they were able to actually using only one model |
|
|
|
41:55.200 --> 41:57.200 |
|
instead of a multi step process, |
|
|
|
41:57.200 --> 42:01.200 |
|
actually generate realistic images of faces and things like that. |
|
|
|
42:01.200 --> 42:05.200 |
|
That was sort of like the beginning |
|
|
|
42:05.200 --> 42:07.200 |
|
of the Cambrian explosion of GANs. |
|
|
|
42:07.200 --> 42:09.200 |
|
Once you had animals that had a backbone, |
|
|
|
42:09.200 --> 42:12.200 |
|
you suddenly got lots of different versions of fish |
|
|
|
42:12.200 --> 42:15.200 |
|
and four legged animals and things like that. |
|
|
|
42:15.200 --> 42:17.200 |
|
So DC GAN became kind of the backbone |
|
|
|
42:17.200 --> 42:19.200 |
|
for many different models that came out. |
|
|
|
42:19.200 --> 42:21.200 |
|
Used as a baseline even still. |
|
|
|
42:21.200 --> 42:23.200 |
|
Yeah, yeah. |
|
|
|
42:23.200 --> 42:26.200 |
|
And so from there, I would say some interesting things we've seen |
|
|
|
42:26.200 --> 42:30.200 |
|
are there's a lot you can say about how just |
|
|
|
42:30.200 --> 42:33.200 |
|
the quality of standard image generation GANs has increased, |
|
|
|
42:33.200 --> 42:36.200 |
|
but what's also maybe more interesting on an intellectual level |
|
|
|
42:36.200 --> 42:40.200 |
|
is how the things you can use GANs for has also changed. |
|
|
|
42:40.200 --> 42:44.200 |
|
One thing is that you can use them to learn classifiers |
|
|
|
42:44.200 --> 42:47.200 |
|
without having to have class labels for every example |
|
|
|
42:47.200 --> 42:49.200 |
|
in your training set. |
|
|
|
42:49.200 --> 42:51.200 |
|
So that's called semi supervised learning. |
|
|
|
42:51.200 --> 42:55.200 |
|
My colleague at OpenAI, Tim Solomon, who's at Brain now, |
|
|
|
42:55.200 --> 42:57.200 |
|
wrote a paper called |
|
|
|
42:57.200 --> 42:59.200 |
|
Improved Techniques for Training GANs. |
|
|
|
42:59.200 --> 43:01.200 |
|
I'm a coauthor on this paper, |
|
|
|
43:01.200 --> 43:03.200 |
|
but I can't claim any credit for this particular part. |
|
|
|
43:03.200 --> 43:05.200 |
|
One thing he showed on the paper is that |
|
|
|
43:05.200 --> 43:09.200 |
|
you can take the GAN discriminator and use it as a classifier |
|
|
|
43:09.200 --> 43:12.200 |
|
that actually tells you this image is a cat, |
|
|
|
43:12.200 --> 43:14.200 |
|
this image is a dog, this image is a car, |
|
|
|
43:14.200 --> 43:16.200 |
|
this image is a truck. |
|
|
|
43:16.200 --> 43:18.200 |
|
And so not just to say whether the image is real or fake, |
|
|
|
43:18.200 --> 43:22.200 |
|
but if it is real to say specifically what kind of object it is. |
|
|
|
43:22.200 --> 43:25.200 |
|
And he found that you can train these classifiers |
|
|
|
43:25.200 --> 43:28.200 |
|
with far fewer labeled examples |
|
|
|
43:28.200 --> 43:30.200 |
|
than traditional classifiers. |
|
|
|
43:30.200 --> 43:33.200 |
|
So if you supervise based on also |
|
|
|
43:33.200 --> 43:35.200 |
|
not just your discrimination ability, |
|
|
|
43:35.200 --> 43:37.200 |
|
but your ability to classify, |
|
|
|
43:37.200 --> 43:40.200 |
|
you're going to converge much faster |
|
|
|
43:40.200 --> 43:43.200 |
|
to being effective at being a discriminator. |
|
|
|
43:43.200 --> 43:44.200 |
|
Yeah. |
|
|
|
43:44.200 --> 43:46.200 |
|
So for example, for the MNIST dataset, |
|
|
|
43:46.200 --> 43:49.200 |
|
you want to look at an image of a handwritten digit |
|
|
|
43:49.200 --> 43:53.200 |
|
and say whether it's a zero, a one, or two, and so on. |
|
|
|
43:53.200 --> 43:57.200 |
|
To get down to less than 1% accuracy, |
|
|
|
43:57.200 --> 44:00.200 |
|
we required around 60,000 examples |
|
|
|
44:00.200 --> 44:03.200 |
|
until maybe about 2014 or so. |
|
|
|
44:03.200 --> 44:07.200 |
|
In 2016, with this semi supervised GAN project, |
|
|
|
44:07.200 --> 44:10.200 |
|
Tim was able to get below 1% error |
|
|
|
44:10.200 --> 44:13.200 |
|
using only 100 labeled examples. |
|
|
|
44:13.200 --> 44:16.200 |
|
So that was about a 600x decrease |
|
|
|
44:16.200 --> 44:18.200 |
|
in the amount of labels that he needed. |
|
|
|
44:18.200 --> 44:21.200 |
|
He's still using more images than that, |
|
|
|
44:21.200 --> 44:23.200 |
|
but he doesn't need to have each of them labeled as, |
|
|
|
44:23.200 --> 44:25.200 |
|
you know, this one's a one, this one's a two, |
|
|
|
44:25.200 --> 44:27.200 |
|
this one's a zero, and so on. |
|
|
|
44:27.200 --> 44:29.200 |
|
Then to be able to, for GANs, |
|
|
|
44:29.200 --> 44:31.200 |
|
to be able to generate recognizable objects, |
|
|
|
44:31.200 --> 44:33.200 |
|
so objects from a particular class, |
|
|
|
44:33.200 --> 44:36.200 |
|
you still need labeled data, |
|
|
|
44:36.200 --> 44:38.200 |
|
because you need to know |
|
|
|
44:38.200 --> 44:41.200 |
|
what it means to be a particular class cat dog. |
|
|
|
44:41.200 --> 44:44.200 |
|
How do you think we can move away from that? |
|
|
|
44:44.200 --> 44:46.200 |
|
Yeah, some researchers at Brain Zurich |
|
|
|
44:46.200 --> 44:49.200 |
|
actually just released a really great paper |
|
|
|
44:49.200 --> 44:51.200 |
|
on semi supervised GANs, |
|
|
|
44:51.200 --> 44:54.200 |
|
where their goal isn't to classify, |
|
|
|
44:54.200 --> 44:56.200 |
|
to make recognizable objects |
|
|
|
44:56.200 --> 44:58.200 |
|
despite not having a lot of labeled data. |
|
|
|
44:58.200 --> 45:02.200 |
|
They were working off of DeepMind's BigGAN project, |
|
|
|
45:02.200 --> 45:04.200 |
|
and they showed that they can match |
|
|
|
45:04.200 --> 45:06.200 |
|
the performance of BigGAN |
|
|
|
45:06.200 --> 45:10.200 |
|
using only 10%, I believe, of the labels. |
|
|
|
45:10.200 --> 45:12.200 |
|
BigGAN was trained on the ImageNet data set, |
|
|
|
45:12.200 --> 45:14.200 |
|
which is about 1.2 million images, |
|
|
|
45:14.200 --> 45:17.200 |
|
and had all of them labeled. |
|
|
|
45:17.200 --> 45:19.200 |
|
This latest project from Brain Zurich |
|
|
|
45:19.200 --> 45:21.200 |
|
shows that they're able to get away with |
|
|
|
45:21.200 --> 45:25.200 |
|
having about 10% of the images labeled. |
|
|
|
45:25.200 --> 45:29.200 |
|
They do that essentially using a clustering algorithm, |
|
|
|
45:29.200 --> 45:32.200 |
|
where the discriminator learns to assign |
|
|
|
45:32.200 --> 45:34.200 |
|
the objects to groups, |
|
|
|
45:34.200 --> 45:38.200 |
|
and then this understanding that objects can be grouped |
|
|
|
45:38.200 --> 45:40.200 |
|
into similar types, |
|
|
|
45:40.200 --> 45:43.200 |
|
helps it to form more realistic ideas |
|
|
|
45:43.200 --> 45:45.200 |
|
of what should be appearing in the image, |
|
|
|
45:45.200 --> 45:47.200 |
|
because it knows that every image it creates |
|
|
|
45:47.200 --> 45:50.200 |
|
has to come from one of these archetypal groups, |
|
|
|
45:50.200 --> 45:53.200 |
|
rather than just being some arbitrary image. |
|
|
|
45:53.200 --> 45:55.200 |
|
If you train again with no class labels, |
|
|
|
45:55.200 --> 45:57.200 |
|
you tend to get things that look sort of like |
|
|
|
45:57.200 --> 46:00.200 |
|
grass or water or brick or dirt, |
|
|
|
46:00.200 --> 46:04.200 |
|
but without necessarily a lot going on in them. |
|
|
|
46:04.200 --> 46:06.200 |
|
I think that's partly because if you look |
|
|
|
46:06.200 --> 46:08.200 |
|
at a large ImageNet image, |
|
|
|
46:08.200 --> 46:11.200 |
|
the object doesn't necessarily occupy the whole image, |
|
|
|
46:11.200 --> 46:15.200 |
|
and so you learn to create realistic sets of pixels, |
|
|
|
46:15.200 --> 46:17.200 |
|
but you don't necessarily learn |
|
|
|
46:17.200 --> 46:19.200 |
|
that the object is the star of the show, |
|
|
|
46:19.200 --> 46:22.200 |
|
and you want it to be in every image you make. |
|
|
|
46:22.200 --> 46:25.200 |
|
Yeah, I've heard you talk about the horse, |
|
|
|
46:25.200 --> 46:27.200 |
|
the zebra cycle, gang mapping, |
|
|
|
46:27.200 --> 46:30.200 |
|
and how it turns out, again, |
|
|
|
46:30.200 --> 46:33.200 |
|
thought provoking that horses are usually on grass, |
|
|
|
46:33.200 --> 46:35.200 |
|
and zebras are usually on drier terrain, |
|
|
|
46:35.200 --> 46:38.200 |
|
so when you're doing that kind of generation, |
|
|
|
46:38.200 --> 46:43.200 |
|
you're going to end up generating greener horses or whatever. |
|
|
|
46:43.200 --> 46:45.200 |
|
So those are connected together. |
|
|
|
46:45.200 --> 46:46.200 |
|
It's not just... |
|
|
|
46:46.200 --> 46:47.200 |
|
Yeah, yeah. |
|
|
|
46:47.200 --> 46:49.200 |
|
You're not able to segment, |
|
|
|
46:49.200 --> 46:52.200 |
|
to be able to generate in a segmental way. |
|
|
|
46:52.200 --> 46:55.200 |
|
So are there other types of games you come across |
|
|
|
46:55.200 --> 47:00.200 |
|
in your mind that neural networks can play with each other |
|
|
|
47:00.200 --> 47:05.200 |
|
to be able to solve problems? |
|
|
|
47:05.200 --> 47:09.200 |
|
Yeah, the one that I spend most of my time on is in security. |
|
|
|
47:09.200 --> 47:13.200 |
|
You can model most interactions as a game |
|
|
|
47:13.200 --> 47:16.200 |
|
where there's attackers trying to break your system |
|
|
|
47:16.200 --> 47:19.200 |
|
or the defender trying to build a resilient system. |
|
|
|
47:19.200 --> 47:22.200 |
|
There's also domain adversarial learning, |
|
|
|
47:22.200 --> 47:25.200 |
|
which is an approach to domain adaptation |
|
|
|
47:25.200 --> 47:27.200 |
|
that looks really a lot like GANs. |
|
|
|
47:27.200 --> 47:31.200 |
|
The authors had the idea before the GAN paper came out. |
|
|
|
47:31.200 --> 47:33.200 |
|
Their paper came out a little bit later, |
|
|
|
47:33.200 --> 47:38.200 |
|
and they were very nice and cited the GAN paper, |
|
|
|
47:38.200 --> 47:41.200 |
|
but I know that they actually had the idea before it came out. |
|
|
|
47:41.200 --> 47:45.200 |
|
Domain adaptation is when you want to train a machine learning model |
|
|
|
47:45.200 --> 47:47.200 |
|
in one setting called a domain, |
|
|
|
47:47.200 --> 47:50.200 |
|
and then deploy it in another domain later, |
|
|
|
47:50.200 --> 47:52.200 |
|
and you would like it to perform well in the new domain, |
|
|
|
47:52.200 --> 47:55.200 |
|
even though the new domain is different from how it was trained. |
|
|
|
47:55.200 --> 47:58.200 |
|
So, for example, you might want to train |
|
|
|
47:58.200 --> 48:01.200 |
|
on a really clean image dataset like ImageNet, |
|
|
|
48:01.200 --> 48:03.200 |
|
but then deploy on users phones, |
|
|
|
48:03.200 --> 48:06.200 |
|
where the user is taking pictures in the dark |
|
|
|
48:06.200 --> 48:08.200 |
|
and pictures while moving quickly |
|
|
|
48:08.200 --> 48:10.200 |
|
and just pictures that aren't really centered |
|
|
|
48:10.200 --> 48:13.200 |
|
or composed all that well. |
|
|
|
48:13.200 --> 48:16.200 |
|
When you take a normal machine learning model, |
|
|
|
48:16.200 --> 48:19.200 |
|
it often degrades really badly when you move to the new domain |
|
|
|
48:19.200 --> 48:22.200 |
|
because it looks so different from what the model was trained on. |
|
|
|
48:22.200 --> 48:25.200 |
|
Domain adaptation algorithms try to smooth out that gap, |
|
|
|
48:25.200 --> 48:28.200 |
|
and the domain adversarial approach is based on |
|
|
|
48:28.200 --> 48:30.200 |
|
training a feature extractor, |
|
|
|
48:30.200 --> 48:32.200 |
|
where the features have the same statistics |
|
|
|
48:32.200 --> 48:35.200 |
|
regardless of which domain you extracted them on. |
|
|
|
48:35.200 --> 48:37.200 |
|
So, in the domain adversarial game, |
|
|
|
48:37.200 --> 48:39.200 |
|
you have one player that's a feature extractor |
|
|
|
48:39.200 --> 48:42.200 |
|
and another player that's a domain recognizer. |
|
|
|
48:42.200 --> 48:44.200 |
|
The domain recognizer wants to look at the output |
|
|
|
48:44.200 --> 48:47.200 |
|
of the feature extractor and guess which of the two domains |
|
|
|
48:47.200 --> 48:49.200 |
|
the features came from. |
|
|
|
48:49.200 --> 48:52.200 |
|
So, it's a lot like the real versus fake discriminator in GANs. |
|
|
|
48:52.200 --> 48:54.200 |
|
And then the feature extractor, |
|
|
|
48:54.200 --> 48:57.200 |
|
you can think of as loosely analogous to the generator in GANs, |
|
|
|
48:57.200 --> 48:59.200 |
|
except what it's trying to do here |
|
|
|
48:59.200 --> 49:02.200 |
|
is both fool the domain recognizer |
|
|
|
49:02.200 --> 49:05.200 |
|
into not knowing which domain the data came from |
|
|
|
49:05.200 --> 49:08.200 |
|
and also extract features that are good for classification. |
|
|
|
49:08.200 --> 49:13.200 |
|
So, at the end of the day, in the cases where it works out, |
|
|
|
49:13.200 --> 49:18.200 |
|
you can actually get features that work about the same |
|
|
|
49:18.200 --> 49:20.200 |
|
in both domains. |
|
|
|
49:20.200 --> 49:22.200 |
|
Sometimes this has a drawback where, |
|
|
|
49:22.200 --> 49:24.200 |
|
in order to make things work the same in both domains, |
|
|
|
49:24.200 --> 49:26.200 |
|
it just gets worse at the first one. |
|
|
|
49:26.200 --> 49:28.200 |
|
But there are a lot of cases where it actually |
|
|
|
49:28.200 --> 49:30.200 |
|
works out well on both. |
|
|
|
49:30.200 --> 49:33.200 |
|
So, do you think of GANs being useful in the context |
|
|
|
49:33.200 --> 49:35.200 |
|
of data augmentation? |
|
|
|
49:35.200 --> 49:37.200 |
|
Yeah, one thing you could hope for with GANs |
|
|
|
49:37.200 --> 49:39.200 |
|
is you could imagine, |
|
|
|
49:39.200 --> 49:41.200 |
|
I've got a limited training set |
|
|
|
49:41.200 --> 49:43.200 |
|
and I'd like to make more training data |
|
|
|
49:43.200 --> 49:46.200 |
|
to train something else like a classifier. |
|
|
|
49:46.200 --> 49:50.200 |
|
You could train the GAN on the training set |
|
|
|
49:50.200 --> 49:52.200 |
|
and then create more data |
|
|
|
49:52.200 --> 49:55.200 |
|
and then maybe the classifier would perform better |
|
|
|
49:55.200 --> 49:58.200 |
|
on the test set after training on this bigger GAN generated data set. |
|
|
|
49:58.200 --> 50:00.200 |
|
So, that's the simplest version |
|
|
|
50:00.200 --> 50:02.200 |
|
of something you might hope would work. |
|
|
|
50:02.200 --> 50:05.200 |
|
I've never heard of that particular approach working, |
|
|
|
50:05.200 --> 50:08.200 |
|
but I think there's some closely related things |
|
|
|
50:08.200 --> 50:11.200 |
|
that I think could work in the future |
|
|
|
50:11.200 --> 50:13.200 |
|
and some that actually already have worked. |
|
|
|
50:13.200 --> 50:15.200 |
|
So, if we think a little bit about what we'd be hoping for |
|
|
|
50:15.200 --> 50:17.200 |
|
if we use the GAN to make more training data, |
|
|
|
50:17.200 --> 50:20.200 |
|
we're hoping that the GAN will generalize |
|
|
|
50:20.200 --> 50:23.200 |
|
to new examples better than the classifier would have |
|
|
|
50:23.200 --> 50:25.200 |
|
generalized if it was trained on the same data. |
|
|
|
50:25.200 --> 50:27.200 |
|
And I don't know of any reason to believe |
|
|
|
50:27.200 --> 50:30.200 |
|
that the GAN would generalize better than the classifier would. |
|
|
|
50:30.200 --> 50:33.200 |
|
But what we might hope for is that the GAN |
|
|
|
50:33.200 --> 50:37.200 |
|
could generalize differently from a specific classifier. |
|
|
|
50:37.200 --> 50:39.200 |
|
So, one thing I think is worth trying |
|
|
|
50:39.200 --> 50:41.200 |
|
that I haven't personally tried, but someone could try is |
|
|
|
50:41.200 --> 50:44.200 |
|
what if you trained a whole lot of different generative models |
|
|
|
50:44.200 --> 50:46.200 |
|
on the same training set, |
|
|
|
50:46.200 --> 50:48.200 |
|
create samples from all of them |
|
|
|
50:48.200 --> 50:50.200 |
|
and then train a classifier on that. |
|
|
|
50:50.200 --> 50:52.200 |
|
Because each of the generative models |
|
|
|
50:52.200 --> 50:54.200 |
|
might generalize in a slightly different way, |
|
|
|
50:54.200 --> 50:56.200 |
|
they might capture many different axes of variation |
|
|
|
50:56.200 --> 50:58.200 |
|
that one individual model wouldn't. |
|
|
|
50:58.200 --> 51:01.200 |
|
And then the classifier can capture all of those ideas |
|
|
|
51:01.200 --> 51:03.200 |
|
by training in all of their data. |
|
|
|
51:03.200 --> 51:06.200 |
|
So, it'd be a little bit like making an ensemble of classifiers. |
|
|
|
51:06.200 --> 51:08.200 |
|
An ensemble of GANs in a way. |
|
|
|
51:08.200 --> 51:10.200 |
|
I think that could generalize better. |
|
|
|
51:10.200 --> 51:12.200 |
|
The other thing that GANs are really good for |
|
|
|
51:12.200 --> 51:16.200 |
|
is not necessarily generating new data |
|
|
|
51:16.200 --> 51:19.200 |
|
that's exactly like what you already have, |
|
|
|
51:19.200 --> 51:23.200 |
|
but by generating new data that has different properties |
|
|
|
51:23.200 --> 51:25.200 |
|
from the data you already had. |
|
|
|
51:25.200 --> 51:27.200 |
|
One thing that you can do is you can create |
|
|
|
51:27.200 --> 51:29.200 |
|
differentially private data. |
|
|
|
51:29.200 --> 51:31.200 |
|
So, suppose that you have something like medical records |
|
|
|
51:31.200 --> 51:34.200 |
|
and you don't want to train a classifier on the medical records |
|
|
|
51:34.200 --> 51:36.200 |
|
and then publish the classifier |
|
|
|
51:36.200 --> 51:38.200 |
|
because someone might be able to reverse engineer |
|
|
|
51:38.200 --> 51:40.200 |
|
some of the medical records you trained on. |
|
|
|
51:40.200 --> 51:42.200 |
|
There's a paper from Casey Green's lab |
|
|
|
51:42.200 --> 51:46.200 |
|
that shows how you can train again using differential privacy. |
|
|
|
51:46.200 --> 51:48.200 |
|
And then the samples from the GAN |
|
|
|
51:48.200 --> 51:51.200 |
|
still have the same differential privacy guarantees |
|
|
|
51:51.200 --> 51:53.200 |
|
as the parameters of the GAN. |
|
|
|
51:53.200 --> 51:55.200 |
|
So, you can make fake patient data |
|
|
|
51:55.200 --> 51:57.200 |
|
for other researchers to use |
|
|
|
51:57.200 --> 51:59.200 |
|
and they can do almost anything they want with that data |
|
|
|
51:59.200 --> 52:02.200 |
|
because it doesn't come from real people. |
|
|
|
52:02.200 --> 52:04.200 |
|
And the differential privacy mechanism |
|
|
|
52:04.200 --> 52:07.200 |
|
gives you clear guarantees on how much |
|
|
|
52:07.200 --> 52:09.200 |
|
the original people's data has been protected. |
|
|
|
52:09.200 --> 52:11.200 |
|
That's really interesting, actually. |
|
|
|
52:11.200 --> 52:13.200 |
|
I haven't heard you talk about that before. |
|
|
|
52:13.200 --> 52:15.200 |
|
In terms of fairness, |
|
|
|
52:15.200 --> 52:19.200 |
|
I've seen from AAAI your talk, |
|
|
|
52:19.200 --> 52:21.200 |
|
how can adversarial machine learning |
|
|
|
52:21.200 --> 52:23.200 |
|
help models be more fair |
|
|
|
52:23.200 --> 52:25.200 |
|
with respect to sensitive variables? |
|
|
|
52:25.200 --> 52:28.200 |
|
Yeah. So, there's a paper from Emma Storky's lab |
|
|
|
52:28.200 --> 52:31.200 |
|
about how to learn machine learning models |
|
|
|
52:31.200 --> 52:34.200 |
|
that are incapable of using specific variables. |
|
|
|
52:34.200 --> 52:36.200 |
|
So, say, for example, you wanted to make predictions |
|
|
|
52:36.200 --> 52:39.200 |
|
that are not affected by gender. |
|
|
|
52:39.200 --> 52:41.200 |
|
It isn't enough to just leave gender |
|
|
|
52:41.200 --> 52:43.200 |
|
out of the input to the model. |
|
|
|
52:43.200 --> 52:45.200 |
|
You can often infer gender from a lot of other characteristics. |
|
|
|
52:45.200 --> 52:47.200 |
|
Like, say that you have the person's name, |
|
|
|
52:47.200 --> 52:49.200 |
|
but you're not told their gender. |
|
|
|
52:49.200 --> 52:53.200 |
|
Well, if their name is Ian, they're kind of obviously a man. |
|
|
|
52:53.200 --> 52:55.200 |
|
So, what you'd like to do is make a machine learning model |
|
|
|
52:55.200 --> 52:58.200 |
|
that can still take in a lot of different attributes |
|
|
|
52:58.200 --> 53:02.200 |
|
and make a really accurate informed prediction, |
|
|
|
53:02.200 --> 53:05.200 |
|
but be confident that it isn't reverse engineering gender |
|
|
|
53:05.200 --> 53:08.200 |
|
or another sensitive variable internally. |
|
|
|
53:08.200 --> 53:10.200 |
|
You can do that using something very similar |
|
|
|
53:10.200 --> 53:12.200 |
|
to the domain adversarial approach, |
|
|
|
53:12.200 --> 53:15.200 |
|
where you have one player that's a feature extractor |
|
|
|
53:15.200 --> 53:18.200 |
|
and another player that's a feature analyzer. |
|
|
|
53:18.200 --> 53:21.200 |
|
And you want to make sure that the feature analyzer |
|
|
|
53:21.200 --> 53:24.200 |
|
is not able to guess the value of the sensitive variable |
|
|
|
53:24.200 --> 53:26.200 |
|
that you're trying to keep private. |
|
|
|
53:26.200 --> 53:29.200 |
|
Right. Yeah, I love this approach. |
|
|
|
53:29.200 --> 53:34.200 |
|
So, with the feature, you're not able to infer |
|
|
|
53:34.200 --> 53:36.200 |
|
the sensitive variables. |
|
|
|
53:36.200 --> 53:39.200 |
|
It's brilliant. It's quite brilliant and simple, actually. |
|
|
|
53:39.200 --> 53:42.200 |
|
Another way I think that GANs in particular |
|
|
|
53:42.200 --> 53:44.200 |
|
could be used for fairness would be |
|
|
|
53:44.200 --> 53:46.200 |
|
to make something like a cycle GAN, |
|
|
|
53:46.200 --> 53:49.200 |
|
where you can take data from one domain |
|
|
|
53:49.200 --> 53:51.200 |
|
and convert it into another. |
|
|
|
53:51.200 --> 53:54.200 |
|
We've seen cycle GAN turning horses into zebras. |
|
|
|
53:54.200 --> 53:59.200 |
|
We've seen other unsupervised GANs made by Mingyu Liu |
|
|
|
53:59.200 --> 54:02.200 |
|
doing things like turning day photos into night photos. |
|
|
|
54:02.200 --> 54:05.200 |
|
I think for fairness, you could imagine |
|
|
|
54:05.200 --> 54:08.200 |
|
taking records for people in one group |
|
|
|
54:08.200 --> 54:11.200 |
|
and transforming them into analogous people in another group |
|
|
|
54:11.200 --> 54:14.200 |
|
and testing to see if they're treated equitably |
|
|
|
54:14.200 --> 54:16.200 |
|
across those two groups. |
|
|
|
54:16.200 --> 54:18.200 |
|
There's a lot of things that would be hard to get right |
|
|
|
54:18.200 --> 54:21.200 |
|
and make sure that the conversion process itself is fair. |
|
|
|
54:21.200 --> 54:24.200 |
|
And I don't think it's anywhere near something |
|
|
|
54:24.200 --> 54:26.200 |
|
that we could actually use yet. |
|
|
|
54:26.200 --> 54:28.200 |
|
But if you could design that conversion process very carefully, |
|
|
|
54:28.200 --> 54:30.200 |
|
it might give you a way of doing audits |
|
|
|
54:30.200 --> 54:33.200 |
|
where you say, what if we took people from this group, |
|
|
|
54:33.200 --> 54:35.200 |
|
converted them into equivalent people in another group? |
|
|
|
54:35.200 --> 54:39.200 |
|
Does the system actually treat them how it ought to? |
|
|
|
54:39.200 --> 54:41.200 |
|
That's also really interesting. |
|
|
|
54:41.200 --> 54:46.200 |
|
You know, in popular press |
|
|
|
54:46.200 --> 54:48.200 |
|
and in general, in our imagination, |
|
|
|
54:48.200 --> 54:51.200 |
|
you think, well, GANs are able to generate data |
|
|
|
54:51.200 --> 54:54.200 |
|
and you start to think about deep fakes |
|
|
|
54:54.200 --> 54:57.200 |
|
or being able to sort of maliciously generate data |
|
|
|
54:57.200 --> 55:00.200 |
|
that fakes the identity of other people. |
|
|
|
55:00.200 --> 55:03.200 |
|
Is this something of a concern to you? |
|
|
|
55:03.200 --> 55:06.200 |
|
Is this something, if you look 10, 20 years into the future, |
|
|
|
55:06.200 --> 55:10.200 |
|
is that something that pops up in your work, |
|
|
|
55:10.200 --> 55:13.200 |
|
in the work of the community that's working on generative models? |
|
|
|
55:13.200 --> 55:15.200 |
|
I'm a lot less concerned about 20 years from now |
|
|
|
55:15.200 --> 55:17.200 |
|
than the next few years. |
|
|
|
55:17.200 --> 55:20.200 |
|
I think there will be a kind of bumpy cultural transition |
|
|
|
55:20.200 --> 55:22.200 |
|
as people encounter this idea |
|
|
|
55:22.200 --> 55:25.200 |
|
that there can be very realistic videos and audio that aren't real. |
|
|
|
55:25.200 --> 55:27.200 |
|
I think 20 years from now, |
|
|
|
55:27.200 --> 55:30.200 |
|
people will mostly understand that you shouldn't believe |
|
|
|
55:30.200 --> 55:33.200 |
|
something is real just because you saw a video of it. |
|
|
|
55:33.200 --> 55:37.200 |
|
People will expect to see that it's been cryptographically signed |
|
|
|
55:37.200 --> 55:41.200 |
|
or have some other mechanism to make them believe |
|
|
|
55:41.200 --> 55:43.200 |
|
that the content is real. |
|
|
|
55:43.200 --> 55:45.200 |
|
There's already people working on this, |
|
|
|
55:45.200 --> 55:47.200 |
|
like there's a startup called TruePick |
|
|
|
55:47.200 --> 55:50.200 |
|
that provides a lot of mechanisms for authenticating |
|
|
|
55:50.200 --> 55:52.200 |
|
that an image is real. |
|
|
|
55:52.200 --> 55:55.200 |
|
They're maybe not quite up to having a state actor |
|
|
|
55:55.200 --> 55:59.200 |
|
try to evade their verification techniques, |
|
|
|
55:59.200 --> 56:02.200 |
|
but it's something that people are already working on |
|
|
|
56:02.200 --> 56:04.200 |
|
and I think will get right eventually. |
|
|
|
56:04.200 --> 56:08.200 |
|
So you think authentication will eventually win out? |
|
|
|
56:08.200 --> 56:11.200 |
|
So being able to authenticate that this is real and this is not? |
|
|
|
56:11.200 --> 56:13.200 |
|
Yeah. |
|
|
|
56:13.200 --> 56:15.200 |
|
As opposed to GANs just getting better and better |
|
|
|
56:15.200 --> 56:18.200 |
|
or generative models being able to get better and better |
|
|
|
56:18.200 --> 56:21.200 |
|
to where the nature of what is real is normal. |
|
|
|
56:21.200 --> 56:25.200 |
|
I don't think we'll ever be able to look at the pixels of a photo |
|
|
|
56:25.200 --> 56:28.200 |
|
and tell you for sure that it's real or not real, |
|
|
|
56:28.200 --> 56:32.200 |
|
and I think it would actually be somewhat dangerous |
|
|
|
56:32.200 --> 56:34.200 |
|
to rely on that approach too much. |
|
|
|
56:34.200 --> 56:36.200 |
|
If you make a really good fake detector |
|
|
|
56:36.200 --> 56:38.200 |
|
and then someone's able to fool your fake detector |
|
|
|
56:38.200 --> 56:41.200 |
|
and your fake detector says this image is not fake, |
|
|
|
56:41.200 --> 56:43.200 |
|
then it's even more credible |
|
|
|
56:43.200 --> 56:46.200 |
|
than if you've never made a fake detector in the first place. |
|
|
|
56:46.200 --> 56:50.200 |
|
What I do think we'll get to is systems |
|
|
|
56:50.200 --> 56:52.200 |
|
that we can kind of use behind the scenes |
|
|
|
56:52.200 --> 56:55.200 |
|
to make estimates of what's going on |
|
|
|
56:55.200 --> 56:59.200 |
|
and maybe not use them in court for a definitive analysis. |
|
|
|
56:59.200 --> 57:04.200 |
|
I also think we will likely get better authentication systems |
|
|
|
57:04.200 --> 57:08.200 |
|
where, imagine that every phone cryptographically |
|
|
|
57:08.200 --> 57:10.200 |
|
signs everything that comes out of it. |
|
|
|
57:10.200 --> 57:12.200 |
|
You wouldn't be able to conclusively tell |
|
|
|
57:12.200 --> 57:14.200 |
|
that an image was real, |
|
|
|
57:14.200 --> 57:18.200 |
|
but you would be able to tell somebody who knew |
|
|
|
57:18.200 --> 57:21.200 |
|
the appropriate private key for this phone |
|
|
|
57:21.200 --> 57:24.200 |
|
was actually able to sign this image |
|
|
|
57:24.200 --> 57:28.200 |
|
and upload it to this server at this time stamp. |
|
|
|
57:28.200 --> 57:31.200 |
|
You could imagine maybe you make phones |
|
|
|
57:31.200 --> 57:35.200 |
|
that have the private keys hardware embedded in them. |
|
|
|
57:35.200 --> 57:37.200 |
|
If a state security agency |
|
|
|
57:37.200 --> 57:39.200 |
|
really wants to infiltrate the company, |
|
|
|
57:39.200 --> 57:42.200 |
|
they could probably plant a private key of their choice |
|
|
|
57:42.200 --> 57:44.200 |
|
or break open the chip |
|
|
|
57:44.200 --> 57:46.200 |
|
and learn the private key or something like that. |
|
|
|
57:46.200 --> 57:48.200 |
|
But it would make it a lot harder |
|
|
|
57:48.200 --> 57:51.200 |
|
for an adversary with fewer resources to fake things. |
|
|
|
57:51.200 --> 57:53.200 |
|
For most of us, it would be okay. |
|
|
|
57:53.200 --> 57:58.200 |
|
You mentioned the beer and the bar and the new ideas. |
|
|
|
57:58.200 --> 58:01.200 |
|
You were able to come up with this new idea |
|
|
|
58:01.200 --> 58:04.200 |
|
pretty quickly and implement it pretty quickly. |
|
|
|
58:04.200 --> 58:06.200 |
|
Do you think there are still many |
|
|
|
58:06.200 --> 58:08.200 |
|
such groundbreaking ideas in deep learning |
|
|
|
58:08.200 --> 58:10.200 |
|
that could be developed so quickly? |
|
|
|
58:10.200 --> 58:13.200 |
|
Yeah, I do think that there are a lot of ideas |
|
|
|
58:13.200 --> 58:15.200 |
|
that can be developed really quickly. |
|
|
|
58:15.200 --> 58:18.200 |
|
GANs were probably a little bit of an outlier |
|
|
|
58:18.200 --> 58:20.200 |
|
on the whole one hour time scale. |
|
|
|
58:20.200 --> 58:24.200 |
|
But just in terms of low resource ideas |
|
|
|
58:24.200 --> 58:26.200 |
|
where you do something really different |
|
|
|
58:26.200 --> 58:29.200 |
|
on a high scale and get a big payback, |
|
|
|
58:29.200 --> 58:32.200 |
|
I think it's not as likely that you'll see that |
|
|
|
58:32.200 --> 58:35.200 |
|
in terms of things like core machine learning technologies |
|
|
|
58:35.200 --> 58:37.200 |
|
like a better classifier |
|
|
|
58:37.200 --> 58:39.200 |
|
or a better reinforcement learning algorithm |
|
|
|
58:39.200 --> 58:41.200 |
|
or a better generative model. |
|
|
|
58:41.200 --> 58:43.200 |
|
If I had the GAN idea today, |
|
|
|
58:43.200 --> 58:45.200 |
|
it would be a lot harder to prove that it was useful |
|
|
|
58:45.200 --> 58:47.200 |
|
than it was back in 2014 |
|
|
|
58:47.200 --> 58:50.200 |
|
because I would need to get it running on something |
|
|
|
58:50.200 --> 58:54.200 |
|
like ImageNet or Celeb A at high resolution. |
|
|
|
58:54.200 --> 58:56.200 |
|
Those take a while to train. |
|
|
|
58:56.200 --> 58:58.200 |
|
You couldn't train it in an hour |
|
|
|
58:58.200 --> 59:01.200 |
|
and know that it was something really new and exciting. |
|
|
|
59:01.200 --> 59:04.200 |
|
Back in 2014, training on MNIST was enough. |
|
|
|
59:04.200 --> 59:07.200 |
|
But there are other areas of machine learning |
|
|
|
59:07.200 --> 59:11.200 |
|
where I think a new idea could actually be developed |
|
|
|
59:11.200 --> 59:13.200 |
|
really quickly with low resources. |
|
|
|
59:13.200 --> 59:15.200 |
|
What's your intuition about what areas |
|
|
|
59:15.200 --> 59:18.200 |
|
of machine learning are ripe for this? |
|
|
|
59:18.200 --> 59:23.200 |
|
Yeah, so I think fairness and interpretability |
|
|
|
59:23.200 --> 59:27.200 |
|
are areas where we just really don't have any idea |
|
|
|
59:27.200 --> 59:29.200 |
|
how anything should be done yet. |
|
|
|
59:29.200 --> 59:31.200 |
|
Like for interpretability, |
|
|
|
59:31.200 --> 59:33.200 |
|
I don't think we even have the right definitions. |
|
|
|
59:33.200 --> 59:36.200 |
|
And even just defining a really useful concept, |
|
|
|
59:36.200 --> 59:38.200 |
|
you don't even need to run any experiments. |
|
|
|
59:38.200 --> 59:40.200 |
|
It could have a huge impact on the field. |
|
|
|
59:40.200 --> 59:43.200 |
|
We've seen that, for example, in differential privacy |
|
|
|
59:43.200 --> 59:45.200 |
|
that Cynthia Dwork and her collaborators |
|
|
|
59:45.200 --> 59:48.200 |
|
made this technical definition of privacy |
|
|
|
59:48.200 --> 59:50.200 |
|
where before a lot of things were really mushy |
|
|
|
59:50.200 --> 59:53.200 |
|
and with that definition, you could actually design |
|
|
|
59:53.200 --> 59:55.200 |
|
randomized algorithms for accessing databases |
|
|
|
59:55.200 --> 59:59.200 |
|
and guarantee that they preserved individual people's privacy |
|
|
|
59:59.200 --> 1:00:03.200 |
|
in a mathematical quantitative sense. |
|
|
|
1:00:03.200 --> 1:00:05.200 |
|
Right now, we all talk a lot about |
|
|
|
1:00:05.200 --> 1:00:07.200 |
|
how interpretable different machine learning algorithms are, |
|
|
|
1:00:07.200 --> 1:00:09.200 |
|
but it's really just people's opinion. |
|
|
|
1:00:09.200 --> 1:00:11.200 |
|
And everybody probably has a different idea |
|
|
|
1:00:11.200 --> 1:00:13.200 |
|
of what interpretability means in their head. |
|
|
|
1:00:13.200 --> 1:00:16.200 |
|
If we could define some concept related to interpretability |
|
|
|
1:00:16.200 --> 1:00:18.200 |
|
that's actually measurable, |
|
|
|
1:00:18.200 --> 1:00:20.200 |
|
that would be a huge leap forward |
|
|
|
1:00:20.200 --> 1:00:24.200 |
|
even without a new algorithm that increases that quantity. |
|
|
|
1:00:24.200 --> 1:00:28.200 |
|
And also, once we had the definition of differential privacy, |
|
|
|
1:00:28.200 --> 1:00:31.200 |
|
it was fast to get the algorithms that guaranteed it. |
|
|
|
1:00:31.200 --> 1:00:33.200 |
|
So you could imagine once we have definitions |
|
|
|
1:00:33.200 --> 1:00:35.200 |
|
of good concepts and interpretability, |
|
|
|
1:00:35.200 --> 1:00:37.200 |
|
we might be able to provide the algorithms |
|
|
|
1:00:37.200 --> 1:00:42.200 |
|
that have the interpretability guarantees quickly, too. |
|
|
|
1:00:42.200 --> 1:00:46.200 |
|
What do you think it takes to build a system |
|
|
|
1:00:46.200 --> 1:00:48.200 |
|
with human level intelligence |
|
|
|
1:00:48.200 --> 1:00:51.200 |
|
as we quickly venture into the philosophical? |
|
|
|
1:00:51.200 --> 1:00:55.200 |
|
So artificial general intelligence, what do you think it takes? |
|
|
|
1:00:55.200 --> 1:01:01.200 |
|
I think that it definitely takes better environments |
|
|
|
1:01:01.200 --> 1:01:03.200 |
|
than we currently have for training agents, |
|
|
|
1:01:03.200 --> 1:01:08.200 |
|
that we want them to have a really wide diversity of experiences. |
|
|
|
1:01:08.200 --> 1:01:11.200 |
|
I also think it's going to take really a lot of computation. |
|
|
|
1:01:11.200 --> 1:01:13.200 |
|
It's hard to imagine exactly how much. |
|
|
|
1:01:13.200 --> 1:01:16.200 |
|
So you're optimistic about simulation, |
|
|
|
1:01:16.200 --> 1:01:19.200 |
|
simulating a variety of environments as the path forward |
|
|
|
1:01:19.200 --> 1:01:21.200 |
|
as opposed to operating in the real world? |
|
|
|
1:01:21.200 --> 1:01:23.200 |
|
I think it's a necessary ingredient. |
|
|
|
1:01:23.200 --> 1:01:27.200 |
|
I don't think that we're going to get to artificial general intelligence |
|
|
|
1:01:27.200 --> 1:01:29.200 |
|
by training on fixed data sets |
|
|
|
1:01:29.200 --> 1:01:32.200 |
|
or by thinking really hard about the problem. |
|
|
|
1:01:32.200 --> 1:01:36.200 |
|
I think that the agent really needs to interact |
|
|
|
1:01:36.200 --> 1:01:41.200 |
|
and have a variety of experiences within the same lifespan. |
|
|
|
1:01:41.200 --> 1:01:45.200 |
|
And today we have many different models that can each do one thing, |
|
|
|
1:01:45.200 --> 1:01:49.200 |
|
and we tend to train them on one dataset or one RL environment. |
|
|
|
1:01:49.200 --> 1:01:53.200 |
|
Sometimes there are actually papers about getting one set of parameters |
|
|
|
1:01:53.200 --> 1:01:56.200 |
|
to perform well in many different RL environments, |
|
|
|
1:01:56.200 --> 1:01:59.200 |
|
but we don't really have anything like an agent |
|
|
|
1:01:59.200 --> 1:02:02.200 |
|
that goes seamlessly from one type of experience to another |
|
|
|
1:02:02.200 --> 1:02:05.200 |
|
and really integrates all the different things that it does |
|
|
|
1:02:05.200 --> 1:02:07.200 |
|
over the course of its life. |
|
|
|
1:02:07.200 --> 1:02:10.200 |
|
When we do see multiagent environments, |
|
|
|
1:02:10.200 --> 1:02:16.200 |
|
they tend to be similar environments. |
|
|
|
1:02:16.200 --> 1:02:19.200 |
|
All of them are playing an action based video game. |
|
|
|
1:02:19.200 --> 1:02:24.200 |
|
We don't really have an agent that goes from playing a video game |
|
|
|
1:02:24.200 --> 1:02:27.200 |
|
to reading the Wall Street Journal |
|
|
|
1:02:27.200 --> 1:02:32.200 |
|
to predicting how effective a molecule will be as a drug or something like that. |
|
|
|
1:02:32.200 --> 1:02:36.200 |
|
What do you think is a good test for intelligence in your view? |
|
|
|
1:02:36.200 --> 1:02:41.200 |
|
There's been a lot of benchmarks started with Alan Turing, |
|
|
|
1:02:41.200 --> 1:02:46.200 |
|
natural conversation being a good benchmark for intelligence. |
|
|
|
1:02:46.200 --> 1:02:53.200 |
|
What would you and good fellows sit back and be really damn impressed |
|
|
|
1:02:53.200 --> 1:02:55.200 |
|
if a system was able to accomplish? |
|
|
|
1:02:55.200 --> 1:02:59.200 |
|
Something that doesn't take a lot of glue from human engineers. |
|
|
|
1:02:59.200 --> 1:03:07.200 |
|
Imagine that instead of having to go to the CIFAR website and download CIFAR 10 |
|
|
|
1:03:07.200 --> 1:03:11.200 |
|
and then write a Python script to parse it and all that, |
|
|
|
1:03:11.200 --> 1:03:16.200 |
|
you could just point an agent at the CIFAR 10 problem |
|
|
|
1:03:16.200 --> 1:03:20.200 |
|
and it downloads and extracts the data and trains a model |
|
|
|
1:03:20.200 --> 1:03:22.200 |
|
and starts giving you predictions. |
|
|
|
1:03:22.200 --> 1:03:28.200 |
|
I feel like something that doesn't need to have every step of the pipeline assembled for it |
|
|
|
1:03:28.200 --> 1:03:30.200 |
|
definitely understands what it's doing. |
|
|
|
1:03:30.200 --> 1:03:34.200 |
|
Is AutoML moving into that direction or are you thinking way even bigger? |
|
|
|
1:03:34.200 --> 1:03:39.200 |
|
AutoML has mostly been moving toward once we've built all the glue, |
|
|
|
1:03:39.200 --> 1:03:44.200 |
|
can the machine learning system design the architecture really well? |
|
|
|
1:03:44.200 --> 1:03:49.200 |
|
I'm more of saying if something knows how to pre process the data |
|
|
|
1:03:49.200 --> 1:03:52.200 |
|
so that it successfully accomplishes the task, |
|
|
|
1:03:52.200 --> 1:03:56.200 |
|
then it would be very hard to argue that it doesn't truly understand the task |
|
|
|
1:03:56.200 --> 1:03:58.200 |
|
in some fundamental sense. |
|
|
|
1:03:58.200 --> 1:04:02.200 |
|
I don't necessarily know that that's the philosophical definition of intelligence, |
|
|
|
1:04:02.200 --> 1:04:05.200 |
|
but that's something that would be really cool to build that would be really useful |
|
|
|
1:04:05.200 --> 1:04:09.200 |
|
and would impress me and would convince me that we've made a step forward in real AI. |
|
|
|
1:04:09.200 --> 1:04:13.200 |
|
You give it the URL for Wikipedia |
|
|
|
1:04:13.200 --> 1:04:18.200 |
|
and then next day expect it to be able to solve CIFAR 10. |
|
|
|
1:04:18.200 --> 1:04:22.200 |
|
Or you type in a paragraph explaining what you want it to do |
|
|
|
1:04:22.200 --> 1:04:28.200 |
|
and it figures out what web searches it should run and downloads all the necessary ingredients. |
|
|
|
1:04:28.200 --> 1:04:37.200 |
|
So you have a very clear, calm way of speaking, no ums, easy to edit. |
|
|
|
1:04:37.200 --> 1:04:44.200 |
|
I've seen comments for both you and I have been identified as both potentially being robots. |
|
|
|
1:04:44.200 --> 1:04:48.200 |
|
If you have to prove to the world that you are indeed human, how would you do it? |
|
|
|
1:04:48.200 --> 1:04:53.200 |
|
I can understand thinking that I'm a robot. |
|
|
|
1:04:53.200 --> 1:04:57.200 |
|
It's the flip side of the Turing test, I think. |
|
|
|
1:04:57.200 --> 1:05:00.200 |
|
Yeah, the prove your human test. |
|
|
|
1:05:00.200 --> 1:05:08.200 |
|
Intellectually, so you have to, is there something that's truly unique in your mind |
|
|
|
1:05:08.200 --> 1:05:13.200 |
|
as it doesn't go back to just natural language again, just being able to talk the way out of it? |
|
|
|
1:05:13.200 --> 1:05:16.200 |
|
So proving that I'm not a robot with today's technology, |
|
|
|
1:05:16.200 --> 1:05:18.200 |
|
that's pretty straightforward. |
|
|
|
1:05:18.200 --> 1:05:25.200 |
|
My conversation today hasn't veered off into talking about the stock market or something because it's my training data. |
|
|
|
1:05:25.200 --> 1:05:31.200 |
|
But I guess more generally trying to prove that something is real from the content alone is incredibly hard. |
|
|
|
1:05:31.200 --> 1:05:37.200 |
|
That's one of the main things I've gotten out of my GAN research, that you can simulate almost anything |
|
|
|
1:05:37.200 --> 1:05:42.200 |
|
and so you have to really step back to a separate channel to prove that something is real. |
|
|
|
1:05:42.200 --> 1:05:48.200 |
|
So I guess I should have had myself stamped on a blockchain when I was born or something, but I didn't do that. |
|
|
|
1:05:48.200 --> 1:05:52.200 |
|
So according to my own research methodology, there's just no way to know at this point. |
|
|
|
1:05:52.200 --> 1:05:59.200 |
|
So what last question, problem stands out for you that you're really excited about challenging in the near future? |
|
|
|
1:05:59.200 --> 1:06:06.200 |
|
I think resistance to adversarial examples, figuring out how to make machine learning secure against an adversary |
|
|
|
1:06:06.200 --> 1:06:11.200 |
|
who wants to interfere it and control it, that is one of the most important things researchers today could solve. |
|
|
|
1:06:11.200 --> 1:06:17.200 |
|
In all domains, image, language, driving and everything. |
|
|
|
1:06:17.200 --> 1:06:22.200 |
|
I guess I'm most concerned about domains we haven't really encountered yet. |
|
|
|
1:06:22.200 --> 1:06:28.200 |
|
Imagine 20 years from now when we're using advanced AIs to do things we haven't even thought of yet. |
|
|
|
1:06:28.200 --> 1:06:37.200 |
|
If you ask people what are the important problems in security of phones in 2002, |
|
|
|
1:06:37.200 --> 1:06:43.200 |
|
I don't think we would have anticipated that we're using them for nearly as many things as we're using them for today. |
|
|
|
1:06:43.200 --> 1:06:47.200 |
|
I think it's going to be like that with AI that you can kind of try to speculate about where it's going, |
|
|
|
1:06:47.200 --> 1:06:53.200 |
|
but really the business opportunities that end up taking off would be hard to predict ahead of time. |
|
|
|
1:06:53.200 --> 1:06:58.200 |
|
What you can predict ahead of time is that almost anything you can do with machine learning, |
|
|
|
1:06:58.200 --> 1:07:04.200 |
|
you would like to make sure that people can't get it to do what they want rather than what you want |
|
|
|
1:07:04.200 --> 1:07:08.200 |
|
just by showing it a funny QR code or a funny input pattern. |
|
|
|
1:07:08.200 --> 1:07:12.200 |
|
You think that the set of methodology to do that can be bigger than any one domain? |
|
|
|
1:07:12.200 --> 1:07:15.200 |
|
I think so, yeah. |
|
|
|
1:07:15.200 --> 1:07:20.200 |
|
One methodology that I think is not a specific methodology, |
|
|
|
1:07:20.200 --> 1:07:25.200 |
|
but a category of solutions that I'm excited about today is making dynamic models |
|
|
|
1:07:25.200 --> 1:07:28.200 |
|
that change every time they make a prediction. |
|
|
|
1:07:28.200 --> 1:07:32.200 |
|
Right now, we tend to train models and then after they're trained, we freeze them. |
|
|
|
1:07:32.200 --> 1:07:37.200 |
|
We just use the same rule to classify everything that comes in from then on. |
|
|
|
1:07:37.200 --> 1:07:40.200 |
|
That's really a sitting duck from a security point of view. |
|
|
|
1:07:40.200 --> 1:07:44.200 |
|
If you always output the same answer for the same input, |
|
|
|
1:07:44.200 --> 1:07:49.200 |
|
then people can just run inputs through until they find a mistake that benefits them, |
|
|
|
1:07:49.200 --> 1:07:53.200 |
|
and then they use the same mistake over and over and over again. |
|
|
|
1:07:53.200 --> 1:08:00.200 |
|
I think having a model that updates its predictions so that it's harder to predict what you're going to get |
|
|
|
1:08:00.200 --> 1:08:04.200 |
|
will make it harder for an adversary to really take control of the system |
|
|
|
1:08:04.200 --> 1:08:06.200 |
|
and make it do what they want it to do. |
|
|
|
1:08:06.200 --> 1:08:10.200 |
|
Yeah, models that maintain a bit of a sense of mystery about them |
|
|
|
1:08:10.200 --> 1:08:12.200 |
|
because they always keep changing. |
|
|
|
1:08:12.200 --> 1:08:14.200 |
|
Ian, thanks so much for talking today. It was awesome. |
|
|
|
1:08:14.200 --> 1:08:36.200 |
|
Thank you for coming in. It's great to see you. |
|
|
|
|