|
WEBVTT |
|
|
|
00:00.000 --> 00:03.080 |
|
The following is a conversation with Rajat Manga. |
|
|
|
00:03.080 --> 00:04.960 |
|
He's an engineering director at Google, |
|
|
|
00:04.960 --> 00:06.960 |
|
leading the TensorFlow team. |
|
|
|
00:06.960 --> 00:09.160 |
|
TensorFlow is an open source library |
|
|
|
00:09.160 --> 00:11.520 |
|
at the center of much of the work going on in the world |
|
|
|
00:11.520 --> 00:14.040 |
|
in deep learning, both the cutting edge research |
|
|
|
00:14.040 --> 00:17.720 |
|
and the large scale application of learning based approaches. |
|
|
|
00:17.720 --> 00:19.480 |
|
But it's quickly becoming much more |
|
|
|
00:19.480 --> 00:20.960 |
|
than a software library. |
|
|
|
00:20.960 --> 00:23.760 |
|
It's now an ecosystem of tools for the deployment |
|
|
|
00:23.760 --> 00:25.720 |
|
of machine learning in the cloud, on the phone, |
|
|
|
00:25.720 --> 00:29.840 |
|
in the browser, on both generic and specialized hardware. |
|
|
|
00:29.840 --> 00:31.920 |
|
TPU, GPU, and so on. |
|
|
|
00:31.920 --> 00:34.200 |
|
Plus, there's a big emphasis on growing |
|
|
|
00:34.200 --> 00:36.600 |
|
a passionate community of developers. |
|
|
|
00:36.600 --> 00:39.760 |
|
Rajat, Jeff Dean, and a large team of engineers at Google |
|
|
|
00:39.760 --> 00:42.720 |
|
Brain are working to define the future of machine learning |
|
|
|
00:42.720 --> 00:46.200 |
|
with TensorFlow 2.0, which is now in alpha. |
|
|
|
00:46.200 --> 00:49.120 |
|
I think the decision to open source TensorFlow |
|
|
|
00:49.120 --> 00:51.720 |
|
is a definitive moment in the tech industry. |
|
|
|
00:51.720 --> 00:54.360 |
|
It showed that open innovation can be successful |
|
|
|
00:54.360 --> 00:56.840 |
|
and inspire many companies to open source their code, |
|
|
|
00:56.840 --> 00:59.640 |
|
to publish, and in general engage in the open exchange |
|
|
|
00:59.640 --> 01:01.160 |
|
of ideas. |
|
|
|
01:01.160 --> 01:03.880 |
|
This conversation is part of the artificial intelligence |
|
|
|
01:03.880 --> 01:05.000 |
|
podcast. |
|
|
|
01:05.000 --> 01:07.760 |
|
If you enjoy it, subscribe on YouTube, iTunes, |
|
|
|
01:07.760 --> 01:09.600 |
|
or simply connect with me on Twitter |
|
|
|
01:09.600 --> 01:12.640 |
|
at Lex Friedman, spelled FRID. |
|
|
|
01:12.640 --> 01:17.880 |
|
And now, here's my conversation with Rajat Manga. |
|
|
|
01:17.880 --> 01:22.440 |
|
You were involved with Google Brain since its start in 2011 |
|
|
|
01:22.440 --> 01:24.800 |
|
with Jeff Dean. |
|
|
|
01:24.800 --> 01:29.160 |
|
It started with disbelief, the proprietary machine learning |
|
|
|
01:29.160 --> 01:32.760 |
|
library, and turned into TensorFlow 2014, |
|
|
|
01:32.760 --> 01:35.760 |
|
the open source library. |
|
|
|
01:35.760 --> 01:39.040 |
|
So what were the early days of Google Brain like? |
|
|
|
01:39.040 --> 01:41.760 |
|
What were the goals, the missions? |
|
|
|
01:41.760 --> 01:45.080 |
|
How do you even proceed forward once there's |
|
|
|
01:45.080 --> 01:47.680 |
|
so much possibilities before you? |
|
|
|
01:47.680 --> 01:50.520 |
|
It was interesting back then when I started, |
|
|
|
01:50.520 --> 01:55.320 |
|
or when you were even just talking about it. |
|
|
|
01:55.320 --> 01:58.800 |
|
The idea of deep learning was interesting |
|
|
|
01:58.800 --> 02:00.400 |
|
and intriguing in some ways. |
|
|
|
02:00.400 --> 02:04.840 |
|
It hadn't yet taken off, but it held some promise. |
|
|
|
02:04.840 --> 02:08.680 |
|
It had shown some very promising and early results. |
|
|
|
02:08.680 --> 02:11.360 |
|
I think the idea where Andrew and Jeff had started |
|
|
|
02:11.360 --> 02:16.160 |
|
was what if we can take this, what people are doing in research, |
|
|
|
02:16.160 --> 02:21.560 |
|
and scale it to what Google has in terms of the compute power, |
|
|
|
02:21.560 --> 02:25.240 |
|
and also put that kind of data together, what does it mean? |
|
|
|
02:25.240 --> 02:28.240 |
|
And so far, the results had been if you scale the computer, |
|
|
|
02:28.240 --> 02:31.480 |
|
scale the data, it does better, and would that work? |
|
|
|
02:31.480 --> 02:33.360 |
|
And so that was the first year or two. |
|
|
|
02:33.360 --> 02:35.080 |
|
Can we prove that outright? |
|
|
|
02:35.080 --> 02:37.440 |
|
And with disbelief, when we started the first year, |
|
|
|
02:37.440 --> 02:40.760 |
|
we got some early wins, which is always great. |
|
|
|
02:40.760 --> 02:41.880 |
|
What were the wins like? |
|
|
|
02:41.880 --> 02:45.240 |
|
What was the wins where there are some problems to this? |
|
|
|
02:45.240 --> 02:46.560 |
|
This is going to be good. |
|
|
|
02:46.560 --> 02:49.640 |
|
I think the two early wins were one was speech |
|
|
|
02:49.640 --> 02:52.200 |
|
that we collaborated very closely with the speech research |
|
|
|
02:52.200 --> 02:54.760 |
|
team, who was also getting interested in this. |
|
|
|
02:54.760 --> 02:57.760 |
|
And the other one was on images where |
|
|
|
02:57.760 --> 03:03.120 |
|
the cat paper, as we call it, that was covered by a lot of folks. |
|
|
|
03:03.120 --> 03:07.440 |
|
And the birth of Google Brain was around neural networks. |
|
|
|
03:07.440 --> 03:09.280 |
|
So it was deep learning from the very beginning. |
|
|
|
03:09.280 --> 03:10.760 |
|
That was the whole mission. |
|
|
|
03:10.760 --> 03:18.960 |
|
So in terms of scale, what was the dream |
|
|
|
03:18.960 --> 03:21.040 |
|
of what this could become? |
|
|
|
03:21.040 --> 03:24.280 |
|
Were there echoes of this open source TensorFlow community |
|
|
|
03:24.280 --> 03:26.240 |
|
that might be brought in? |
|
|
|
03:26.240 --> 03:28.640 |
|
Was there a sense of TPUs? |
|
|
|
03:28.640 --> 03:31.120 |
|
Was there a sense of machine learning |
|
|
|
03:31.120 --> 03:33.680 |
|
is now going to be at the core of the entire company? |
|
|
|
03:33.680 --> 03:36.040 |
|
Is going to grow into that direction? |
|
|
|
03:36.040 --> 03:38.320 |
|
Yeah, I think so that was interesting. |
|
|
|
03:38.320 --> 03:41.320 |
|
And if I think back to 2012 or 2011, |
|
|
|
03:41.320 --> 03:45.240 |
|
and first was can we scale it in the year or so, |
|
|
|
03:45.240 --> 03:47.520 |
|
we had started scaling it to hundreds and thousands |
|
|
|
03:47.520 --> 03:48.080 |
|
of machines. |
|
|
|
03:48.080 --> 03:51.040 |
|
In fact, we had some runs even going to 10,000 machines. |
|
|
|
03:51.040 --> 03:53.840 |
|
And all of those shows great promise. |
|
|
|
03:53.840 --> 03:56.760 |
|
In terms of machine learning at Google, |
|
|
|
03:56.760 --> 03:58.760 |
|
the good thing was Google's been doing machine learning |
|
|
|
03:58.760 --> 04:00.200 |
|
for a long time. |
|
|
|
04:00.200 --> 04:02.120 |
|
Deep learning was new. |
|
|
|
04:02.120 --> 04:05.000 |
|
But as we scale this up, we showed that, yes, that was |
|
|
|
04:05.000 --> 04:07.840 |
|
possible, and it was going to impact lots of things. |
|
|
|
04:07.840 --> 04:11.160 |
|
Like, we started seeing real products wanting to use this. |
|
|
|
04:11.160 --> 04:12.720 |
|
Again, speech was the first. |
|
|
|
04:12.720 --> 04:15.120 |
|
There were image things that photos came out of |
|
|
|
04:15.120 --> 04:17.360 |
|
in many other products as well. |
|
|
|
04:17.360 --> 04:20.120 |
|
So that was exciting. |
|
|
|
04:20.120 --> 04:23.120 |
|
As we went into with that a couple of years, |
|
|
|
04:23.120 --> 04:25.760 |
|
externally also academia started to, |
|
|
|
04:25.760 --> 04:27.760 |
|
there was lots of push on, OK, deep learning's |
|
|
|
04:27.760 --> 04:30.520 |
|
interesting, we should be doing more, and so on. |
|
|
|
04:30.520 --> 04:35.560 |
|
And so by 2014, we were looking at, OK, this is a big thing. |
|
|
|
04:35.560 --> 04:36.680 |
|
It's going to grow. |
|
|
|
04:36.680 --> 04:39.400 |
|
And not just internally, externally as well. |
|
|
|
04:39.400 --> 04:42.240 |
|
Yes, maybe Google's ahead of where everybody is, |
|
|
|
04:42.240 --> 04:43.600 |
|
but there's a lot to do. |
|
|
|
04:43.600 --> 04:46.640 |
|
So a lot of this start to make sense and come together. |
|
|
|
04:46.640 --> 04:51.080 |
|
So the decision to open source, I was just chatting with Chris |
|
|
|
04:51.080 --> 04:53.360 |
|
Flattner about this, the decision to go open source |
|
|
|
04:53.360 --> 04:57.040 |
|
with TensorFlow, I would say for me personally, |
|
|
|
04:57.040 --> 05:00.000 |
|
seems to be one of the big seminal moments in all |
|
|
|
05:00.000 --> 05:01.680 |
|
of software engineering ever. |
|
|
|
05:01.680 --> 05:04.600 |
|
I think that when a large company like Google |
|
|
|
05:04.600 --> 05:08.680 |
|
decides to take a large project that many lawyers might argue |
|
|
|
05:08.680 --> 05:12.840 |
|
has a lot of IP, just decide to go open source with it. |
|
|
|
05:12.840 --> 05:15.200 |
|
And in so doing, lead the entire world in saying, |
|
|
|
05:15.200 --> 05:19.280 |
|
you know what, open innovation is a pretty powerful thing. |
|
|
|
05:19.280 --> 05:22.320 |
|
And it's OK to do. |
|
|
|
05:22.320 --> 05:26.400 |
|
That was, I mean, that's an incredible moment in time. |
|
|
|
05:26.400 --> 05:29.280 |
|
So do you remember those discussions happening? |
|
|
|
05:29.280 --> 05:31.320 |
|
Are there open source should be happening? |
|
|
|
05:31.320 --> 05:32.600 |
|
What was that like? |
|
|
|
05:32.600 --> 05:36.840 |
|
I would say, I think, so the initial idea came from Jeff, |
|
|
|
05:36.840 --> 05:39.320 |
|
who was a big proponent of this. |
|
|
|
05:39.320 --> 05:42.400 |
|
I think it came off of two big things. |
|
|
|
05:42.400 --> 05:46.280 |
|
One was research wise, we were a research group. |
|
|
|
05:46.280 --> 05:50.240 |
|
We were putting all our research out there if you wanted to. |
|
|
|
05:50.240 --> 05:51.680 |
|
We were building on other's research, |
|
|
|
05:51.680 --> 05:54.920 |
|
and we wanted to push the state of the art forward. |
|
|
|
05:54.920 --> 05:56.800 |
|
And part of that was to share the research. |
|
|
|
05:56.800 --> 05:58.920 |
|
That's how I think deep learning and machine learning |
|
|
|
05:58.920 --> 06:01.360 |
|
has really grown so fast. |
|
|
|
06:01.360 --> 06:04.280 |
|
So the next step was, OK, now word software |
|
|
|
06:04.280 --> 06:05.280 |
|
help for that. |
|
|
|
06:05.280 --> 06:09.720 |
|
And it seemed like they were existing a few libraries |
|
|
|
06:09.720 --> 06:12.160 |
|
out there, Tiano being one, Torch being another, |
|
|
|
06:12.160 --> 06:13.960 |
|
and a few others. |
|
|
|
06:13.960 --> 06:15.400 |
|
But they were all done by academia, |
|
|
|
06:15.400 --> 06:19.000 |
|
and so the level was significantly different. |
|
|
|
06:19.000 --> 06:22.040 |
|
The other one was, from a software perspective, |
|
|
|
06:22.040 --> 06:27.120 |
|
Google had done lots of software that we used internally. |
|
|
|
06:27.120 --> 06:29.120 |
|
And we published papers. |
|
|
|
06:29.120 --> 06:31.680 |
|
Often there was an open source project |
|
|
|
06:31.680 --> 06:33.600 |
|
that came out of that, that somebody else |
|
|
|
06:33.600 --> 06:35.440 |
|
picked up that paper and implemented, |
|
|
|
06:35.440 --> 06:38.280 |
|
and they were very successful. |
|
|
|
06:38.280 --> 06:40.920 |
|
Back then, it was like, OK, there's |
|
|
|
06:40.920 --> 06:44.200 |
|
Hadoop, which has come off of tech that we've built. |
|
|
|
06:44.200 --> 06:46.240 |
|
We know that tech we've built is way better |
|
|
|
06:46.240 --> 06:47.880 |
|
for a number of different reasons. |
|
|
|
06:47.880 --> 06:51.680 |
|
We've invested a lot of effort in that. |
|
|
|
06:51.680 --> 06:54.320 |
|
And turns out, we have Google Cloud, |
|
|
|
06:54.320 --> 06:57.520 |
|
and we are now not really providing our tech, |
|
|
|
06:57.520 --> 07:00.520 |
|
but we are saying, OK, we have Bigtable, which |
|
|
|
07:00.520 --> 07:02.080 |
|
is the original thing. |
|
|
|
07:02.080 --> 07:05.280 |
|
We are going to now provide HBase APIs on top of that, which |
|
|
|
07:05.280 --> 07:07.480 |
|
isn't as good, but that's what everybody's used to. |
|
|
|
07:07.480 --> 07:10.960 |
|
So there's like, can we make something that is better |
|
|
|
07:10.960 --> 07:12.320 |
|
and really just provide? |
|
|
|
07:12.320 --> 07:14.320 |
|
Helps the community in lots of ways, |
|
|
|
07:14.320 --> 07:18.320 |
|
but it also helps push the right, a good standard forward. |
|
|
|
07:18.320 --> 07:19.960 |
|
So how does Cloud fit into that? |
|
|
|
07:19.960 --> 07:22.680 |
|
There's a TensorFlow open source library. |
|
|
|
07:22.680 --> 07:25.800 |
|
And how does the fact that you can |
|
|
|
07:25.800 --> 07:28.240 |
|
use so many of the resources that Google provides |
|
|
|
07:28.240 --> 07:31.480 |
|
and the Cloud fit into that strategy? |
|
|
|
07:31.480 --> 07:34.920 |
|
So TensorFlow itself is open, and you can use it anywhere. |
|
|
|
07:34.920 --> 07:38.360 |
|
And we want to make sure that continues to be the case. |
|
|
|
07:38.360 --> 07:42.080 |
|
On Google Cloud, we do make sure that there's |
|
|
|
07:42.080 --> 07:43.800 |
|
lots of integrations with everything else, |
|
|
|
07:43.800 --> 07:47.280 |
|
and we want to make sure that it works really, really well there. |
|
|
|
07:47.280 --> 07:50.080 |
|
You're leading the TensorFlow effort. |
|
|
|
07:50.080 --> 07:52.360 |
|
Can you tell me the history and the timeline of TensorFlow |
|
|
|
07:52.360 --> 07:55.880 |
|
project in terms of major design decisions, |
|
|
|
07:55.880 --> 08:01.240 |
|
like the open source decision, but really, what to include |
|
|
|
08:01.240 --> 08:01.600 |
|
and not? |
|
|
|
08:01.600 --> 08:03.600 |
|
There's this incredible ecosystem that I'd |
|
|
|
08:03.600 --> 08:05.680 |
|
like to talk about, there's all these parts. |
|
|
|
08:05.680 --> 08:12.120 |
|
But if you just some sample moments that |
|
|
|
08:12.120 --> 08:15.960 |
|
defined what TensorFlow eventually became through its, |
|
|
|
08:15.960 --> 08:19.400 |
|
I don't know if you were allowed to say history when it's just, |
|
|
|
08:19.400 --> 08:21.240 |
|
but in deep learning, everything moves so fast |
|
|
|
08:21.240 --> 08:23.400 |
|
in just a few years, it's already history. |
|
|
|
08:23.400 --> 08:24.880 |
|
Yes, yes. |
|
|
|
08:24.880 --> 08:29.760 |
|
So looking back, we were building TensorFlow. |
|
|
|
08:29.760 --> 08:34.240 |
|
I guess we open sourced it in 2015, November 2015. |
|
|
|
08:34.240 --> 08:39.800 |
|
We started on it in summer of 2014, I guess. |
|
|
|
08:39.800 --> 08:42.960 |
|
And somewhere like three to six late 2014, |
|
|
|
08:42.960 --> 08:45.320 |
|
by then we had decided that, OK, there's |
|
|
|
08:45.320 --> 08:47.080 |
|
a high likelihood we'll open source it. |
|
|
|
08:47.080 --> 08:49.560 |
|
So we started thinking about that and making sure |
|
|
|
08:49.560 --> 08:51.320 |
|
that we're heading down that path. |
|
|
|
08:53.960 --> 08:56.280 |
|
At that point, by that point, we'd |
|
|
|
08:56.280 --> 08:59.280 |
|
seen a few lots of different use cases at Google. |
|
|
|
08:59.280 --> 09:01.200 |
|
So there were things like, OK, yes, |
|
|
|
09:01.200 --> 09:04.160 |
|
you want to run in at large scale in the data center. |
|
|
|
09:04.160 --> 09:07.480 |
|
Yes, we need to support different kind of hardware. |
|
|
|
09:07.480 --> 09:09.400 |
|
We had GPUs at that point. |
|
|
|
09:09.400 --> 09:11.880 |
|
We had our first GPU at that point |
|
|
|
09:11.880 --> 09:15.760 |
|
or was about to come out roughly around that time. |
|
|
|
09:15.760 --> 09:18.640 |
|
So the design included those. |
|
|
|
09:18.640 --> 09:21.760 |
|
We had started to push on mobile. |
|
|
|
09:21.760 --> 09:24.880 |
|
So we were running models on mobile. |
|
|
|
09:24.880 --> 09:28.080 |
|
At that point, people were customizing code. |
|
|
|
09:28.080 --> 09:30.280 |
|
So we wanted to make sure TensorFlow could support that |
|
|
|
09:30.280 --> 09:34.120 |
|
as well so that that became part of that overall |
|
|
|
09:34.120 --> 09:35.200 |
|
design. |
|
|
|
09:35.200 --> 09:38.040 |
|
When you say mobile, you mean like pretty complicated |
|
|
|
09:38.040 --> 09:39.960 |
|
algorithms of running on the phone? |
|
|
|
09:39.960 --> 09:40.480 |
|
That's correct. |
|
|
|
09:40.480 --> 09:42.680 |
|
So when you have a model that you |
|
|
|
09:42.680 --> 09:45.200 |
|
deploy on the phone and run it there, right? |
|
|
|
09:45.200 --> 09:47.800 |
|
So already at that time, there was ideas of running machine |
|
|
|
09:47.800 --> 09:48.720 |
|
learning on the phone. |
|
|
|
09:48.720 --> 09:49.240 |
|
That's correct. |
|
|
|
09:49.240 --> 09:51.360 |
|
We already had a couple of products |
|
|
|
09:51.360 --> 09:53.240 |
|
that were doing that by then. |
|
|
|
09:53.240 --> 09:55.480 |
|
And in those cases, we had basically |
|
|
|
09:55.480 --> 09:59.280 |
|
customized handcrafted code or some internal libraries |
|
|
|
09:59.280 --> 10:00.080 |
|
that we're using. |
|
|
|
10:00.080 --> 10:03.280 |
|
So I was actually at Google during this time in a parallel, |
|
|
|
10:03.280 --> 10:04.440 |
|
I guess, universe. |
|
|
|
10:04.440 --> 10:09.240 |
|
But we were using Theano and CAFE. |
|
|
|
10:09.240 --> 10:11.560 |
|
Was there some degree to which you were bouncing, |
|
|
|
10:11.560 --> 10:15.440 |
|
like trying to see what CAFE was offering people, |
|
|
|
10:15.440 --> 10:17.920 |
|
trying to see what Theano was offering |
|
|
|
10:17.920 --> 10:21.320 |
|
that you want to make sure you're delivering on whatever that |
|
|
|
10:21.320 --> 10:23.680 |
|
is, perhaps the Python part of thing. |
|
|
|
10:23.680 --> 10:27.440 |
|
Maybe did that influence any design decisions? |
|
|
|
10:27.440 --> 10:27.880 |
|
Totally. |
|
|
|
10:27.880 --> 10:30.840 |
|
So when we built this belief, and some of that |
|
|
|
10:30.840 --> 10:32.920 |
|
was in parallel with some of these libraries |
|
|
|
10:32.920 --> 10:36.600 |
|
coming up, I mean, Theano itself is older. |
|
|
|
10:36.600 --> 10:41.080 |
|
But we were building this belief focused on our internal thing |
|
|
|
10:41.080 --> 10:42.880 |
|
because our systems were very different. |
|
|
|
10:42.880 --> 10:44.480 |
|
By the time we got to this, we looked |
|
|
|
10:44.480 --> 10:47.040 |
|
at a number of libraries that were out there. |
|
|
|
10:47.040 --> 10:49.240 |
|
Theano, there were folks in the group |
|
|
|
10:49.240 --> 10:52.080 |
|
who had experience with Torch, with Lua. |
|
|
|
10:52.080 --> 10:54.720 |
|
There were folks here who had seen CAFE. |
|
|
|
10:54.720 --> 10:58.800 |
|
I mean, actually, Yang Cheng was here as well. |
|
|
|
10:58.800 --> 11:02.960 |
|
There's what other libraries? |
|
|
|
11:02.960 --> 11:04.880 |
|
I think we looked at a number of things. |
|
|
|
11:04.880 --> 11:06.800 |
|
Might even have looked at Jane and her back then. |
|
|
|
11:06.800 --> 11:09.320 |
|
I'm trying to remember if it was there. |
|
|
|
11:09.320 --> 11:12.240 |
|
In fact, yeah, we did discuss ideas around, OK, |
|
|
|
11:12.240 --> 11:15.280 |
|
should we have a graph or not? |
|
|
|
11:15.280 --> 11:19.280 |
|
And they were supporting all these together |
|
|
|
11:19.280 --> 11:21.880 |
|
was definitely, you know, there were key decisions |
|
|
|
11:21.880 --> 11:22.560 |
|
that we wanted. |
|
|
|
11:22.560 --> 11:28.680 |
|
We had seen limitations in our prior disbelief things. |
|
|
|
11:28.680 --> 11:31.320 |
|
A few of them were just in terms of research |
|
|
|
11:31.320 --> 11:32.280 |
|
was moving so fast. |
|
|
|
11:32.280 --> 11:34.520 |
|
We wanted the flexibility. |
|
|
|
11:34.520 --> 11:36.280 |
|
We want the hardware was changing fast. |
|
|
|
11:36.280 --> 11:39.160 |
|
We expected to change that so that those probably were two |
|
|
|
11:39.160 --> 11:41.400 |
|
things. |
|
|
|
11:41.400 --> 11:43.320 |
|
And yeah, I think the flexibility in terms |
|
|
|
11:43.320 --> 11:45.280 |
|
of being able to express all kinds of crazy things |
|
|
|
11:45.280 --> 11:46.840 |
|
was definitely a big one then. |
|
|
|
11:46.840 --> 11:48.920 |
|
So what the graph decisions, though, |
|
|
|
11:48.920 --> 11:53.720 |
|
with moving towards TensorFlow 2.0, there's more, |
|
|
|
11:53.720 --> 11:56.680 |
|
by default, there'll be eager execution. |
|
|
|
11:56.680 --> 11:59.160 |
|
So sort of hiding the graph a little bit |
|
|
|
11:59.160 --> 12:02.120 |
|
because it's less intuitive in terms of the way |
|
|
|
12:02.120 --> 12:03.520 |
|
people develop and so on. |
|
|
|
12:03.520 --> 12:06.720 |
|
What was that discussion like with in terms of using graphs? |
|
|
|
12:06.720 --> 12:09.320 |
|
It seemed it's kind of the theano way. |
|
|
|
12:09.320 --> 12:11.600 |
|
Did it seem the obvious choice? |
|
|
|
12:11.600 --> 12:15.720 |
|
So I think where it came from was our disbelief, |
|
|
|
12:15.720 --> 12:18.560 |
|
had a graph like thing as well. |
|
|
|
12:18.560 --> 12:19.720 |
|
It wasn't a general graph. |
|
|
|
12:19.720 --> 12:23.160 |
|
It was more like a straight line thing. |
|
|
|
12:23.160 --> 12:25.000 |
|
More like what you might think of Cafe, |
|
|
|
12:25.000 --> 12:28.840 |
|
I guess, in that sense. |
|
|
|
12:28.840 --> 12:31.080 |
|
And we always cared about the production stuff. |
|
|
|
12:31.080 --> 12:33.480 |
|
Even with disbelief, we were deploying a whole bunch of stuff |
|
|
|
12:33.480 --> 12:34.440 |
|
in production. |
|
|
|
12:34.440 --> 12:37.960 |
|
So graph did come from that when we thought of, OK, |
|
|
|
12:37.960 --> 12:40.800 |
|
should we do that in Python and we experimented with some ideas |
|
|
|
12:40.800 --> 12:44.680 |
|
where it looked a lot simpler to use, |
|
|
|
12:44.680 --> 12:47.880 |
|
but not having a graph meant, OK, how do you deploy now? |
|
|
|
12:47.880 --> 12:51.080 |
|
So that was probably what tilted the balance for us. |
|
|
|
12:51.080 --> 12:52.880 |
|
And eventually, we ended up with the graph. |
|
|
|
12:52.880 --> 12:55.320 |
|
And I guess the question there is, did you? |
|
|
|
12:55.320 --> 12:59.800 |
|
I mean, production seems to be the really good thing to focus on. |
|
|
|
12:59.800 --> 13:02.400 |
|
But did you even anticipate the other side of it |
|
|
|
13:02.400 --> 13:04.560 |
|
where there could be, what is it? |
|
|
|
13:04.560 --> 13:05.240 |
|
What are the numbers? |
|
|
|
13:05.240 --> 13:08.920 |
|
Something crazy, 41 million downloads? |
|
|
|
13:08.920 --> 13:09.400 |
|
Yep. |
|
|
|
13:12.680 --> 13:16.240 |
|
I mean, was that even like a possibility in your mind |
|
|
|
13:16.240 --> 13:19.120 |
|
that it would be as popular as it became? |
|
|
|
13:19.120 --> 13:24.880 |
|
So I think we did see a need for this a lot |
|
|
|
13:24.880 --> 13:30.000 |
|
from the research perspective and early days of deep learning |
|
|
|
13:30.000 --> 13:32.280 |
|
in some ways. |
|
|
|
13:32.280 --> 13:33.040 |
|
41 million? |
|
|
|
13:33.040 --> 13:37.640 |
|
No, I don't think I imagine this number then. |
|
|
|
13:37.640 --> 13:42.760 |
|
It seemed like there's a potential future where lots more people |
|
|
|
13:42.760 --> 13:43.760 |
|
would be doing this. |
|
|
|
13:43.760 --> 13:45.640 |
|
And how do we enable that? |
|
|
|
13:45.640 --> 13:49.560 |
|
I would say this kind of growth, I probably |
|
|
|
13:49.560 --> 13:53.680 |
|
started seeing somewhat after the open sourcing where it was |
|
|
|
13:53.680 --> 13:56.240 |
|
like, OK, deep learning is actually |
|
|
|
13:56.240 --> 13:59.200 |
|
growing way faster for a lot of different reasons. |
|
|
|
13:59.200 --> 14:02.720 |
|
And we are in just the right place to push on that |
|
|
|
14:02.720 --> 14:06.040 |
|
and leverage that and deliver on lots of things |
|
|
|
14:06.040 --> 14:07.440 |
|
that people want. |
|
|
|
14:07.440 --> 14:09.760 |
|
So what changed once the open source? |
|
|
|
14:09.760 --> 14:13.320 |
|
Like how this incredible amount of attention |
|
|
|
14:13.320 --> 14:16.120 |
|
from a global population of developers, |
|
|
|
14:16.120 --> 14:18.200 |
|
how did the projects start changing? |
|
|
|
14:18.200 --> 14:21.880 |
|
I don't even actually remember it during those times. |
|
|
|
14:21.880 --> 14:24.560 |
|
I know looking now, there's really good documentation. |
|
|
|
14:24.560 --> 14:26.560 |
|
There's an ecosystem of tools. |
|
|
|
14:26.560 --> 14:31.080 |
|
There's a YouTube channel now. |
|
|
|
14:31.080 --> 14:33.760 |
|
It's very community driven. |
|
|
|
14:33.760 --> 14:38.800 |
|
Back then, I guess 0.1 version. |
|
|
|
14:38.800 --> 14:39.760 |
|
Is that the version? |
|
|
|
14:39.760 --> 14:42.680 |
|
I think we called it 0.6 or 5, something like that. |
|
|
|
14:42.680 --> 14:43.720 |
|
Something like that. |
|
|
|
14:43.720 --> 14:47.200 |
|
What changed leading into 1.0? |
|
|
|
14:47.200 --> 14:48.480 |
|
It's interesting. |
|
|
|
14:48.480 --> 14:51.640 |
|
I think we've gone through a few things there. |
|
|
|
14:51.640 --> 14:53.680 |
|
When we started out, when we first came out, |
|
|
|
14:53.680 --> 14:56.080 |
|
people loved the documentation we have. |
|
|
|
14:56.080 --> 14:58.800 |
|
Because it was just a huge step up from everything else. |
|
|
|
14:58.800 --> 15:01.920 |
|
Because all of those were academic projects, people |
|
|
|
15:01.920 --> 15:04.560 |
|
don't think about documentation. |
|
|
|
15:04.560 --> 15:08.040 |
|
I think what that changed was instead of deep learning |
|
|
|
15:08.040 --> 15:12.560 |
|
being a research thing, some people who were just developers |
|
|
|
15:12.560 --> 15:15.080 |
|
could now suddenly take this out and do |
|
|
|
15:15.080 --> 15:16.920 |
|
some interesting things with it. |
|
|
|
15:16.920 --> 15:20.720 |
|
Who had no clue what machine learning was before then. |
|
|
|
15:20.720 --> 15:22.520 |
|
And that, I think, really changed |
|
|
|
15:22.520 --> 15:27.880 |
|
how things started to scale up in some ways and pushed on it. |
|
|
|
15:27.880 --> 15:30.400 |
|
Over the next few months, as we looked at, |
|
|
|
15:30.400 --> 15:31.960 |
|
how do we stabilize things? |
|
|
|
15:31.960 --> 15:33.840 |
|
As we look at not just researchers, |
|
|
|
15:33.840 --> 15:34.880 |
|
now we want stability. |
|
|
|
15:34.880 --> 15:36.480 |
|
People want to deploy things. |
|
|
|
15:36.480 --> 15:38.960 |
|
That's how we started planning for 1.0. |
|
|
|
15:38.960 --> 15:42.240 |
|
And there are certain needs for that perspective. |
|
|
|
15:42.240 --> 15:45.320 |
|
And so, again, documentation comes up, |
|
|
|
15:45.320 --> 15:49.480 |
|
designs, more kinds of things to put that together. |
|
|
|
15:49.480 --> 15:53.120 |
|
And so that was exciting to get that to a stage where |
|
|
|
15:53.120 --> 15:56.400 |
|
more and more enterprises wanted to buy in and really |
|
|
|
15:56.400 --> 15:58.720 |
|
get behind that. |
|
|
|
15:58.720 --> 16:02.640 |
|
And I think post 1.0 and with the next few releases, |
|
|
|
16:02.640 --> 16:05.240 |
|
their enterprise adoption also started to take off. |
|
|
|
16:05.240 --> 16:08.000 |
|
I would say between the initial release and 1.0, |
|
|
|
16:08.000 --> 16:11.000 |
|
it was, OK, researchers, of course. |
|
|
|
16:11.000 --> 16:13.720 |
|
Then a lot of hobbies and early interest, |
|
|
|
16:13.720 --> 16:15.920 |
|
people excited about this who started to get on board. |
|
|
|
16:15.920 --> 16:19.000 |
|
And then over the 1.x thing, lots of enterprises. |
|
|
|
16:19.000 --> 16:23.760 |
|
I imagine anything that's below 1.0 |
|
|
|
16:23.760 --> 16:27.160 |
|
gets pressured to be enterprise problem or something |
|
|
|
16:27.160 --> 16:28.000 |
|
that's stable. |
|
|
|
16:28.000 --> 16:28.800 |
|
Exactly. |
|
|
|
16:28.800 --> 16:33.360 |
|
And do you have a sense now that TensorFlow is stable? |
|
|
|
16:33.360 --> 16:35.520 |
|
It feels like deep learning, in general, |
|
|
|
16:35.520 --> 16:37.800 |
|
is extremely dynamic field. |
|
|
|
16:37.800 --> 16:39.680 |
|
So much is changing. |
|
|
|
16:39.680 --> 16:43.400 |
|
Do you have a, and TensorFlow has been growing incredibly. |
|
|
|
16:43.400 --> 16:46.720 |
|
Do you have a sense of stability at the helm of this? |
|
|
|
16:46.720 --> 16:48.360 |
|
I mean, I know you're in the midst of it. |
|
|
|
16:48.360 --> 16:50.360 |
|
Yeah. |
|
|
|
16:50.360 --> 16:54.000 |
|
I think in the midst of it, it's often easy to forget what |
|
|
|
16:54.000 --> 16:58.160 |
|
an enterprise wants and what some of the people on that side |
|
|
|
16:58.160 --> 16:58.760 |
|
want. |
|
|
|
16:58.760 --> 17:00.360 |
|
There are still people running models |
|
|
|
17:00.360 --> 17:02.640 |
|
that are three years old, four years old. |
|
|
|
17:02.640 --> 17:06.000 |
|
So inception is still used by tons of people. |
|
|
|
17:06.000 --> 17:08.880 |
|
Even less than 50 is what, a couple of years old now or more. |
|
|
|
17:08.880 --> 17:12.200 |
|
But there are tons of people who use that, and they're fine. |
|
|
|
17:12.200 --> 17:16.200 |
|
They don't need the last couple of bits of performance or quality. |
|
|
|
17:16.200 --> 17:19.600 |
|
They want some stability in things that just work. |
|
|
|
17:19.600 --> 17:22.720 |
|
And so there is value in providing that with that kind |
|
|
|
17:22.720 --> 17:25.160 |
|
of stability and making it really simpler, |
|
|
|
17:25.160 --> 17:27.800 |
|
because that allows a lot more people to access it. |
|
|
|
17:27.800 --> 17:31.640 |
|
And then there's the research crowd, which wants, OK, |
|
|
|
17:31.640 --> 17:33.680 |
|
they want to do these crazy things exactly like you're |
|
|
|
17:33.680 --> 17:37.000 |
|
saying, not just deep learning in the straight up models |
|
|
|
17:37.000 --> 17:38.400 |
|
that used to be there. |
|
|
|
17:38.400 --> 17:41.920 |
|
They want RNNs, and even RNNs are maybe old. |
|
|
|
17:41.920 --> 17:45.520 |
|
They are transformers now, and now it |
|
|
|
17:45.520 --> 17:48.720 |
|
needs to combine with RL and GANs and so on. |
|
|
|
17:48.720 --> 17:52.160 |
|
So there's definitely that area, the boundary that's |
|
|
|
17:52.160 --> 17:55.120 |
|
shifting and pushing the state of the art. |
|
|
|
17:55.120 --> 17:57.120 |
|
But I think there's more and more of the past |
|
|
|
17:57.120 --> 17:59.680 |
|
that's much more stable. |
|
|
|
17:59.680 --> 18:02.680 |
|
And even stuff that was two, three years old |
|
|
|
18:02.680 --> 18:04.920 |
|
is very, very usable by lots of people. |
|
|
|
18:04.920 --> 18:07.440 |
|
So that part makes it a lot easier. |
|
|
|
18:07.440 --> 18:09.800 |
|
So I imagine maybe you can correct me if I'm wrong. |
|
|
|
18:09.800 --> 18:12.440 |
|
One of the biggest use cases is essentially |
|
|
|
18:12.440 --> 18:15.160 |
|
taking something like ResNet 50 and doing |
|
|
|
18:15.160 --> 18:18.520 |
|
some kind of transfer learning on a very particular problem |
|
|
|
18:18.520 --> 18:19.600 |
|
that you have. |
|
|
|
18:19.600 --> 18:24.480 |
|
It's basically probably what majority of the world does. |
|
|
|
18:24.480 --> 18:27.040 |
|
And you want to make that as easy as possible. |
|
|
|
18:27.040 --> 18:30.400 |
|
So I would say, for the hobbyist perspective, |
|
|
|
18:30.400 --> 18:32.800 |
|
that's the most common case. |
|
|
|
18:32.800 --> 18:34.800 |
|
In fact, the apps on phones and stuff |
|
|
|
18:34.800 --> 18:37.680 |
|
that you'll see, the early ones, that's the most common case. |
|
|
|
18:37.680 --> 18:40.320 |
|
I would say there are a couple of reasons for that. |
|
|
|
18:40.320 --> 18:44.400 |
|
One is that everybody talks about that. |
|
|
|
18:44.400 --> 18:46.120 |
|
It looks great on slides. |
|
|
|
18:46.120 --> 18:48.120 |
|
That's a great presentation. |
|
|
|
18:48.120 --> 18:50.040 |
|
Exactly. |
|
|
|
18:50.040 --> 18:53.120 |
|
What enterprises want is that is part of it, |
|
|
|
18:53.120 --> 18:54.480 |
|
but that's not the big thing. |
|
|
|
18:54.480 --> 18:56.760 |
|
Enterprises really have data that they |
|
|
|
18:56.760 --> 18:58.040 |
|
want to make predictions on. |
|
|
|
18:58.040 --> 19:01.160 |
|
This is often what they used to do with the people who |
|
|
|
19:01.160 --> 19:03.600 |
|
were doing ML was just regression models, |
|
|
|
19:03.600 --> 19:06.440 |
|
linear regression, logistic regression, linear models, |
|
|
|
19:06.440 --> 19:09.800 |
|
or maybe gradient booster trees and so on. |
|
|
|
19:09.800 --> 19:11.760 |
|
Some of them still benefit from deep learning, |
|
|
|
19:11.760 --> 19:14.440 |
|
but they weren't that that's the bread and butter, |
|
|
|
19:14.440 --> 19:16.280 |
|
like the structured data and so on. |
|
|
|
19:16.280 --> 19:18.200 |
|
So depending on the audience you look at, |
|
|
|
19:18.200 --> 19:19.520 |
|
they're a little bit different. |
|
|
|
19:19.520 --> 19:23.320 |
|
And they just have, I mean, the best of enterprise |
|
|
|
19:23.320 --> 19:26.480 |
|
probably just has a very large data set |
|
|
|
19:26.480 --> 19:28.640 |
|
where deep learning can probably shine. |
|
|
|
19:28.640 --> 19:29.360 |
|
That's correct. |
|
|
|
19:29.360 --> 19:30.320 |
|
That's right. |
|
|
|
19:30.320 --> 19:32.240 |
|
And then I think the other pieces |
|
|
|
19:32.240 --> 19:34.560 |
|
that they wanted, again, to point out |
|
|
|
19:34.560 --> 19:36.400 |
|
that the developer summit we put together |
|
|
|
19:36.400 --> 19:38.200 |
|
is that the whole TensorFlow Extended |
|
|
|
19:38.200 --> 19:40.600 |
|
piece, which is the entire pipeline, |
|
|
|
19:40.600 --> 19:43.560 |
|
they care about stability across doing their entire thing. |
|
|
|
19:43.560 --> 19:46.200 |
|
They want simplicity across the entire thing. |
|
|
|
19:46.200 --> 19:47.680 |
|
I don't need to just train a model. |
|
|
|
19:47.680 --> 19:51.280 |
|
I need to do that every day again, over and over again. |
|
|
|
19:51.280 --> 19:54.720 |
|
I wonder to which degree you have a role in, I don't know. |
|
|
|
19:54.720 --> 19:57.040 |
|
So I teach a course on deep learning. |
|
|
|
19:57.040 --> 20:01.320 |
|
I have people like lawyers come up to me and say, |
|
|
|
20:01.320 --> 20:04.200 |
|
when is machine learning going to enter legal, |
|
|
|
20:04.200 --> 20:05.560 |
|
the legal realm? |
|
|
|
20:05.560 --> 20:11.720 |
|
The same thing in all kinds of disciplines, immigration, |
|
|
|
20:11.720 --> 20:13.800 |
|
insurance. |
|
|
|
20:13.800 --> 20:17.400 |
|
Often when I see what it boils down to is these companies |
|
|
|
20:17.400 --> 20:19.760 |
|
are often a little bit old school in the way |
|
|
|
20:19.760 --> 20:20.840 |
|
they organize the data. |
|
|
|
20:20.840 --> 20:22.800 |
|
So the data is just not ready yet. |
|
|
|
20:22.800 --> 20:24.040 |
|
It's not digitized. |
|
|
|
20:24.040 --> 20:28.160 |
|
Do you also find yourself being in the role of an evangelist |
|
|
|
20:28.160 --> 20:33.040 |
|
for let's organize your data, folks, |
|
|
|
20:33.040 --> 20:35.440 |
|
and then you'll get the big benefit of TensorFlow? |
|
|
|
20:35.440 --> 20:38.000 |
|
Do you have those conversations? |
|
|
|
20:38.000 --> 20:45.160 |
|
Yeah, I get all kinds of questions there from, OK, |
|
|
|
20:45.160 --> 20:49.000 |
|
what do I need to make this work, right? |
|
|
|
20:49.000 --> 20:50.800 |
|
Do we really need deep learning? |
|
|
|
20:50.800 --> 20:52.240 |
|
I mean, there are all these things. |
|
|
|
20:52.240 --> 20:54.000 |
|
I already used this linear model. |
|
|
|
20:54.000 --> 20:55.160 |
|
Why would this help? |
|
|
|
20:55.160 --> 20:57.160 |
|
I don't have enough data, let's say. |
|
|
|
20:57.160 --> 20:59.960 |
|
Or I want to use machine learning, |
|
|
|
20:59.960 --> 21:01.760 |
|
but I have no clue where to start. |
|
|
|
21:01.760 --> 21:04.920 |
|
So it's a great start to all the way to the experts |
|
|
|
21:04.920 --> 21:08.520 |
|
who wise were very specific things, so it's interesting. |
|
|
|
21:08.520 --> 21:09.600 |
|
Is there a good answer? |
|
|
|
21:09.600 --> 21:12.480 |
|
It boils down to oftentimes digitizing data. |
|
|
|
21:12.480 --> 21:15.240 |
|
So whatever you want automated, whatever data |
|
|
|
21:15.240 --> 21:17.480 |
|
you want to make prediction based on, |
|
|
|
21:17.480 --> 21:21.240 |
|
you have to make sure that it's in an organized form. |
|
|
|
21:21.240 --> 21:23.920 |
|
Like with an intensive flow ecosystem, |
|
|
|
21:23.920 --> 21:26.080 |
|
there's now you're providing more and more data |
|
|
|
21:26.080 --> 21:28.960 |
|
sets and more and more pretrained models. |
|
|
|
21:28.960 --> 21:32.400 |
|
Are you finding yourself also the organizer of data sets? |
|
|
|
21:32.400 --> 21:34.480 |
|
Yes, I think with TensorFlow data sets |
|
|
|
21:34.480 --> 21:38.360 |
|
that we just released, that's definitely come up where people |
|
|
|
21:38.360 --> 21:39.200 |
|
want these data sets. |
|
|
|
21:39.200 --> 21:41.560 |
|
Can we organize them and can we make that easier? |
|
|
|
21:41.560 --> 21:45.320 |
|
So that's definitely one important thing. |
|
|
|
21:45.320 --> 21:47.680 |
|
The other related thing I would say is I often tell people, |
|
|
|
21:47.680 --> 21:50.960 |
|
you know what, don't think of the most fanciest thing |
|
|
|
21:50.960 --> 21:53.320 |
|
that the newest model that you see. |
|
|
|
21:53.320 --> 21:55.480 |
|
Make something very basic work, and then |
|
|
|
21:55.480 --> 21:56.360 |
|
you can improve it. |
|
|
|
21:56.360 --> 21:58.840 |
|
There's just lots of things you can do with it. |
|
|
|
21:58.840 --> 22:00.080 |
|
Yeah, start with the basics. |
|
|
|
22:00.080 --> 22:00.580 |
|
Sure. |
|
|
|
22:00.580 --> 22:03.760 |
|
One of the big things that makes TensorFlow even more |
|
|
|
22:03.760 --> 22:06.440 |
|
accessible was the appearance, whenever |
|
|
|
22:06.440 --> 22:12.400 |
|
that happened, of Keras, the Keras standard outside of TensorFlow. |
|
|
|
22:12.400 --> 22:18.200 |
|
I think it was Keras on top of Tiano at first only, |
|
|
|
22:18.200 --> 22:22.480 |
|
and then Keras became on top of TensorFlow. |
|
|
|
22:22.480 --> 22:29.840 |
|
Do you know when Keras chose to also add TensorFlow as a back end, |
|
|
|
22:29.840 --> 22:33.960 |
|
who was it just the community that drove that initially? |
|
|
|
22:33.960 --> 22:37.000 |
|
Do you know if there was discussions, conversations? |
|
|
|
22:37.000 --> 22:40.920 |
|
Yeah, so Franco started the Keras project |
|
|
|
22:40.920 --> 22:44.560 |
|
before he was at Google, and the first thing was Tiano. |
|
|
|
22:44.560 --> 22:47.120 |
|
I don't remember if that was after TensorFlow |
|
|
|
22:47.120 --> 22:49.640 |
|
was created or way before. |
|
|
|
22:49.640 --> 22:52.000 |
|
And then at some point, when TensorFlow |
|
|
|
22:52.000 --> 22:54.160 |
|
started becoming popular, there were enough similarities |
|
|
|
22:54.160 --> 22:56.320 |
|
that he decided to create this interface |
|
|
|
22:56.320 --> 22:59.200 |
|
and put TensorFlow as a back end. |
|
|
|
22:59.200 --> 23:03.320 |
|
I believe that might still have been before he joined Google. |
|
|
|
23:03.320 --> 23:06.720 |
|
So we weren't really talking about that. |
|
|
|
23:06.720 --> 23:09.720 |
|
He decided on his own and thought that was interesting |
|
|
|
23:09.720 --> 23:12.760 |
|
and relevant to the community. |
|
|
|
23:12.760 --> 23:17.080 |
|
In fact, I didn't find out about him being at Google |
|
|
|
23:17.080 --> 23:19.680 |
|
until a few months after he was here. |
|
|
|
23:19.680 --> 23:21.840 |
|
He was working on some research ideas. |
|
|
|
23:21.840 --> 23:24.480 |
|
And doing Keras and his nights and weekends project and stuff. |
|
|
|
23:24.480 --> 23:25.280 |
|
I wish this thing. |
|
|
|
23:25.280 --> 23:28.480 |
|
So he wasn't part of the TensorFlow. |
|
|
|
23:28.480 --> 23:29.680 |
|
He didn't join initially. |
|
|
|
23:29.680 --> 23:32.240 |
|
He joined research, and he was doing some amazing research. |
|
|
|
23:32.240 --> 23:35.440 |
|
He has some papers on that and research. |
|
|
|
23:35.440 --> 23:38.400 |
|
He's a great researcher as well. |
|
|
|
23:38.400 --> 23:42.400 |
|
And at some point, we realized, oh, he's doing this good stuff. |
|
|
|
23:42.400 --> 23:45.480 |
|
People seem to like the API, and he's right here. |
|
|
|
23:45.480 --> 23:48.280 |
|
So we talked to him, and he said, OK, |
|
|
|
23:48.280 --> 23:50.600 |
|
why don't I come over to your team |
|
|
|
23:50.600 --> 23:52.800 |
|
and work with you for a quarter? |
|
|
|
23:52.800 --> 23:55.440 |
|
And let's make that integration happen. |
|
|
|
23:55.440 --> 23:57.200 |
|
And we talked to his manager, and he said, sure, |
|
|
|
23:57.200 --> 23:59.720 |
|
what, quarter's fine. |
|
|
|
23:59.720 --> 24:03.320 |
|
And that quarter's been something like two years now. |
|
|
|
24:03.320 --> 24:05.040 |
|
So he's fully on this. |
|
|
|
24:05.040 --> 24:12.000 |
|
So Keras got integrated into TensorFlow in a deep way. |
|
|
|
24:12.000 --> 24:15.920 |
|
And now with TensorFlow 2.0, Keras |
|
|
|
24:15.920 --> 24:19.400 |
|
is kind of the recommended way for a beginner |
|
|
|
24:19.400 --> 24:21.960 |
|
to interact with TensorFlow, which |
|
|
|
24:21.960 --> 24:24.640 |
|
makes that initial sort of transfer learning |
|
|
|
24:24.640 --> 24:28.040 |
|
or the basic use cases, even for an enterprise, |
|
|
|
24:28.040 --> 24:29.320 |
|
super simple, right? |
|
|
|
24:29.320 --> 24:29.920 |
|
That's correct. |
|
|
|
24:29.920 --> 24:30.440 |
|
That's right. |
|
|
|
24:30.440 --> 24:32.040 |
|
So what was that decision like? |
|
|
|
24:32.040 --> 24:38.640 |
|
That seems like it's kind of a bold decision as well. |
|
|
|
24:38.640 --> 24:41.200 |
|
We did spend a lot of time thinking about that one. |
|
|
|
24:41.200 --> 24:46.000 |
|
We had a bunch of APIs some bit by us. |
|
|
|
24:46.000 --> 24:48.760 |
|
There was a parallel layers API that we were building |
|
|
|
24:48.760 --> 24:51.560 |
|
and when we decided to do Keras in parallel, |
|
|
|
24:51.560 --> 24:54.400 |
|
so they were like, OK, two things that we are looking at. |
|
|
|
24:54.400 --> 24:55.960 |
|
And the first thing we was trying to do |
|
|
|
24:55.960 --> 25:00.080 |
|
is just have them look similar, be as integrated as possible, |
|
|
|
25:00.080 --> 25:02.200 |
|
share all of that stuff. |
|
|
|
25:02.200 --> 25:05.800 |
|
There were also three other APIs that others had built over time |
|
|
|
25:05.800 --> 25:09.000 |
|
because we didn't have a standard one. |
|
|
|
25:09.000 --> 25:12.080 |
|
But one of the messages that we kept hearing from the community, |
|
|
|
25:12.080 --> 25:13.200 |
|
OK, which one do we use? |
|
|
|
25:13.200 --> 25:15.560 |
|
And they kept seeing, OK, here's a model in this one, |
|
|
|
25:15.560 --> 25:18.840 |
|
and here's a model in this one, which should I pick? |
|
|
|
25:18.840 --> 25:22.680 |
|
So that's sort of like, OK, we had to address that |
|
|
|
25:22.680 --> 25:24.000 |
|
straight on with 2.0. |
|
|
|
25:24.000 --> 25:26.320 |
|
The whole idea was we need to simplify. |
|
|
|
25:26.320 --> 25:28.600 |
|
We had to pick one. |
|
|
|
25:28.600 --> 25:34.600 |
|
Based on where we were, we were like, OK, let's see what |
|
|
|
25:34.600 --> 25:35.640 |
|
are the people like. |
|
|
|
25:35.640 --> 25:39.280 |
|
And Keras was clearly one that lots of people loved. |
|
|
|
25:39.280 --> 25:41.600 |
|
There were lots of great things about it. |
|
|
|
25:41.600 --> 25:43.880 |
|
So we settled on that. |
|
|
|
25:43.880 --> 25:44.680 |
|
Organically. |
|
|
|
25:44.680 --> 25:46.560 |
|
That's kind of the best way to do it. |
|
|
|
25:46.560 --> 25:47.160 |
|
It was great. |
|
|
|
25:47.160 --> 25:48.720 |
|
But it was surprising, nevertheless, |
|
|
|
25:48.720 --> 25:51.120 |
|
to sort of bring in and outside. |
|
|
|
25:51.120 --> 25:54.440 |
|
I mean, there was a feeling like Keras might be almost |
|
|
|
25:54.440 --> 25:58.000 |
|
like a competitor in a certain kind of a two tensor flow. |
|
|
|
25:58.000 --> 26:01.320 |
|
And in a sense, it became an empowering element |
|
|
|
26:01.320 --> 26:02.200 |
|
of tensor flow. |
|
|
|
26:02.200 --> 26:03.280 |
|
That's right. |
|
|
|
26:03.280 --> 26:07.200 |
|
Yeah, it's interesting how you can put two things together |
|
|
|
26:07.200 --> 26:08.280 |
|
which can align right. |
|
|
|
26:08.280 --> 26:11.760 |
|
And in this case, I think Francois, the team, |
|
|
|
26:11.760 --> 26:15.480 |
|
and a bunch of us have chatted and I think we all |
|
|
|
26:15.480 --> 26:17.320 |
|
want to see the same kind of things. |
|
|
|
26:17.320 --> 26:20.360 |
|
We all care about making it easier for the huge set |
|
|
|
26:20.360 --> 26:21.440 |
|
of developers out there. |
|
|
|
26:21.440 --> 26:23.440 |
|
And that makes a difference. |
|
|
|
26:23.440 --> 26:27.280 |
|
So Python has Guido van Rossum, who |
|
|
|
26:27.280 --> 26:30.320 |
|
until recently held the position of benevolent |
|
|
|
26:30.320 --> 26:31.960 |
|
dictator for life. |
|
|
|
26:31.960 --> 26:36.040 |
|
Right, so there's a huge successful open source |
|
|
|
26:36.040 --> 26:37.320 |
|
project like tensor flow. |
|
|
|
26:37.320 --> 26:40.680 |
|
Need one person who makes a final decision. |
|
|
|
26:40.680 --> 26:45.480 |
|
So you did a pretty successful tensor flow Dev Summit |
|
|
|
26:45.480 --> 26:47.520 |
|
just now, last couple of days. |
|
|
|
26:47.520 --> 26:51.080 |
|
There's clearly a lot of different new features |
|
|
|
26:51.080 --> 26:55.480 |
|
being incorporated in amazing ecosystem, so on. |
|
|
|
26:55.480 --> 26:57.320 |
|
How are those design decisions made? |
|
|
|
26:57.320 --> 27:00.960 |
|
Is there a BDFL in tensor flow? |
|
|
|
27:00.960 --> 27:05.800 |
|
And or is it more distributed and organic? |
|
|
|
27:05.800 --> 27:09.880 |
|
I think it's somewhat different, I would say. |
|
|
|
27:09.880 --> 27:16.160 |
|
I've always been involved in the key design directions. |
|
|
|
27:16.160 --> 27:17.560 |
|
But there are lots of things that |
|
|
|
27:17.560 --> 27:20.960 |
|
are distributed where their number of people, Martin |
|
|
|
27:20.960 --> 27:24.760 |
|
Wick being one who has really driven a lot of our open source |
|
|
|
27:24.760 --> 27:27.360 |
|
stuff, a lot of the APIs. |
|
|
|
27:27.360 --> 27:29.200 |
|
And there are a number of other people |
|
|
|
27:29.200 --> 27:32.720 |
|
who have been pushed and been responsible |
|
|
|
27:32.720 --> 27:35.240 |
|
for different parts of it. |
|
|
|
27:35.240 --> 27:37.840 |
|
We do have regular design reviews. |
|
|
|
27:37.840 --> 27:40.680 |
|
Over the last year, we've really spent a lot of time opening up |
|
|
|
27:40.680 --> 27:44.160 |
|
to the community and adding transparency. |
|
|
|
27:44.160 --> 27:45.880 |
|
We're setting more processes in place, |
|
|
|
27:45.880 --> 27:49.600 |
|
so RFCs, special interest groups, really |
|
|
|
27:49.600 --> 27:53.560 |
|
grow that community and scale that. |
|
|
|
27:53.560 --> 27:57.680 |
|
I think the kind of scale that ecosystem is in, |
|
|
|
27:57.680 --> 28:00.240 |
|
I don't think we could scale with having me as the lone |
|
|
|
28:00.240 --> 28:02.320 |
|
point of decision maker. |
|
|
|
28:02.320 --> 28:03.440 |
|
I got it. |
|
|
|
28:03.440 --> 28:05.880 |
|
So yeah, the growth of that ecosystem, |
|
|
|
28:05.880 --> 28:08.040 |
|
maybe you can talk about it a little bit. |
|
|
|
28:08.040 --> 28:10.720 |
|
First of all, when I started with Andre Karpathi |
|
|
|
28:10.720 --> 28:13.640 |
|
when he first did ComNet.js, the fact |
|
|
|
28:13.640 --> 28:15.360 |
|
that you can train in your own network |
|
|
|
28:15.360 --> 28:18.480 |
|
and the browser in JavaScript was incredible. |
|
|
|
28:18.480 --> 28:21.000 |
|
So now TensorFlow.js is really making |
|
|
|
28:21.000 --> 28:26.920 |
|
that a serious, a legit thing, a way |
|
|
|
28:26.920 --> 28:29.560 |
|
to operate, whether it's in the back end or the front end. |
|
|
|
28:29.560 --> 28:32.720 |
|
Then there's the TensorFlow Extended, like you mentioned. |
|
|
|
28:32.720 --> 28:35.360 |
|
There's TensorFlow Lite for mobile. |
|
|
|
28:35.360 --> 28:37.480 |
|
And all of it, as far as I can tell, |
|
|
|
28:37.480 --> 28:39.640 |
|
it's really converging towards being |
|
|
|
28:39.640 --> 28:43.440 |
|
able to save models in the same kind of way. |
|
|
|
28:43.440 --> 28:46.680 |
|
You can move around, you can train on the desktop, |
|
|
|
28:46.680 --> 28:48.800 |
|
and then move it to mobile, and so on. |
|
|
|
28:48.800 --> 28:49.280 |
|
That's right. |
|
|
|
28:49.280 --> 28:52.320 |
|
So this is that cohesiveness. |
|
|
|
28:52.320 --> 28:55.240 |
|
So can you maybe give me whatever |
|
|
|
28:55.240 --> 28:58.840 |
|
I missed, a bigger overview of the mission of the ecosystem |
|
|
|
28:58.840 --> 29:02.120 |
|
that's trying to be built, and where is it moving forward? |
|
|
|
29:02.120 --> 29:02.800 |
|
Yeah. |
|
|
|
29:02.800 --> 29:05.720 |
|
So in short, the way I like to think of this |
|
|
|
29:05.720 --> 29:09.760 |
|
is our goals to enable machine learning. |
|
|
|
29:09.760 --> 29:13.320 |
|
And in a couple of ways, one is we |
|
|
|
29:13.320 --> 29:16.560 |
|
have lots of exciting things going on in ML today. |
|
|
|
29:16.560 --> 29:18.160 |
|
We started with deep learning, but we now |
|
|
|
29:18.160 --> 29:21.400 |
|
support a bunch of other algorithms too. |
|
|
|
29:21.400 --> 29:23.760 |
|
So one is to, on the research side, |
|
|
|
29:23.760 --> 29:25.360 |
|
keep pushing on the state of the art. |
|
|
|
29:25.360 --> 29:27.240 |
|
Can we, how do we enable researchers |
|
|
|
29:27.240 --> 29:28.960 |
|
to build the next amazing thing? |
|
|
|
29:28.960 --> 29:31.800 |
|
So BERT came out recently. |
|
|
|
29:31.800 --> 29:34.000 |
|
It's great that people are able to do new kinds of research. |
|
|
|
29:34.000 --> 29:35.400 |
|
There are lots of amazing research |
|
|
|
29:35.400 --> 29:37.600 |
|
that happens across the world. |
|
|
|
29:37.600 --> 29:38.880 |
|
So that's one direction. |
|
|
|
29:38.880 --> 29:41.400 |
|
The other is, how do you take that |
|
|
|
29:41.400 --> 29:45.200 |
|
across all the people outside who want to take that research |
|
|
|
29:45.200 --> 29:47.400 |
|
and do some great things with it and integrate it |
|
|
|
29:47.400 --> 29:51.800 |
|
to build real products, to have a real impact on people? |
|
|
|
29:51.800 --> 29:56.720 |
|
And so if that's the other axes in some ways. |
|
|
|
29:56.720 --> 29:58.520 |
|
And a high level, one way I think about it |
|
|
|
29:58.520 --> 30:02.480 |
|
is there are a crazy number of computer devices |
|
|
|
30:02.480 --> 30:04.240 |
|
across the world. |
|
|
|
30:04.240 --> 30:08.440 |
|
And we often used to think of ML and training and all of this |
|
|
|
30:08.440 --> 30:10.800 |
|
as, OK, something you do either in the workstation |
|
|
|
30:10.800 --> 30:13.600 |
|
or the data center or cloud. |
|
|
|
30:13.600 --> 30:15.720 |
|
But we see things running on the phones. |
|
|
|
30:15.720 --> 30:17.640 |
|
We see things running on really tiny chips. |
|
|
|
30:17.640 --> 30:20.760 |
|
And we had some demos at the developer summit. |
|
|
|
30:20.760 --> 30:25.160 |
|
And so the way I think about this ecosystem |
|
|
|
30:25.160 --> 30:30.280 |
|
is, how do we help get machine learning on every device that |
|
|
|
30:30.280 --> 30:32.520 |
|
has a compute capability? |
|
|
|
30:32.520 --> 30:33.760 |
|
And that continues to grow. |
|
|
|
30:33.760 --> 30:37.240 |
|
And so in some ways, this ecosystem |
|
|
|
30:37.240 --> 30:40.280 |
|
has looked at various aspects of that |
|
|
|
30:40.280 --> 30:42.440 |
|
and grown over time to cover more of those. |
|
|
|
30:42.440 --> 30:44.640 |
|
And we continue to push the boundaries. |
|
|
|
30:44.640 --> 30:48.640 |
|
In some areas, we've built more tooling and things |
|
|
|
30:48.640 --> 30:50.040 |
|
around that to help you. |
|
|
|
30:50.040 --> 30:52.800 |
|
I mean, the first tool we started was TensorBoard. |
|
|
|
30:52.800 --> 30:56.920 |
|
You want to learn just the training piece, the effects |
|
|
|
30:56.920 --> 30:59.840 |
|
for TensorFlow Extended to really do your entire ML |
|
|
|
30:59.840 --> 31:04.760 |
|
pipelines if you care about all that production stuff, |
|
|
|
31:04.760 --> 31:09.520 |
|
but then going to the edge, going to different kinds of things. |
|
|
|
31:09.520 --> 31:11.800 |
|
And it's not just us now. |
|
|
|
31:11.800 --> 31:15.120 |
|
We are a place where there are lots of libraries being built |
|
|
|
31:15.120 --> 31:15.840 |
|
on top. |
|
|
|
31:15.840 --> 31:18.440 |
|
So there are some for research, maybe things |
|
|
|
31:18.440 --> 31:21.240 |
|
like TensorFlow Agents or TensorFlow Probability that |
|
|
|
31:21.240 --> 31:23.480 |
|
started as research things or for researchers |
|
|
|
31:23.480 --> 31:26.160 |
|
for focusing on certain kinds of algorithms, |
|
|
|
31:26.160 --> 31:30.280 |
|
but they're also being deployed or reduced by production folks. |
|
|
|
31:30.280 --> 31:34.000 |
|
And some have come from within Google, just teams |
|
|
|
31:34.000 --> 31:37.040 |
|
across Google who wanted to do the build these things. |
|
|
|
31:37.040 --> 31:39.680 |
|
Others have come from just the community |
|
|
|
31:39.680 --> 31:41.840 |
|
because there are different pieces |
|
|
|
31:41.840 --> 31:44.640 |
|
that different parts of the community care about. |
|
|
|
31:44.640 --> 31:49.520 |
|
And I see our goal as enabling even that. |
|
|
|
31:49.520 --> 31:53.240 |
|
It's not we cannot and won't build every single thing. |
|
|
|
31:53.240 --> 31:54.840 |
|
That just doesn't make sense. |
|
|
|
31:54.840 --> 31:57.320 |
|
But if we can enable others to build the things |
|
|
|
31:57.320 --> 32:00.640 |
|
that they care about, and there's a broader community that |
|
|
|
32:00.640 --> 32:02.880 |
|
cares about that, and we can help encourage that, |
|
|
|
32:02.880 --> 32:05.240 |
|
and that's great. |
|
|
|
32:05.240 --> 32:08.600 |
|
That really helps the entire ecosystem, not just those. |
|
|
|
32:08.600 --> 32:11.280 |
|
One of the big things about 2.0 that we're pushing on |
|
|
|
32:11.280 --> 32:14.680 |
|
is, OK, we have these so many different pieces, right? |
|
|
|
32:14.680 --> 32:18.440 |
|
How do we help make all of them work well together? |
|
|
|
32:18.440 --> 32:21.960 |
|
There are a few key pieces there that we're pushing on, |
|
|
|
32:21.960 --> 32:23.840 |
|
one being the core format in there |
|
|
|
32:23.840 --> 32:27.480 |
|
and how we share the models themselves through SAVE model |
|
|
|
32:27.480 --> 32:30.440 |
|
and what TensorFlow Hub and so on. |
|
|
|
32:30.440 --> 32:34.000 |
|
And a few of the pieces that we really put this together. |
|
|
|
32:34.000 --> 32:37.240 |
|
I was very skeptical that that's, when TensorFlow.js came out, |
|
|
|
32:37.240 --> 32:40.120 |
|
it didn't seem or deep learning.js. |
|
|
|
32:40.120 --> 32:41.680 |
|
Yeah, that was the first. |
|
|
|
32:41.680 --> 32:45.040 |
|
It seems like technically very difficult project. |
|
|
|
32:45.040 --> 32:47.040 |
|
As a standalone, it's not as difficult. |
|
|
|
32:47.040 --> 32:49.920 |
|
But as a thing that integrates into the ecosystem, |
|
|
|
32:49.920 --> 32:51.240 |
|
it seems very difficult. |
|
|
|
32:51.240 --> 32:53.200 |
|
So I mean, there's a lot of aspects of this |
|
|
|
32:53.200 --> 32:54.200 |
|
you're making look easy. |
|
|
|
32:54.200 --> 32:58.160 |
|
But on the technical side, how many challenges |
|
|
|
32:58.160 --> 33:00.560 |
|
have to be overcome here? |
|
|
|
33:00.560 --> 33:01.520 |
|
A lot. |
|
|
|
33:01.520 --> 33:03.080 |
|
And still have to be overcome. |
|
|
|
33:03.080 --> 33:04.840 |
|
That's the question here, too. |
|
|
|
33:04.840 --> 33:06.160 |
|
There are lots of steps to it. |
|
|
|
33:06.160 --> 33:08.160 |
|
I think we've iterated over the last few years, |
|
|
|
33:08.160 --> 33:10.720 |
|
so there's a lot we've learned. |
|
|
|
33:10.720 --> 33:14.200 |
|
I, yeah, and often when things come together well, |
|
|
|
33:14.200 --> 33:15.080 |
|
things look easy. |
|
|
|
33:15.080 --> 33:16.400 |
|
And that's exactly the point. |
|
|
|
33:16.400 --> 33:18.280 |
|
It should be easy for the end user. |
|
|
|
33:18.280 --> 33:21.320 |
|
But there are lots of things that go behind that. |
|
|
|
33:21.320 --> 33:25.320 |
|
If I think about still challenges ahead, |
|
|
|
33:25.320 --> 33:32.880 |
|
there are we have a lot more devices coming on board, |
|
|
|
33:32.880 --> 33:35.280 |
|
for example, from the hardware perspective. |
|
|
|
33:35.280 --> 33:37.640 |
|
How do we make it really easy for these vendors |
|
|
|
33:37.640 --> 33:42.040 |
|
to integrate with something like TensorFlow? |
|
|
|
33:42.040 --> 33:43.640 |
|
So there's a lot of compiler stuff |
|
|
|
33:43.640 --> 33:45.320 |
|
that others are working on. |
|
|
|
33:45.320 --> 33:48.320 |
|
There are things we can do in terms of our APIs |
|
|
|
33:48.320 --> 33:50.520 |
|
and so on that we can do. |
|
|
|
33:50.520 --> 33:55.840 |
|
As we, TensorFlow started as a very monolithic system. |
|
|
|
33:55.840 --> 33:57.680 |
|
And to some extent, it still is. |
|
|
|
33:57.680 --> 33:59.400 |
|
There are less lots of tools around it, |
|
|
|
33:59.400 --> 34:02.960 |
|
but the core is still pretty large and monolithic. |
|
|
|
34:02.960 --> 34:05.760 |
|
One of the key challenges for us to scale that out |
|
|
|
34:05.760 --> 34:10.440 |
|
is how do we break that apart with clear interfaces? |
|
|
|
34:10.440 --> 34:13.720 |
|
It's, in some ways, it's software engineering one |
|
|
|
34:13.720 --> 34:18.520 |
|
one, but for a system that's now four years old, I guess, |
|
|
|
34:18.520 --> 34:21.600 |
|
or more, and that's still rapidly evolving |
|
|
|
34:21.600 --> 34:24.000 |
|
and that we're not slowing down with, |
|
|
|
34:24.000 --> 34:28.240 |
|
it's hard to change and modify and really break apart. |
|
|
|
34:28.240 --> 34:29.880 |
|
It's sort of like, as people say, right, |
|
|
|
34:29.880 --> 34:32.560 |
|
it's like changing the engine with a car running |
|
|
|
34:32.560 --> 34:33.560 |
|
or fixed benefits. |
|
|
|
34:33.560 --> 34:35.200 |
|
That's exactly what we're trying to do. |
|
|
|
34:35.200 --> 34:39.960 |
|
So there's a challenge here, because the downside |
|
|
|
34:39.960 --> 34:43.840 |
|
of so many people being excited about TensorFlow |
|
|
|
34:43.840 --> 34:48.600 |
|
and becoming to rely on it in many other applications |
|
|
|
34:48.600 --> 34:52.200 |
|
is that you're kind of responsible. |
|
|
|
34:52.200 --> 34:53.520 |
|
It's the technical debt. |
|
|
|
34:53.520 --> 34:55.640 |
|
You're responsible for previous versions |
|
|
|
34:55.640 --> 34:57.720 |
|
to some degree still working. |
|
|
|
34:57.720 --> 34:59.920 |
|
So when you're trying to innovate, |
|
|
|
34:59.920 --> 35:03.760 |
|
I mean, it's probably easier to just start from scratch |
|
|
|
35:03.760 --> 35:05.800 |
|
every few months. |
|
|
|
35:05.800 --> 35:07.160 |
|
Absolutely. |
|
|
|
35:07.160 --> 35:10.880 |
|
So do you feel the pain of that? |
|
|
|
35:10.880 --> 35:15.360 |
|
2.0 does break some back compatibility, but not too much. |
|
|
|
35:15.360 --> 35:18.120 |
|
It seems like the conversion is pretty straightforward. |
|
|
|
35:18.120 --> 35:20.240 |
|
Do you think that's still important, |
|
|
|
35:20.240 --> 35:22.880 |
|
given how quickly deep learning is changing? |
|
|
|
35:22.880 --> 35:26.360 |
|
Can you just, the things that you've learned, |
|
|
|
35:26.360 --> 35:27.440 |
|
can you just start over? |
|
|
|
35:27.440 --> 35:30.120 |
|
Or is there pressure to not? |
|
|
|
35:30.120 --> 35:31.640 |
|
It's a tricky balance. |
|
|
|
35:31.640 --> 35:36.840 |
|
So if it was just a researcher writing a paper who |
|
|
|
35:36.840 --> 35:39.400 |
|
a year later will not look at that code again, |
|
|
|
35:39.400 --> 35:41.560 |
|
sure, it doesn't matter. |
|
|
|
35:41.560 --> 35:43.440 |
|
There are a lot of production systems |
|
|
|
35:43.440 --> 35:45.480 |
|
that rely on TensorFlow, both at Google |
|
|
|
35:45.480 --> 35:47.240 |
|
and across the world. |
|
|
|
35:47.240 --> 35:49.760 |
|
And people worry about this. |
|
|
|
35:49.760 --> 35:53.400 |
|
I mean, these systems run for a long time. |
|
|
|
35:53.400 --> 35:57.240 |
|
So it is important to keep that compatibility and so on. |
|
|
|
35:57.240 --> 36:00.960 |
|
And yes, it does come with a huge cost. |
|
|
|
36:00.960 --> 36:02.920 |
|
We have to think about a lot of things |
|
|
|
36:02.920 --> 36:06.960 |
|
as we do new things and make new changes. |
|
|
|
36:06.960 --> 36:09.120 |
|
I think it's a trade off, right? |
|
|
|
36:09.120 --> 36:12.960 |
|
You can, you might slow certain kinds of things down, |
|
|
|
36:12.960 --> 36:15.440 |
|
but the overall value you're bringing because of that |
|
|
|
36:15.440 --> 36:18.440 |
|
is much bigger because it's not just |
|
|
|
36:18.440 --> 36:20.520 |
|
about breaking the person yesterday. |
|
|
|
36:20.520 --> 36:24.840 |
|
It's also about telling the person tomorrow that, you know what? |
|
|
|
36:24.840 --> 36:26.320 |
|
This is how we do things. |
|
|
|
36:26.320 --> 36:28.520 |
|
We're not going to break you when you come on board |
|
|
|
36:28.520 --> 36:30.320 |
|
because there are lots of new people who are also |
|
|
|
36:30.320 --> 36:32.880 |
|
going to come on board. |
|
|
|
36:32.880 --> 36:34.680 |
|
So one way I like to think about this, |
|
|
|
36:34.680 --> 36:37.960 |
|
and I always push the team to think about it as well, |
|
|
|
36:37.960 --> 36:39.640 |
|
when you want to do new things, you |
|
|
|
36:39.640 --> 36:42.000 |
|
want to start with a clean slate, |
|
|
|
36:42.000 --> 36:44.880 |
|
design with a clean slate in mind, |
|
|
|
36:44.880 --> 36:48.160 |
|
and then we'll figure out how to make sure all the other things |
|
|
|
36:48.160 --> 36:48.640 |
|
work. |
|
|
|
36:48.640 --> 36:52.160 |
|
And yes, we do make compromises occasionally. |
|
|
|
36:52.160 --> 36:55.200 |
|
But unless you design with the clean slate |
|
|
|
36:55.200 --> 36:58.400 |
|
and not worry about that, you'll never get to a good place. |
|
|
|
36:58.400 --> 36:59.120 |
|
That's brilliant. |
|
|
|
36:59.120 --> 37:04.080 |
|
So even if you are responsible in the idea stage, |
|
|
|
37:04.080 --> 37:07.680 |
|
when you're thinking of new, just put all that behind you. |
|
|
|
37:07.680 --> 37:09.600 |
|
OK, that's really well put. |
|
|
|
37:09.600 --> 37:12.480 |
|
So I have to ask this because a lot of students, developers, |
|
|
|
37:12.480 --> 37:16.280 |
|
asked me how I feel about PyTorch versus TensorFlow. |
|
|
|
37:16.280 --> 37:19.720 |
|
So I've recently completely switched my research group |
|
|
|
37:19.720 --> 37:20.920 |
|
to TensorFlow. |
|
|
|
37:20.920 --> 37:23.280 |
|
I wish everybody would just use the same thing. |
|
|
|
37:23.280 --> 37:26.960 |
|
And TensorFlow is as close to that, I believe, as we have. |
|
|
|
37:26.960 --> 37:32.000 |
|
But do you enjoy competition? |
|
|
|
37:32.000 --> 37:35.800 |
|
So TensorFlow is leading in many ways, many dimensions |
|
|
|
37:35.800 --> 37:39.000 |
|
in terms of the ecosystem, in terms of the number of users, |
|
|
|
37:39.000 --> 37:41.200 |
|
momentum power, production level, so on. |
|
|
|
37:41.200 --> 37:46.000 |
|
But a lot of researchers are now also using PyTorch. |
|
|
|
37:46.000 --> 37:47.520 |
|
Do you enjoy that kind of competition, |
|
|
|
37:47.520 --> 37:49.440 |
|
or do you just ignore it and focus |
|
|
|
37:49.440 --> 37:52.320 |
|
on making TensorFlow the best that it can be? |
|
|
|
37:52.320 --> 37:55.480 |
|
So just like research or anything people are doing, |
|
|
|
37:55.480 --> 37:58.120 |
|
it's great to get different kinds of ideas. |
|
|
|
37:58.120 --> 38:01.440 |
|
And when we started with TensorFlow, |
|
|
|
38:01.440 --> 38:05.480 |
|
like I was saying earlier, it was very important for us |
|
|
|
38:05.480 --> 38:07.440 |
|
to also have production in mind. |
|
|
|
38:07.440 --> 38:08.960 |
|
We didn't want just research, right? |
|
|
|
38:08.960 --> 38:11.280 |
|
And that's why we chose certain things. |
|
|
|
38:11.280 --> 38:13.480 |
|
Now PyTorch came along and said, you know what? |
|
|
|
38:13.480 --> 38:14.880 |
|
I only care about research. |
|
|
|
38:14.880 --> 38:16.320 |
|
This is what I'm trying to do. |
|
|
|
38:16.320 --> 38:18.400 |
|
What's the best thing I can do for this? |
|
|
|
38:18.400 --> 38:21.120 |
|
And it started iterating and said, OK, |
|
|
|
38:21.120 --> 38:22.520 |
|
I don't need to worry about graphs. |
|
|
|
38:22.520 --> 38:25.200 |
|
Let me just run things. |
|
|
|
38:25.200 --> 38:27.440 |
|
I don't care if it's not as fast as it can be, |
|
|
|
38:27.440 --> 38:30.480 |
|
but let me just make this part easy. |
|
|
|
38:30.480 --> 38:32.560 |
|
And there are things you can learn from that, right? |
|
|
|
38:32.560 --> 38:36.720 |
|
They, again, had the benefit of seeing what had come before, |
|
|
|
38:36.720 --> 38:40.520 |
|
but also exploring certain different kinds of spaces. |
|
|
|
38:40.520 --> 38:43.560 |
|
And they had some good things there, |
|
|
|
38:43.560 --> 38:46.680 |
|
building on, say, things like Jainer and so on before that. |
|
|
|
38:46.680 --> 38:49.320 |
|
So competition is definitely interesting. |
|
|
|
38:49.320 --> 38:51.040 |
|
It made us, you know, this is an area |
|
|
|
38:51.040 --> 38:53.720 |
|
that we had thought about, like I said, very early on. |
|
|
|
38:53.720 --> 38:56.600 |
|
Over time, we had revisited this a couple of times. |
|
|
|
38:56.600 --> 38:59.000 |
|
Should we add this again? |
|
|
|
38:59.000 --> 39:00.480 |
|
At some point, we said, you know what, |
|
|
|
39:00.480 --> 39:02.920 |
|
here's it seems like this can be done well. |
|
|
|
39:02.920 --> 39:04.280 |
|
So let's try it again. |
|
|
|
39:04.280 --> 39:07.680 |
|
And that's how we started pushing on eager execution. |
|
|
|
39:07.680 --> 39:09.880 |
|
How do we combine those two together, |
|
|
|
39:09.880 --> 39:13.080 |
|
which has finally come very well together in 2.0, |
|
|
|
39:13.080 --> 39:15.720 |
|
but it took us a while to get all the things together |
|
|
|
39:15.720 --> 39:16.320 |
|
and so on. |
|
|
|
39:16.320 --> 39:19.320 |
|
So let me, I mean, ask, put another way. |
|
|
|
39:19.320 --> 39:21.800 |
|
I think eager execution is a really powerful thing, |
|
|
|
39:21.800 --> 39:22.680 |
|
those added. |
|
|
|
39:22.680 --> 39:24.320 |
|
Do you think he wouldn't have been, |
|
|
|
39:25.840 --> 39:28.400 |
|
you know, Muhammad Ali versus Frazier, right? |
|
|
|
39:28.400 --> 39:31.200 |
|
Do you think it wouldn't have been added as quickly |
|
|
|
39:31.200 --> 39:33.760 |
|
if PyTorch wasn't there? |
|
|
|
39:33.760 --> 39:35.440 |
|
It might have taken longer. |
|
|
|
39:35.440 --> 39:36.280 |
|
No longer. |
|
|
|
39:36.280 --> 39:38.960 |
|
It was, I mean, we had tried some variants of that before. |
|
|
|
39:38.960 --> 39:40.920 |
|
So I'm sure it would have happened, |
|
|
|
39:40.920 --> 39:42.240 |
|
but it might have taken longer. |
|
|
|
39:42.240 --> 39:44.800 |
|
I'm grateful that TensorFlow is part of the way they did. |
|
|
|
39:44.800 --> 39:47.760 |
|
That's doing some incredible work last couple of years. |
|
|
|
39:47.760 --> 39:49.640 |
|
What other things that we didn't talk about? |
|
|
|
39:49.640 --> 39:51.520 |
|
Are you looking forward in 2.0? |
|
|
|
39:51.520 --> 39:54.040 |
|
That comes to mind. |
|
|
|
39:54.040 --> 39:56.520 |
|
So we talked about some of the ecosystem stuff, |
|
|
|
39:56.520 --> 40:01.440 |
|
making it easily accessible to Keras, eager execution. |
|
|
|
40:01.440 --> 40:02.880 |
|
Is there other things that we missed? |
|
|
|
40:02.880 --> 40:07.480 |
|
Yeah, so I would say one is just where 2.0 is, |
|
|
|
40:07.480 --> 40:10.760 |
|
and, you know, with all the things that we've talked about, |
|
|
|
40:10.760 --> 40:13.760 |
|
I think as we think beyond that, |
|
|
|
40:13.760 --> 40:16.640 |
|
there are lots of other things that it enables us to do |
|
|
|
40:16.640 --> 40:18.760 |
|
and that we're excited about. |
|
|
|
40:18.760 --> 40:20.720 |
|
So what it's setting us up for, |
|
|
|
40:20.720 --> 40:22.520 |
|
okay, here are these really clean APIs. |
|
|
|
40:22.520 --> 40:25.640 |
|
We've cleaned up the surface for what the users want. |
|
|
|
40:25.640 --> 40:28.320 |
|
What it also allows us to do a whole bunch of stuff |
|
|
|
40:28.320 --> 40:31.600 |
|
behind the scenes once we are ready with 2.0. |
|
|
|
40:31.600 --> 40:36.600 |
|
So for example, in TensorFlow with graphs |
|
|
|
40:36.760 --> 40:37.720 |
|
and all the things you could do, |
|
|
|
40:37.720 --> 40:40.600 |
|
you could always get a lot of good performance |
|
|
|
40:40.600 --> 40:43.280 |
|
if you spent the time to tune it, right? |
|
|
|
40:43.280 --> 40:47.720 |
|
And we've clearly shown that, lots of people do that. |
|
|
|
40:47.720 --> 40:52.720 |
|
With 2.0, with these APIs where we are, |
|
|
|
40:53.040 --> 40:55.120 |
|
we can give you a lot of performance |
|
|
|
40:55.120 --> 40:57.040 |
|
just with whatever you do. |
|
|
|
40:57.040 --> 41:01.400 |
|
You know, because we see these, it's much cleaner. |
|
|
|
41:01.400 --> 41:03.720 |
|
We know most people are gonna do things this way. |
|
|
|
41:03.720 --> 41:05.520 |
|
We can really optimize for that |
|
|
|
41:05.520 --> 41:09.040 |
|
and get a lot of those things out of the box. |
|
|
|
41:09.040 --> 41:10.400 |
|
And it really allows us, you know, |
|
|
|
41:10.400 --> 41:13.880 |
|
both for single machine and distributed and so on, |
|
|
|
41:13.880 --> 41:17.200 |
|
to really explore other spaces behind the scenes |
|
|
|
41:17.200 --> 41:19.680 |
|
after 2.0 in the future versions as well. |
|
|
|
41:19.680 --> 41:23.000 |
|
So right now, the team's really excited about that, |
|
|
|
41:23.000 --> 41:25.800 |
|
that over time, I think we'll see that. |
|
|
|
41:25.800 --> 41:27.720 |
|
The other piece that I was talking about |
|
|
|
41:27.720 --> 41:31.600 |
|
in terms of just restructuring the monolithic thing |
|
|
|
41:31.600 --> 41:34.320 |
|
into more pieces and making it more modular, |
|
|
|
41:34.320 --> 41:36.800 |
|
I think that's gonna be really important |
|
|
|
41:36.800 --> 41:41.800 |
|
for a lot of the other people in the ecosystem, |
|
|
|
41:41.800 --> 41:44.760 |
|
other organizations and so on that wanted to build things. |
|
|
|
41:44.760 --> 41:46.360 |
|
Can you elaborate a little bit what you mean |
|
|
|
41:46.360 --> 41:50.680 |
|
by making TensorFlow more ecosystem or modular? |
|
|
|
41:50.680 --> 41:55.000 |
|
So the way it's organized today is there's one, |
|
|
|
41:55.000 --> 41:56.280 |
|
there are lots of repositories |
|
|
|
41:56.280 --> 41:58.320 |
|
in the TensorFlow organization at GitHub, |
|
|
|
41:58.320 --> 42:01.080 |
|
the core one where we have TensorFlow, |
|
|
|
42:01.080 --> 42:04.080 |
|
it has the execution engine, |
|
|
|
42:04.080 --> 42:08.280 |
|
it has, you know, the key backends for CPUs and GPUs, |
|
|
|
42:08.280 --> 42:12.560 |
|
it has the work to do distributed stuff. |
|
|
|
42:12.560 --> 42:14.360 |
|
And all of these just work together |
|
|
|
42:14.360 --> 42:17.240 |
|
in a single library or binary, |
|
|
|
42:17.240 --> 42:18.800 |
|
there's no way to split them apart easily. |
|
|
|
42:18.800 --> 42:19.960 |
|
I mean, there are some interfaces, |
|
|
|
42:19.960 --> 42:21.600 |
|
but they're not very clean. |
|
|
|
42:21.600 --> 42:24.800 |
|
In a perfect world, you would have clean interfaces where, |
|
|
|
42:24.800 --> 42:27.720 |
|
okay, I wanna run it on my fancy cluster |
|
|
|
42:27.720 --> 42:29.360 |
|
with some custom networking, |
|
|
|
42:29.360 --> 42:30.960 |
|
just implement this and do that. |
|
|
|
42:30.960 --> 42:32.640 |
|
I mean, we kind of support that, |
|
|
|
42:32.640 --> 42:34.560 |
|
but it's hard for people today. |
|
|
|
42:35.480 --> 42:38.160 |
|
I think as we are starting to see more interesting things |
|
|
|
42:38.160 --> 42:39.400 |
|
in some of these spaces, |
|
|
|
42:39.400 --> 42:42.280 |
|
having that clean separation will really start to help. |
|
|
|
42:42.280 --> 42:47.280 |
|
And again, going to the large size of the ecosystem |
|
|
|
42:47.360 --> 42:50.120 |
|
and the different groups involved there, |
|
|
|
42:50.120 --> 42:53.440 |
|
enabling people to evolve and push on things |
|
|
|
42:53.440 --> 42:56.040 |
|
more independently just allows it to scale better. |
|
|
|
42:56.040 --> 42:59.080 |
|
And by people, you mean individual developers and? |
|
|
|
42:59.080 --> 42:59.920 |
|
And organizations. |
|
|
|
42:59.920 --> 43:00.920 |
|
And organizations. |
|
|
|
43:00.920 --> 43:01.760 |
|
That's right. |
|
|
|
43:01.760 --> 43:04.200 |
|
So the hope is that everybody sort of major, |
|
|
|
43:04.200 --> 43:06.880 |
|
I don't know, Pepsi or something uses, |
|
|
|
43:06.880 --> 43:11.040 |
|
like major corporations go to TensorFlow to this kind of. |
|
|
|
43:11.040 --> 43:13.640 |
|
Yeah, if you look at enterprise like Pepsi or these, |
|
|
|
43:13.640 --> 43:15.520 |
|
I mean, a lot of them are already using TensorFlow. |
|
|
|
43:15.520 --> 43:18.960 |
|
They are not the ones that do the development |
|
|
|
43:18.960 --> 43:20.360 |
|
or changes in the core. |
|
|
|
43:20.360 --> 43:21.920 |
|
Some of them do, but a lot of them don't. |
|
|
|
43:21.920 --> 43:23.720 |
|
I mean, they touch small pieces. |
|
|
|
43:23.720 --> 43:26.400 |
|
There are lots of these, some of them being, |
|
|
|
43:26.400 --> 43:28.200 |
|
let's say hardware vendors who are building |
|
|
|
43:28.200 --> 43:30.840 |
|
their custom hardware and they want their own pieces. |
|
|
|
43:30.840 --> 43:34.160 |
|
Or some of them being bigger companies, say IBM. |
|
|
|
43:34.160 --> 43:37.320 |
|
I mean, they're involved in some of our special interest |
|
|
|
43:37.320 --> 43:39.960 |
|
groups and they see a lot of users |
|
|
|
43:39.960 --> 43:42.640 |
|
who want certain things and they want to optimize for that. |
|
|
|
43:42.640 --> 43:44.480 |
|
So folks like that often. |
|
|
|
43:44.480 --> 43:46.400 |
|
Autonomous vehicle companies, perhaps. |
|
|
|
43:46.400 --> 43:48.200 |
|
Exactly, yes. |
|
|
|
43:48.200 --> 43:50.520 |
|
So yeah, like I mentioned, TensorFlow |
|
|
|
43:50.520 --> 43:54.120 |
|
has been down on it 41 million times, 50,000 commits, |
|
|
|
43:54.120 --> 43:58.360 |
|
almost 10,000 pull requests, 1,800 contributors. |
|
|
|
43:58.360 --> 44:02.160 |
|
So I'm not sure if you can explain it, |
|
|
|
44:02.160 --> 44:06.840 |
|
but what does it take to build a community like that? |
|
|
|
44:06.840 --> 44:09.200 |
|
In retrospect, what do you think? |
|
|
|
44:09.200 --> 44:12.080 |
|
What is the critical thing that allowed for this growth |
|
|
|
44:12.080 --> 44:14.600 |
|
to happen and how does that growth continue? |
|
|
|
44:14.600 --> 44:17.920 |
|
Yeah, that's an interesting question. |
|
|
|
44:17.920 --> 44:20.240 |
|
I wish I had all the answers there, I guess, |
|
|
|
44:20.240 --> 44:22.520 |
|
so you could replicate it. |
|
|
|
44:22.520 --> 44:25.520 |
|
I think there are a number of things |
|
|
|
44:25.520 --> 44:27.880 |
|
that need to come together, right? |
|
|
|
44:27.880 --> 44:33.720 |
|
One, just like any new thing, there's |
|
|
|
44:33.720 --> 44:37.960 |
|
a sweet spot of timing, what's needed, |
|
|
|
44:37.960 --> 44:39.520 |
|
does it grow with what's needed. |
|
|
|
44:39.520 --> 44:41.960 |
|
So in this case, for example, TensorFlow |
|
|
|
44:41.960 --> 44:43.640 |
|
is not just grown because it has a good tool, |
|
|
|
44:43.640 --> 44:46.640 |
|
it's also grown with the growth of deep learning itself. |
|
|
|
44:46.640 --> 44:49.000 |
|
So those factors come into play. |
|
|
|
44:49.000 --> 44:53.120 |
|
Other than that, though, I think just |
|
|
|
44:53.120 --> 44:55.560 |
|
hearing, listening to the community, what they're |
|
|
|
44:55.560 --> 44:58.400 |
|
doing, what they need, being open to, |
|
|
|
44:58.400 --> 45:01.080 |
|
like in terms of external contributions, |
|
|
|
45:01.080 --> 45:04.520 |
|
we've spent a lot of time in making sure |
|
|
|
45:04.520 --> 45:06.840 |
|
we can accept those contributions well, |
|
|
|
45:06.840 --> 45:09.400 |
|
we can help the contributors in adding those, |
|
|
|
45:09.400 --> 45:11.240 |
|
putting the right process in place, |
|
|
|
45:11.240 --> 45:13.320 |
|
getting the right kind of community, |
|
|
|
45:13.320 --> 45:16.120 |
|
welcoming them, and so on. |
|
|
|
45:16.120 --> 45:19.000 |
|
Like over the last year, we've really pushed on transparency. |
|
|
|
45:19.000 --> 45:22.200 |
|
That's important for an open source project. |
|
|
|
45:22.200 --> 45:23.760 |
|
People want to know where things are going, |
|
|
|
45:23.760 --> 45:26.400 |
|
and we're like, OK, here's a process for you. |
|
|
|
45:26.400 --> 45:29.320 |
|
You can do that, here are our seasons, and so on. |
|
|
|
45:29.320 --> 45:32.880 |
|
So thinking through, there are lots of community aspects |
|
|
|
45:32.880 --> 45:36.400 |
|
that come into that you can really work on. |
|
|
|
45:36.400 --> 45:38.720 |
|
As a small project, it's maybe easy to do, |
|
|
|
45:38.720 --> 45:42.240 |
|
because there's two developers, and you can do those. |
|
|
|
45:42.240 --> 45:46.960 |
|
As you grow, putting more of these processes in place, |
|
|
|
45:46.960 --> 45:49.080 |
|
thinking about the documentation, |
|
|
|
45:49.080 --> 45:51.400 |
|
thinking about what two developers |
|
|
|
45:51.400 --> 45:55.080 |
|
care about, what kind of tools would they want to use, |
|
|
|
45:55.080 --> 45:56.840 |
|
all of these come into play, I think. |
|
|
|
45:56.840 --> 45:58.400 |
|
So one of the big things, I think, |
|
|
|
45:58.400 --> 46:02.560 |
|
that feeds the TensorFlow fire is people building something |
|
|
|
46:02.560 --> 46:07.680 |
|
on TensorFlow, and implement a particular architecture |
|
|
|
46:07.680 --> 46:09.480 |
|
that does something cool and useful, |
|
|
|
46:09.480 --> 46:11.080 |
|
and they put that on GitHub. |
|
|
|
46:11.080 --> 46:15.640 |
|
And so it just feeds this growth. |
|
|
|
46:15.640 --> 46:19.560 |
|
Do you have a sense that with 2.0 and 1.0, |
|
|
|
46:19.560 --> 46:21.880 |
|
that there may be a little bit of a partitioning like there |
|
|
|
46:21.880 --> 46:26.040 |
|
is with Python 2 and 3, that there'll be a code base |
|
|
|
46:26.040 --> 46:28.320 |
|
in the older versions of TensorFlow |
|
|
|
46:28.320 --> 46:31.120 |
|
that will not be as compatible easily, |
|
|
|
46:31.120 --> 46:35.600 |
|
or are you pretty confident that this kind of conversion |
|
|
|
46:35.600 --> 46:37.960 |
|
is pretty natural and easy to do? |
|
|
|
46:37.960 --> 46:41.480 |
|
So we're definitely working hard to make that very easy to do. |
|
|
|
46:41.480 --> 46:44.040 |
|
There's lots of tooling that we talked about at the developer |
|
|
|
46:44.040 --> 46:46.480 |
|
summit this week, and we'll continue |
|
|
|
46:46.480 --> 46:48.280 |
|
to invest in that tooling. |
|
|
|
46:48.280 --> 46:52.560 |
|
It's when you think of these significant version changes, |
|
|
|
46:52.560 --> 46:55.720 |
|
that's always a risk, and we are really pushing hard |
|
|
|
46:55.720 --> 46:59.160 |
|
to make that transition very, very smooth. |
|
|
|
46:59.160 --> 47:03.000 |
|
I think, so at some level, people |
|
|
|
47:03.000 --> 47:05.520 |
|
want to move when they see the value in the new thing. |
|
|
|
47:05.520 --> 47:07.640 |
|
They don't want to move just because it's a new thing. |
|
|
|
47:07.640 --> 47:11.400 |
|
And some people do, but most people want a really good thing. |
|
|
|
47:11.400 --> 47:13.760 |
|
And I think over the next few months, |
|
|
|
47:13.760 --> 47:15.400 |
|
as people start to see the value, |
|
|
|
47:15.400 --> 47:17.640 |
|
we'll definitely see that shift happening. |
|
|
|
47:17.640 --> 47:20.080 |
|
So I'm pretty excited and confident that we |
|
|
|
47:20.080 --> 47:22.440 |
|
will see people moving. |
|
|
|
47:22.440 --> 47:24.680 |
|
As you said earlier, this field is also moving rapidly, |
|
|
|
47:24.680 --> 47:26.720 |
|
so that'll help because we can do more things. |
|
|
|
47:26.720 --> 47:28.520 |
|
And all the new things will clearly |
|
|
|
47:28.520 --> 47:32.280 |
|
happen in 2.x, so people will have lots of good reasons to move. |
|
|
|
47:32.280 --> 47:36.160 |
|
So what do you think TensorFlow 3.0 looks like? |
|
|
|
47:36.160 --> 47:40.320 |
|
Is there things happening so crazily |
|
|
|
47:40.320 --> 47:42.520 |
|
that even at the end of this year, |
|
|
|
47:42.520 --> 47:45.320 |
|
seems impossible to plan for? |
|
|
|
47:45.320 --> 47:49.440 |
|
Or is it possible to plan for the next five years? |
|
|
|
47:49.440 --> 47:50.800 |
|
I think it's tricky. |
|
|
|
47:50.800 --> 47:55.760 |
|
There are some things that we can expect in terms of, OK, |
|
|
|
47:55.760 --> 47:59.720 |
|
change, yes, change is going to happen. |
|
|
|
47:59.720 --> 48:01.680 |
|
Are there some things going to stick around |
|
|
|
48:01.680 --> 48:03.720 |
|
and some things not going to stick around? |
|
|
|
48:03.720 --> 48:08.160 |
|
I would say the basics of deep learning, |
|
|
|
48:08.160 --> 48:12.680 |
|
the convolutional models or the basic kind of things, |
|
|
|
48:12.680 --> 48:16.280 |
|
they'll probably be around in some form still in five years. |
|
|
|
48:16.280 --> 48:21.160 |
|
Will Aurel and Gans stay very likely based on where they are? |
|
|
|
48:21.160 --> 48:22.840 |
|
Will we have new things? |
|
|
|
48:22.840 --> 48:24.680 |
|
Probably, but those are hard to predict. |
|
|
|
48:24.680 --> 48:29.080 |
|
And some directionally, some things that we can see |
|
|
|
48:29.080 --> 48:32.800 |
|
is in things that we're starting to do |
|
|
|
48:32.800 --> 48:36.560 |
|
with some of our projects right now is just |
|
|
|
48:36.560 --> 48:39.120 |
|
to point out combining eager execution and graphs, |
|
|
|
48:39.120 --> 48:42.240 |
|
where we're starting to make it more like just your natural |
|
|
|
48:42.240 --> 48:43.160 |
|
programming language. |
|
|
|
48:43.160 --> 48:45.640 |
|
You're not trying to program something else. |
|
|
|
48:45.640 --> 48:47.240 |
|
Similarly, with Swift for TensorFlow, |
|
|
|
48:47.240 --> 48:48.280 |
|
we're taking that approach. |
|
|
|
48:48.280 --> 48:50.040 |
|
Can you do something round up? |
|
|
|
48:50.040 --> 48:52.080 |
|
So some of those ideas seem like, OK, |
|
|
|
48:52.080 --> 48:55.000 |
|
that's the right direction in five years |
|
|
|
48:55.000 --> 48:58.360 |
|
we expect to see more in that area. |
|
|
|
48:58.360 --> 49:01.760 |
|
Other things we don't know is, will hardware accelerators |
|
|
|
49:01.760 --> 49:03.200 |
|
be the same? |
|
|
|
49:03.200 --> 49:09.000 |
|
Will we be able to train with four bits instead of 32 bits? |
|
|
|
49:09.000 --> 49:11.440 |
|
And I think the TPU side of things is exploring. |
|
|
|
49:11.440 --> 49:13.960 |
|
I mean, TPU is already on version three. |
|
|
|
49:13.960 --> 49:17.520 |
|
It seems that the evolution of TPU and TensorFlow |
|
|
|
49:17.520 --> 49:24.080 |
|
are coevolving in terms of both their learning |
|
|
|
49:24.080 --> 49:25.720 |
|
from each other and from the community |
|
|
|
49:25.720 --> 49:29.720 |
|
and from the applications where the biggest benefit is achieved. |
|
|
|
49:29.720 --> 49:30.560 |
|
That's right. |
|
|
|
49:30.560 --> 49:33.320 |
|
You've been trying with eager with Keras |
|
|
|
49:33.320 --> 49:36.480 |
|
to make TensorFlow as accessible and easy to use as possible. |
|
|
|
49:36.480 --> 49:39.040 |
|
What do you think for beginners is the biggest thing |
|
|
|
49:39.040 --> 49:40.000 |
|
they struggle with? |
|
|
|
49:40.000 --> 49:42.080 |
|
Have you encountered that? |
|
|
|
49:42.080 --> 49:44.280 |
|
Or is basically what Keras is solving |
|
|
|
49:44.280 --> 49:48.680 |
|
is that eager, like we talked about TensorFlow? |
|
|
|
49:48.680 --> 49:51.480 |
|
For some of them, like you said, the beginners |
|
|
|
49:51.480 --> 49:54.840 |
|
want to just be able to take some image model. |
|
|
|
49:54.840 --> 49:58.040 |
|
They don't care if it's inception or rest net or something else |
|
|
|
49:58.040 --> 50:00.760 |
|
and do some training or transfer learning |
|
|
|
50:00.760 --> 50:02.440 |
|
on their kind of model. |
|
|
|
50:02.440 --> 50:04.400 |
|
Being able to make that easy is important. |
|
|
|
50:04.400 --> 50:08.560 |
|
So in some ways, if you do that by providing them |
|
|
|
50:08.560 --> 50:11.360 |
|
simple models with, say, in Hub or so on, |
|
|
|
50:11.360 --> 50:13.680 |
|
they don't care about what's inside that box, |
|
|
|
50:13.680 --> 50:15.120 |
|
but they want to be able to use it. |
|
|
|
50:15.120 --> 50:17.600 |
|
So we're pushing on, I think, different levels. |
|
|
|
50:17.600 --> 50:20.120 |
|
If you look at just a component that you get, which |
|
|
|
50:20.120 --> 50:22.800 |
|
has the layers already smushed in, |
|
|
|
50:22.800 --> 50:25.200 |
|
the beginners probably just want that. |
|
|
|
50:25.200 --> 50:27.360 |
|
Then the next step is, OK, look at building |
|
|
|
50:27.360 --> 50:29.000 |
|
layers with Keras. |
|
|
|
50:29.000 --> 50:30.600 |
|
If you go out to research, then they |
|
|
|
50:30.600 --> 50:33.120 |
|
are probably writing custom layers themselves |
|
|
|
50:33.120 --> 50:34.360 |
|
or doing their own loops. |
|
|
|
50:34.360 --> 50:36.320 |
|
So there's a whole spectrum there. |
|
|
|
50:36.320 --> 50:38.600 |
|
And then providing the preentrain models |
|
|
|
50:38.600 --> 50:44.760 |
|
seems to really decrease the time from you trying to start. |
|
|
|
50:44.760 --> 50:46.800 |
|
So you could basically, in a Colab notebook, |
|
|
|
50:46.800 --> 50:49.080 |
|
achieve what you need. |
|
|
|
50:49.080 --> 50:51.280 |
|
So I'm basically answering my own question, |
|
|
|
50:51.280 --> 50:54.240 |
|
because I think what TensorFlow delivered on recently |
|
|
|
50:54.240 --> 50:57.000 |
|
is trivial for beginners. |
|
|
|
50:57.000 --> 51:00.760 |
|
So I was just wondering if there was other pain points |
|
|
|
51:00.760 --> 51:02.480 |
|
you're trying to ease, but I'm not sure there would. |
|
|
|
51:02.480 --> 51:04.240 |
|
No, those are probably the big ones. |
|
|
|
51:04.240 --> 51:07.080 |
|
I mean, I see high schoolers doing a whole bunch of things |
|
|
|
51:07.080 --> 51:08.840 |
|
now, which is pretty amazing. |
|
|
|
51:08.840 --> 51:11.360 |
|
It's both amazing and terrifying. |
|
|
|
51:11.360 --> 51:12.640 |
|
Yes. |
|
|
|
51:12.640 --> 51:16.920 |
|
In a sense that when they grow up, |
|
|
|
51:16.920 --> 51:19.280 |
|
some incredible ideas will be coming from them. |
|
|
|
51:19.280 --> 51:21.800 |
|
So there's certainly a technical aspect to your work, |
|
|
|
51:21.800 --> 51:24.600 |
|
but you also have a management aspect |
|
|
|
51:24.600 --> 51:28.000 |
|
to your role with TensorFlow, leading the project, |
|
|
|
51:28.000 --> 51:31.080 |
|
a large number of developers and people. |
|
|
|
51:31.080 --> 51:34.680 |
|
So what do you look for in a good team? |
|
|
|
51:34.680 --> 51:37.400 |
|
What do you think Google has been at the forefront |
|
|
|
51:37.400 --> 51:40.440 |
|
of exploring what it takes to build a good team? |
|
|
|
51:40.440 --> 51:45.520 |
|
And TensorFlow is one of the most cutting edge technologies |
|
|
|
51:45.520 --> 51:46.120 |
|
in the world. |
|
|
|
51:46.120 --> 51:48.080 |
|
So in this context, what do you think |
|
|
|
51:48.080 --> 51:50.480 |
|
makes for a good team? |
|
|
|
51:50.480 --> 51:53.200 |
|
It's definitely something I think a fair bit about. |
|
|
|
51:53.200 --> 51:59.560 |
|
I think in terms of the team being |
|
|
|
51:59.560 --> 52:02.120 |
|
able to deliver something well, one of the things that's |
|
|
|
52:02.120 --> 52:05.800 |
|
important is a cohesion across the team. |
|
|
|
52:05.800 --> 52:10.400 |
|
So being able to execute together and doing things, |
|
|
|
52:10.400 --> 52:11.440 |
|
it's not an end. |
|
|
|
52:11.440 --> 52:14.120 |
|
Like at this scale, an individual engineer |
|
|
|
52:14.120 --> 52:15.400 |
|
can only do so much. |
|
|
|
52:15.400 --> 52:18.200 |
|
There's a lot more that they can do together, |
|
|
|
52:18.200 --> 52:21.640 |
|
even though we have some amazing superstars across Google |
|
|
|
52:21.640 --> 52:22.600 |
|
and in the team. |
|
|
|
52:22.600 --> 52:26.200 |
|
But there's often the way I see it |
|
|
|
52:26.200 --> 52:28.360 |
|
is the product of what the team generates |
|
|
|
52:28.360 --> 52:34.440 |
|
is way larger than the whole individual put together. |
|
|
|
52:34.440 --> 52:37.320 |
|
And so how do we have all of them work together, |
|
|
|
52:37.320 --> 52:40.000 |
|
the culture of the team itself? |
|
|
|
52:40.000 --> 52:43.000 |
|
Hiring good people is important. |
|
|
|
52:43.000 --> 52:45.600 |
|
But part of that is it's not just that, OK, |
|
|
|
52:45.600 --> 52:48.120 |
|
we hire a bunch of smart people and throw them together |
|
|
|
52:48.120 --> 52:49.720 |
|
and let them do things. |
|
|
|
52:49.720 --> 52:52.920 |
|
It's also people have to care about what they're building. |
|
|
|
52:52.920 --> 52:57.320 |
|
People have to be motivated for the right kind of things. |
|
|
|
52:57.320 --> 53:01.400 |
|
That's often an important factor. |
|
|
|
53:01.400 --> 53:04.600 |
|
And finally, how do you put that together |
|
|
|
53:04.600 --> 53:08.840 |
|
with a somewhat unified vision of where we want to go? |
|
|
|
53:08.840 --> 53:11.200 |
|
So are we all looking in the same direction |
|
|
|
53:11.200 --> 53:13.520 |
|
or just going all over? |
|
|
|
53:13.520 --> 53:16.040 |
|
And sometimes it's a mix. |
|
|
|
53:16.040 --> 53:21.400 |
|
Google's a very bottom up organization in some sense. |
|
|
|
53:21.400 --> 53:24.680 |
|
Also research even more so. |
|
|
|
53:24.680 --> 53:26.320 |
|
And that's how we started. |
|
|
|
53:26.320 --> 53:30.840 |
|
But as we've become this larger product and ecosystem, |
|
|
|
53:30.840 --> 53:35.040 |
|
I think it's also important to combine that well with a mix |
|
|
|
53:35.040 --> 53:37.920 |
|
of, OK, here's the direction we want to go in. |
|
|
|
53:37.920 --> 53:39.880 |
|
There is exploration we'll do around that. |
|
|
|
53:39.880 --> 53:43.320 |
|
But let's keep staying in that direction, not just |
|
|
|
53:43.320 --> 53:44.360 |
|
all over the place. |
|
|
|
53:44.360 --> 53:46.880 |
|
And is there a way you monitor the health of the team? |
|
|
|
53:46.880 --> 53:51.920 |
|
Sort of like, is there a way you know you did a good job? |
|
|
|
53:51.920 --> 53:53.000 |
|
The team is good. |
|
|
|
53:53.000 --> 53:56.960 |
|
I mean, you're saying nice things, but it's sometimes |
|
|
|
53:56.960 --> 54:01.120 |
|
difficult to determine how aligned. |
|
|
|
54:01.120 --> 54:04.480 |
|
Because it's not binary, it's not like there's tensions |
|
|
|
54:04.480 --> 54:06.680 |
|
and complexities and so on. |
|
|
|
54:06.680 --> 54:09.400 |
|
And the other element of this is the mesh of superstars. |
|
|
|
54:09.400 --> 54:12.880 |
|
There's so much, even at Google, such a large percentage |
|
|
|
54:12.880 --> 54:16.000 |
|
of work is done by individual superstars too. |
|
|
|
54:16.000 --> 54:19.920 |
|
So there's a, and sometimes those superstars |
|
|
|
54:19.920 --> 54:25.120 |
|
could be against the dynamic of a team and those tensions. |
|
|
|
54:25.120 --> 54:27.320 |
|
I mean, I'm sure TensorFlow might be a little bit easier |
|
|
|
54:27.320 --> 54:31.720 |
|
because the mission of the project is so beautiful. |
|
|
|
54:31.720 --> 54:34.760 |
|
You're at the cutting edge, so it's exciting. |
|
|
|
54:34.760 --> 54:36.640 |
|
But have you had struggle with that? |
|
|
|
54:36.640 --> 54:38.360 |
|
Has there been challenges? |
|
|
|
54:38.360 --> 54:39.800 |
|
There are always people challenges |
|
|
|
54:39.800 --> 54:41.240 |
|
in different kinds of ways. |
|
|
|
54:41.240 --> 54:44.520 |
|
That said, I think we've been what's |
|
|
|
54:44.520 --> 54:49.320 |
|
good about getting people who care and have |
|
|
|
54:49.320 --> 54:51.440 |
|
the same kind of culture, and that's Google in general |
|
|
|
54:51.440 --> 54:53.480 |
|
to a large extent. |
|
|
|
54:53.480 --> 54:56.760 |
|
But also, like you said, given that the project has had |
|
|
|
54:56.760 --> 54:59.160 |
|
so many exciting things to do, there's |
|
|
|
54:59.160 --> 55:02.080 |
|
been room for lots of people to do different kinds of things |
|
|
|
55:02.080 --> 55:06.440 |
|
and grow, which does make the problem a bit easier, I guess. |
|
|
|
55:06.440 --> 55:09.920 |
|
And it allows people, depending on what they're doing, |
|
|
|
55:09.920 --> 55:13.120 |
|
if there's room around them, then that's fine. |
|
|
|
55:13.120 --> 55:19.160 |
|
But yes, we do care about whether a superstar or not |
|
|
|
55:19.160 --> 55:22.560 |
|
that they need to work well with the team across Google. |
|
|
|
55:22.560 --> 55:23.760 |
|
That's interesting to hear. |
|
|
|
55:23.760 --> 55:27.960 |
|
So it's like superstar or not, the productivity broadly |
|
|
|
55:27.960 --> 55:30.520 |
|
is about the team. |
|
|
|
55:30.520 --> 55:31.520 |
|
Yeah. |
|
|
|
55:31.520 --> 55:32.960 |
|
I mean, they might add a lot of value, |
|
|
|
55:32.960 --> 55:35.720 |
|
but if they're hurting the team, then that's a problem. |
|
|
|
55:35.720 --> 55:38.720 |
|
So in hiring engineers, it's so interesting, right? |
|
|
|
55:38.720 --> 55:41.840 |
|
The high rank process, what do you look for? |
|
|
|
55:41.840 --> 55:44.240 |
|
How do you determine a good developer |
|
|
|
55:44.240 --> 55:47.280 |
|
or a good member of a team from just a few minutes |
|
|
|
55:47.280 --> 55:50.320 |
|
or hours together? |
|
|
|
55:50.320 --> 55:51.920 |
|
Again, no magic answers, I'm sure. |
|
|
|
55:51.920 --> 55:52.760 |
|
Yeah. |
|
|
|
55:52.760 --> 55:56.240 |
|
And Google has a hiring process that we've refined |
|
|
|
55:56.240 --> 56:00.880 |
|
over the last 20 years, I guess, and that you've probably |
|
|
|
56:00.880 --> 56:02.200 |
|
heard and seen a lot about. |
|
|
|
56:02.200 --> 56:05.280 |
|
So we do work with the same hiring process in that. |
|
|
|
56:05.280 --> 56:08.280 |
|
That's really helped. |
|
|
|
56:08.280 --> 56:10.880 |
|
For me in particular, I would say, |
|
|
|
56:10.880 --> 56:14.200 |
|
in addition to the core technical skills, |
|
|
|
56:14.200 --> 56:17.560 |
|
what does matter is their motivation |
|
|
|
56:17.560 --> 56:19.560 |
|
in what they want to do. |
|
|
|
56:19.560 --> 56:22.960 |
|
Because if that doesn't align well with where we want to go, |
|
|
|
56:22.960 --> 56:25.320 |
|
that's not going to lead to long term success |
|
|
|
56:25.320 --> 56:27.640 |
|
for either them or the team. |
|
|
|
56:27.640 --> 56:30.640 |
|
And I think that becomes more important the more senior |
|
|
|
56:30.640 --> 56:33.520 |
|
the person is, but it's important at every level. |
|
|
|
56:33.520 --> 56:34.920 |
|
Like even the junior most engineer, |
|
|
|
56:34.920 --> 56:37.680 |
|
if they're not motivated to do well at what they're trying to do, |
|
|
|
56:37.680 --> 56:39.080 |
|
however smart they are, it's going |
|
|
|
56:39.080 --> 56:40.320 |
|
to be hard for them to succeed. |
|
|
|
56:40.320 --> 56:44.520 |
|
Does the Google hiring process touch on that passion? |
|
|
|
56:44.520 --> 56:46.440 |
|
So like trying to determine. |
|
|
|
56:46.440 --> 56:48.440 |
|
Because I think as far as I understand, |
|
|
|
56:48.440 --> 56:52.000 |
|
maybe you can speak to it that the Google hiring process sort |
|
|
|
56:52.000 --> 56:56.360 |
|
of helps the initial like determines the skill set there, |
|
|
|
56:56.360 --> 56:59.840 |
|
is your puzzle solving ability, problem solving ability good. |
|
|
|
56:59.840 --> 57:05.000 |
|
But I'm not sure, but it seems that the determining |
|
|
|
57:05.000 --> 57:07.560 |
|
whether the person is like fire inside them |
|
|
|
57:07.560 --> 57:09.840 |
|
that burns to do anything really doesn't really matter. |
|
|
|
57:09.840 --> 57:11.520 |
|
It's just some cool stuff. |
|
|
|
57:11.520 --> 57:15.320 |
|
I'm going to do it that I don't know. |
|
|
|
57:15.320 --> 57:17.000 |
|
Is that something that ultimately ends up |
|
|
|
57:17.000 --> 57:18.840 |
|
when they have a conversation with you |
|
|
|
57:18.840 --> 57:22.600 |
|
or once it gets closer to the team? |
|
|
|
57:22.600 --> 57:25.400 |
|
So one of the things we do have as part of the process |
|
|
|
57:25.400 --> 57:28.600 |
|
is just a culture fit, like part of the interview process |
|
|
|
57:28.600 --> 57:31.040 |
|
itself, in addition to just the technical skills. |
|
|
|
57:31.040 --> 57:34.240 |
|
And each engineer or whoever the interviewer is, |
|
|
|
57:34.240 --> 57:38.800 |
|
is supposed to rate the person on the culture and the culture |
|
|
|
57:38.800 --> 57:39.960 |
|
fit with Google and so on. |
|
|
|
57:39.960 --> 57:42.160 |
|
So that is definitely part of the process. |
|
|
|
57:42.160 --> 57:45.800 |
|
Now, there are various kinds of projects |
|
|
|
57:45.800 --> 57:46.960 |
|
and different kinds of things. |
|
|
|
57:46.960 --> 57:50.040 |
|
So there might be variants in the kind of culture |
|
|
|
57:50.040 --> 57:51.320 |
|
you want there and so on. |
|
|
|
57:51.320 --> 57:52.720 |
|
And yes, that does vary. |
|
|
|
57:52.720 --> 57:54.920 |
|
So for example, TensorFlow has always |
|
|
|
57:54.920 --> 57:56.920 |
|
been a fast moving project. |
|
|
|
57:56.920 --> 58:00.920 |
|
And we want people who are comfortable with that. |
|
|
|
58:00.920 --> 58:02.640 |
|
But at the same time now, for example, |
|
|
|
58:02.640 --> 58:05.200 |
|
we are at a place where we are also very full fledged product. |
|
|
|
58:05.200 --> 58:08.440 |
|
And we want to make sure things that work really, really |
|
|
|
58:08.440 --> 58:09.320 |
|
work right. |
|
|
|
58:09.320 --> 58:11.680 |
|
You can't cut corners all the time. |
|
|
|
58:11.680 --> 58:14.320 |
|
So balancing that out and finding the people |
|
|
|
58:14.320 --> 58:17.560 |
|
who are the right fit for those is important. |
|
|
|
58:17.560 --> 58:19.720 |
|
And I think those kind of things do vary a bit |
|
|
|
58:19.720 --> 58:23.200 |
|
across projects and teams and product areas across Google. |
|
|
|
58:23.200 --> 58:25.240 |
|
And so you'll see some differences there |
|
|
|
58:25.240 --> 58:27.640 |
|
in the final checklist. |
|
|
|
58:27.640 --> 58:29.600 |
|
But a lot of the core culture, it |
|
|
|
58:29.600 --> 58:32.200 |
|
comes along with just the engineering, excellence, |
|
|
|
58:32.200 --> 58:34.720 |
|
and so on. |
|
|
|
58:34.720 --> 58:39.680 |
|
What is the hardest part of your job? |
|
|
|
58:39.680 --> 58:41.920 |
|
I'll take your pick, I guess. |
|
|
|
58:41.920 --> 58:44.440 |
|
It's fun, I would say. |
|
|
|
58:44.440 --> 58:45.520 |
|
Hard, yes. |
|
|
|
58:45.520 --> 58:47.240 |
|
I mean, lots of things at different times. |
|
|
|
58:47.240 --> 58:49.160 |
|
I think that does vary. |
|
|
|
58:49.160 --> 58:52.640 |
|
So let me clarify that difficult things are fun |
|
|
|
58:52.640 --> 58:55.720 |
|
when you solve them, right? |
|
|
|
58:55.720 --> 58:57.480 |
|
It's fun in that sense. |
|
|
|
58:57.480 --> 59:02.600 |
|
I think the key to a successful thing across the board, |
|
|
|
59:02.600 --> 59:05.320 |
|
and in this case, it's a large ecosystem now, |
|
|
|
59:05.320 --> 59:09.800 |
|
but even a small product, is striking that fine balance |
|
|
|
59:09.800 --> 59:12.000 |
|
across different aspects of it. |
|
|
|
59:12.000 --> 59:17.000 |
|
Sometimes it's how fast you go versus how perfect it is. |
|
|
|
59:17.000 --> 59:21.400 |
|
Sometimes it's how do you involve this huge community? |
|
|
|
59:21.400 --> 59:22.360 |
|
Who do you involve? |
|
|
|
59:22.360 --> 59:25.440 |
|
Or do you decide, OK, now is not a good time to involve them |
|
|
|
59:25.440 --> 59:30.160 |
|
because it's not the right fit? |
|
|
|
59:30.160 --> 59:33.640 |
|
Sometimes it's saying no to certain kinds of things. |
|
|
|
59:33.640 --> 59:36.880 |
|
Those are often the hard decisions. |
|
|
|
59:36.880 --> 59:41.000 |
|
Some of them you make quickly because you don't have the time. |
|
|
|
59:41.000 --> 59:43.200 |
|
Some of them you get time to think about them, |
|
|
|
59:43.200 --> 59:44.480 |
|
but they're always hard. |
|
|
|
59:44.480 --> 59:49.200 |
|
So both choices are pretty good, those decisions. |
|
|
|
59:49.200 --> 59:50.360 |
|
What about deadlines? |
|
|
|
59:50.360 --> 59:58.200 |
|
Is this defined TensorFlow to be driven by deadlines |
|
|
|
59:58.200 --> 1:00:00.360 |
|
to a degree that a product might? |
|
|
|
1:00:00.360 --> 1:00:04.920 |
|
Or is there still a balance to where it's less deadline? |
|
|
|
1:00:04.920 --> 1:00:08.920 |
|
You had the Dev Summit, they came together incredibly. |
|
|
|
1:00:08.920 --> 1:00:11.440 |
|
Looked like there's a lot of moving pieces and so on. |
|
|
|
1:00:11.440 --> 1:00:15.080 |
|
So did that deadline make people rise to the occasion, |
|
|
|
1:00:15.080 --> 1:00:18.360 |
|
releasing TensorFlow 2.0 Alpha? |
|
|
|
1:00:18.360 --> 1:00:20.360 |
|
I'm sure that was done last minute as well. |
|
|
|
1:00:20.360 --> 1:00:25.600 |
|
I mean, up to the last point. |
|
|
|
1:00:25.600 --> 1:00:28.600 |
|
Again, it's one of those things that you |
|
|
|
1:00:28.600 --> 1:00:29.960 |
|
need to strike the good balance. |
|
|
|
1:00:29.960 --> 1:00:32.040 |
|
There's some value that deadlines bring |
|
|
|
1:00:32.040 --> 1:00:33.920 |
|
that does bring a sense of urgency |
|
|
|
1:00:33.920 --> 1:00:35.720 |
|
to get the right things together. |
|
|
|
1:00:35.720 --> 1:00:38.280 |
|
Instead of getting the perfect thing out, |
|
|
|
1:00:38.280 --> 1:00:41.280 |
|
you need something that's good and works well. |
|
|
|
1:00:41.280 --> 1:00:43.720 |
|
And the team definitely did a great job in putting that |
|
|
|
1:00:43.720 --> 1:00:46.560 |
|
together, so it was very amazed and excited by everything, |
|
|
|
1:00:46.560 --> 1:00:48.680 |
|
how that came together. |
|
|
|
1:00:48.680 --> 1:00:50.640 |
|
That said, across the year, we try not |
|
|
|
1:00:50.640 --> 1:00:52.520 |
|
to put out official deadlines. |
|
|
|
1:00:52.520 --> 1:00:56.960 |
|
We focus on key things that are important, |
|
|
|
1:00:56.960 --> 1:01:00.600 |
|
figure out how much of it's important, |
|
|
|
1:01:00.600 --> 1:01:05.760 |
|
and we are developing in the open, internally and externally, |
|
|
|
1:01:05.760 --> 1:01:07.920 |
|
everything's available to everybody. |
|
|
|
1:01:07.920 --> 1:01:11.120 |
|
So you can pick and look at where things are. |
|
|
|
1:01:11.120 --> 1:01:13.160 |
|
We do releases at a regular cadence, |
|
|
|
1:01:13.160 --> 1:01:16.320 |
|
so fine if something doesn't necessarily end up with this |
|
|
|
1:01:16.320 --> 1:01:19.600 |
|
month, it'll end up in the next release in a month or two. |
|
|
|
1:01:19.600 --> 1:01:22.840 |
|
And that's OK, but we want to keep moving |
|
|
|
1:01:22.840 --> 1:01:26.520 |
|
as fast as we can in these different areas. |
|
|
|
1:01:26.520 --> 1:01:30.080 |
|
Because we can iterate and improve on things, sometimes |
|
|
|
1:01:30.080 --> 1:01:32.920 |
|
it's OK to put things out that aren't fully ready. |
|
|
|
1:01:32.920 --> 1:01:35.640 |
|
If you make sure it's clear that, OK, this is experimental, |
|
|
|
1:01:35.640 --> 1:01:37.960 |
|
but it's out there if you want to try and give feedback. |
|
|
|
1:01:37.960 --> 1:01:39.400 |
|
That's very, very useful. |
|
|
|
1:01:39.400 --> 1:01:43.560 |
|
I think that quick cycle and quick iteration is important. |
|
|
|
1:01:43.560 --> 1:01:47.200 |
|
That's what we often focus on rather than here's |
|
|
|
1:01:47.200 --> 1:01:49.200 |
|
a deadline where you get everything else. |
|
|
|
1:01:49.200 --> 1:01:52.880 |
|
It's 2.0, is there pressure to make that stable? |
|
|
|
1:01:52.880 --> 1:01:57.760 |
|
Or like, for example, WordPress 5.0 just came out, |
|
|
|
1:01:57.760 --> 1:02:01.760 |
|
and there was no pressure to, it was a lot of build updates |
|
|
|
1:02:01.760 --> 1:02:04.960 |
|
that delivered way too late. |
|
|
|
1:02:04.960 --> 1:02:06.440 |
|
And they said, OK, well, we're going |
|
|
|
1:02:06.440 --> 1:02:09.680 |
|
to release a lot of updates really quickly to improve it. |
|
|
|
1:02:09.680 --> 1:02:12.240 |
|
Do you see TensorFlow 2.0 in that same kind of way, |
|
|
|
1:02:12.240 --> 1:02:15.240 |
|
or is there this pressure to once it hits 2.0, |
|
|
|
1:02:15.240 --> 1:02:16.760 |
|
once you get to the release candidate, |
|
|
|
1:02:16.760 --> 1:02:19.440 |
|
and then you get to the final, that's |
|
|
|
1:02:19.440 --> 1:02:22.480 |
|
going to be the stable thing? |
|
|
|
1:02:22.480 --> 1:02:26.680 |
|
So it's going to be stable in just like 1.0X |
|
|
|
1:02:26.680 --> 1:02:32.080 |
|
was where every API that's there is going to remain in work. |
|
|
|
1:02:32.080 --> 1:02:34.800 |
|
It doesn't mean we can't change things under the covers. |
|
|
|
1:02:34.800 --> 1:02:36.720 |
|
It doesn't mean we can't add things. |
|
|
|
1:02:36.720 --> 1:02:39.200 |
|
So there's still a lot more for us to do, |
|
|
|
1:02:39.200 --> 1:02:41.080 |
|
and we continue to have more releases. |
|
|
|
1:02:41.080 --> 1:02:42.920 |
|
So in that sense, there's still, I |
|
|
|
1:02:42.920 --> 1:02:44.680 |
|
don't think we'd be done in like two months |
|
|
|
1:02:44.680 --> 1:02:46.160 |
|
when we release this. |
|
|
|
1:02:46.160 --> 1:02:49.880 |
|
I don't know if you can say, but is there, you know, |
|
|
|
1:02:49.880 --> 1:02:53.680 |
|
there's not external deadlines for TensorFlow 2.0, |
|
|
|
1:02:53.680 --> 1:02:58.520 |
|
but is there internal deadlines, artificial or otherwise, |
|
|
|
1:02:58.520 --> 1:03:00.840 |
|
that you're trying to set for yourself, |
|
|
|
1:03:00.840 --> 1:03:03.080 |
|
or is it whenever it's ready? |
|
|
|
1:03:03.080 --> 1:03:05.680 |
|
So we want it to be a great product, right? |
|
|
|
1:03:05.680 --> 1:03:09.880 |
|
And that's a big, important piece for us. |
|
|
|
1:03:09.880 --> 1:03:11.160 |
|
TensorFlow is already out there. |
|
|
|
1:03:11.160 --> 1:03:13.720 |
|
We have 41 million downloads for 1.x, |
|
|
|
1:03:13.720 --> 1:03:15.880 |
|
so it's not like we have to have this. |
|
|
|
1:03:15.880 --> 1:03:17.280 |
|
Yeah, exactly. |
|
|
|
1:03:17.280 --> 1:03:19.320 |
|
So it's not like a lot of the features |
|
|
|
1:03:19.320 --> 1:03:22.080 |
|
that we've really polishing and putting them together |
|
|
|
1:03:22.080 --> 1:03:26.240 |
|
are there, we don't have to rush that just because. |
|
|
|
1:03:26.240 --> 1:03:28.040 |
|
So in that sense, we want to get it right |
|
|
|
1:03:28.040 --> 1:03:29.920 |
|
and really focus on that. |
|
|
|
1:03:29.920 --> 1:03:31.520 |
|
That said, we have said that we are |
|
|
|
1:03:31.520 --> 1:03:33.520 |
|
looking to get this out in the next few months, |
|
|
|
1:03:33.520 --> 1:03:37.120 |
|
in the next quarter, and as far as possible, |
|
|
|
1:03:37.120 --> 1:03:40.000 |
|
we'll definitely try to make that happen. |
|
|
|
1:03:40.000 --> 1:03:44.360 |
|
Yeah, my favorite line was, spring is a relative concept. |
|
|
|
1:03:44.360 --> 1:03:45.960 |
|
I love it. |
|
|
|
1:03:45.960 --> 1:03:47.680 |
|
Spoken like a true developer. |
|
|
|
1:03:47.680 --> 1:03:50.200 |
|
So something I'm really interested in, |
|
|
|
1:03:50.200 --> 1:03:53.840 |
|
and your previous line of work is, before TensorFlow, |
|
|
|
1:03:53.840 --> 1:03:57.720 |
|
you let a team and Google on search ads. |
|
|
|
1:03:57.720 --> 1:04:02.840 |
|
I think this is a very interesting topic on every level, |
|
|
|
1:04:02.840 --> 1:04:07.200 |
|
on a technical level, because if their best ads connect people |
|
|
|
1:04:07.200 --> 1:04:10.080 |
|
to the things they want and need, |
|
|
|
1:04:10.080 --> 1:04:12.280 |
|
and that they're worse, they're just these things |
|
|
|
1:04:12.280 --> 1:04:15.840 |
|
that annoy the heck out of you to the point of ruining |
|
|
|
1:04:15.840 --> 1:04:20.240 |
|
the entire user experience of whatever you're actually doing. |
|
|
|
1:04:20.240 --> 1:04:23.600 |
|
So they have a bad rep, I guess. |
|
|
|
1:04:23.600 --> 1:04:28.080 |
|
And on the other end, so that this connecting users |
|
|
|
1:04:28.080 --> 1:04:32.120 |
|
to the thing they need to want is a beautiful opportunity |
|
|
|
1:04:32.120 --> 1:04:35.360 |
|
for machine learning to shine, like huge amounts of data |
|
|
|
1:04:35.360 --> 1:04:36.720 |
|
that's personalized, and you've got |
|
|
|
1:04:36.720 --> 1:04:40.400 |
|
to map to the thing they actually won't get annoyed. |
|
|
|
1:04:40.400 --> 1:04:43.760 |
|
So what have you learned from this Google that's |
|
|
|
1:04:43.760 --> 1:04:45.160 |
|
leading the world in this aspect? |
|
|
|
1:04:45.160 --> 1:04:47.560 |
|
What have you learned from that experience? |
|
|
|
1:04:47.560 --> 1:04:51.520 |
|
And what do you think is the future of ads? |
|
|
|
1:04:51.520 --> 1:04:54.040 |
|
Take you back to the end of that. |
|
|
|
1:04:54.040 --> 1:04:59.720 |
|
Yes, it's been a while, but I totally agree with what you said. |
|
|
|
1:04:59.720 --> 1:05:03.200 |
|
I think the search ads, the way it was always looked at, |
|
|
|
1:05:03.200 --> 1:05:05.520 |
|
and I believe it still is, is it's |
|
|
|
1:05:05.520 --> 1:05:08.240 |
|
an extension of what search is trying to do. |
|
|
|
1:05:08.240 --> 1:05:10.560 |
|
The goal is to make the information |
|
|
|
1:05:10.560 --> 1:05:14.680 |
|
and make the world's information accessible. |
|
|
|
1:05:14.680 --> 1:05:17.120 |
|
With ads, it's not just information, |
|
|
|
1:05:17.120 --> 1:05:19.120 |
|
but it may be products or other things |
|
|
|
1:05:19.120 --> 1:05:20.800 |
|
that people care about. |
|
|
|
1:05:20.800 --> 1:05:23.360 |
|
And so it's really important for them |
|
|
|
1:05:23.360 --> 1:05:26.480 |
|
to align with what the users need. |
|
|
|
1:05:26.480 --> 1:05:30.920 |
|
And in search ads, there's a minimum quality level |
|
|
|
1:05:30.920 --> 1:05:32.320 |
|
before that ad would be shown. |
|
|
|
1:05:32.320 --> 1:05:34.200 |
|
If we don't have an ad that hits that quality bar, |
|
|
|
1:05:34.200 --> 1:05:35.960 |
|
it will not be shown, even if we have it. |
|
|
|
1:05:35.960 --> 1:05:38.080 |
|
And OK, maybe we lose some money there. |
|
|
|
1:05:38.080 --> 1:05:39.560 |
|
That's fine. |
|
|
|
1:05:39.560 --> 1:05:41.200 |
|
That is really, really important, |
|
|
|
1:05:41.200 --> 1:05:43.000 |
|
and I think that that is something I really |
|
|
|
1:05:43.000 --> 1:05:45.040 |
|
liked about being there. |
|
|
|
1:05:45.040 --> 1:05:48.120 |
|
Advertising is a key part. |
|
|
|
1:05:48.120 --> 1:05:51.680 |
|
I mean, as a model, it's been around for ages, right? |
|
|
|
1:05:51.680 --> 1:05:52.920 |
|
It's not a new model. |
|
|
|
1:05:52.920 --> 1:05:57.440 |
|
It's been adapted to the web and became a core part of search |
|
|
|
1:05:57.440 --> 1:06:02.120 |
|
and in many other search engines across the world. |
|
|
|
1:06:02.120 --> 1:06:05.920 |
|
I do hope, like I said, there are aspects of ads |
|
|
|
1:06:05.920 --> 1:06:06.680 |
|
that are annoying. |
|
|
|
1:06:06.680 --> 1:06:09.600 |
|
And I go to a website, and if it just |
|
|
|
1:06:09.600 --> 1:06:12.160 |
|
keeps popping an ad in my face, not to let me read, |
|
|
|
1:06:12.160 --> 1:06:13.800 |
|
that's going to be annoying clearly. |
|
|
|
1:06:13.800 --> 1:06:22.080 |
|
So I hope we can strike that balance between showing a good |
|
|
|
1:06:22.080 --> 1:06:25.040 |
|
ad where it's valuable to the user |
|
|
|
1:06:25.040 --> 1:06:30.960 |
|
and provides the monetization to the service. |
|
|
|
1:06:30.960 --> 1:06:32.000 |
|
And this might be search. |
|
|
|
1:06:32.000 --> 1:06:33.680 |
|
This might be a website. |
|
|
|
1:06:33.680 --> 1:06:37.320 |
|
All of these, they do need the monetization for them |
|
|
|
1:06:37.320 --> 1:06:39.640 |
|
to provide that service. |
|
|
|
1:06:39.640 --> 1:06:45.720 |
|
But if it's done in a good balance between showing |
|
|
|
1:06:45.720 --> 1:06:48.040 |
|
just some random stuff that's distracting |
|
|
|
1:06:48.040 --> 1:06:50.920 |
|
versus showing something that's actually valuable. |
|
|
|
1:06:50.920 --> 1:06:55.360 |
|
So do you see it moving forward as to continue |
|
|
|
1:06:55.360 --> 1:07:00.960 |
|
being a model that funds businesses like Google? |
|
|
|
1:07:00.960 --> 1:07:05.160 |
|
That's a significant revenue stream. |
|
|
|
1:07:05.160 --> 1:07:08.080 |
|
Because that's one of the most exciting things, |
|
|
|
1:07:08.080 --> 1:07:09.680 |
|
but also limiting things on the internet |
|
|
|
1:07:09.680 --> 1:07:12.200 |
|
is nobody wants to pay for anything. |
|
|
|
1:07:12.200 --> 1:07:15.360 |
|
And advertisements, again, coupled at their best |
|
|
|
1:07:15.360 --> 1:07:17.360 |
|
are actually really useful and not annoying. |
|
|
|
1:07:17.360 --> 1:07:22.320 |
|
Do you see that continuing and growing and improving? |
|
|
|
1:07:22.320 --> 1:07:26.680 |
|
Or is there GC sort of more Netflix type models |
|
|
|
1:07:26.680 --> 1:07:28.960 |
|
where you have to start to pay for content? |
|
|
|
1:07:28.960 --> 1:07:31.000 |
|
I think it's a mix. |
|
|
|
1:07:31.000 --> 1:07:32.840 |
|
I think it's going to take a long while for everything |
|
|
|
1:07:32.840 --> 1:07:35.320 |
|
to be paid on the internet, if at all. |
|
|
|
1:07:35.320 --> 1:07:36.160 |
|
Probably not. |
|
|
|
1:07:36.160 --> 1:07:37.400 |
|
I mean, I think there's always going |
|
|
|
1:07:37.400 --> 1:07:40.760 |
|
to be things that are sort of monetized with things like ads. |
|
|
|
1:07:40.760 --> 1:07:42.800 |
|
But over the last few years, I would say |
|
|
|
1:07:42.800 --> 1:07:44.760 |
|
we've definitely seen that transition |
|
|
|
1:07:44.760 --> 1:07:48.560 |
|
towards more paid services across the web |
|
|
|
1:07:48.560 --> 1:07:50.360 |
|
and people are willing to pay for them |
|
|
|
1:07:50.360 --> 1:07:51.760 |
|
because they do see the value. |
|
|
|
1:07:51.760 --> 1:07:53.600 |
|
I mean, Netflix is a great example. |
|
|
|
1:07:53.600 --> 1:07:56.520 |
|
I mean, we have YouTube doing things. |
|
|
|
1:07:56.520 --> 1:07:59.720 |
|
People pay for the apps they buy, more people |
|
|
|
1:07:59.720 --> 1:08:03.120 |
|
they find are willing to pay for newspaper content, |
|
|
|
1:08:03.120 --> 1:08:07.240 |
|
for the good news websites across the web. |
|
|
|
1:08:07.240 --> 1:08:11.040 |
|
That wasn't the case even a few years ago, I would say. |
|
|
|
1:08:11.040 --> 1:08:13.280 |
|
And I just see that change in myself as well |
|
|
|
1:08:13.280 --> 1:08:14.840 |
|
and just lots of people around me. |
|
|
|
1:08:14.840 --> 1:08:19.240 |
|
So definitely hopeful that we'll transition to that mix model |
|
|
|
1:08:19.240 --> 1:08:23.400 |
|
where maybe you get to try something out for free, |
|
|
|
1:08:23.400 --> 1:08:24.120 |
|
maybe with ads. |
|
|
|
1:08:24.120 --> 1:08:27.080 |
|
But then there is a more clear revenue model |
|
|
|
1:08:27.080 --> 1:08:30.600 |
|
that sort of helps go beyond that. |
|
|
|
1:08:30.600 --> 1:08:34.760 |
|
So speaking of revenue, how is it |
|
|
|
1:08:34.760 --> 1:08:39.400 |
|
that a person can use the TPU in a Google Colab for free? |
|
|
|
1:08:39.400 --> 1:08:43.920 |
|
So what's the, I guess, the question is, |
|
|
|
1:08:43.920 --> 1:08:48.880 |
|
what's the future of TensorFlow in terms of empowering, |
|
|
|
1:08:48.880 --> 1:08:51.880 |
|
say, a class of 300 students? |
|
|
|
1:08:51.880 --> 1:08:55.920 |
|
And I'm asked by MIT, what is going |
|
|
|
1:08:55.920 --> 1:08:58.640 |
|
to be the future of them being able to do their homework |
|
|
|
1:08:58.640 --> 1:09:00.200 |
|
in TensorFlow? |
|
|
|
1:09:00.200 --> 1:09:02.800 |
|
Where are they going to train these networks, right? |
|
|
|
1:09:02.800 --> 1:09:07.720 |
|
What's that future look like with TPUs, with cloud services, |
|
|
|
1:09:07.720 --> 1:09:08.920 |
|
and so on? |
|
|
|
1:09:08.920 --> 1:09:10.240 |
|
I think a number of things there. |
|
|
|
1:09:10.240 --> 1:09:12.600 |
|
I mean, any TensorFlow open source, |
|
|
|
1:09:12.600 --> 1:09:13.640 |
|
you can run it wherever. |
|
|
|
1:09:13.640 --> 1:09:15.880 |
|
You can run it on your desktop, and your desktops |
|
|
|
1:09:15.880 --> 1:09:19.480 |
|
always keep getting more powerful, so maybe you can do more. |
|
|
|
1:09:19.480 --> 1:09:22.040 |
|
My phone is like, I don't know how many times more powerful |
|
|
|
1:09:22.040 --> 1:09:23.520 |
|
than my first desktop. |
|
|
|
1:09:23.520 --> 1:09:25.200 |
|
You'll probably train it on your phone, though. |
|
|
|
1:09:25.200 --> 1:09:26.200 |
|
Yeah, that's true. |
|
|
|
1:09:26.200 --> 1:09:28.080 |
|
Right, so in that sense, the power |
|
|
|
1:09:28.080 --> 1:09:31.440 |
|
you have in your hand is a lot more. |
|
|
|
1:09:31.440 --> 1:09:34.400 |
|
Clouds are actually very interesting from, say, |
|
|
|
1:09:34.400 --> 1:09:37.840 |
|
students or courses perspective, because they |
|
|
|
1:09:37.840 --> 1:09:40.040 |
|
make it very easy to get started. |
|
|
|
1:09:40.040 --> 1:09:42.040 |
|
I mean, Colab, the great thing about it |
|
|
|
1:09:42.040 --> 1:09:45.120 |
|
is go to a website, and it just works. |
|
|
|
1:09:45.120 --> 1:09:47.560 |
|
No installation needed, nothing to, you know, |
|
|
|
1:09:47.560 --> 1:09:49.960 |
|
you're just there, and things are working. |
|
|
|
1:09:49.960 --> 1:09:52.280 |
|
That's really the power of cloud, as well. |
|
|
|
1:09:52.280 --> 1:09:55.320 |
|
And so I do expect that to grow. |
|
|
|
1:09:55.320 --> 1:09:57.920 |
|
Again, Colab is a free service. |
|
|
|
1:09:57.920 --> 1:10:00.840 |
|
It's great to get started, to play with things, |
|
|
|
1:10:00.840 --> 1:10:03.080 |
|
to explore things. |
|
|
|
1:10:03.080 --> 1:10:08.200 |
|
That said, with free, you can only get so much, maybe. |
|
|
|
1:10:08.200 --> 1:10:11.080 |
|
So just like we were talking about free versus paid, |
|
|
|
1:10:11.080 --> 1:10:15.280 |
|
and there are services you can pay for and get a lot more. |
|
|
|
1:10:15.280 --> 1:10:16.000 |
|
Great. |
|
|
|
1:10:16.000 --> 1:10:18.480 |
|
So if I'm a complete beginner interested in machine |
|
|
|
1:10:18.480 --> 1:10:21.560 |
|
learning and TensorFlow, what should I do? |
|
|
|
1:10:21.560 --> 1:10:24.240 |
|
Probably start with going to a website and playing there. |
|
|
|
1:10:24.240 --> 1:10:26.560 |
|
Just go to TensorFlow.org and start clicking on things. |
|
|
|
1:10:26.560 --> 1:10:28.440 |
|
Yep, check out tutorials and guides. |
|
|
|
1:10:28.440 --> 1:10:30.680 |
|
There's stuff you can just click there and go to Colab |
|
|
|
1:10:30.680 --> 1:10:31.320 |
|
and do things. |
|
|
|
1:10:31.320 --> 1:10:32.360 |
|
No installation needed. |
|
|
|
1:10:32.360 --> 1:10:34.040 |
|
You can get started right there. |
|
|
|
1:10:34.040 --> 1:10:34.840 |
|
OK, awesome. |
|
|
|
1:10:34.840 --> 1:10:36.720 |
|
Roger, thank you so much for talking today. |
|
|
|
1:10:36.720 --> 1:10:37.440 |
|
Thank you, Lex. |
|
|
|
1:10:37.440 --> 1:10:46.680 |
|
Have fun this week. |
|
|
|
|