lexicap / vtt /episode_023_small.vtt
Shubham Gupta
Add readme and files
a3be5d0
raw
history blame
83.7 kB
WEBVTT
00:00.000 --> 00:04.720
The following is a conversation with Gavin Miller, he's the head of Adobe Research.
00:04.720 --> 00:08.960
Adobe has empowered artists, designers, and creative minds from all professions,
00:08.960 --> 00:14.320
working in the digital medium for over 30 years with software such as Photoshop, Illustrator,
00:14.320 --> 00:20.560
Premiere, After Effects, InDesign, Audition, Software that work with images, video, and audio.
00:21.200 --> 00:25.920
Adobe Research is working to define the future evolution of these products in a way
00:25.920 --> 00:31.360
that makes the life of creatives easier, automates the tedious tasks, and gives more and more time
00:31.360 --> 00:36.880
to operate in the idea space instead of pixel space. This is where the cutting edge, deep
00:36.880 --> 00:41.360
learning methods of the past decade can really shine more than perhaps any other application.
00:42.240 --> 00:47.840
Gavin is the embodiment of combining tech and creativity. Outside of Adobe Research,
00:47.840 --> 00:53.600
he writes poetry and builds robots, both things that are near and dear to my heart as well.
00:53.600 --> 00:59.200
This conversation is part of the Artificial Intelligence Podcast. If you enjoy it, subscribe
00:59.200 --> 01:05.360
on YouTube, iTunes, or simply connect with me on Twitter at Lex Friedman's spelled F R I D.
01:06.000 --> 01:09.600
And now here's my conversation with Gavin Miller.
01:11.120 --> 01:15.920
You're head of Adobe Research, leading a lot of innovative efforts and applications of AI,
01:15.920 --> 01:23.200
creating images, video, audio, language, but you're also yourself an artist, a poet,
01:23.200 --> 01:28.640
a writer, and even a roboticist. So while I promise to everyone listening,
01:28.640 --> 01:32.880
that I will not spend the entire time we have together reading your poetry, which I love.
01:33.440 --> 01:39.200
I have to sprinkle it in at least a little bit. So some of them are pretty deep and profound,
01:39.200 --> 01:43.520
and some are light and silly. Let's start with a few lines from the silly variety.
01:43.520 --> 01:56.800
You write in a beautiful parody, both Edith Piaf's and my web at Frank Sinatra.
01:56.800 --> 02:06.400
So it opens with, and now dessert is near. It's time to pay the final total. I've tried to slim
02:06.400 --> 02:14.800
all year, but my diets have been anecdotal. So where does that love for poetry come from
02:14.800 --> 02:20.880
for you? And if we dissect your mind, how does it all fit together in the bigger puzzle of Dr.
02:20.880 --> 02:27.440
Gavin Miller? Well, interesting you chose that one. That was a poem I wrote when I'd been to
02:27.440 --> 02:32.400
my doctor and he said, you really need to lose some weight and go on a diet. And whilst the
02:32.400 --> 02:37.200
rational part of my brain wanted to do that, the irrational part of my brain was protesting and
02:37.200 --> 02:42.400
sort of embraced the opposite idea. I regret nothing, hence. Yes, exactly. Taken to an extreme,
02:42.400 --> 02:49.600
I thought it would be funny. Obviously, it's a serious topic for some people. But I think,
02:49.600 --> 02:53.920
for me, I've always been interested in writing since I was in high school, as well as doing
02:53.920 --> 02:58.960
technology and invention. And sometimes the parallel strands in your life that carry on,
02:58.960 --> 03:05.120
and one is more about your private life and one's more about your technological career.
03:05.680 --> 03:10.640
And then at sort of happy moments along the way, sometimes the two things touch, one idea informs
03:10.640 --> 03:17.040
the other. And we can talk about that as we go. Do you think you're writing the art, the poetry
03:17.040 --> 03:23.440
contribute indirectly or directly to your research, to your work in Adobe? Well, sometimes it does if
03:23.440 --> 03:30.000
I say, imagine a future in a science fiction kind of way. And then once it exists on paper,
03:30.000 --> 03:37.520
I think, well, why shouldn't I just build that? There was an example where when realistic voice
03:37.520 --> 03:42.800
synthesis first started in the 90s at Apple, where I worked in research. I was done by a friend of mine.
03:44.000 --> 03:48.640
I sort of sat down and started writing a poem which each line I would enter into the voice
03:48.640 --> 03:54.240
synthesizer and see how it sounded and sort of wrote it for that voice. And at the time,
03:55.040 --> 04:00.160
the agents weren't very sophisticated. So they'd sort of add random intonation. And I kind of made
04:00.160 --> 04:06.560
up the poem to sort of match the tone of the voice. And it sounded slightly sad and depressed. So I
04:06.560 --> 04:12.720
pretended it was a poem written by an intelligent agent, sort of telling the user to go home and
04:12.720 --> 04:16.560
leave them alone. But at the same time, they were lonely and wanted to have company and learn from
04:16.560 --> 04:21.760
what the user was saying. And at the time, it was way beyond anything that AI could possibly do.
04:21.760 --> 04:27.520
But, you know, since then, it's becoming more within the bounds of possibility.
04:29.040 --> 04:34.800
And then at the same time, I had a project at home where I did sort of a smart home. This was
04:34.800 --> 04:40.960
probably 93, 94. And I had the talking voice who'd remind me when I walked in the door of what
04:40.960 --> 04:45.600
things I had to do. I had buttons on my washing machine because I was a bachelor and I'd leave
04:45.600 --> 04:49.920
the clothes in there for three days and they'd go moldy. So as I got up in the morning, I would say,
04:49.920 --> 04:56.640
don't forget the washing and so on. I made photographic photo albums that used light
04:56.640 --> 05:01.040
sensors to know which page you were looking at would send that over wireless radio to the agent
05:01.040 --> 05:05.760
who would then play sounds that matched the image she were looking at in the book. So I was kind of
05:05.760 --> 05:10.480
in love with this idea of magical realism and whether it was possible to do that with technology.
05:10.480 --> 05:16.080
So that was a case where the sort of the agent sort of intrigued me from a literary point of
05:16.080 --> 05:22.880
view and became a personality. I think more recently, I've also written plays and when
05:22.880 --> 05:27.440
plays you write dialogue and obviously you write a fixed set of dialogue that follows a linear
05:27.440 --> 05:33.360
narrative. But with modern agents, as you design a personality or a capability for conversation,
05:33.360 --> 05:37.760
you're sort of thinking of, I kind of have imaginary dialogue in my head. And then I think,
05:37.760 --> 05:43.680
what would it take not only to have that be real, but for it to really know what it's talking about.
05:44.240 --> 05:49.440
So it's easy to fall into the uncanny valley with AI where it says something it doesn't really
05:49.440 --> 05:54.560
understand, but it sounds good to the person. But you rapidly realize that it's kind of just
05:55.520 --> 06:00.000
stimulus response. It doesn't really have real world knowledge about the thing it's describing.
06:00.640 --> 06:06.320
And so when you get to that point, it really needs to have multiple ways of talking about
06:06.320 --> 06:10.560
the same concept. So it sounds as though it really understands it. Now, what really understanding
06:10.560 --> 06:16.160
means is in the eye of the beholder, right? But if it only has one way of referring to something,
06:16.160 --> 06:21.200
it feels like it's a canned response. But if it can reason about it, or you can go at it from
06:21.200 --> 06:25.600
multiple angles and give a similar kind of response that people would, then it starts to
06:26.400 --> 06:30.480
seem more like there's something there that's sentient.
06:31.040 --> 06:35.600
You can say the same thing, multiple things from different perspectives. I mean, with the
06:35.600 --> 06:40.000
automatic image captioning that I've seen the work that you're doing, there's elements of that,
06:40.000 --> 06:46.000
right? Being able to generate different kinds of... Right. So one in my team, there's a lot of work on
06:46.640 --> 06:52.000
turning a medium from one form to another, whether it's auto tagging imagery or making up full
06:52.000 --> 06:57.840
sentences about what's in the image, then changing the sentence, finding another image that matches
06:57.840 --> 07:04.720
the new sentence or vice versa. And in the modern world of GANs, you sort of give it a description
07:04.720 --> 07:11.360
and it synthesizes an asset that matches the description. So I've sort of gone on a journey.
07:11.360 --> 07:16.560
My early days in my career were about 3D computer graphics, the sort of pioneering work sort of
07:16.560 --> 07:22.400
before movies had special effects done with 3D graphics and sort of rode that revolution. And
07:22.400 --> 07:26.720
that was very much like the renaissance where people would model light and color and shape
07:26.720 --> 07:32.160
and everything. And now we're kind of in another wave where it's more impressionistic and it's
07:32.160 --> 07:38.240
sort of the idea of something can be used to generate an image directly, which is sort of the
07:38.240 --> 07:45.520
new frontier in computer image generation using AI algorithms. So the creative process is more in
07:45.520 --> 07:49.840
the space of ideas or becoming more in the space of ideas versus in the raw pixels?
07:50.720 --> 07:55.280
Well, it's interesting. It depends. I think at Adobe, we really want to span the entire range
07:55.280 --> 08:01.040
from really, really good, what you might call low level tools by low level, as close to say analog
08:01.040 --> 08:07.040
workflows as possible. So what we do there is we make up systems that do really realistic oil
08:07.040 --> 08:11.600
paint and watercolor simulation. So if you want every bristle to behave as it would in the real
08:11.600 --> 08:17.760
world and leave a beautiful analog trail of water and then flow after you've made the brushstroke,
08:17.760 --> 08:21.600
you can do that. And that's really important for people who want to create something
08:22.720 --> 08:28.160
really expressive or really novel because they have complete control. And then a certain other
08:28.160 --> 08:35.600
task become automated. It frees the artists up to focus on the inspiration and less of the perspiration.
08:35.600 --> 08:43.920
So thinking about different ideas, obviously, once you finish the design, there's a lot of work to
08:43.920 --> 08:49.840
say do it for all the different aspect ratio of phones or websites and so on. And that used to
08:49.840 --> 08:55.040
take up an awful lot of time for artists. It still does for many, what we call content velocity.
08:55.040 --> 09:01.360
And one of the targets of AI is actually to reason about from the first example of what are the
09:01.360 --> 09:06.960
likely intent for these other formats, maybe if you change the language to German and the words
09:06.960 --> 09:12.160
are longer, how do you reflow everything so that it looks nicely artistic in that way.
09:12.160 --> 09:17.360
And so the person can focus on the really creative bit in the middle, which is what is the look and
09:17.360 --> 09:21.200
style and feel and what's the message and what's the story and the human element.
09:21.200 --> 09:27.920
So I think creativity is changing. So that's one way in which we're trying to just make it easier
09:27.920 --> 09:32.880
and faster and cheaper to do so that there can be more of it, more demand, because it's less
09:32.880 --> 09:39.200
expensive. So everyone wants beautiful artwork for everything from a school website to Hollywood movie.
09:40.800 --> 09:46.480
On the other side, as some of these things have automatic versions of them, people will
09:46.480 --> 09:52.480
possibly change role from being the hands on artist and to being either the art director or
09:52.480 --> 09:57.920
the conceptual artist. And then the computer will be a partner to help create polished examples of
09:57.920 --> 10:02.720
the idea that they're exploring. Let's talk about Adobe products versus AI and Adobe products.
10:04.000 --> 10:11.200
Just so you know where I'm coming from, I'm a huge fan of Photoshop for images premiere for video,
10:11.200 --> 10:17.200
audition for audio. I'll probably use Photoshop to create the thumbnail for this video, premiere
10:17.200 --> 10:24.800
to edit the video, audition to do the audio. That said, everything I do is really manually. And I
10:24.800 --> 10:30.640
set up, I use this old school kinesis keyboard and I have auto hotkey that just it's really about
10:30.640 --> 10:37.280
optimizing the flow of just making sure there's as few clicks as possible. So just being extremely
10:37.280 --> 10:43.920
efficient. It's something you started to speak to. So before we get into the fun, sort of awesome
10:43.920 --> 10:48.880
deep learning things, where does AI, if you could speak a little more to it AI or just
10:48.880 --> 10:57.280
automation in general, do you see in the coming months and years or in general prior in 2018
10:58.560 --> 11:03.840
fitting into making the life, the low level pixel work flow easier?
11:03.840 --> 11:10.160
Yeah, that's a great question. So we have a very rich array of algorithms already in Photoshop,
11:10.160 --> 11:17.600
just classical procedural algorithms as well as ones based on data. In some cases, they end up
11:17.600 --> 11:22.560
with a large number of sliders and degrees of freedom. So one way in which AI can help is just
11:22.560 --> 11:27.520
an auto button which comes up with default settings based on the content itself rather than
11:27.520 --> 11:34.080
default values for the tool. At that point, you then start tweaking. So that's that's a very kind of
11:34.080 --> 11:39.600
make life easier for people whilst making use of common sense from other example images.
11:39.600 --> 11:40.960
So like smart defaults.
11:40.960 --> 11:47.760
Smart defaults, absolutely. Another one is something we've spent a lot of work over the last
11:47.760 --> 11:53.360
20 years. I've been at Adobe 19 thinking about selection, for instance, where
11:53.360 --> 11:58.720
you know, with a quick select, you would look at color boundaries and figure out how to sort of
11:58.720 --> 12:02.480
flood fill into regions that you thought were physically connected in the real world.
12:03.360 --> 12:08.080
But that algorithm had no visual common sense about what a cat looks like or a dog. It would just do
12:08.080 --> 12:14.080
it based on rules of thumb, which were applied to graph theory. And it was a big improvement over
12:14.080 --> 12:19.600
the previous work we had sort of almost click every everything by hand or if it just did similar
12:19.600 --> 12:24.880
colors, it would do little tiny regions that wouldn't be connected. But in the future,
12:24.880 --> 12:31.120
using neural nets to actually do a great job with say a single click or even in the case of
12:31.120 --> 12:36.160
well known categories like people or animals, no click, where you just say select the object and
12:36.160 --> 12:41.440
it just knows the dominant object as a person in the middle of the photograph. Those kinds of things
12:41.440 --> 12:47.520
are really valuable if they can be robust enough to give you good quality results.
12:47.520 --> 12:50.720
Or they can be a great start for like tweaking it.
12:50.720 --> 12:56.080
So for example, background removal, like one thing I'll, in a thumbnail,
12:56.080 --> 13:00.240
I'll take a picture of you right now and essentially remove the background behind you.
13:00.240 --> 13:06.480
And I want to make that as easy as possible. You don't have flowing hair, like rich at the
13:06.480 --> 13:11.040
moment. Rich sort of. I had it in the past, it may come again in the future, but for now.
13:12.320 --> 13:16.160
So that sometimes makes it a little more challenging to remove the background.
13:16.160 --> 13:23.680
How difficult do you think is that problem for AI for basically making the quick selection tool
13:23.680 --> 13:26.960
smarter and smarter and smarter? Well, we have a lot of research on that already.
13:28.400 --> 13:34.160
If you want a sort of quick, cheap and cheerful, look, I'm pretending I'm in Hawaii,
13:34.160 --> 13:38.400
but it's sort of a joke, then you don't need perfect boundaries. And you can do that today
13:38.400 --> 13:44.960
with a single click for the algorithms we have. We have other algorithms where with a little bit
13:44.960 --> 13:48.560
more guidance on the boundaries, like you might need to touch it up a little bit.
13:49.920 --> 13:56.480
We have other algorithms that can pull a nice mat from a crude selection. So we have combinations
13:56.480 --> 14:04.000
of tools that can do all of that. And at our recent Max conference at AB Max, we demonstrated how
14:04.880 --> 14:09.680
very quickly just by drawing a simple polygon around the object of interest, we could not
14:09.680 --> 14:17.920
only do it for a single still, but we could pull at least a selection mask from a moving target,
14:17.920 --> 14:23.360
like a person dancing in front of a brick wall or something. And so it's going from hours to
14:23.360 --> 14:29.360
a few seconds for workflows that are really nice. And then you might go in and touch up a little.
14:30.240 --> 14:33.040
So that's a really interesting question. You mentioned the word robust.
14:33.040 --> 14:40.000
You know, there's like a journey for an idea, right? And what you presented probably at Max
14:41.360 --> 14:46.320
has elements of just sort of it inspires the concept, it can work pretty well in a majority
14:46.320 --> 14:51.520
of cases. But how do you make something that works? Well, in majority of cases, how do you make
14:51.520 --> 14:56.480
something that works, maybe in all cases, or it becomes a robust tool?
14:56.480 --> 15:01.760
There are a couple of things. So that really touches on the difference between academic research
15:01.760 --> 15:06.640
and industrial research. So in academic research, it's really about who's the person to have the
15:06.640 --> 15:12.480
great new idea that shows promise. And we certainly love to be those people too. But
15:13.120 --> 15:17.840
we have sort of two forms of publishing. One is academic peer review, which we do a lot of,
15:17.840 --> 15:24.720
and we have great success there as much as some universities. But then we also have shipping,
15:24.720 --> 15:30.160
which is a different type of, and then we get customer review, as well as, you know, product
15:30.160 --> 15:37.840
critics. And that might be a case where it's not about being perfect every single time, but
15:37.840 --> 15:43.280
perfect enough at the time, plus a mechanism to intervene and recover where you do have mistakes.
15:43.280 --> 15:47.120
So we have the luxury of very talented customers. We don't want them to be
15:48.720 --> 15:55.200
overly taxed doing it every time. But if they can go in and just take it from 99 to 100,
15:55.200 --> 16:01.440
100 with the touch of a mouse or something, then for the professional end, that's something
16:01.440 --> 16:06.320
that we definitely want to support as well. And for them, it went from having to do that
16:06.320 --> 16:13.360
tedious task all the time to much less often. So I think that gives us an out. If it had to be
16:13.360 --> 16:18.640
100% automatic all the time, then that would delay the time at which we could get to market.
16:18.640 --> 16:26.000
So on that thread, maybe you can untangle something. Again, I'm sort of just speaking to
16:26.000 --> 16:36.000
my own experience. Maybe that is the most useful idea. So I think Photoshop, as an example or premiere,
16:37.680 --> 16:44.240
has a lot of amazing features that I haven't touched. And so what's the, in terms of AI,
16:44.240 --> 16:54.080
helping make my life or the life of creatives easier? How this collaboration between human
16:54.080 --> 16:58.960
and machine, how do you learn to collaborate better? How do you learn the new algorithms?
17:00.000 --> 17:04.240
Is it something that where you have to watch tutorials and you have to watch videos and so
17:04.240 --> 17:10.400
on? Or do you ever think, do you think about the experience itself through exploration being
17:10.400 --> 17:18.320
the teacher? We absolutely do. So I'm glad that you brought this up. We sort of think about
17:18.320 --> 17:22.080
two things. One is helping the person in the moment to do the task that they need to do. But
17:22.080 --> 17:26.960
the other is thinking more holistically about their journey learning a tool. And when it's like,
17:26.960 --> 17:31.120
think of it as Adobe University, where you use the tool long enough, you become an expert.
17:31.120 --> 17:34.960
And not necessarily an expert in everything. It's like living in a city. You don't necessarily
17:34.960 --> 17:40.160
know every street, but you know, the important ones you need to get to. So we have projects in
17:40.160 --> 17:45.360
research, which actually look at the thousands of hours of tutorials online and try to understand
17:45.360 --> 17:51.040
what's being taught in them. And then we had one publication at CHI where it was looking at,
17:52.560 --> 17:57.360
given the last three or four actions you did, what did other people in tutorials do next?
17:57.360 --> 18:01.680
So if you want some inspiration for what you might do next, or you just want to watch the
18:01.680 --> 18:06.800
tutorial and see, learn from people who are doing similar workflows to you, you can without having
18:06.800 --> 18:13.360
to go and search on keywords and everything. So really trying to use the context of your use of
18:13.360 --> 18:18.640
the app to make intelligent suggestions, either about choices that you might make,
18:20.800 --> 18:26.480
or in a more assistive way where it could say, if you did this next, we could show you. And that's
18:26.480 --> 18:31.360
basically the frontier that we're exploring now, which is, if we really deeply understand the
18:31.360 --> 18:38.480
domain in which designers and creative people work, can we combine that with AI and pattern
18:38.480 --> 18:47.520
matching of behavior to make intelligent suggestions, either through verbal possibilities or just
18:47.520 --> 18:53.840
showing the results of if you try this. And that's really the sort of, I was in a meeting today
18:53.840 --> 18:58.880
thinking about these things. So it's still a grand challenge. We'd all love
18:58.880 --> 19:05.920
an artist over one shoulder and a teacher over the other, right? And we hope to get there. And
19:05.920 --> 19:10.640
the right thing to do is to give enough at each stage that it's useful in itself, but it builds
19:10.640 --> 19:19.120
a foundation for the next level of expectation. Are you aware of this gigantic medium of YouTube
19:19.120 --> 19:26.240
that's creating just a bunch of creative people, both artists and teachers of different kinds?
19:26.240 --> 19:31.440
Absolutely. And the more we can understand those media types, both visually and in terms of
19:31.440 --> 19:37.200
transcripts and words, the more we can bring the wisdom that they embody into the guidance that's
19:37.200 --> 19:42.960
embedded in the tool. That would be brilliant to remove the barrier from having to yourself type
19:42.960 --> 19:49.600
in the keyword, searching, so on. Absolutely. And then in the longer term, an interesting
19:49.600 --> 19:54.000
discussion is, does it ultimately not just assist with learning the interface we have,
19:54.000 --> 19:59.200
but does it modify the interface to be simpler? Or do you fragment into a variety of tools,
19:59.200 --> 20:05.600
each of which has a different level of visibility of the functionality? I like to say that if you
20:05.600 --> 20:12.640
add a feature to a GUI, you have to have yet more visual complexity confronting the new user.
20:12.640 --> 20:17.120
Whereas if you have an assistant with a new skill, if you know they have it, so you know
20:17.120 --> 20:23.280
to ask for it, then it's sort of additive without being more intimidating. So we definitely think
20:23.280 --> 20:29.120
about new users and how to onboard them. Many actually value the idea of being able to master
20:29.120 --> 20:34.720
that complex interface and keyboard shortcuts, like you were talking about earlier, because
20:35.520 --> 20:39.680
with great familiarity, it becomes a musical instrument for expressing your visual ideas.
20:40.480 --> 20:45.840
And other people just want to get something done quickly in the simplest way possible,
20:45.840 --> 20:50.400
and that's where a more assistive version of the same technology might be useful,
20:50.400 --> 20:54.560
maybe on a different class of device, which is more in context for capture, say,
20:55.920 --> 21:01.680
whereas somebody who's in a deep post production workflow maybe want to be on a laptop or a big
21:01.680 --> 21:10.560
screen desktop and have more knobs and dials to really express the subtlety of what they want to do.
21:12.160 --> 21:16.320
So there's so many exciting applications of computer vision and machine learning
21:16.320 --> 21:21.920
that Adobe is working on, like scene stitching, sky replacement, foreground,
21:21.920 --> 21:26.880
background removal, spatial object based image search, automatic image captioning,
21:26.880 --> 21:31.280
like we mentioned, project cloak, project deep fill filling in parts of the images,
21:31.920 --> 21:38.640
project scribbler, style transfer video, style transfer faces and video with Project Puppetron,
21:38.640 --> 21:49.040
best name ever. Can you talk through a favorite or some of them or examples that pop in mind?
21:49.040 --> 21:54.800
I'm sure I'll be able to provide links to other ones we don't talk about because there's visual
21:54.800 --> 22:00.640
elements to all of them that are exciting. Why they're interesting for different reasons might
22:00.640 --> 22:06.640
be a good way to go. So I think sky replace is interesting because we talked about selection
22:06.640 --> 22:11.440
being sort of an atomic operation. It's almost like a, if you think of an assembly language,
22:11.440 --> 22:17.840
it's like a single instruction. Whereas sky replace is a compound action where you automatically
22:17.840 --> 22:22.720
select the sky, you look for stock content that matches the geometry of the scene.
22:24.000 --> 22:27.600
You try to have variety in your choices so that you do coverage of different moods.
22:28.160 --> 22:35.600
It then mats in the sky behind the foreground, but then importantly it uses the foreground
22:35.600 --> 22:40.560
of the other image that you just searched on to recolor the foreground of the image that
22:40.560 --> 22:47.840
you're editing. So if you say go from a midday sky to an evening sky, it will actually add
22:47.840 --> 22:53.760
sort of an orange glow to the foreground objects as well. I was a big fan in college of Magritte
22:53.760 --> 22:59.600
and he has a number of paintings where it's surrealism because he'll like do a composite,
22:59.600 --> 23:03.440
but the foreground building will be at night and the sky will be during the day. There's one
23:03.440 --> 23:09.120
called The Empire of Light which was on my wall in college and we're trying not to do surrealism.
23:09.120 --> 23:15.680
It can be a choice, but we'd rather have it be natural by default rather than it looking fake
23:15.680 --> 23:19.760
and then you have to do a whole bunch of post production to fix it. So that's a case where
23:19.760 --> 23:25.120
we're kind of capturing an entire workflow into a single action and doing it in about a second
23:25.120 --> 23:30.720
rather than a minute or two. And when you do that, you can not just do it once, but you can do it
23:30.720 --> 23:36.960
for say like 10 different backgrounds and then you're almost back to this inspiration idea of
23:36.960 --> 23:41.840
I don't know quite what I want, but I'll know it when I see it. And you can just explore the
23:41.840 --> 23:47.200
design space as close to final production value as possible. And then when you really pick one,
23:47.200 --> 23:51.120
you might go back and slightly tweak the selection mask just to make it perfect and
23:51.120 --> 23:54.320
do that kind of polish that professionals like to bring to their work.
23:54.320 --> 24:01.040
So then there's this idea of, as you mentioned, the sky replacing it to different stock images of
24:01.040 --> 24:07.040
the sky. In general, you have this idea or it could be on your disk or whatever. But making even
24:07.040 --> 24:12.400
more intelligent choices about ways to search stock images which is really interesting. It's
24:12.400 --> 24:19.120
kind of spatial. Absolutely. Right. So that was something we called Concept Canvas. So normally
24:19.120 --> 24:26.240
when you do say an image search, I assume it's just based on text. You would give the keywords
24:26.240 --> 24:30.080
of the things you want to be in the image and it would find the nearest one that had those tags.
24:32.720 --> 24:37.200
For many tasks, you really want to be able to say I want a big person in the middle or in a
24:37.200 --> 24:41.280
dog to the right and umbrella above the left because you want to leave space for the text or
24:41.280 --> 24:48.640
whatever. And so Concept Canvas lets you assign spatial regions to the keywords. And then we've
24:48.640 --> 24:54.560
already preindexed the images to know where the important concepts are in the picture. So we then
24:54.560 --> 25:01.200
go through that index matching to assets. And even though it's just another form of search,
25:01.200 --> 25:06.480
because you're doing spatial design or layout, it starts to feel like design. You sort of feel
25:06.480 --> 25:13.120
oddly responsible for the image that comes back as if you invented it a little bit. So it's a good
25:13.120 --> 25:18.960
example where giving enough control starts to make people have a sense of ownership over the
25:18.960 --> 25:23.280
outcome of the event. And then we also have technologies in Photoshop where you physically
25:23.280 --> 25:29.440
can move the dog in post as well. But for Concept Canvas, it was just a very fast way to sort of
25:29.440 --> 25:38.560
loop through and be able to lay things out. In terms of being able to remove objects from a scene
25:38.560 --> 25:45.920
and fill in the background automatically. So that's extremely exciting. And that's
25:45.920 --> 25:51.200
so neural networks are stepping in there. I just talked this week with Ian Goodfellow.
25:51.200 --> 25:55.360
So the GANS for doing that is definitely one approach.
25:55.360 --> 25:59.440
So is that a really difficult problem? Is it as difficult as it looks,
25:59.440 --> 26:06.080
again, to take it to a robust product level? Well, there are certain classes of image for
26:06.080 --> 26:10.800
which the traditional algorithms like Content Aware Fill work really well. Like if you have
26:10.800 --> 26:15.200
a naturalistic texture like a gravel path or something, because it's patch based, it will
26:15.200 --> 26:19.840
make up a very plausible looking intermediate thing and fill in the hole. And then we use some
26:20.960 --> 26:25.200
algorithms to sort of smooth out the lighting so you don't see any brightness contrasts in that
26:25.200 --> 26:29.680
region or you've gradually ramped from dark to light if it straddles the boundary.
26:29.680 --> 26:37.600
Where it gets complicated is if you have to infer invisible structure behind the person in front.
26:37.600 --> 26:41.920
And that really requires a common sense knowledge of the world to know what,
26:42.480 --> 26:47.040
if I see three quarters of a house, do I have a rough sense of what the rest of the house looks
26:47.040 --> 26:51.840
like? If you just fill it in with patches, it can end up sort of doing things that make sense
26:51.840 --> 26:56.480
locally. But you look at the global structure and it looks like it's just sort of crumpled or messed
26:56.480 --> 27:02.800
up. And so what GANs and neural nets bring to the table is this common sense learned from the
27:02.800 --> 27:10.640
training set. And the challenge right now is that the generative methods that can make up
27:10.640 --> 27:14.960
missing holes using that kind of technology are still only stable at low resolutions.
27:15.520 --> 27:19.680
And so you either need to then go from a low resolution to a high resolution using some other
27:19.680 --> 27:24.720
algorithm or we need to push the state of the art and it's still in research to get to that point.
27:24.720 --> 27:30.720
Right. Of course, if you show it something, say it's trained on houses and then you show it in
27:30.720 --> 27:37.360
octopus, it's not going to do a very good job of showing common sense about octopuses. So
27:39.360 --> 27:45.600
again, you're asking about how you know that it's ready for prime time. You really need a very
27:45.600 --> 27:52.880
diverse training set of images. And ultimately, that may be a case where you put it out there
27:52.880 --> 28:00.400
with some guard rails where you might do a detector which looks at the image and
28:00.960 --> 28:05.280
sort of estimates its own competence of how well a job could this algorithm do.
28:05.920 --> 28:10.400
So eventually, there may be this idea of what we call an ensemble of experts where
28:11.120 --> 28:15.440
any particular expert is specialized in certain things and then there's sort of a
28:15.440 --> 28:19.360
either they vote to say how confident they are about what to do. This is sort of more future
28:19.360 --> 28:24.080
looking or there's some dispatcher which says you're good at houses, you're good at trees.
28:27.120 --> 28:31.520
So I mean, all this adds up to a lot of work because each of those models will be a whole
28:31.520 --> 28:38.320
bunch of work. But I think over time, you'd gradually fill out the set and initially focus
28:38.320 --> 28:41.520
on certain workflows and then sort of branch out as you get more capable.
28:41.520 --> 28:48.640
So you mentioned workflows and have you considered maybe looking far into the future?
28:50.000 --> 28:57.680
First of all, using the fact that there is a huge amount of people that use Photoshop,
28:57.680 --> 29:03.520
for example, they have certain workflows, being able to collect the information by which
29:04.880 --> 29:09.440
they basically get information about their workflows, about what they need,
29:09.440 --> 29:15.120
what can the ways to help them, whether it is houses or octopus that people work on more.
29:16.000 --> 29:23.440
Basically getting a beat on what kind of data is needed to be annotated and collected for people
29:23.440 --> 29:26.320
to build tools that actually work well for people.
29:26.320 --> 29:31.680
Absolutely. And this is a big topic and the whole world of AI is what data can you gather and why.
29:33.200 --> 29:39.120
At one level, the way to think about it is we not only want to train our customers in how to use
29:39.120 --> 29:44.160
our products, but we want them to teach us what's important and what's useful. At the same time,
29:44.160 --> 29:51.120
we want to respect their privacy and obviously we wouldn't do things without their explicit permission.
29:52.800 --> 29:57.440
And I think the modern spirit of the age around this is you have to demonstrate to somebody
29:57.440 --> 30:02.720
how they're benefiting from sharing their data with the tool. Either it's helping in the short
30:02.720 --> 30:07.120
term to understand their intent so you can make better recommendations or if they're
30:07.120 --> 30:11.840
there friendly to your cause or you're tall or they want to help you evolve quickly because
30:11.840 --> 30:16.320
they depend on you for their livelihood, they may be willing to share some of their
30:17.360 --> 30:23.360
workflows or choices with the dataset to be then trained.
30:24.720 --> 30:30.560
There are technologies for looking at learning without necessarily storing all the information
30:30.560 --> 30:36.080
permanently so that you can learn on the fly but not keep a record of what somebody did.
30:36.080 --> 30:38.720
So, we're definitely exploring all of those possibilities.
30:38.720 --> 30:45.520
And I think Adobe exists in a space where Photoshop, if I look at the data I've created
30:45.520 --> 30:51.840
in OWN, I'm less comfortable sharing data with social networks than I am with Adobe because
30:51.840 --> 31:01.360
there's just exactly as you said, there's an obvious benefit for sharing the data that I use
31:01.360 --> 31:05.440
to create in Photoshop because it's helping improve the workflow in the future.
31:05.440 --> 31:06.080
Right.
31:06.080 --> 31:09.440
As opposed to it's not clear what the benefit is in social networks.
31:10.000 --> 31:14.000
It's nice of you to say that. I mean, I think there are some professional workflows where
31:14.000 --> 31:17.360
people might be very protective of what they're doing such as if I was preparing
31:18.240 --> 31:22.640
evidence for a legal case, I wouldn't want any of that, you know,
31:22.640 --> 31:25.440
phoning home to help train the algorithm or anything.
31:26.560 --> 31:30.720
There may be other cases where people say having a trial version or they're doing some,
31:30.720 --> 31:35.280
I'm not saying we're doing this today, but there's a future scenario where somebody has a more
31:35.280 --> 31:40.400
permissive relationship with Adobe where they explicitly say, I'm fine, I'm only doing hobby
31:40.400 --> 31:48.000
projects or things which are non confidential and in exchange for some benefit tangible or
31:48.000 --> 31:51.200
otherwise, I'm willing to share very fine grain data.
31:51.840 --> 31:57.920
So, another possible scenario is to capture relatively crude high level things from more
31:57.920 --> 32:02.160
people and then more detailed knowledge from people who are willing to participate.
32:02.160 --> 32:07.280
We do that today with explicit customer studies where, you know, we go and visit somebody and
32:07.280 --> 32:10.640
ask them to try the tool and we human observe what they're doing.
32:12.000 --> 32:15.760
In the future, to be able to do that enough to be able to train an algorithm,
32:16.320 --> 32:20.240
we'd need a more systematic process, but we'd have to do it very consciously because
32:21.200 --> 32:24.560
one of the things people treasure about Adobe is a sense of trust
32:24.560 --> 32:28.880
and we don't want to endanger that through overly aggressive data collection.
32:28.880 --> 32:35.520
So, we have a Chief Privacy Officer and it's definitely front and center of thinking about AI
32:35.520 --> 32:39.920
rather than an afterthought. Well, when you start that program, sign me up.
32:39.920 --> 32:41.040
Okay, happy to.
32:42.800 --> 32:48.640
Is there other projects that you wanted to mention that I didn't perhaps that pop into mind?
32:48.640 --> 32:51.840
Well, you covered the number. I think you mentioned Project Puppetron.
32:51.840 --> 32:58.480
I think that one is interesting because you might think of Adobe as only thinking in 2D
32:59.760 --> 33:04.800
and that's a good example where we're actually thinking more three dimensionally about how to
33:04.800 --> 33:09.440
assign features to faces so that we can, you know, if you take, so what Puppetron does,
33:09.440 --> 33:16.320
it takes either a still or a video of a person talking and then it can take a painting of somebody
33:16.320 --> 33:20.160
else and then apply the style of the painting to the person who's talking in the video.
33:20.160 --> 33:29.600
And it's unlike a sort of screen door post filter effect that you sometimes see online.
33:30.320 --> 33:36.080
It really looks as though it's sort of somehow attached or reflecting the motion of the face.
33:36.080 --> 33:40.320
And so that's the case where even to do a 2D workflow like stylization,
33:40.320 --> 33:44.160
you really need to infer more about the 3D structure of the world.
33:44.160 --> 33:48.560
And I think as 3D computer vision algorithms get better,
33:48.560 --> 33:52.960
initially they'll focus on particular domains like faces where you have a lot of
33:52.960 --> 33:57.680
prior knowledge about structure and you can maybe have a parameterized template that you fit to the
33:57.680 --> 34:03.600
image. But over time, this should be possible for more general content. And it might even be
34:03.600 --> 34:09.360
invisible to the user that you're doing 3D reconstruction but under the hood, but it might
34:09.360 --> 34:15.200
then let you do edits much more reliably or correctly than you would otherwise.
34:15.200 --> 34:20.800
And you know, the face is a very important application, right?
34:20.800 --> 34:22.480
So making things work.
34:22.480 --> 34:26.640
And a very sensitive one. If you do something uncanny, it's very disturbing.
34:26.640 --> 34:30.080
That's right. You have to get it. You have to get it right.
34:30.080 --> 34:39.040
So in the space of augmented reality and virtual reality, what do you think is the role of AR and
34:39.040 --> 34:45.360
VR and in the content we consume as people as consumers and the content we create as creators?
34:45.360 --> 34:47.920
No, that's a great question. Let me think about this a lot too.
34:48.720 --> 34:55.360
So I think VR and AR serve slightly different purposes. So VR can really transport you to an
34:55.360 --> 35:02.400
entire immersive world, no matter what your personal situation is. To that extent, it's a bit like
35:02.400 --> 35:06.720
a really, really widescreen television where it sort of snaps you out of your context and puts you
35:06.720 --> 35:13.360
in a new one. And I think it's still evolving in terms of the hardware I actually worked on,
35:13.360 --> 35:17.680
VR in the 90s, trying to solve the latency and sort of nausea problem, which we did,
35:17.680 --> 35:23.040
but it was very expensive and a bit early. There's a new wave of that now. I think
35:23.600 --> 35:27.600
and increasingly those devices are becoming all in one rather than something that's tethered to a
35:27.600 --> 35:34.480
box. I think the market seems to be bifurcating into things for consumers and things for professional
35:34.480 --> 35:40.080
use cases, like for architects and people designing where your product is a building and you really
35:40.080 --> 35:45.200
want to experience it better than looking at a scale model or a drawing, I think,
35:45.920 --> 35:50.320
or even than a video. So I think for that, where you need a sense of scale and spatial
35:50.320 --> 35:59.600
relationships, it's great. I think AR holds the promise of sort of taking digital assets off the
35:59.600 --> 36:03.600
screen and putting them in context in the real world on the table in front of you on the wall
36:03.600 --> 36:10.400
behind you. And that has the corresponding need that the assets need to adapt to the physical
36:10.400 --> 36:15.680
context in which they're being placed. I mean, it's a bit like having a live theater troupe come
36:15.680 --> 36:20.880
to your house and put on Hamlet. My mother had a friend who used to do this at Stately Homes in
36:20.880 --> 36:26.640
England for the National Trust. And they would adapt the scenes and even they'd walk the audience
36:26.640 --> 36:32.880
through the rooms to see the action based on the country house they found themselves in for two
36:32.880 --> 36:38.720
days. And I think AR will have the same issue that if you have a tiny table in a big living room
36:38.720 --> 36:44.960
or something, it'll try to figure out what can you change and what's fixed. And there's a little
36:44.960 --> 36:52.560
bit of a tension between fidelity, where if you captured Senior Eye of doing a fantastic ballet,
36:52.560 --> 36:56.640
you'd want it to be sort of exactly reproduced. And maybe all you could do is scale it down.
36:56.640 --> 37:03.760
Whereas somebody telling you a story might be walking around the room doing some gestures
37:03.760 --> 37:07.040
and that could adapt to the room in which they were telling the story.
37:07.760 --> 37:12.800
And do you think fidelity is that important in that space or is it more about the storytelling?
37:12.800 --> 37:17.840
I think it may depend on the characteristic of the media. If it's a famous celebrity,
37:17.840 --> 37:22.480
then it may be that you want to catch every nuance and they don't want to be reanimated by some
37:22.480 --> 37:30.800
algorithm. It could be that if it's really a loveable frog telling you a story and it's
37:30.800 --> 37:34.320
about a princess and a frog, then it doesn't matter if the frog moves in a different way.
37:35.520 --> 37:38.640
I think a lot of the ideas that have sort of grown up in the game world will
37:39.520 --> 37:45.120
now come into the broader commercial sphere once they're needing adaptive characters in AR.
37:45.120 --> 37:51.920
Are you thinking of engineering tools that allow creators to create in the augmented world,
37:52.480 --> 37:56.000
basically making a Photoshop for the augmented world?
37:57.360 --> 38:02.560
Well, we have shown a few demos of sort of taking a Photoshop layer stack and then expanding it into
38:02.560 --> 38:08.640
3D. That's actually been shown publicly as one example in AR. Where we're particularly excited
38:08.640 --> 38:17.120
at the moment is in 3D. 3D design is still a very challenging space. We believe that it's
38:17.120 --> 38:22.800
a worthwhile experiment to try to figure out if AR or immersive makes 3D design more spontaneous.
38:23.360 --> 38:26.960
Can you give me an example of 3D design just like applications?
38:26.960 --> 38:32.080
Well, literally, a simple one would be laying out objects, right? On a conventional screen,
38:32.080 --> 38:35.680
you'd sort of have a plan view and a side view and a perspective view and you sort of be dragging
38:35.680 --> 38:39.440
it around with the mouse and if you're not careful, it would go through the wall and all that.
38:39.440 --> 38:46.400
Whereas if you were really laying out objects, say in a VR headset, you could literally move
38:46.400 --> 38:50.720
your head to see a different viewpoint. They'd be in stereo, so you'd have a sense of depth
38:50.720 --> 38:57.040
because you're already wearing the depth glasses, right? So it would be those sort of big gross
38:57.040 --> 39:01.120
motor, move things around, kind of skills seem much more spontaneous just like they are in the
39:01.120 --> 39:08.320
real world. The frontier for us, I think, is whether that same medium can be used to do fine
39:08.320 --> 39:14.880
grain design tasks, like very accurate constraints on, say, a CAD model or something. That may be
39:14.880 --> 39:19.120
better done on a desktop, but it may just be a matter of inventing the right UI.
39:20.160 --> 39:27.600
So we're hopeful that because there will be this potential explosion of demand for 3D assets
39:27.600 --> 39:32.560
that's driven by AR and more real time animation on conventional screens,
39:34.880 --> 39:40.800
those tools will also help with, or those devices will help with designing the content as well.
39:40.800 --> 39:46.480
You've mentioned quite a few interesting sort of new ideas. At the same time, there's old
39:46.480 --> 39:51.280
timers like me that are stuck in their old ways. I think I'm the old timer.
39:51.280 --> 39:58.560
Okay. All right. But the opposed all change at all costs. When you're thinking about
39:58.560 --> 40:05.440
creating new interfaces, do you feel the burden of just this giant user base that loves the
40:05.440 --> 40:13.760
current product? So anything new you do that any new idea comes at a cost that you'll be resisted?
40:13.760 --> 40:21.200
Well, I think if you have to trade off control for convenience, then our existing user base
40:21.200 --> 40:27.680
would definitely be offended by that. I think if there are some things where you have more convenience
40:27.680 --> 40:34.160
and just as much control, that may be more welcome. We do think about not breaking well known
40:34.160 --> 40:40.640
metaphors for things. So things should sort of make sense. Photoshop has never been a static
40:40.640 --> 40:46.640
target. It's always been evolving and growing. And to some extent, there's been a lot of brilliant
40:46.640 --> 40:50.640
thought along the way of how it works today. So we don't want to just throw all that out.
40:51.840 --> 40:55.600
If there's a fundamental breakthrough, like a single click is good enough to select an object
40:55.600 --> 41:01.760
rather than having to do lots of strokes, that actually fits in quite nicely to the existing
41:01.760 --> 41:07.840
tool set, either as an optional mode or as a starting point. I think where we're looking at
41:07.840 --> 41:13.440
radical simplicity where you could encapsulate an entire workflow with a much simpler UI,
41:14.000 --> 41:18.640
then sometimes that's easier to do in the context of either a different device like a
41:18.640 --> 41:25.120
mobile device where the affordances are naturally different or in a tool that's targeted at a
41:25.120 --> 41:31.040
different workflow where it's about spontaneity and velocity rather than precision. And we have
41:31.040 --> 41:36.720
projects like Rush, which can let you do professional quality video editing for a certain class of
41:36.720 --> 41:47.440
media output that is targeted very differently in terms of users and the experience. And ideally,
41:47.440 --> 41:54.160
people would go, if I'm feeling like doing Premiere, big project, I'm doing a four part
41:54.160 --> 41:59.200
television series. That's definitely a premier thing. But if I want to do something to show my
41:59.200 --> 42:05.200
recent vacation, maybe I'll just use Rush because I can do it in the half an hour. I have free at
42:05.200 --> 42:12.480
home rather than the four hours I need to do it at work. And for the use cases which we can do well,
42:12.480 --> 42:16.960
it really is much faster to get the same output. But the more professional tools obviously have
42:16.960 --> 42:21.520
a much richer toolkit and more flexibility in what they can do.
42:21.520 --> 42:26.400
And then at the same time, with the flexibility and control, I like this idea of smart defaults,
42:27.040 --> 42:33.520
of using AI to coach you to like what Google has, I'm feeling lucky button.
42:33.520 --> 42:37.360
Right. Or one button kind of gives you a pretty good
42:38.160 --> 42:41.440
set of settings. And then you almost, that's almost an educational tool.
42:42.000 --> 42:43.360
Absolutely. Yeah.
42:43.360 --> 42:49.920
To show, because sometimes when you have all this control, you're not sure about the
42:51.040 --> 42:55.600
correlation between the different bars that control different elements of the image and so on.
42:55.600 --> 43:00.400
And sometimes there's a degree of, you don't know what the optimal is.
43:00.400 --> 43:05.360
And then some things are sort of on demand like help, right?
43:05.360 --> 43:06.160
Help, yeah.
43:06.160 --> 43:09.600
I'm stuck. I need to know what to look for. I'm not quite sure what it's called.
43:10.400 --> 43:13.920
And something that was proactively making helpful suggestions or,
43:14.800 --> 43:20.640
you know, you could imagine a make a suggestion button where you'd use all of that knowledge
43:20.640 --> 43:25.120
of workflows and everything to maybe suggest something to go and learn about or just to try
43:25.120 --> 43:31.280
or show the answer. And maybe it's not one intelligent to pothole, but it's like a variety
43:31.280 --> 43:33.760
of defaults. And then you go, oh, I like that one.
43:33.760 --> 43:34.960
Yeah. Yeah.
43:34.960 --> 43:35.680
Several options.
43:37.200 --> 43:39.520
So back to poetry.
43:39.520 --> 43:40.640
Ah, yes.
43:40.640 --> 43:42.400
We're going to interleave.
43:43.440 --> 43:48.000
So first few lines of a recent poem of yours before I ask the next question.
43:48.000 --> 43:55.840
Yeah. This is about the smartphone. Today left my phone at home and went down to the sea.
43:57.120 --> 44:01.600
The sand was soft, the ocean glass, but I was still just me.
44:02.480 --> 44:08.240
So this is a poem about you leaving your phone behind and feeling quite liberated because of it.
44:08.960 --> 44:14.800
So this is kind of a difficult topic and let's see if we can talk about it, figure it out.
44:14.800 --> 44:21.120
But so with the help of AI, more and more, we can create versions of ourselves, versions of
44:21.120 --> 44:31.920
reality that are in some ways more beautiful than actual reality. And some of the creative effort
44:31.920 --> 44:38.880
there is part of creating this illusion. So of course, this is inevitable, but how do you think
44:38.880 --> 44:44.320
we should adjust this human beings to live in this digital world that's partly artificial,
44:44.320 --> 44:51.520
that's better than the world that we lived in a hundred years ago when you didn't have
44:51.520 --> 44:55.840
Instagram and Facebook versions of ourselves and the online.
44:55.840 --> 44:58.880
Oh, this is sort of showing off better versions of ourselves.
44:58.880 --> 45:04.880
We're using the tooling of modifying the images or even with artificial intelligence
45:04.880 --> 45:12.880
ideas of deep fakes and creating adjusted or fake versions of ourselves and reality.
45:14.080 --> 45:17.600
I think it's an interesting question. You're all sort of historical bent on this.
45:19.360 --> 45:24.720
I actually wonder if 18th century aristocrats who commissioned famous painters to paint portraits
45:24.720 --> 45:28.960
of them had portraits that were slightly nicer than they actually looked in practice.
45:28.960 --> 45:29.680
Well played, sir.
45:29.680 --> 45:34.720
So human desire to put your best foot forward has always been true.
45:37.440 --> 45:42.240
I think it's interesting. You sort of framed it in two ways. One is if we can imagine alternate
45:42.240 --> 45:47.280
realities and visualize them, is that a good or bad thing? In the old days, you do it with
45:47.280 --> 45:52.560
storytelling and words and poetry, which still resides sometimes on websites. But
45:52.560 --> 46:00.640
we've become a very visual culture in particular. In the 19th century, we were very much a text
46:00.640 --> 46:07.280
based culture. People would read long tracks. Political speeches were very long. Nowadays,
46:07.280 --> 46:15.280
everything's very kind of quick and visual and snappy. I think it depends on how harmless your
46:15.280 --> 46:23.120
intent. A lot of it's about intent. So if you have a somewhat flattering photo that you pick
46:23.120 --> 46:30.960
out of the photos that you have in your inbox to say, this is what I look like, it's probably fine.
46:32.160 --> 46:37.040
If someone's going to judge you by how you look, then they'll decide soon enough when they meet
46:37.040 --> 46:45.120
you whether the reality. I think where it can be harmful is if people hold themselves up to an
46:45.120 --> 46:49.520
impossible standard, which they then feel bad about themselves for not meeting. I think that's
46:49.520 --> 46:58.400
definitely can be an issue. But I think the ability to imagine and visualize an alternate
46:58.400 --> 47:04.800
reality, which sometimes which you then go off and build later, can be a wonderful thing too.
47:04.800 --> 47:10.720
People can imagine architectural styles, which they then have a startup, make a fortune and then
47:10.720 --> 47:17.680
build a house that looks like their favorite video game. Is that a terrible thing? I think
47:18.720 --> 47:24.560
I used to worry about exploration actually, that part of the joy of going to the moon
47:24.560 --> 47:30.320
when I was a tiny child, I remember it, and grainy black and white, was to know what it would look
47:30.320 --> 47:35.120
like when you got there. And I think now we have such good graphics for knowing, for visualizing
47:35.120 --> 47:40.960
experience before it happens, that I slightly worry that it may take the edge off actually wanting
47:40.960 --> 47:46.160
to go. Because we've seen it on TV, we kind of, oh, by the time we finally get to Mars,
47:46.160 --> 47:53.200
we're going, oh yeah, it's Mars, that's what it looks like. But then the outer exploration,
47:53.200 --> 47:58.480
I mean, I think Pluto was a fantastic recent discovery where nobody had any idea what it
47:58.480 --> 48:04.800
looked like and it was just breathtakingly varied and beautiful. So I think expanding
48:04.800 --> 48:10.800
the ability of the human toolkit to imagine and communicate on balance is a good thing.
48:10.800 --> 48:15.920
I think there are abuses, we definitely take them seriously and try to discourage them.
48:17.440 --> 48:22.960
I think there's a parallel side where the public needs to know what's possible through events like
48:22.960 --> 48:30.720
this, right? So that you don't believe everything you read and print anymore, and it may over time
48:30.720 --> 48:35.440
become true of images as well. Or you need multiple sets of evidence to really believe
48:35.440 --> 48:40.240
something rather than a single media asset. So I think it's a constantly evolving thing.
48:40.240 --> 48:45.680
It's been true forever. There's a famous story about Anne of Cleves and Henry VIII where,
48:47.760 --> 48:52.720
luckily for Anne, they didn't get married, right? So, or they got married and
48:53.840 --> 48:54.560
What's the story?
48:54.560 --> 48:59.200
Oh, so Holbein went and painted a picture and then Henry VIII wasn't pleased and, you know,
48:59.200 --> 49:03.520
history doesn't record whether Anne was pleased, but I think she was pleased not
49:03.520 --> 49:07.920
to be married more than a day or something. So I mean, this has gone on for a long time,
49:07.920 --> 49:13.280
but I think it's just part of the magnification of human capability.
49:14.640 --> 49:20.560
You've kind of built up an amazing research environment here, research culture, research
49:20.560 --> 49:25.600
lab, and you've written that the secret to a thriving research lab is interns. Can you unpack
49:25.600 --> 49:29.520
that a little bit? Oh, absolutely. So a couple of reasons.
49:31.360 --> 49:36.080
As you see, looking at my personal history, there are certain ideas you bond with at a certain
49:36.080 --> 49:40.880
stage of your career and you tend to keep revisiting them through time. If you're lucky,
49:40.880 --> 49:44.880
you pick one that doesn't just get solved in the next five years, and then you're sort of out of
49:44.880 --> 49:49.840
luck. So I think a constant influx of new people brings new ideas with it.
49:49.840 --> 49:56.800
From the point of view of industrial research, because a big part of what we do is really taking
49:56.800 --> 50:03.360
those ideas to the point where they can ship us very robust features, you end up investing a lot
50:03.360 --> 50:08.640
in a particular idea. And if you're not careful, people can get too conservative in what they
50:08.640 --> 50:15.280
choose to do next, knowing that the product teams will want it. And interns let you explore the more
50:15.280 --> 50:22.160
fanciful or unproven ideas in a relatively lightweight way, ideally leading to new publications for
50:22.160 --> 50:28.080
the intern and for the researcher. And it gives you then a portfolio from which to draw which idea
50:28.080 --> 50:32.720
am I going to then try to take all the way through to being robust in the next year or two to ship.
50:34.000 --> 50:37.520
So it sort of becomes part of the funnel. It's also a great way for us to
50:38.160 --> 50:42.640
identify future full time researchers, many of our greatest researchers were former interns.
50:42.640 --> 50:48.800
It builds a bridge to university departments so we can get to know and build an enduring relationship
50:48.800 --> 50:53.840
with the professors and we often do academic give funds to as well as an acknowledgement of the
50:53.840 --> 51:01.360
value the interns add and their own collaborations. So it's sort of a virtuous cycle. And then the
51:01.360 --> 51:06.960
long term legacy of a great research lab hopefully will be not only the people who stay but the ones
51:06.960 --> 51:11.520
who move through and then go off and carry that same model to other companies.
51:11.520 --> 51:17.440
And so we believe strongly in industrial research and how it can complement academia and
51:17.440 --> 51:21.920
we hope that this model will continue to propagate and be invested in by other companies,
51:21.920 --> 51:26.800
which makes it harder for us to recruit, of course, but you know, that's the sign of success
51:26.800 --> 51:30.320
and a rising tide lifts all ships in that sense.
51:31.040 --> 51:38.080
And where's the idea of born with the interns? Is there brainstorming? Is there discussions
51:38.080 --> 51:46.240
about, you know, like what the ideas come from? Yeah, as I'm asking the question, I
51:46.240 --> 51:51.760
realized how dumb it is, but I'm hoping you have a better answer than a question I ask at the
51:51.760 --> 51:59.280
beginning of every summer. So what will happen is we'll send out a call for interns. They'll
52:00.000 --> 52:04.240
we'll have a number of resumes come in, people will contact the candidates, talk to them about
52:04.240 --> 52:09.920
their interests. They'll usually try to find some somebody who has a reasonably good match to what
52:09.920 --> 52:14.320
they're already doing, or just has a really interesting domain that they've been pursuing in
52:14.320 --> 52:20.640
their PhD. And we think we'd love to do one of those projects too. And then the intern stays in
52:20.640 --> 52:27.360
touch with the mentors, we call them. And then they come and in the first at the end of two weeks,
52:27.360 --> 52:32.000
they have to decide. So they'll often have a general sense by the time they arrive.
52:32.000 --> 52:36.080
And we'll have internal discussions about what are all the general
52:36.800 --> 52:41.120
ideas that we're wanting to pursue to see whether two people have the same idea and maybe they
52:41.120 --> 52:47.040
should talk and all that. But then once the intern actually arrives, sometimes the idea goes linearly
52:47.040 --> 52:51.120
and sometimes it takes a giant left turn and we go, that sounded good. But when we thought about
52:51.120 --> 52:54.880
it, there's this other project or it's already been done and we found this paper that we were
52:54.880 --> 53:02.240
scooped. But we have this other great idea. So it's pretty flexible at the beginning. One of the
53:02.240 --> 53:08.080
questions for research labs is who's deciding what to do, and then who's to blame if it goes
53:08.080 --> 53:15.600
wrong, who gets the credit if it goes right. And so in Adobe, we push the needle very much towards
53:15.600 --> 53:22.960
freedom of choice of projects by the researchers and the interns. But then we reward people based
53:22.960 --> 53:28.000
on impact. So if the projects ultimately end up impacting the products and having papers and so
53:28.000 --> 53:34.400
on. And so your alternative model just to be clear is that you have one lab director who thinks he's
53:34.400 --> 53:38.720
a genius and tells everybody what to do, takes all the credit if it goes well, blames everybody
53:38.720 --> 53:44.800
else if it goes badly. So we don't want that model. And this helps new ideas percolate up.
53:45.440 --> 53:49.840
The art of running such a lab is that there are strategic priorities for the company
53:49.840 --> 53:55.520
and there are areas where we do want to invest in pressing problems. And so it's a little bit of a
53:55.520 --> 54:01.360
trickle down and filter up meets in the middle. And so you don't tell people you have to do X,
54:01.360 --> 54:07.360
but you say X would be particularly appreciated this year. And then people reinterpret X through
54:07.360 --> 54:12.720
the filter of things they want to do and they're interested in. And miraculously, it usually comes
54:12.720 --> 54:18.640
together very well. One thing that really helps is Adobe has a really broad portfolio of products.
54:18.640 --> 54:24.960
So if we have a good idea, there's usually a product team that is intrigued or interested.
54:26.000 --> 54:31.520
So it means we don't have to qualify things too much ahead of time. Once in a while, the product
54:31.520 --> 54:36.880
teams sponsor an extra intern because they have a particular problem that they really care about,
54:36.880 --> 54:41.440
in which case it's a little bit more, we really need one of these. And then we sort of say,
54:41.440 --> 54:45.920
great, I get an extra intern. We find an intern who thinks that's a great problem. But that's not
54:45.920 --> 54:49.520
the typical model. That's sort of the icing on the cake as far as the budget's concerned.
54:51.440 --> 54:55.920
And all of the above end up being important. It's really hard to predict at the beginning of the
54:55.920 --> 55:01.200
summer, which we all have high hopes of all of the intern projects. But ultimately, some of them
55:01.200 --> 55:06.480
pay off and some of them sort of are a nice paper, but don't turn into a feature. Others turn out
55:06.480 --> 55:12.080
not to be as novel as we thought, but they'd be a great feature, but not a paper. And then others,
55:12.080 --> 55:16.400
we make a little bit of progress and we realize how much we don't know. And maybe we revisit that
55:16.400 --> 55:22.320
problem several years in a row until it finally we have a breakthrough. And then it becomes more
55:22.320 --> 55:30.320
on track to impact a product. Jumping back to a big overall view of Adobe Research, what are you
55:30.320 --> 55:37.360
looking forward to in 2019 and beyond? What is, you mentioned there's a giant suite of products,
55:37.360 --> 55:45.920
giant suite of products, giant suite of ideas, new interns, a large team of researchers.
55:46.960 --> 55:54.400
Where do you think the future holds? In terms of the technological breakthroughs?
55:54.400 --> 56:00.960
Technological breakthroughs, especially ones that will make it into product will get to
56:00.960 --> 56:05.920
impact the world. So I think the creative or the analytics assistance that we talked about where
56:05.920 --> 56:10.320
they're constantly trying to figure out what you're trying to do and how can they be helpful and make
56:10.320 --> 56:16.160
useful suggestions is a really hot topic. And it's very unpredictable as to when it'll be ready,
56:16.160 --> 56:20.720
but I'm really looking forward to seeing how much progress we make against that. I think
56:22.480 --> 56:28.640
some of the core technologies like generative adversarial networks are immensely promising
56:28.640 --> 56:34.720
and seeing how quickly those become practical for mainstream use cases at high resolution with
56:34.720 --> 56:39.840
really good quality is also exciting. And they also have this sort of strange way of even the
56:39.840 --> 56:45.120
things they do oddly are odd in an interesting way. So it can look like dreaming or something.
56:46.160 --> 56:55.040
So that's fascinating. I think internally we have a Sensei platform, which is a way in which
56:55.040 --> 57:01.760
we're pooling our neural net and other intelligence models into a central platform, which can then be
57:01.760 --> 57:07.040
leveraged by multiple product teams at once. So we're in the middle of transitioning from a,
57:07.040 --> 57:11.040
you know, once you have a good idea, you pick a product team to work with and you sort of hand
57:11.040 --> 57:17.520
design it for that use case to a more sort of Henry Ford, stand it up in a standard way, which
57:17.520 --> 57:22.880
can be accessed in a standard way, which should mean that the time between a good idea and impacting
57:22.880 --> 57:28.960
our products will be greatly shortened. And when one product has a good idea, many of the other
57:28.960 --> 57:34.240
products can just leverage it too. So it's sort of an economy of scale. So that's more about the
57:34.240 --> 57:39.680
how then the what, but that combination of this sort of renaissance in AI, there's a comparable
57:39.680 --> 57:45.280
one in graphics with real time ray tracing and other really exciting emerging technologies.
57:45.280 --> 57:49.920
And when these all come together, you'll sort of basically be dancing with light, right? Where
57:49.920 --> 57:56.080
you'll have real time shadows, reflections, and as if it's a real world in front of you, but then
57:56.080 --> 58:00.880
with all these magical properties brought by AI where it sort of anticipates or modifies itself
58:00.880 --> 58:05.040
in ways that make sense based on how it understands the creative task you're trying to do.
58:06.320 --> 58:12.320
That's a really exciting future for creative for myself too, the creator. So first of all,
58:12.320 --> 58:17.760
I work in autonomous vehicles. I'm a roboticist. I love robots. And I think you have a fascination
58:17.760 --> 58:23.920
with snakes, both natural and artificial robots. I share your fascination. I mean, their movement
58:23.920 --> 58:31.760
is beautiful, adaptable. The adaptability is fascinating. There are, I looked it up, 2900
58:31.760 --> 58:37.200
species of snakes in the world. Wow. The 175 venomous, some are tiny, some are huge.
58:38.880 --> 58:44.560
Saw that there's one that's 25 feet in some cases. So what's the most interesting thing
58:44.560 --> 58:52.000
that you connect with in terms of snakes, both natural and artificial? Why, what was the connection
58:52.000 --> 58:58.000
with robotics AI in this particular form of a robot? Well, it actually came out of my work
58:58.000 --> 59:02.880
in the 80s on computer animation, where I started doing things like cloth simulation and
59:02.880 --> 59:07.360
other kind of soft body simulation. And you'd sort of drop it and it would bounce,
59:07.360 --> 59:10.880
then it would just sort of stop moving. And I thought, well, what if you animate the spring
59:10.880 --> 59:16.160
lengths and simulate muscles? And the simplest object I could do that for was an earthworm.
59:16.160 --> 59:21.680
So I actually did a paper in 1988 on called the motion dynamics of snakes and worms. And I
59:21.680 --> 59:27.760
read the physiology literature on both Hale snakes and worms move and then did some of the early
59:27.760 --> 59:34.640
computer animation examples of that. So your interest in robotics started with graphics?
59:34.640 --> 59:40.960
Came out of simulation and graphics. When I moved from Alias to Apple, we actually did a
59:40.960 --> 59:46.160
movie called Her Majesty's Secret Serpent, which is about a secret agent snake that parachutes in
59:46.160 --> 59:50.320
and captures a film canister from a satellite, which tells you how old fashioned we were thinking
59:50.320 --> 59:57.760
back then, sort of classic 19 sort of 50s or 60s Bond movie kind of thing. And at the same time,
59:57.760 --> 1:00:03.120
I'd always made radio control ships when I was a child and from scratch. And I thought, well,
1:00:03.120 --> 1:00:08.800
how can it be to build a real one? And so then started what turned out to be like a 15 year
1:00:08.800 --> 1:00:14.320
obsession with trying to build better snake robots. And the first one that I built just sort of
1:00:14.320 --> 1:00:19.520
slithered sideways, but didn't actually go forward, then added wheels and building things in real
1:00:19.520 --> 1:00:25.840
life makes you honest about the friction. The thing that appeals to me is I love creating the
1:00:25.840 --> 1:00:30.800
illusion of life, which is what drove me to drove me to animation. And if you have a robot with
1:00:30.800 --> 1:00:36.320
enough degrees of coordinated freedom that move in a kind of biological way, then it starts to
1:00:36.320 --> 1:00:42.080
cross the Yankani Valley and to see me like a creature rather than a thing. And I certainly got
1:00:42.080 --> 1:00:50.320
that with the early snakes by S3, I had it able to sidewind as well as go directly forward. My
1:00:50.320 --> 1:00:54.240
wife to be suggested that it would be the ring bearer at our wedding. So it actually went down
1:00:54.240 --> 1:00:59.360
the aisle carrying the rings and got in the local paper for that, which was really fun.
1:01:00.160 --> 1:01:07.200
And this was all done as a hobby. And then I at the time that can onboard compute was incredibly
1:01:07.200 --> 1:01:11.840
limited. It was sort of yes, you should explain that these snakes, the whole idea is that you would
1:01:11.840 --> 1:01:18.640
you're trying to run it autonomously. Autonomously, on board right. And so
1:01:19.760 --> 1:01:25.280
the very first one, I actually built the controller from discrete logic, because I used to do LSI,
1:01:25.280 --> 1:01:30.640
you know, circuits and things when I was a teenager. And then the second and third one,
1:01:30.640 --> 1:01:35.120
the 8 bit microprocessors were available with like a whole 256 bytes of RAM,
1:01:36.000 --> 1:01:39.840
which you could just about squeeze in. So they were radio controlled rather than autonomous.
1:01:39.840 --> 1:01:44.480
And really, we're more about the physic physicality and coordinated motion.
1:01:46.560 --> 1:01:51.520
I've occasionally taken a side step into if only I could make it cheaply enough,
1:01:51.520 --> 1:01:59.040
bake a great toy, which has been a lesson in how clockwork is its own magical realm that
1:01:59.040 --> 1:02:03.680
you venture into and learn things about backlash and other things you don't take into account as
1:02:03.680 --> 1:02:07.600
a computer scientist, which is why what seemed like a good idea doesn't work. So it's quite
1:02:07.600 --> 1:02:14.160
humbling. And then more recently, I've been building S9, which is a much better engineered
1:02:14.160 --> 1:02:17.760
version of S3 where the motors wore out and it doesn't work anymore. And you can't buy
1:02:17.760 --> 1:02:24.640
replacements, which is sad given that it was such a meaningful one. S5 was about twice as long and
1:02:25.760 --> 1:02:32.960
look much more biologically inspired. I, unlike the typical roboticist, I taper my snakes.
1:02:33.520 --> 1:02:37.040
There are good mechanical reasons to do that, but it also makes them look more biological,
1:02:37.040 --> 1:02:43.280
although it means every segment's unique rather than a repetition, which is why most engineers
1:02:43.280 --> 1:02:49.840
don't do it. It actually saves weight and leverage and everything. And that one is currently on
1:02:49.840 --> 1:02:54.480
display at the International Spy Museum in Washington, DC. None of it has done any spying.
1:02:56.080 --> 1:03:00.000
It was on YouTube and it got its own conspiracy theory where people thought that it wasn't real
1:03:00.000 --> 1:03:04.160
because they work at Adobe, it must be fake graphics. And people would write to me, tell me
1:03:04.160 --> 1:03:11.200
it's real. They say the background doesn't move and it's like, it's on a tripod. So that one,
1:03:11.200 --> 1:03:16.960
but you can see the real thing. So it really is true. And then the latest one is the first one
1:03:16.960 --> 1:03:21.280
where I could put a Raspberry Pi, which leads to all sorts of terrible jokes about pythons and
1:03:21.280 --> 1:03:29.920
things. But this one can have onboard compute. And then where my hobby work and my work worker
1:03:29.920 --> 1:03:36.400
converging is you can now add vision accelerator chips, which can evaluate neural nets and do
1:03:36.400 --> 1:03:41.600
object recognition and everything. So both for the snakes and more recently for the spider that
1:03:41.600 --> 1:03:48.640
I've been working on, having desktop level compute is now opening up a whole world of
1:03:49.200 --> 1:03:54.880
true autonomy with onboard compute, onboard batteries, and still having that sort of
1:03:54.880 --> 1:04:01.680
biomimetic quality that appeals to children in particular. They are really drawn to them and
1:04:01.680 --> 1:04:08.960
adults think they look creepy, but children actually think they look charming. And I gave a
1:04:08.960 --> 1:04:14.880
series of lectures at Girls Who Code to encourage people to take an interest in technology. And
1:04:14.880 --> 1:04:19.200
at the moment, I'd say they're still more expensive than the value that they add,
1:04:19.200 --> 1:04:22.560
which is why they're a great hobby for me, but they're not really a great product.
1:04:22.560 --> 1:04:30.000
It makes me think about doing that very early thing I did at Alias with changing the muscle
1:04:30.000 --> 1:04:35.760
rest lengths. If I could do that with a real artificial muscle material, then the next snake
1:04:35.760 --> 1:04:40.960
ideally would use that rather than motors and gearboxes and everything. It would be lighter,
1:04:40.960 --> 1:04:49.200
much stronger, and more continuous and smooth. So I like to say being in research as a license
1:04:49.200 --> 1:04:54.640
to be curious, and I have the same feeling with my hobby yet. It forced me to read biology and
1:04:54.640 --> 1:04:59.680
be curious about things that otherwise would have just been natural geographic special.
1:04:59.680 --> 1:05:04.320
Suddenly, I'm thinking, how does that snake move? Can I copy it? I look at the trails that
1:05:04.320 --> 1:05:08.640
side winding snakes leave in sand and see if my snake robots would do the same thing.
1:05:10.240 --> 1:05:13.840
Out of something inanimate, I like why you put a try to bring life into it and beauty.
1:05:13.840 --> 1:05:18.240
Absolutely. And then ultimately, give it a personality, which is where the intelligent
1:05:18.240 --> 1:05:25.040
agent research will converge with the vision and voice synthesis to give it a sense of having
1:05:25.040 --> 1:05:29.840
not necessarily human level intelligence. I think the Turing test is such a high bar, it's
1:05:30.480 --> 1:05:36.080
a little bit self defeating, but having one that you can have a meaningful conversation with,
1:05:36.880 --> 1:05:43.040
especially if you have a reasonably good sense of what you can say. So not trying to have it so
1:05:43.040 --> 1:05:50.080
a stranger could walk up and have one, but so as a pet owner or a robot pet owner, you could
1:05:50.080 --> 1:05:53.200
know what it thinks about and what it can reason about.
1:05:53.200 --> 1:05:58.800
Or sometimes just meaningful interaction. If you have the kind of interaction you have with a dog,
1:05:58.800 --> 1:06:02.080
sometimes you might have a conversation, but it's usually one way.
1:06:02.080 --> 1:06:02.640
Absolutely.
1:06:02.640 --> 1:06:06.000
And nevertheless, it feels like a meaningful connection.
1:06:06.800 --> 1:06:12.720
And one of the things that I'm trying to do in the sample audio that will play you is beginning
1:06:12.720 --> 1:06:17.680
to get towards the point where the reasoning system can explain why it knows something or
1:06:17.680 --> 1:06:22.320
why it thinks something. And that again, creates the sense that it really does know what it's
1:06:22.320 --> 1:06:29.840
talking about, but also for debugging. As you get more and more elaborate behavior, it's like,
1:06:29.840 --> 1:06:37.120
why did you decide to do that? How do you know that? I think the robot's really
1:06:37.120 --> 1:06:42.560
my muse for helping me think about the future of AI and what to invent next.
1:06:42.560 --> 1:06:47.040
So even at Adobe, that's mostly operating in the digital world.
1:06:47.040 --> 1:06:47.440
Correct.
1:06:47.440 --> 1:06:54.480
Do you ever, do you see a future where Adobe even expands into the more physical world perhaps?
1:06:54.480 --> 1:07:02.080
So bringing life not into animations, but bringing life into physical objects with whether it's,
1:07:02.080 --> 1:07:08.640
well, I'd have to say at the moment it's a twinkle in my eye. I think the more likely thing is that
1:07:08.640 --> 1:07:14.240
we will bring virtual objects into the physical world through augmented reality.
1:07:14.240 --> 1:07:21.360
And many of the ideas that might take five years to build a robot to do, you can do in a few weeks
1:07:21.360 --> 1:07:29.120
with digital assets. So I think when really intelligent robots finally become commonplace,
1:07:29.120 --> 1:07:34.000
they won't be that surprising because we'll have been living with those personalities in the virtual
1:07:34.000 --> 1:07:38.640
sphere for a long time. And then they'll just say, oh, it's Siri with legs or Alexa,
1:07:39.520 --> 1:07:47.760
Alexa on hooves or something. So I can see that welcoming. And for now, it's still an adventure
1:07:47.760 --> 1:07:53.760
and we don't know quite what the experience will be like. And it's really exciting to sort of see
1:07:53.760 --> 1:07:59.680
all of these different strands of my career converge. Yeah, in interesting ways. And it is
1:07:59.680 --> 1:08:07.920
definitely a fun adventure. So let me end with my favorite poem, the last few lines of my favorite
1:08:07.920 --> 1:08:14.560
poem of yours that ponders mortality. And in some sense, immortality, as our ideas live through
1:08:14.560 --> 1:08:21.120
the ideas of others through the work of others, it ends with, do not weep or mourn. It was enough
1:08:21.120 --> 1:08:28.000
the little atomies permitted just a single dance. Scatter them as deep as your eyes can see. I'm
1:08:28.000 --> 1:08:34.320
content they'll have another chance, sweeping more centered parts along to join a jostling,
1:08:34.320 --> 1:08:41.440
lifting throng as others danced in me. Beautiful poem. Beautiful way to end it. Gavin, thank you
1:08:41.440 --> 1:08:45.920
so much for talking today. And thank you for inspiring and empowering millions of people
1:08:45.920 --> 1:08:50.960
like myself for creating amazing stuff. Oh, thank you. It's a great conversation.