|
WEBVTT |
|
|
|
00:00.000 --> 00:02.960 |
|
The following is a conversation with Kai Fu Li. |
|
|
|
00:02.960 --> 00:06.520 |
|
He's the chairman and CEO of Sinovation Ventures |
|
|
|
00:06.520 --> 00:10.560 |
|
that manages a $2 billion dual currency investment fund |
|
|
|
00:10.560 --> 00:13.160 |
|
with a focus on developing the next generation |
|
|
|
00:13.160 --> 00:15.440 |
|
of Chinese high tech companies. |
|
|
|
00:15.440 --> 00:17.840 |
|
He's the former president of Google China |
|
|
|
00:17.840 --> 00:20.880 |
|
and the founder of what is now called Microsoft Research |
|
|
|
00:20.880 --> 00:24.160 |
|
Asia, an institute that trained many |
|
|
|
00:24.160 --> 00:26.520 |
|
of the artificial intelligence leaders in China, |
|
|
|
00:26.520 --> 00:32.080 |
|
including CTOs or AI execs at Baidu, Tencent, Alibaba, |
|
|
|
00:32.080 --> 00:34.840 |
|
Lenovo, and Huawei. |
|
|
|
00:34.840 --> 00:38.520 |
|
He was named one of the 100 most influential people |
|
|
|
00:38.520 --> 00:40.680 |
|
in the world by Time Magazine. |
|
|
|
00:40.680 --> 00:43.880 |
|
He's the author of seven bestselling books in Chinese |
|
|
|
00:43.880 --> 00:47.080 |
|
and most recently, the New York Times bestseller called |
|
|
|
00:47.080 --> 00:50.600 |
|
AI Superpowers, China, Silicon Valley, |
|
|
|
00:50.600 --> 00:52.760 |
|
and the New World Order. |
|
|
|
00:52.760 --> 00:57.200 |
|
He has unparalleled experience in working across major tech |
|
|
|
00:57.200 --> 01:00.120 |
|
companies and governments on applications of AI. |
|
|
|
01:00.120 --> 01:02.440 |
|
And so he has a unique perspective |
|
|
|
01:02.440 --> 01:05.080 |
|
on global innovation in the future of AI |
|
|
|
01:05.080 --> 01:09.000 |
|
that I think is important to listen to and think about. |
|
|
|
01:09.000 --> 01:11.960 |
|
This is the Artificial Intelligence Podcast. |
|
|
|
01:11.960 --> 01:15.240 |
|
If you enjoy it, subscribe on YouTube and iTunes, |
|
|
|
01:15.240 --> 01:18.880 |
|
support it on Patreon, or simply connect with me on Twitter |
|
|
|
01:18.880 --> 01:21.040 |
|
at Lex Freedman. |
|
|
|
01:21.040 --> 01:26.120 |
|
And now, here's my conversation with Kaifu Li. |
|
|
|
01:26.120 --> 01:29.440 |
|
I immigrated from Russia to US when I was 13. |
|
|
|
01:29.440 --> 01:32.480 |
|
You immigrated to US at about the same age. |
|
|
|
01:32.480 --> 01:35.920 |
|
The Russian people, the American people, the Chinese people, |
|
|
|
01:35.920 --> 01:39.440 |
|
each have a certain soul, a spirit, |
|
|
|
01:39.440 --> 01:42.080 |
|
that permeates throughout the generations. |
|
|
|
01:42.080 --> 01:45.120 |
|
So maybe it's a little bit of a poetic question, |
|
|
|
01:45.120 --> 01:49.240 |
|
but could you describe your sense of what |
|
|
|
01:49.240 --> 01:52.080 |
|
defines the Chinese soul? |
|
|
|
01:52.080 --> 01:56.160 |
|
I think the Chinese soul of people today, right, |
|
|
|
01:56.160 --> 02:02.000 |
|
we're talking about people who have had centuries of burden |
|
|
|
02:02.000 --> 02:05.240 |
|
because of the poverty that the country has gone through |
|
|
|
02:05.240 --> 02:10.560 |
|
and suddenly shined with hope of prosperity |
|
|
|
02:10.560 --> 02:13.440 |
|
in the past 40 years as China opened up |
|
|
|
02:13.440 --> 02:16.440 |
|
and embraced market economy. |
|
|
|
02:16.440 --> 02:20.200 |
|
And undoubtedly, there are two sets of pressures |
|
|
|
02:20.200 --> 02:24.160 |
|
on the people, that of the tradition, |
|
|
|
02:24.160 --> 02:28.040 |
|
that of facing difficult situations, |
|
|
|
02:28.040 --> 02:31.160 |
|
and that of hope of wanting to be the first |
|
|
|
02:31.160 --> 02:33.840 |
|
to become successful and wealthy, |
|
|
|
02:33.840 --> 02:38.360 |
|
so that it's a very strong hunger and strong desire |
|
|
|
02:38.360 --> 02:41.160 |
|
and strong work ethic that drives China forward. |
|
|
|
02:41.160 --> 02:43.960 |
|
And is there roots to not just this generation, |
|
|
|
02:43.960 --> 02:47.880 |
|
but before, that's deeper than just |
|
|
|
02:47.880 --> 02:50.080 |
|
the new economic developments? |
|
|
|
02:50.080 --> 02:52.520 |
|
Is there something that's unique to China |
|
|
|
02:52.520 --> 02:54.960 |
|
that you could speak to that's in the people? |
|
|
|
02:54.960 --> 02:56.000 |
|
Yeah. |
|
|
|
02:56.000 --> 03:00.280 |
|
Well, the Chinese tradition is about excellence, |
|
|
|
03:00.280 --> 03:02.680 |
|
dedication, and results. |
|
|
|
03:02.680 --> 03:07.240 |
|
And the Chinese exams and study subjects in schools |
|
|
|
03:07.240 --> 03:11.080 |
|
have traditionally started from memorizing 10,000 characters, |
|
|
|
03:11.080 --> 03:13.600 |
|
not an easy task to start with. |
|
|
|
03:13.600 --> 03:17.640 |
|
And further by memorizing historic philosophers, |
|
|
|
03:17.640 --> 03:19.000 |
|
literature, poetry. |
|
|
|
03:19.000 --> 03:22.480 |
|
So it really is probably the strongest road |
|
|
|
03:22.480 --> 03:26.920 |
|
learning mechanism created to make sure people had good memory |
|
|
|
03:26.920 --> 03:30.080 |
|
and remembered things extremely well. |
|
|
|
03:30.080 --> 03:33.720 |
|
That, I think, at the same time suppresses |
|
|
|
03:33.720 --> 03:37.360 |
|
the breakthrough innovation. |
|
|
|
03:37.360 --> 03:42.520 |
|
And also enhances the speed execution get results. |
|
|
|
03:42.520 --> 03:47.400 |
|
And that, I think, characterizes the historic basis of China. |
|
|
|
03:47.400 --> 03:49.160 |
|
That's interesting, because there's echoes of that |
|
|
|
03:49.160 --> 03:52.080 |
|
in Russian education as well as rote memorization. |
|
|
|
03:52.080 --> 03:53.800 |
|
So you memorize a lot of poetry. |
|
|
|
03:53.800 --> 03:59.240 |
|
I mean, there's just an emphasis on perfection in all forms |
|
|
|
03:59.240 --> 04:02.240 |
|
that's not conducive to perhaps what you're speaking to, |
|
|
|
04:02.240 --> 04:03.640 |
|
which is creativity. |
|
|
|
04:03.640 --> 04:05.640 |
|
But you think that kind of education |
|
|
|
04:05.640 --> 04:09.040 |
|
holds back the innovative spirit that you |
|
|
|
04:09.040 --> 04:10.960 |
|
might see in the United States? |
|
|
|
04:10.960 --> 04:14.840 |
|
Well, it holds back the breakthrough innovative spirit |
|
|
|
04:14.840 --> 04:16.480 |
|
that we see in the United States. |
|
|
|
04:16.480 --> 04:21.880 |
|
But it does not hold back the valuable execution oriented, |
|
|
|
04:21.880 --> 04:26.320 |
|
result oriented value creating engines, which we see China |
|
|
|
04:26.320 --> 04:27.960 |
|
being very successful. |
|
|
|
04:27.960 --> 04:32.320 |
|
So is there a difference between a Chinese AI engineer |
|
|
|
04:32.320 --> 04:35.600 |
|
today and an American AI engineer perhaps rooted |
|
|
|
04:35.600 --> 04:38.320 |
|
in the culture that we just talked about or the education |
|
|
|
04:38.320 --> 04:41.160 |
|
or the very soul of the people or no? |
|
|
|
04:41.160 --> 04:43.720 |
|
And what would your advice be to each |
|
|
|
04:43.720 --> 04:45.520 |
|
if there's a difference? |
|
|
|
04:45.520 --> 04:47.120 |
|
Well, there's a lot that's similar, |
|
|
|
04:47.120 --> 04:51.240 |
|
because AI is about mastering sciences, |
|
|
|
04:51.240 --> 04:54.880 |
|
about using known technologies and trying new things. |
|
|
|
04:54.880 --> 04:59.760 |
|
But it's also about picking from many parts of possible networks |
|
|
|
04:59.760 --> 05:02.920 |
|
to use and different types of parameters to tune. |
|
|
|
05:02.920 --> 05:05.280 |
|
And that part is somewhat rote. |
|
|
|
05:05.280 --> 05:09.040 |
|
And it is also, as anyone who's built AI products, |
|
|
|
05:09.040 --> 05:12.680 |
|
can tell you a lot about cleansing the data. |
|
|
|
05:12.680 --> 05:15.200 |
|
Because AI runs better with more data. |
|
|
|
05:15.200 --> 05:20.160 |
|
And data is generally unstructured, errorful, |
|
|
|
05:20.160 --> 05:22.360 |
|
and unclean. |
|
|
|
05:22.360 --> 05:26.280 |
|
And the effort to clean the data is immense. |
|
|
|
05:26.280 --> 05:32.280 |
|
So I think the better part of the American AI engineering |
|
|
|
05:32.280 --> 05:36.840 |
|
process is to try new things, to do things people haven't done |
|
|
|
05:36.840 --> 05:41.840 |
|
before, and to use technology to solve most, if not all, |
|
|
|
05:41.840 --> 05:43.480 |
|
problems. |
|
|
|
05:43.480 --> 05:47.160 |
|
So to make the algorithm work despite not so great data, |
|
|
|
05:47.160 --> 05:50.680 |
|
find error tolerant ways to deal with the data. |
|
|
|
05:50.680 --> 05:55.960 |
|
The Chinese way would be to basically enumerate, |
|
|
|
05:55.960 --> 05:58.560 |
|
to the fullest extent, all the possible ways |
|
|
|
05:58.560 --> 06:01.000 |
|
by a lot of machines, try lots of different ways |
|
|
|
06:01.000 --> 06:05.320 |
|
to get it to work, and spend a lot of resources and money |
|
|
|
06:05.320 --> 06:07.720 |
|
and time cleaning up data. |
|
|
|
06:07.720 --> 06:11.880 |
|
That means the AI engineer may be writing data cleansing |
|
|
|
06:11.880 --> 06:15.600 |
|
algorithms, working with thousands of people |
|
|
|
06:15.600 --> 06:19.160 |
|
who label or correct or do things with the data. |
|
|
|
06:19.160 --> 06:21.920 |
|
That is the incredible hard work that |
|
|
|
06:21.920 --> 06:24.040 |
|
might lead to better results. |
|
|
|
06:24.040 --> 06:28.240 |
|
So the Chinese engineer would rely on and ask for more and more |
|
|
|
06:28.240 --> 06:31.120 |
|
data and find ways to cleanse them and make them work |
|
|
|
06:31.120 --> 06:34.200 |
|
in the system, and probably less time thinking |
|
|
|
06:34.200 --> 06:39.320 |
|
about new algorithms that can overcome data or other issues. |
|
|
|
06:39.320 --> 06:40.560 |
|
So where's your intuition? |
|
|
|
06:40.560 --> 06:43.160 |
|
What do you think the biggest impact the next 10 years |
|
|
|
06:43.160 --> 06:43.920 |
|
lies? |
|
|
|
06:43.920 --> 06:47.120 |
|
Is it in some breakthrough algorithms? |
|
|
|
06:47.120 --> 06:53.920 |
|
Or is it in just this at scale rigor, a rigorous approach |
|
|
|
06:53.920 --> 06:57.120 |
|
to data, cleaning data, organizing data |
|
|
|
06:57.120 --> 06:58.440 |
|
onto the same algorithms? |
|
|
|
06:58.440 --> 07:02.600 |
|
What do you think the big impact in the applied world is? |
|
|
|
07:02.600 --> 07:04.560 |
|
Well, if you're really in the company |
|
|
|
07:04.560 --> 07:08.400 |
|
and you have to deliver results, using known techniques |
|
|
|
07:08.400 --> 07:12.240 |
|
and enhancing data seems like the more expedient approach |
|
|
|
07:12.240 --> 07:15.640 |
|
that's very low risk and likely to generate |
|
|
|
07:15.640 --> 07:17.200 |
|
better and better results. |
|
|
|
07:17.200 --> 07:20.520 |
|
And that's why the Chinese approach has done quite well. |
|
|
|
07:20.520 --> 07:24.240 |
|
Now, there are a lot of more challenging startups |
|
|
|
07:24.240 --> 07:28.440 |
|
and problems, such as autonomous vehicles, |
|
|
|
07:28.440 --> 07:32.560 |
|
medical diagnosis, that existing algorithms probably |
|
|
|
07:32.560 --> 07:34.240 |
|
won't solve. |
|
|
|
07:34.240 --> 07:38.680 |
|
And that would put the Chinese approach more challenged |
|
|
|
07:38.680 --> 07:43.720 |
|
and give them more breakthrough innovation approach, more |
|
|
|
07:43.720 --> 07:45.440 |
|
of an edge on those kinds of problems. |
|
|
|
07:45.440 --> 07:47.040 |
|
So let me talk to that a little more. |
|
|
|
07:47.040 --> 07:50.960 |
|
So my intuition, personally, is that data |
|
|
|
07:50.960 --> 07:53.680 |
|
can take us extremely far. |
|
|
|
07:53.680 --> 07:56.480 |
|
So you brought up autonomous vehicles and medical diagnosis. |
|
|
|
07:56.480 --> 08:00.080 |
|
So your intuition is that huge amounts of data |
|
|
|
08:00.080 --> 08:04.000 |
|
might not be able to completely help us solve that problem. |
|
|
|
08:04.000 --> 08:04.600 |
|
Right. |
|
|
|
08:04.600 --> 08:08.080 |
|
So breaking that down further, autonomous vehicle, |
|
|
|
08:08.080 --> 08:10.080 |
|
I think huge amounts of data probably |
|
|
|
08:10.080 --> 08:13.360 |
|
will solve trucks driving on highways, which |
|
|
|
08:13.360 --> 08:15.640 |
|
will deliver significant value. |
|
|
|
08:15.640 --> 08:19.320 |
|
And China will probably lead in that. |
|
|
|
08:19.320 --> 08:24.880 |
|
And full L5 autonomous is likely to require new technologies |
|
|
|
08:24.880 --> 08:26.320 |
|
we don't yet know. |
|
|
|
08:26.320 --> 08:30.320 |
|
And that might require academia and great industrial research, |
|
|
|
08:30.320 --> 08:32.480 |
|
both innovating and working together. |
|
|
|
08:32.480 --> 08:35.360 |
|
And in that case, US has an advantage. |
|
|
|
08:35.360 --> 08:37.040 |
|
So the interesting question there is, |
|
|
|
08:37.040 --> 08:39.280 |
|
I don't know if you're familiar on the autonomous vehicle |
|
|
|
08:39.280 --> 08:43.480 |
|
space and the developments with Tesla and Elon Musk, |
|
|
|
08:43.480 --> 08:49.400 |
|
where they are, in fact, a full steam ahead |
|
|
|
08:49.400 --> 08:53.480 |
|
into this mysterious, complex world of full autonomy, L5, |
|
|
|
08:53.480 --> 08:55.080 |
|
L4, L5. |
|
|
|
08:55.080 --> 08:58.800 |
|
And they're trying to solve that purely with data. |
|
|
|
08:58.800 --> 09:00.800 |
|
So the same kind of thing that you're saying |
|
|
|
09:00.800 --> 09:03.200 |
|
is just for highway, which is what a lot of people |
|
|
|
09:03.200 --> 09:07.200 |
|
share your intuition, they're trying to solve with data. |
|
|
|
09:07.200 --> 09:09.320 |
|
It's just to linger on that moment further. |
|
|
|
09:09.320 --> 09:13.600 |
|
Do you think possible for them to achieve success |
|
|
|
09:13.600 --> 09:17.040 |
|
with simply just a huge amount of this training |
|
|
|
09:17.040 --> 09:20.440 |
|
on edge cases, on difficult cases in urban environments, |
|
|
|
09:20.440 --> 09:22.840 |
|
not just highway and so on? |
|
|
|
09:22.840 --> 09:24.480 |
|
I think they'll be very hard. |
|
|
|
09:24.480 --> 09:27.680 |
|
One could characterize Tesla's approach as kind |
|
|
|
09:27.680 --> 09:31.600 |
|
of a Chinese strength approach, gather all the data you can, |
|
|
|
09:31.600 --> 09:34.000 |
|
and hope that will overcome the problems. |
|
|
|
09:34.000 --> 09:38.480 |
|
But in autonomous driving, clearly a lot of the decisions |
|
|
|
09:38.480 --> 09:41.480 |
|
aren't merely solved by aggregating data |
|
|
|
09:41.480 --> 09:43.520 |
|
and having feedback loop. |
|
|
|
09:43.520 --> 09:48.040 |
|
There are things that are more akin to human thinking. |
|
|
|
09:48.040 --> 09:51.680 |
|
And how would those be integrated and built? |
|
|
|
09:51.680 --> 09:54.000 |
|
There has not yet been a lot of success |
|
|
|
09:54.000 --> 09:57.200 |
|
integrating human intelligence or, you know, |
|
|
|
09:57.200 --> 09:58.800 |
|
colored expert systems, if you will, |
|
|
|
09:58.800 --> 10:02.960 |
|
even though that's a taboo word with the machine learning. |
|
|
|
10:02.960 --> 10:05.600 |
|
And the integration of the two types of thinking |
|
|
|
10:05.600 --> 10:07.840 |
|
hasn't yet been demonstrated. |
|
|
|
10:07.840 --> 10:09.600 |
|
And the question is, how much can you |
|
|
|
10:09.600 --> 10:12.440 |
|
push a purely machine learning approach? |
|
|
|
10:12.440 --> 10:15.480 |
|
And of course, Tesla also has an additional constraint |
|
|
|
10:15.480 --> 10:18.520 |
|
that they don't have all the sensors. |
|
|
|
10:18.520 --> 10:21.120 |
|
I know that they think it's foolish to use LIDARS, |
|
|
|
10:21.120 --> 10:25.920 |
|
but that's clearly a one less, very valuable and reliable |
|
|
|
10:25.920 --> 10:29.200 |
|
source of input that they're foregoing, which |
|
|
|
10:29.200 --> 10:32.440 |
|
may also have consequences. |
|
|
|
10:32.440 --> 10:33.840 |
|
I think the advantage, of course, |
|
|
|
10:33.840 --> 10:37.040 |
|
is capturing data that no one has ever seen before. |
|
|
|
10:37.040 --> 10:41.040 |
|
And in some cases, such as computer vision and speech |
|
|
|
10:41.040 --> 10:44.800 |
|
recognition, I have seen Chinese companies accumulate data |
|
|
|
10:44.800 --> 10:47.320 |
|
that's not seen anywhere in the Western world, |
|
|
|
10:47.320 --> 10:50.200 |
|
and they have delivered superior results. |
|
|
|
10:50.200 --> 10:53.720 |
|
But then speech recognition and object recognition |
|
|
|
10:53.720 --> 10:57.080 |
|
are relatively suitable problems for deep learning |
|
|
|
10:57.080 --> 11:02.440 |
|
and don't have the potentially need for the human intelligence |
|
|
|
11:02.440 --> 11:04.440 |
|
analytical planning elements. |
|
|
|
11:04.440 --> 11:06.400 |
|
And the same on the speech recognition side, |
|
|
|
11:06.400 --> 11:09.440 |
|
your intuition that speech recognition and the machine |
|
|
|
11:09.440 --> 11:11.440 |
|
learning approaches to speech recognition |
|
|
|
11:11.440 --> 11:14.600 |
|
won't take us to a conversational system that |
|
|
|
11:14.600 --> 11:19.160 |
|
can pass the Turing test, which is maybe akin to what |
|
|
|
11:19.160 --> 11:20.040 |
|
driving is. |
|
|
|
11:20.040 --> 11:25.120 |
|
So it needs to have something more than just simply simple |
|
|
|
11:25.120 --> 11:27.480 |
|
language understanding, simple language generation. |
|
|
|
11:27.480 --> 11:32.000 |
|
Roughly right, I would say that based on purely machine |
|
|
|
11:32.000 --> 11:35.160 |
|
learning approaches, it's hard to imagine. |
|
|
|
11:35.160 --> 11:40.520 |
|
It could lead to a full conversational experience |
|
|
|
11:40.520 --> 11:44.600 |
|
across arbitrary domains, which is akin to L5. |
|
|
|
11:44.600 --> 11:46.920 |
|
I'm a little hesitant to use the word Turing test, |
|
|
|
11:46.920 --> 11:50.280 |
|
because the original definition was probably too easy. |
|
|
|
11:50.280 --> 11:52.320 |
|
We probably do that. |
|
|
|
11:52.320 --> 11:55.280 |
|
The spirit of the Turing test is what I was referring to. |
|
|
|
11:55.280 --> 11:56.520 |
|
Of course. |
|
|
|
11:56.520 --> 11:59.400 |
|
So you've had major leadership research positions |
|
|
|
11:59.400 --> 12:01.640 |
|
at Apple, Microsoft, Google. |
|
|
|
12:01.640 --> 12:06.320 |
|
So continuing on the discussion of America, Russia, Chinese soul |
|
|
|
12:06.320 --> 12:10.520 |
|
and culture and so on, what is the culture of Silicon |
|
|
|
12:10.520 --> 12:16.400 |
|
Valley in contrast to China and maybe US broadly? |
|
|
|
12:16.400 --> 12:19.920 |
|
And what is the unique culture of each of these three |
|
|
|
12:19.920 --> 12:22.040 |
|
major companies, in your view? |
|
|
|
12:22.040 --> 12:25.120 |
|
I think in aggregate, Silicon Valley companies, |
|
|
|
12:25.120 --> 12:27.200 |
|
we could probably include Microsoft in that, |
|
|
|
12:27.200 --> 12:29.120 |
|
even though they're not in the Valley, |
|
|
|
12:29.120 --> 12:33.960 |
|
is really dream big and have visionary goals |
|
|
|
12:33.960 --> 12:37.920 |
|
and believe that technology will conquer all |
|
|
|
12:37.920 --> 12:42.240 |
|
and also the self confidence and the self entitlement |
|
|
|
12:42.240 --> 12:45.440 |
|
that whatever they produce, the whole world should use |
|
|
|
12:45.440 --> 12:47.240 |
|
and must use. |
|
|
|
12:47.240 --> 12:54.080 |
|
And those are historically important, I think. |
|
|
|
12:54.080 --> 12:59.120 |
|
Steve Jobs's famous quote that he doesn't do focus groups. |
|
|
|
12:59.120 --> 13:02.360 |
|
He looks in the mirror and asks the person in the mirror, |
|
|
|
13:02.360 --> 13:03.520 |
|
what do you want? |
|
|
|
13:03.520 --> 13:07.000 |
|
And that really is an inspirational comment |
|
|
|
13:07.000 --> 13:10.480 |
|
that says the great company shouldn't just ask users |
|
|
|
13:10.480 --> 13:13.240 |
|
what they want, but develop something |
|
|
|
13:13.240 --> 13:16.200 |
|
that users will know they want when they see it, |
|
|
|
13:16.200 --> 13:18.960 |
|
but they could never come up with themselves. |
|
|
|
13:18.960 --> 13:23.880 |
|
I think that is probably the most exhilarating description |
|
|
|
13:23.880 --> 13:26.560 |
|
of what the essence of Silicon Valley is, |
|
|
|
13:26.560 --> 13:31.840 |
|
that this brilliant idea could cause you to build something |
|
|
|
13:31.840 --> 13:35.520 |
|
that couldn't come out of the focus groups or A.B. tests. |
|
|
|
13:35.520 --> 13:38.040 |
|
And iPhone would be an example of that. |
|
|
|
13:38.040 --> 13:40.560 |
|
No one in the age of BlackBerry would write down |
|
|
|
13:40.560 --> 13:43.720 |
|
they want an iPhone or multi touch, a browser, |
|
|
|
13:43.720 --> 13:44.800 |
|
might be another example. |
|
|
|
13:44.800 --> 13:47.520 |
|
No one would say they want that in the days of FTP, |
|
|
|
13:47.520 --> 13:49.440 |
|
but once they see it, they want it. |
|
|
|
13:49.440 --> 13:55.680 |
|
So I think that is what Silicon Valley is best at. |
|
|
|
13:55.680 --> 13:58.920 |
|
But it also came with a lot of success. |
|
|
|
13:58.920 --> 14:01.960 |
|
These products became global platforms, |
|
|
|
14:01.960 --> 14:05.080 |
|
and there were basically no competitors anywhere. |
|
|
|
14:05.080 --> 14:08.400 |
|
And that has also led to a belief |
|
|
|
14:08.400 --> 14:13.240 |
|
that these are the only things that one should do, |
|
|
|
14:13.240 --> 14:17.960 |
|
that companies should not tread on other companies territory, |
|
|
|
14:17.960 --> 14:24.040 |
|
so that a Groupon and a Yelp and an OpenTable |
|
|
|
14:24.040 --> 14:26.240 |
|
and the Grubhub would each feel, |
|
|
|
14:26.240 --> 14:28.520 |
|
okay, I'm not going to do the other companies business |
|
|
|
14:28.520 --> 14:33.280 |
|
because that would not be the pride of innovating |
|
|
|
14:33.280 --> 14:36.920 |
|
what each of these four companies have innovated. |
|
|
|
14:36.920 --> 14:42.720 |
|
But I think the Chinese approach is do whatever it takes to win. |
|
|
|
14:42.720 --> 14:45.000 |
|
And it's a winner take all market. |
|
|
|
14:45.000 --> 14:47.200 |
|
And in fact, in the internet space, |
|
|
|
14:47.200 --> 14:50.840 |
|
the market leader will get predominantly all the value |
|
|
|
14:50.840 --> 14:53.320 |
|
extracted out of the system. |
|
|
|
14:53.320 --> 14:59.600 |
|
And the system isn't just defined as one narrow category, |
|
|
|
14:59.600 --> 15:01.360 |
|
but gets broader and broader. |
|
|
|
15:01.360 --> 15:07.960 |
|
So it's amazing ambition for success and domination |
|
|
|
15:07.960 --> 15:11.760 |
|
of increasingly larger product categories |
|
|
|
15:11.760 --> 15:15.080 |
|
leading to clear market winner status |
|
|
|
15:15.080 --> 15:19.120 |
|
and the opportunity to extract tremendous value. |
|
|
|
15:19.120 --> 15:25.840 |
|
And that develops a practical, result oriented, |
|
|
|
15:25.840 --> 15:31.520 |
|
ultra ambitious winner take all gladiatorial mentality. |
|
|
|
15:31.520 --> 15:37.400 |
|
And if what it takes is to build what the competitors built, |
|
|
|
15:37.400 --> 15:41.920 |
|
essentially a copycat, that can be done without infringing laws. |
|
|
|
15:41.920 --> 15:46.280 |
|
If what it takes is to satisfy a foreign country's need |
|
|
|
15:46.280 --> 15:48.480 |
|
by forking the code base and building something |
|
|
|
15:48.480 --> 15:51.440 |
|
that looks really ugly and different, they'll do it. |
|
|
|
15:51.440 --> 15:56.280 |
|
So it's contrasted very sharply with the Silicon Valley approach. |
|
|
|
15:56.280 --> 16:00.080 |
|
And I think the flexibility and the speed and execution |
|
|
|
16:00.080 --> 16:01.960 |
|
has helped the Chinese approach. |
|
|
|
16:01.960 --> 16:05.040 |
|
And I think the Silicon Valley approach |
|
|
|
16:05.040 --> 16:10.280 |
|
is potentially challenged if every Chinese entrepreneur is |
|
|
|
16:10.280 --> 16:13.200 |
|
learning from the whole world, US and China, |
|
|
|
16:13.200 --> 16:16.280 |
|
and the American entrepreneurs only look internally |
|
|
|
16:16.280 --> 16:19.600 |
|
and write off China as a copycat. |
|
|
|
16:19.600 --> 16:22.880 |
|
And the second part of your question about the three |
|
|
|
16:22.880 --> 16:23.520 |
|
companies. |
|
|
|
16:23.520 --> 16:26.000 |
|
The unique elements of the three companies, perhaps. |
|
|
|
16:26.000 --> 16:26.840 |
|
Yeah. |
|
|
|
16:26.840 --> 16:33.080 |
|
I think Apple represents, while the user, please the user, |
|
|
|
16:33.080 --> 16:38.520 |
|
and the essence of design and brand, |
|
|
|
16:38.520 --> 16:44.080 |
|
and it's the one company and perhaps the only tech company |
|
|
|
16:44.080 --> 16:49.920 |
|
that draws people with a strong, serious desire |
|
|
|
16:49.920 --> 16:53.560 |
|
for the product and the willingness to pay a premium |
|
|
|
16:53.560 --> 16:57.160 |
|
because of the halo effect of the brand, which |
|
|
|
16:57.160 --> 17:00.960 |
|
came from the attention to detail and great respect |
|
|
|
17:00.960 --> 17:03.360 |
|
for user needs. |
|
|
|
17:03.360 --> 17:09.200 |
|
Microsoft represents a platform approach |
|
|
|
17:09.200 --> 17:14.280 |
|
that builds giant products that become very strong modes |
|
|
|
17:14.280 --> 17:17.680 |
|
that others can't do because it's |
|
|
|
17:17.680 --> 17:21.480 |
|
well architected at the bottom level |
|
|
|
17:21.480 --> 17:26.640 |
|
and the work is efficiently delegated to individuals |
|
|
|
17:26.640 --> 17:30.360 |
|
and then the whole product is built |
|
|
|
17:30.360 --> 17:33.560 |
|
by adding small parts that sum together. |
|
|
|
17:33.560 --> 17:37.760 |
|
So it's probably the most effective high tech assembly |
|
|
|
17:37.760 --> 17:40.480 |
|
line that builds a very difficult product |
|
|
|
17:40.480 --> 17:44.800 |
|
that the whole process of doing that |
|
|
|
17:44.800 --> 17:50.800 |
|
is kind of a differentiation and something competitors |
|
|
|
17:50.800 --> 17:52.480 |
|
can't easily repeat. |
|
|
|
17:52.480 --> 17:54.800 |
|
Are there elements of the Chinese approach |
|
|
|
17:54.800 --> 17:59.280 |
|
in the way Microsoft went about assembling those little pieces |
|
|
|
17:59.280 --> 18:03.920 |
|
and essentially dominating the market for a long time? |
|
|
|
18:03.920 --> 18:05.640 |
|
Or do you see those as distinct? |
|
|
|
18:05.640 --> 18:08.240 |
|
I think there are elements that are the same. |
|
|
|
18:08.240 --> 18:10.440 |
|
I think the three American companies |
|
|
|
18:10.440 --> 18:13.880 |
|
that had or have Chinese characteristics, |
|
|
|
18:13.880 --> 18:16.080 |
|
and obviously as well as American characteristics, |
|
|
|
18:16.080 --> 18:20.400 |
|
are Microsoft, Facebook, and Amazon. |
|
|
|
18:20.400 --> 18:21.720 |
|
Yes, that's right, Amazon. |
|
|
|
18:21.720 --> 18:25.560 |
|
Because these are companies that will tenaciously |
|
|
|
18:25.560 --> 18:31.320 |
|
go after adjacent markets, build up strong product offering, |
|
|
|
18:31.320 --> 18:38.200 |
|
and find ways to extract greater value from a sphere that's |
|
|
|
18:38.200 --> 18:39.960 |
|
ever increasing. |
|
|
|
18:39.960 --> 18:43.520 |
|
And they understand the value of the platforms. |
|
|
|
18:43.520 --> 18:45.600 |
|
So that's the similarity. |
|
|
|
18:45.600 --> 18:53.760 |
|
And then with Google, I think it's a genuinely value oriented |
|
|
|
18:53.760 --> 18:56.960 |
|
company that does have a heart and soul |
|
|
|
18:56.960 --> 18:59.760 |
|
and that wants to do great things for the world |
|
|
|
18:59.760 --> 19:06.040 |
|
by connecting information and that has also |
|
|
|
19:06.040 --> 19:13.280 |
|
very strong technology genes and wants to use technology |
|
|
|
19:13.280 --> 19:19.080 |
|
and has found out of the box ways to use technology |
|
|
|
19:19.080 --> 19:23.680 |
|
to deliver incredible value to the end user. |
|
|
|
19:23.680 --> 19:25.240 |
|
We can look at Google, for example. |
|
|
|
19:25.240 --> 19:28.040 |
|
You mentioned heart and soul. |
|
|
|
19:28.040 --> 19:31.840 |
|
There seems to be an element where Google |
|
|
|
19:31.840 --> 19:34.840 |
|
is after making the world better. |
|
|
|
19:34.840 --> 19:36.520 |
|
There's a more positive view. |
|
|
|
19:36.520 --> 19:38.960 |
|
I mean, they used to have the slogan, don't be evil. |
|
|
|
19:38.960 --> 19:43.120 |
|
And Facebook a little bit more has a negative tend to it, |
|
|
|
19:43.120 --> 19:46.000 |
|
at least in the perception of privacy and so on. |
|
|
|
19:46.000 --> 19:51.280 |
|
Do you have a sense of how these different companies can |
|
|
|
19:51.280 --> 19:53.400 |
|
achieve, because you've talked about how much |
|
|
|
19:53.400 --> 19:55.600 |
|
we can make the world better in all these kinds of ways |
|
|
|
19:55.600 --> 19:59.360 |
|
with AI, what is it about a company that can make, |
|
|
|
19:59.360 --> 20:03.200 |
|
give it a heart and soul, gain the trust of the public, |
|
|
|
20:03.200 --> 20:08.000 |
|
and just actually just not be evil and do good for the world? |
|
|
|
20:08.000 --> 20:09.000 |
|
It's really hard. |
|
|
|
20:09.000 --> 20:13.120 |
|
And I think Google has struggled with that. |
|
|
|
20:13.120 --> 20:15.160 |
|
First, they don't do evil. |
|
|
|
20:15.160 --> 20:18.880 |
|
Mantra is very dangerous, because every employee's |
|
|
|
20:18.880 --> 20:20.800 |
|
definition of evil is different. |
|
|
|
20:20.800 --> 20:23.800 |
|
And that has led to some difficult employee situations |
|
|
|
20:23.800 --> 20:25.240 |
|
for them. |
|
|
|
20:25.240 --> 20:29.520 |
|
So I don't necessarily think that's a good value statement. |
|
|
|
20:29.520 --> 20:31.840 |
|
But just watching the kinds of things |
|
|
|
20:31.840 --> 20:36.440 |
|
Google or its parent company Alphabet does in new areas |
|
|
|
20:36.440 --> 20:40.440 |
|
like health care, like eradicating mosquitoes, |
|
|
|
20:40.440 --> 20:42.360 |
|
things that are really not in the business |
|
|
|
20:42.360 --> 20:45.040 |
|
of a internet tech company, I think |
|
|
|
20:45.040 --> 20:47.200 |
|
that shows that there is a heart and soul |
|
|
|
20:47.200 --> 20:53.920 |
|
and desire to do good and willingness to put in the resources |
|
|
|
20:53.920 --> 20:58.280 |
|
to do something when they see it's good, they will pursue it. |
|
|
|
20:58.280 --> 21:00.640 |
|
That doesn't necessarily mean it has |
|
|
|
21:00.640 --> 21:02.520 |
|
all the trust of the users. |
|
|
|
21:02.520 --> 21:06.400 |
|
I realize while most people would view Facebook |
|
|
|
21:06.400 --> 21:09.760 |
|
as the primary target of their recent unhappiness |
|
|
|
21:09.760 --> 21:12.720 |
|
about Silicon Valley companies, many would put Google |
|
|
|
21:12.720 --> 21:14.080 |
|
in that category. |
|
|
|
21:14.080 --> 21:16.800 |
|
And some have named Google's business practices |
|
|
|
21:16.800 --> 21:19.840 |
|
as predatory also. |
|
|
|
21:19.840 --> 21:24.240 |
|
So it's kind of difficult to have the two parts of a body. |
|
|
|
21:24.240 --> 21:28.080 |
|
The brain wants to do what it's supposed to do for a shareholder, |
|
|
|
21:28.080 --> 21:29.280 |
|
maximize profit. |
|
|
|
21:29.280 --> 21:30.880 |
|
And then the heart and soul wants |
|
|
|
21:30.880 --> 21:36.120 |
|
to do good things that may run against what the brain wants to do. |
|
|
|
21:36.120 --> 21:40.320 |
|
So in this complex balancing that these companies have to do, |
|
|
|
21:40.320 --> 21:44.520 |
|
you've mentioned that you're concerned about a future where |
|
|
|
21:44.520 --> 21:47.360 |
|
too few companies like Google, Facebook, Amazon |
|
|
|
21:47.360 --> 21:51.560 |
|
are controlling our data or are controlling too much |
|
|
|
21:51.560 --> 21:53.360 |
|
of our digital lives. |
|
|
|
21:53.360 --> 21:55.400 |
|
Can you elaborate on this concern? |
|
|
|
21:55.400 --> 21:58.640 |
|
Perhaps do you have a better way forward? |
|
|
|
21:58.640 --> 22:05.000 |
|
I think I'm hardly the most vocal complainer of this. |
|
|
|
22:05.000 --> 22:07.280 |
|
There are a lot louder complainers out there. |
|
|
|
22:07.280 --> 22:11.840 |
|
I do observe that having a lot of data |
|
|
|
22:11.840 --> 22:16.120 |
|
does perpetuate their strength and limits |
|
|
|
22:16.120 --> 22:19.400 |
|
competition in many spaces. |
|
|
|
22:19.400 --> 22:24.200 |
|
But I also believe AI is much broader than the internet space. |
|
|
|
22:24.200 --> 22:26.280 |
|
So the entrepreneurial opportunities |
|
|
|
22:26.280 --> 22:30.480 |
|
still exists in using AI to empower |
|
|
|
22:30.480 --> 22:34.160 |
|
financial, retail, manufacturing, education, |
|
|
|
22:34.160 --> 22:35.480 |
|
applications. |
|
|
|
22:35.480 --> 22:39.800 |
|
So I don't think it's quite a case of full monopolistic dominance |
|
|
|
22:39.800 --> 22:43.960 |
|
that totally stifles innovation. |
|
|
|
22:43.960 --> 22:46.400 |
|
But I do believe in their areas of strength |
|
|
|
22:46.400 --> 22:49.760 |
|
it's hard to dislodge them. |
|
|
|
22:49.760 --> 22:53.280 |
|
I don't know if I have a good solution. |
|
|
|
22:53.280 --> 22:57.160 |
|
Probably the best solution is let the entrepreneurial VC |
|
|
|
22:57.160 --> 23:00.840 |
|
ecosystem work well and find all the places that |
|
|
|
23:00.840 --> 23:04.200 |
|
can create the next Google, the next Facebook. |
|
|
|
23:04.200 --> 23:08.560 |
|
So there will always be increasing number of challengers. |
|
|
|
23:08.560 --> 23:11.360 |
|
In some sense, that has happened a little bit. |
|
|
|
23:11.360 --> 23:15.760 |
|
You see Uber, Airbnb having emerged despite the strength |
|
|
|
23:15.760 --> 23:19.040 |
|
of the big three. |
|
|
|
23:19.040 --> 23:22.400 |
|
And I think China as an environment |
|
|
|
23:22.400 --> 23:25.280 |
|
may be more interesting for the emergence. |
|
|
|
23:25.280 --> 23:28.920 |
|
Because if you look at companies between, let's say, |
|
|
|
23:28.920 --> 23:36.320 |
|
$50 to $300 billion, China has emerged more of such companies |
|
|
|
23:36.320 --> 23:39.880 |
|
than the US in the last three to four years. |
|
|
|
23:39.880 --> 23:42.120 |
|
Because of the larger marketplace, |
|
|
|
23:42.120 --> 23:47.000 |
|
because of the more fearless nature of the entrepreneurs. |
|
|
|
23:47.000 --> 23:50.840 |
|
And the Chinese giants are just as powerful as American ones. |
|
|
|
23:50.840 --> 23:52.920 |
|
Tencent Alibaba are very strong. |
|
|
|
23:52.920 --> 23:57.040 |
|
But Bytes Dance has emerged worth $75 billion. |
|
|
|
23:57.040 --> 24:00.120 |
|
And financial, while it's Alibaba affiliated, |
|
|
|
24:00.120 --> 24:03.920 |
|
it's nevertheless independent and worth $150 billion. |
|
|
|
24:03.920 --> 24:08.280 |
|
And so I do think if we start to extend |
|
|
|
24:08.280 --> 24:12.640 |
|
to traditional businesses, we will see very valuable companies. |
|
|
|
24:12.640 --> 24:18.120 |
|
So it's probably not the case that in five or 10 years, |
|
|
|
24:18.120 --> 24:20.920 |
|
we'll still see the whole world with these five companies |
|
|
|
24:20.920 --> 24:22.680 |
|
having such dominance. |
|
|
|
24:22.680 --> 24:26.040 |
|
So you've mentioned a couple of times |
|
|
|
24:26.040 --> 24:27.840 |
|
this fascinating world of entrepreneurship |
|
|
|
24:27.840 --> 24:31.080 |
|
in China of the fearless nature of the entrepreneurs. |
|
|
|
24:31.080 --> 24:32.640 |
|
So can you maybe talk a little bit |
|
|
|
24:32.640 --> 24:35.520 |
|
about what it takes to be an entrepreneur in China? |
|
|
|
24:35.520 --> 24:38.240 |
|
What are the strategies that are undertaken? |
|
|
|
24:38.240 --> 24:41.120 |
|
What are the ways that you success? |
|
|
|
24:41.120 --> 24:43.960 |
|
What is the dynamic of VCF funding, |
|
|
|
24:43.960 --> 24:46.480 |
|
of the way the government helps companies, and so on? |
|
|
|
24:46.480 --> 24:49.520 |
|
What are the interesting aspects here that are distinct from, |
|
|
|
24:49.520 --> 24:52.880 |
|
that are different from the Silicon Valley world |
|
|
|
24:52.880 --> 24:55.240 |
|
of entrepreneurship? |
|
|
|
24:55.240 --> 24:58.080 |
|
Well, many of the listeners probably |
|
|
|
24:58.080 --> 25:03.000 |
|
still would brand Chinese entrepreneur as copycats. |
|
|
|
25:03.000 --> 25:06.120 |
|
And no doubt, 10 years ago, that would not |
|
|
|
25:06.120 --> 25:09.080 |
|
be an inaccurate description. |
|
|
|
25:09.080 --> 25:12.320 |
|
Back 10 years ago, an entrepreneur probably |
|
|
|
25:12.320 --> 25:14.840 |
|
could not get funding if he or she could not |
|
|
|
25:14.840 --> 25:20.400 |
|
describe what product he or she is copying from the US. |
|
|
|
25:20.400 --> 25:23.520 |
|
The first question is, who has proven this business model, |
|
|
|
25:23.520 --> 25:27.200 |
|
which is a nice way of asking, who are you copying? |
|
|
|
25:27.200 --> 25:29.520 |
|
And that reason is understandable, |
|
|
|
25:29.520 --> 25:34.840 |
|
because China had a much lower internet penetration |
|
|
|
25:34.840 --> 25:40.920 |
|
and didn't have enough indigenous experience |
|
|
|
25:40.920 --> 25:43.200 |
|
to build innovative products. |
|
|
|
25:43.200 --> 25:47.600 |
|
And secondly, internet was emerging. |
|
|
|
25:47.600 --> 25:49.800 |
|
Link startup was the way to do things, |
|
|
|
25:49.800 --> 25:52.920 |
|
building a first minimally viable product, |
|
|
|
25:52.920 --> 25:55.320 |
|
and then expanding was the right way to go. |
|
|
|
25:55.320 --> 25:59.480 |
|
And the American successes have given a shortcut |
|
|
|
25:59.480 --> 26:02.840 |
|
that if you build your minimally viable product based |
|
|
|
26:02.840 --> 26:05.040 |
|
on an American product, it's guaranteed |
|
|
|
26:05.040 --> 26:06.720 |
|
to be a decent starting point. |
|
|
|
26:06.720 --> 26:08.400 |
|
Then you tweak it afterwards. |
|
|
|
26:08.400 --> 26:11.720 |
|
So as long as there are no IP infringement, which, |
|
|
|
26:11.720 --> 26:15.080 |
|
as far as I know, there hasn't been in the mobile and AI |
|
|
|
26:15.080 --> 26:19.360 |
|
spaces, that's a much better shortcut. |
|
|
|
26:19.360 --> 26:23.720 |
|
And I think Silicon Valley would view that as still not |
|
|
|
26:23.720 --> 26:29.200 |
|
very honorable, because that's not your own idea to start with. |
|
|
|
26:29.200 --> 26:32.600 |
|
But you can't really, at the same time, |
|
|
|
26:32.600 --> 26:35.160 |
|
believe every idea must be your own |
|
|
|
26:35.160 --> 26:38.120 |
|
and believe in the link startup methodology, |
|
|
|
26:38.120 --> 26:41.880 |
|
because link startup is intended to try many, many things |
|
|
|
26:41.880 --> 26:44.240 |
|
and then converge when that works. |
|
|
|
26:44.240 --> 26:46.720 |
|
And it's meant to be iterated and changed. |
|
|
|
26:46.720 --> 26:51.240 |
|
So finding a decent starting point without legal violations, |
|
|
|
26:51.240 --> 26:55.520 |
|
there should be nothing morally dishonorable about that. |
|
|
|
26:55.520 --> 26:57.080 |
|
So just a quick pause on that. |
|
|
|
26:57.080 --> 27:01.920 |
|
It's fascinating that that's why is that not honorable, right? |
|
|
|
27:01.920 --> 27:04.680 |
|
It's exactly as you formulated. |
|
|
|
27:04.680 --> 27:08.040 |
|
It seems like a perfect start for business |
|
|
|
27:08.040 --> 27:12.440 |
|
is to take a look at Amazon and say, OK, |
|
|
|
27:12.440 --> 27:14.560 |
|
we'll do exactly what Amazon is doing. |
|
|
|
27:14.560 --> 27:16.800 |
|
Let's start there in this particular market. |
|
|
|
27:16.800 --> 27:20.520 |
|
And then let's out innovate them from that starting point. |
|
|
|
27:20.520 --> 27:22.200 |
|
Yes. Come up with new ways. |
|
|
|
27:22.200 --> 27:26.520 |
|
I mean, is it wrong to be, except the word copycat just |
|
|
|
27:26.520 --> 27:28.800 |
|
sounds bad, but is it wrong to be a copycat? |
|
|
|
27:28.800 --> 27:31.640 |
|
It just seems like a smart strategy. |
|
|
|
27:31.640 --> 27:35.800 |
|
But yes, doesn't have a heroic nature to it |
|
|
|
27:35.800 --> 27:42.280 |
|
that Steve Jobs, Elon Musk, sort of in something completely |
|
|
|
27:42.280 --> 27:43.880 |
|
coming up with something completely new. |
|
|
|
27:43.880 --> 27:45.480 |
|
Yeah, I like the way you describe it. |
|
|
|
27:45.480 --> 27:50.440 |
|
It's a nonheroic, acceptable way to start the company. |
|
|
|
27:50.440 --> 27:52.840 |
|
And maybe more expedient. |
|
|
|
27:52.840 --> 27:58.920 |
|
So that's, I think, a baggage for Silicon Valley, |
|
|
|
27:58.920 --> 28:01.320 |
|
that if it doesn't let go, then it |
|
|
|
28:01.320 --> 28:05.160 |
|
may limit the ultimate ceiling of the company. |
|
|
|
28:05.160 --> 28:07.200 |
|
Take Snapchat as an example. |
|
|
|
28:07.200 --> 28:09.840 |
|
I think Evan's brilliant. |
|
|
|
28:09.840 --> 28:11.480 |
|
He built a great product. |
|
|
|
28:11.480 --> 28:14.160 |
|
But he's very proud that he wants |
|
|
|
28:14.160 --> 28:16.800 |
|
to build his own features, not copy others. |
|
|
|
28:16.800 --> 28:21.000 |
|
While Facebook was more willing to copy his features, |
|
|
|
28:21.000 --> 28:23.440 |
|
and you see what happens in the competition. |
|
|
|
28:23.440 --> 28:27.440 |
|
So I think putting that handcuff on the company |
|
|
|
28:27.440 --> 28:31.560 |
|
would limit its ability to reach the maximum potential. |
|
|
|
28:31.560 --> 28:33.800 |
|
So back to the Chinese environment, |
|
|
|
28:33.800 --> 28:38.400 |
|
copying was merely a way to learn from the American masters. |
|
|
|
28:38.400 --> 28:43.480 |
|
Just like if we learned to play piano or painting, |
|
|
|
28:43.480 --> 28:44.560 |
|
you start by copying. |
|
|
|
28:44.560 --> 28:46.160 |
|
You don't start by innovating when |
|
|
|
28:46.160 --> 28:48.200 |
|
you don't have the basic skill sets. |
|
|
|
28:48.200 --> 28:51.040 |
|
So very amazingly, the Chinese entrepreneurs |
|
|
|
28:51.040 --> 28:56.160 |
|
about six years ago started to branch off |
|
|
|
28:56.160 --> 28:59.520 |
|
with these lean startups built on American ideas |
|
|
|
28:59.520 --> 29:02.280 |
|
to build better products than American products. |
|
|
|
29:02.280 --> 29:04.960 |
|
But they did start from the American idea. |
|
|
|
29:04.960 --> 29:08.600 |
|
And today, WeChat is better than WhatsApp. |
|
|
|
29:08.600 --> 29:10.520 |
|
Weibo is better than Twitter. |
|
|
|
29:10.520 --> 29:12.920 |
|
Zihu is better than Quora and so on. |
|
|
|
29:12.920 --> 29:17.000 |
|
So that, I think, is Chinese entrepreneurs |
|
|
|
29:17.000 --> 29:18.480 |
|
going to step two. |
|
|
|
29:18.480 --> 29:21.760 |
|
And then step three is once these entrepreneurs have |
|
|
|
29:21.760 --> 29:23.720 |
|
done one or two of these companies, |
|
|
|
29:23.720 --> 29:27.400 |
|
they now look at the Chinese market and the opportunities |
|
|
|
29:27.400 --> 29:30.600 |
|
and come up with ideas that didn't exist elsewhere. |
|
|
|
29:30.600 --> 29:36.320 |
|
So products like and financial under which includes Alipay, |
|
|
|
29:36.320 --> 29:42.080 |
|
which is mobile payments, and also the financial products |
|
|
|
29:42.080 --> 29:48.560 |
|
for loans built on that, and also in education, VIP kid, |
|
|
|
29:48.560 --> 29:54.880 |
|
and in social video, social network, TikTok, |
|
|
|
29:54.880 --> 29:58.640 |
|
and in social eCommerce, Pinduoduo, |
|
|
|
29:58.640 --> 30:01.720 |
|
and then in ride sharing, Mobike. |
|
|
|
30:01.720 --> 30:05.640 |
|
These are all Chinese innovative products |
|
|
|
30:05.640 --> 30:08.720 |
|
that now are being copied elsewhere. |
|
|
|
30:08.720 --> 30:13.040 |
|
So an additional interesting observation |
|
|
|
30:13.040 --> 30:16.000 |
|
is some of these products are built on unique Chinese |
|
|
|
30:16.000 --> 30:19.360 |
|
demographics, which may not work in the US, |
|
|
|
30:19.360 --> 30:23.160 |
|
but may work very well in Southeast Asia, Africa, |
|
|
|
30:23.160 --> 30:27.840 |
|
and other developing worlds that are a few years behind China. |
|
|
|
30:27.840 --> 30:31.040 |
|
And a few of these products maybe are universal |
|
|
|
30:31.040 --> 30:33.760 |
|
and are getting traction even in the United States, |
|
|
|
30:33.760 --> 30:35.360 |
|
such as TikTok. |
|
|
|
30:35.360 --> 30:42.080 |
|
So this whole ecosystem is supported by VCs |
|
|
|
30:42.080 --> 30:44.920 |
|
as a virtuous cycle, because a large market |
|
|
|
30:44.920 --> 30:49.400 |
|
with innovative entrepreneurs will draw a lot of money |
|
|
|
30:49.400 --> 30:51.560 |
|
and then invest in these companies. |
|
|
|
30:51.560 --> 30:54.480 |
|
As the market gets larger and larger, |
|
|
|
30:54.480 --> 30:58.400 |
|
China market is easily three, four times larger than the US. |
|
|
|
30:58.400 --> 31:01.120 |
|
They will create greater value and greater returns |
|
|
|
31:01.120 --> 31:05.400 |
|
for the VCs, thereby raising even more money. |
|
|
|
31:05.400 --> 31:10.000 |
|
So at Sinovation Ventures, our first fund was $15 million. |
|
|
|
31:10.000 --> 31:12.040 |
|
Our last fund was $500 million. |
|
|
|
31:12.040 --> 31:16.520 |
|
So it reflects the valuation of the companies |
|
|
|
31:16.520 --> 31:19.840 |
|
and our us going multi stage and things like that. |
|
|
|
31:19.840 --> 31:23.840 |
|
It also has government support, but not |
|
|
|
31:23.840 --> 31:26.080 |
|
in the way most Americans would think of it. |
|
|
|
31:26.080 --> 31:29.520 |
|
The government actually leaves the entrepreneurial space |
|
|
|
31:29.520 --> 31:33.200 |
|
as a private enterprise, so the self regulating. |
|
|
|
31:33.200 --> 31:36.200 |
|
And the government would build infrastructures |
|
|
|
31:36.200 --> 31:39.320 |
|
that would around it to make it work better. |
|
|
|
31:39.320 --> 31:41.960 |
|
For example, the mass entrepreneur mass innovation |
|
|
|
31:41.960 --> 31:44.880 |
|
plan builds 8,000 incubators. |
|
|
|
31:44.880 --> 31:48.360 |
|
So the pipeline is very strong to the VCs |
|
|
|
31:48.360 --> 31:49.680 |
|
for autonomous vehicles. |
|
|
|
31:49.680 --> 31:53.280 |
|
The Chinese government is building smart highways |
|
|
|
31:53.280 --> 31:56.680 |
|
with sensors, smart cities that separate pedestrians |
|
|
|
31:56.680 --> 32:01.560 |
|
from cars that may allow initially an inferior autonomous |
|
|
|
32:01.560 --> 32:05.760 |
|
vehicle company to launch a car without increasing, |
|
|
|
32:05.760 --> 32:11.520 |
|
with lower casualty, because the roads or the city is smart. |
|
|
|
32:11.520 --> 32:13.800 |
|
And the Chinese government at local levels |
|
|
|
32:13.800 --> 32:17.360 |
|
would have these guiding funds acting as LPs, |
|
|
|
32:17.360 --> 32:19.400 |
|
passive LPs to funds. |
|
|
|
32:19.400 --> 32:23.240 |
|
And when the fund makes money, part of the money made |
|
|
|
32:23.240 --> 32:27.280 |
|
is given back to the GPs and potentially other LPs |
|
|
|
32:27.280 --> 32:31.960 |
|
to increase everybody's return at the expense |
|
|
|
32:31.960 --> 32:33.680 |
|
of the government's return. |
|
|
|
32:33.680 --> 32:36.360 |
|
So that's an interesting incentive |
|
|
|
32:36.360 --> 32:41.640 |
|
that entrusts the task of choosing entrepreneurs to VCs |
|
|
|
32:41.640 --> 32:43.800 |
|
who are better at it than the government |
|
|
|
32:43.800 --> 32:46.680 |
|
by letting some of the profits move that way. |
|
|
|
32:46.680 --> 32:48.720 |
|
So this is really fascinating, right? |
|
|
|
32:48.720 --> 32:51.800 |
|
So I look at the Russian government as a case study |
|
|
|
32:51.800 --> 32:54.480 |
|
where, let me put it this way, there |
|
|
|
32:54.480 --> 32:58.520 |
|
is no such government driven, large scale |
|
|
|
32:58.520 --> 33:00.840 |
|
support of entrepreneurship. |
|
|
|
33:00.840 --> 33:04.000 |
|
And probably the same is true in the United States. |
|
|
|
33:04.000 --> 33:07.640 |
|
But the entrepreneurs themselves kind of find a way. |
|
|
|
33:07.640 --> 33:11.680 |
|
So maybe in a form of advice or explanation, |
|
|
|
33:11.680 --> 33:15.560 |
|
how did the Chinese government arrive to be this way, |
|
|
|
33:15.560 --> 33:17.680 |
|
so supportive on entrepreneurship, |
|
|
|
33:17.680 --> 33:21.520 |
|
to be in this particular way so forward thinking |
|
|
|
33:21.520 --> 33:23.120 |
|
at such a large scale? |
|
|
|
33:23.120 --> 33:28.280 |
|
And also perhaps, how can we copy it in other countries? |
|
|
|
33:28.280 --> 33:29.800 |
|
How can we encourage other governments, |
|
|
|
33:29.800 --> 33:31.600 |
|
like even the United States government, |
|
|
|
33:31.600 --> 33:33.760 |
|
to support infrastructure for autonomous vehicles |
|
|
|
33:33.760 --> 33:36.040 |
|
in that same kind of way, perhaps? |
|
|
|
33:36.040 --> 33:36.680 |
|
Yes. |
|
|
|
33:36.680 --> 33:44.440 |
|
So these techniques are the result of several key things, |
|
|
|
33:44.440 --> 33:46.480 |
|
some of which may be learnable, some of which |
|
|
|
33:46.480 --> 33:48.440 |
|
may be very hard. |
|
|
|
33:48.440 --> 33:51.080 |
|
One is just trial and error and watching |
|
|
|
33:51.080 --> 33:52.960 |
|
what everyone else is doing. |
|
|
|
33:52.960 --> 33:54.960 |
|
I think it's important to be humble and not |
|
|
|
33:54.960 --> 33:56.920 |
|
feel like you know all the answers. |
|
|
|
33:56.920 --> 33:59.480 |
|
The guiding funds idea came from Singapore, |
|
|
|
33:59.480 --> 34:01.440 |
|
which came from Israel. |
|
|
|
34:01.440 --> 34:06.080 |
|
And China made a few tweaks and turned it into a, |
|
|
|
34:06.080 --> 34:09.600 |
|
because the Chinese cities and government officials kind |
|
|
|
34:09.600 --> 34:11.320 |
|
of compete with each other. |
|
|
|
34:11.320 --> 34:14.640 |
|
Because they all want to make their city more successful, |
|
|
|
34:14.640 --> 34:20.280 |
|
so they can get the next level in their political career. |
|
|
|
34:20.280 --> 34:22.320 |
|
And it's somewhat competitive. |
|
|
|
34:22.320 --> 34:25.200 |
|
So the central government made it a bit of a competition. |
|
|
|
34:25.200 --> 34:26.840 |
|
Everybody has a budget. |
|
|
|
34:26.840 --> 34:29.840 |
|
They can put it on AI, or they can put it on bio, |
|
|
|
34:29.840 --> 34:32.200 |
|
or they can put it on energy. |
|
|
|
34:32.200 --> 34:35.040 |
|
And then whoever gets the results, the city shines, |
|
|
|
34:35.040 --> 34:38.000 |
|
the people are better off, the mayor gets a promotion. |
|
|
|
34:38.000 --> 34:41.680 |
|
So the tools is kind of almost like an entrepreneurial |
|
|
|
34:41.680 --> 34:44.840 |
|
environment for local governments |
|
|
|
34:44.840 --> 34:47.480 |
|
to see who can do a better job. |
|
|
|
34:47.480 --> 34:52.440 |
|
And also, many of them tried different experiments. |
|
|
|
34:52.440 --> 34:58.440 |
|
Some have given award to very smart researchers, |
|
|
|
34:58.440 --> 35:00.840 |
|
just give them money and hope they'll start a company. |
|
|
|
35:00.840 --> 35:05.840 |
|
Some have given money to academic research labs, |
|
|
|
35:05.840 --> 35:08.440 |
|
maybe government research labs, to see |
|
|
|
35:08.440 --> 35:11.920 |
|
if they can spin off some companies from the science |
|
|
|
35:11.920 --> 35:14.040 |
|
lab or something like that. |
|
|
|
35:14.040 --> 35:17.080 |
|
Some have tried to recruit overseas Chinese |
|
|
|
35:17.080 --> 35:18.960 |
|
to come back and start companies. |
|
|
|
35:18.960 --> 35:20.960 |
|
And they've had mixed results. |
|
|
|
35:20.960 --> 35:23.400 |
|
The one that worked the best was the guiding funds. |
|
|
|
35:23.400 --> 35:25.840 |
|
So it's almost like a lean startup idea |
|
|
|
35:25.840 --> 35:29.160 |
|
where people try different things in what works, sticks, |
|
|
|
35:29.160 --> 35:30.600 |
|
and everybody copies. |
|
|
|
35:30.600 --> 35:32.880 |
|
So now every city has a guiding fund. |
|
|
|
35:32.880 --> 35:35.680 |
|
So that's how that came about. |
|
|
|
35:35.680 --> 35:40.400 |
|
The autonomous vehicle and the massive spending |
|
|
|
35:40.400 --> 35:46.080 |
|
in highways and smart cities, that's a Chinese way. |
|
|
|
35:46.080 --> 35:49.480 |
|
It's about building infrastructure to facilitate. |
|
|
|
35:49.480 --> 35:52.840 |
|
It's a clear division of the government's responsibility |
|
|
|
35:52.840 --> 35:55.400 |
|
from the market. |
|
|
|
35:55.400 --> 36:00.560 |
|
The market should do everything in a private freeway. |
|
|
|
36:00.560 --> 36:02.920 |
|
But there are things the market can't afford to do, |
|
|
|
36:02.920 --> 36:04.520 |
|
like infrastructure. |
|
|
|
36:04.520 --> 36:08.000 |
|
So the government always appropriates |
|
|
|
36:08.000 --> 36:12.000 |
|
large amounts of money for infrastructure building. |
|
|
|
36:12.000 --> 36:16.880 |
|
This happens with not only autonomous vehicle and AI, |
|
|
|
36:16.880 --> 36:20.840 |
|
but happened with the 3G and 4G. |
|
|
|
36:20.840 --> 36:25.320 |
|
You'll find that the Chinese wireless reception |
|
|
|
36:25.320 --> 36:28.760 |
|
is better than the US, because massive spending that |
|
|
|
36:28.760 --> 36:30.720 |
|
tries to cover the whole country. |
|
|
|
36:30.720 --> 36:34.360 |
|
Whereas in the US, it may be a little spotty. |
|
|
|
36:34.360 --> 36:36.160 |
|
It's a government driven, because I think |
|
|
|
36:36.160 --> 36:44.120 |
|
they view the coverage of cell access and 3G, 4G access |
|
|
|
36:44.120 --> 36:47.080 |
|
to be a governmental infrastructure spending, |
|
|
|
36:47.080 --> 36:49.880 |
|
as opposed to capitalistic. |
|
|
|
36:49.880 --> 36:52.160 |
|
So of course, the state or enterprise |
|
|
|
36:52.160 --> 36:55.000 |
|
is also publicly traded, but they also |
|
|
|
36:55.000 --> 36:57.720 |
|
carry a government responsibility |
|
|
|
36:57.720 --> 37:00.240 |
|
to deliver infrastructure to all. |
|
|
|
37:00.240 --> 37:01.880 |
|
So it's a different way of thinking |
|
|
|
37:01.880 --> 37:05.400 |
|
that may be very hard to inject into Western countries |
|
|
|
37:05.400 --> 37:09.280 |
|
to say starting tomorrow, bandwidth infrastructure |
|
|
|
37:09.280 --> 37:13.840 |
|
and highways are going to be governmental spending |
|
|
|
37:13.840 --> 37:16.240 |
|
with some characteristics. |
|
|
|
37:16.240 --> 37:18.240 |
|
What's your sense, and sorry to interrupt, |
|
|
|
37:18.240 --> 37:21.680 |
|
but because it's such a fascinating point, |
|
|
|
37:21.680 --> 37:25.600 |
|
do you think on the autonomous vehicle space |
|
|
|
37:25.600 --> 37:30.120 |
|
it's possible to solve the problem of full autonomy |
|
|
|
37:30.120 --> 37:34.040 |
|
without significant investment in infrastructure? |
|
|
|
37:34.040 --> 37:36.400 |
|
Well, that's really hard to speculate. |
|
|
|
37:36.400 --> 37:38.960 |
|
I think it's not a yes, no question, |
|
|
|
37:38.960 --> 37:41.920 |
|
but how long does it take question? |
|
|
|
37:41.920 --> 37:45.120 |
|
15 years, 30 years, 45 years. |
|
|
|
37:45.120 --> 37:48.960 |
|
Clearly with infrastructure augmentation, |
|
|
|
37:48.960 --> 37:52.320 |
|
where there's road, the city, or whole city planning, |
|
|
|
37:52.320 --> 37:56.440 |
|
building a new city, I'm sure that will accelerate |
|
|
|
37:56.440 --> 37:59.040 |
|
the day of the L5. |
|
|
|
37:59.040 --> 38:01.520 |
|
I'm not knowledgeable enough, and it's |
|
|
|
38:01.520 --> 38:03.920 |
|
hard to predict even when we're knowledgeable, |
|
|
|
38:03.920 --> 38:07.120 |
|
because a lot of it is speculative. |
|
|
|
38:07.120 --> 38:09.800 |
|
But in the US, I don't think people |
|
|
|
38:09.800 --> 38:13.240 |
|
would consider building a new city the size of Chicago |
|
|
|
38:13.240 --> 38:15.920 |
|
to make it the AI slash autonomous city. |
|
|
|
38:15.920 --> 38:18.840 |
|
There are smaller ones being built, I'm aware of that. |
|
|
|
38:18.840 --> 38:21.280 |
|
But is infrastructure spend really |
|
|
|
38:21.280 --> 38:23.720 |
|
impossible for US or Western countries? |
|
|
|
38:23.720 --> 38:25.680 |
|
I don't think so. |
|
|
|
38:25.680 --> 38:28.920 |
|
The US highway system was built. |
|
|
|
38:28.920 --> 38:31.960 |
|
Was that during President Eisenhower or Kennedy? |
|
|
|
38:31.960 --> 38:33.160 |
|
Eisenhower, yeah. |
|
|
|
38:33.160 --> 38:38.960 |
|
So maybe historians can study how the President Eisenhower |
|
|
|
38:38.960 --> 38:42.960 |
|
get the resources to build this massive infrastructure that |
|
|
|
38:42.960 --> 38:47.560 |
|
surely gave US a tremendous amount of prosperity |
|
|
|
38:47.560 --> 38:50.800 |
|
over the next decade, if not century. |
|
|
|
38:50.800 --> 38:53.240 |
|
If I may comment on that, then, it |
|
|
|
38:53.240 --> 38:54.880 |
|
takes us to artificial intelligence |
|
|
|
38:54.880 --> 38:58.080 |
|
a little bit, because in order to build infrastructure, |
|
|
|
38:58.080 --> 39:00.520 |
|
it creates a lot of jobs. |
|
|
|
39:00.520 --> 39:02.840 |
|
So I'll be actually interested if you |
|
|
|
39:02.840 --> 39:06.120 |
|
would say that you're talking in your book about all kinds |
|
|
|
39:06.120 --> 39:08.960 |
|
of jobs that could and could not be automated. |
|
|
|
39:08.960 --> 39:12.000 |
|
I wonder if building infrastructure |
|
|
|
39:12.000 --> 39:15.720 |
|
is one of the jobs that would not be easily automated, |
|
|
|
39:15.720 --> 39:18.160 |
|
something you can think about, because I think you've mentioned |
|
|
|
39:18.160 --> 39:21.160 |
|
somewhere in a talk, or that there |
|
|
|
39:21.160 --> 39:24.280 |
|
might be, as jobs are being automated, |
|
|
|
39:24.280 --> 39:28.160 |
|
a role for government to create jobs that can't be automated. |
|
|
|
39:28.160 --> 39:31.040 |
|
Yes, I think that's a possibility. |
|
|
|
39:31.040 --> 39:34.280 |
|
Back in the last financial crisis, |
|
|
|
39:34.280 --> 39:40.320 |
|
China put a lot of money to basically give this economy |
|
|
|
39:40.320 --> 39:45.520 |
|
a boost, and a lot of it went into infrastructure building. |
|
|
|
39:45.520 --> 39:49.920 |
|
And I think that's a legitimate way, at the government level, |
|
|
|
39:49.920 --> 39:55.680 |
|
to deal with the employment issues as well as build out |
|
|
|
39:55.680 --> 39:58.960 |
|
the infrastructure, as long as the infrastructures are truly |
|
|
|
39:58.960 --> 40:03.160 |
|
needed, and as long as there is an employment problem, which |
|
|
|
40:03.160 --> 40:04.960 |
|
we don't know. |
|
|
|
40:04.960 --> 40:07.920 |
|
So maybe taking a little step back, |
|
|
|
40:07.920 --> 40:12.840 |
|
if you've been a leader and a researcher in AI |
|
|
|
40:12.840 --> 40:16.200 |
|
for several decades, at least 30 years, |
|
|
|
40:16.200 --> 40:21.040 |
|
so how has AI changed in the West and the East |
|
|
|
40:21.040 --> 40:23.120 |
|
as you've observed, as you've been deep in it |
|
|
|
40:23.120 --> 40:25.120 |
|
over the past 30 years? |
|
|
|
40:25.120 --> 40:28.520 |
|
Well, AI began as the pursuit of understanding |
|
|
|
40:28.520 --> 40:34.160 |
|
human intelligence, and the term itself represents that. |
|
|
|
40:34.160 --> 40:37.680 |
|
But it kind of drifted into the one subarea that |
|
|
|
40:37.680 --> 40:40.880 |
|
worked extremely well, which is machine intelligence. |
|
|
|
40:40.880 --> 40:45.080 |
|
And that's actually more using pattern recognition techniques |
|
|
|
40:45.080 --> 40:51.280 |
|
to basically do incredibly well on a limited domain, |
|
|
|
40:51.280 --> 40:54.840 |
|
large amount of data, but relatively simple kinds |
|
|
|
40:54.840 --> 40:58.720 |
|
of planning, tasks, and not very creative. |
|
|
|
40:58.720 --> 41:02.480 |
|
So we didn't end up building human intelligence. |
|
|
|
41:02.480 --> 41:04.760 |
|
We built a different machine that |
|
|
|
41:04.760 --> 41:08.040 |
|
was a lot better than us, some problems, |
|
|
|
41:08.040 --> 41:11.840 |
|
but nowhere close to us on other problems. |
|
|
|
41:11.840 --> 41:14.200 |
|
So today, I think a lot of people still |
|
|
|
41:14.200 --> 41:18.080 |
|
misunderstand when we say artificial intelligence |
|
|
|
41:18.080 --> 41:20.720 |
|
and what various products can do. |
|
|
|
41:20.720 --> 41:24.160 |
|
People still think it's about replicating human intelligence. |
|
|
|
41:24.160 --> 41:26.160 |
|
But the products out there really |
|
|
|
41:26.160 --> 41:31.680 |
|
are closer to having invented the internet or the spreadsheet |
|
|
|
41:31.680 --> 41:35.360 |
|
or the database and getting broader adoption. |
|
|
|
41:35.360 --> 41:38.400 |
|
And speaking further to the fears, near term fears |
|
|
|
41:38.400 --> 41:41.240 |
|
that people have about AI, so you're commenting |
|
|
|
41:41.240 --> 41:45.680 |
|
on the general intelligence that people |
|
|
|
41:45.680 --> 41:48.040 |
|
in the popular culture from sci fi movies |
|
|
|
41:48.040 --> 41:50.920 |
|
have a sense about AI, but there's practical fears |
|
|
|
41:50.920 --> 41:54.800 |
|
about AI, the kind of narrow AI that you're talking about |
|
|
|
41:54.800 --> 41:57.280 |
|
of automating particular kinds of jobs, |
|
|
|
41:57.280 --> 41:59.400 |
|
and you talk about them in the book. |
|
|
|
41:59.400 --> 42:01.520 |
|
So what are the kinds of jobs in your view |
|
|
|
42:01.520 --> 42:04.840 |
|
that you see in the next five, 10 years beginning |
|
|
|
42:04.840 --> 42:09.240 |
|
to be automated by AI systems algorithms? |
|
|
|
42:09.240 --> 42:13.000 |
|
Yes, this is also maybe a little bit counterintuitive |
|
|
|
42:13.000 --> 42:15.440 |
|
because it's the routine jobs that |
|
|
|
42:15.440 --> 42:18.360 |
|
will be displaced the soonest. |
|
|
|
42:18.360 --> 42:23.120 |
|
And they may not be displaced entirely, maybe 50%, 80% |
|
|
|
42:23.120 --> 42:26.320 |
|
of a job, but when the workload drops by that much, |
|
|
|
42:26.320 --> 42:28.760 |
|
employment will come down. |
|
|
|
42:28.760 --> 42:31.520 |
|
And also another part of misunderstanding |
|
|
|
42:31.520 --> 42:35.720 |
|
is most people think of AI replacing routine jobs, |
|
|
|
42:35.720 --> 42:38.760 |
|
then they think of the assembly line, the workers. |
|
|
|
42:38.760 --> 42:40.960 |
|
Well, that will have some effects, |
|
|
|
42:40.960 --> 42:44.600 |
|
but it's actually the routine white collar workers that's |
|
|
|
42:44.600 --> 42:49.280 |
|
easiest to replace because to replace a white collar worker, |
|
|
|
42:49.280 --> 42:50.720 |
|
you just need software. |
|
|
|
42:50.720 --> 42:53.120 |
|
To replace a blue collar worker, |
|
|
|
42:53.120 --> 42:57.200 |
|
you need robotics, mechanical excellence, |
|
|
|
42:57.200 --> 43:01.880 |
|
and the ability to deal with dexterity, |
|
|
|
43:01.880 --> 43:05.640 |
|
and maybe even unknown environments, very, very difficult. |
|
|
|
43:05.640 --> 43:11.200 |
|
So if we were to categorize the most dangerous white collar |
|
|
|
43:11.200 --> 43:15.600 |
|
jobs, they would be things like back office, |
|
|
|
43:15.600 --> 43:20.800 |
|
people who copy and paste and deal with simple computer |
|
|
|
43:20.800 --> 43:25.560 |
|
programs and data, and maybe paper and OCR, |
|
|
|
43:25.560 --> 43:29.000 |
|
and they don't make strategic decisions, |
|
|
|
43:29.000 --> 43:32.040 |
|
they basically facilitate the process. |
|
|
|
43:32.040 --> 43:34.680 |
|
These software and paper systems don't work, |
|
|
|
43:34.680 --> 43:40.520 |
|
so you have people dealing with new employee orientation, |
|
|
|
43:40.520 --> 43:45.400 |
|
searching for past lawsuits and financial documents, |
|
|
|
43:45.400 --> 43:49.800 |
|
and doing reference check, so basic searching and management |
|
|
|
43:49.800 --> 43:52.800 |
|
of data that's the most in danger of being lost. |
|
|
|
43:52.800 --> 43:56.440 |
|
In addition to the white collar repetitive work, |
|
|
|
43:56.440 --> 43:59.360 |
|
a lot of simple interaction work can also |
|
|
|
43:59.360 --> 44:02.840 |
|
be taken care of, such as tele sales, telemarketing, |
|
|
|
44:02.840 --> 44:07.280 |
|
customer service, as well as many physical jobs |
|
|
|
44:07.280 --> 44:09.880 |
|
that are in the same location and don't |
|
|
|
44:09.880 --> 44:12.240 |
|
require a high degree of dexterity, |
|
|
|
44:12.240 --> 44:17.840 |
|
so fruit picking, dishwashing, assembly line, inspection, |
|
|
|
44:17.840 --> 44:20.360 |
|
our jobs in that category. |
|
|
|
44:20.360 --> 44:25.440 |
|
So altogether, back office is a big part, |
|
|
|
44:25.440 --> 44:29.840 |
|
and the other, the blue collar may be smaller initially, |
|
|
|
44:29.840 --> 44:32.560 |
|
but over time, AI will get better. |
|
|
|
44:32.560 --> 44:36.880 |
|
And when we start to get to over the next 15, 20 years, |
|
|
|
44:36.880 --> 44:39.120 |
|
the ability to actually have the dexterity |
|
|
|
44:39.120 --> 44:42.600 |
|
of doing assembly line, that's a huge chunk of jobs. |
|
|
|
44:42.600 --> 44:44.760 |
|
And when autonomous vehicles start |
|
|
|
44:44.760 --> 44:47.400 |
|
to work initially starting with truck drivers, |
|
|
|
44:47.400 --> 44:49.640 |
|
but eventually to all drivers, that's |
|
|
|
44:49.640 --> 44:52.040 |
|
another huge group of workers. |
|
|
|
44:52.040 --> 44:55.560 |
|
So I see modest numbers in the next five years, |
|
|
|
44:55.560 --> 44:58.080 |
|
but increasing rapidly after that. |
|
|
|
44:58.080 --> 45:01.240 |
|
On the worry of the jobs that are in danger |
|
|
|
45:01.240 --> 45:04.320 |
|
and the gradual loss of jobs, I'm not |
|
|
|
45:04.320 --> 45:06.680 |
|
sure if you're familiar with Andrew Yang. |
|
|
|
45:06.680 --> 45:07.800 |
|
Yes, I am. |
|
|
|
45:07.800 --> 45:10.560 |
|
So there's a candidate for president of the United States |
|
|
|
45:10.560 --> 45:14.960 |
|
whose platform, Andrew Yang, is based around, in part, |
|
|
|
45:14.960 --> 45:17.680 |
|
around job loss due to automation, |
|
|
|
45:17.680 --> 45:21.120 |
|
and also, in addition, the need, perhaps, |
|
|
|
45:21.120 --> 45:26.120 |
|
of universal basic income to support jobs that are folks who |
|
|
|
45:26.120 --> 45:28.560 |
|
lose their job due to automation and so on, |
|
|
|
45:28.560 --> 45:31.960 |
|
and in general, support people under complex, |
|
|
|
45:31.960 --> 45:34.320 |
|
unstable job market. |
|
|
|
45:34.320 --> 45:36.720 |
|
So what are your thoughts about his concerns, |
|
|
|
45:36.720 --> 45:40.000 |
|
him as a candidate, his ideas in general? |
|
|
|
45:40.000 --> 45:44.600 |
|
I think his thinking is generally in the right direction, |
|
|
|
45:44.600 --> 45:48.440 |
|
but his approach as a presidential candidate |
|
|
|
45:48.440 --> 45:52.240 |
|
may be a little bit ahead at the time. |
|
|
|
45:52.240 --> 45:56.080 |
|
I think the displacements will happen, |
|
|
|
45:56.080 --> 45:58.280 |
|
but will they happen soon enough for people |
|
|
|
45:58.280 --> 46:00.480 |
|
to agree to vote for him? |
|
|
|
46:00.480 --> 46:03.760 |
|
The unemployment numbers are not very high yet. |
|
|
|
46:03.760 --> 46:07.600 |
|
And I think he and I have the same challenge. |
|
|
|
46:07.600 --> 46:11.520 |
|
If I want to theoretically convince people this is an issue |
|
|
|
46:11.520 --> 46:13.880 |
|
and he wants to become the president, |
|
|
|
46:13.880 --> 46:17.760 |
|
people have to see how can this be the case when |
|
|
|
46:17.760 --> 46:19.680 |
|
unemployment numbers are low. |
|
|
|
46:19.680 --> 46:21.360 |
|
So that is the challenge. |
|
|
|
46:21.360 --> 46:27.360 |
|
And I think I do agree with him on the displacement issue, |
|
|
|
46:27.360 --> 46:32.280 |
|
on universal basic income, at a very vanilla level. |
|
|
|
46:32.280 --> 46:36.800 |
|
I don't agree with it because I think the main issue |
|
|
|
46:36.800 --> 46:38.320 |
|
is retraining. |
|
|
|
46:38.320 --> 46:43.200 |
|
So people need to be incented not by just giving a monthly |
|
|
|
46:43.200 --> 46:47.160 |
|
$2,000 check or $1,000 check and do whatever they want |
|
|
|
46:47.160 --> 46:50.920 |
|
because they don't have the know how |
|
|
|
46:50.920 --> 46:56.840 |
|
to know what to retrain to go into what type of a job |
|
|
|
46:56.840 --> 46:58.640 |
|
and guidance is needed. |
|
|
|
46:58.640 --> 47:01.720 |
|
And retraining is needed because historically |
|
|
|
47:01.720 --> 47:05.080 |
|
in technology revolutions, when routine jobs were displaced, |
|
|
|
47:05.080 --> 47:06.920 |
|
new routine jobs came up. |
|
|
|
47:06.920 --> 47:09.400 |
|
So there was always room for that. |
|
|
|
47:09.400 --> 47:12.640 |
|
But with AI and automation, the whole point |
|
|
|
47:12.640 --> 47:15.320 |
|
is replacing all routine jobs eventually. |
|
|
|
47:15.320 --> 47:17.840 |
|
So there will be fewer and fewer routine jobs. |
|
|
|
47:17.840 --> 47:22.640 |
|
And AI will create jobs, but it won't create routine jobs |
|
|
|
47:22.640 --> 47:24.840 |
|
because if it creates routine jobs, |
|
|
|
47:24.840 --> 47:26.880 |
|
why wouldn't AI just do it? |
|
|
|
47:26.880 --> 47:30.360 |
|
So therefore, the people who are losing the jobs |
|
|
|
47:30.360 --> 47:32.280 |
|
are losing routine jobs. |
|
|
|
47:32.280 --> 47:35.720 |
|
The jobs that are becoming available are nonroutine jobs. |
|
|
|
47:35.720 --> 47:39.320 |
|
So the social stipend needs to be put in place |
|
|
|
47:39.320 --> 47:42.040 |
|
is for the routine workers who lost their jobs |
|
|
|
47:42.040 --> 47:46.120 |
|
to be retrained maybe in six months, maybe in three years. |
|
|
|
47:46.120 --> 47:48.560 |
|
Takes a while to retrain on the nonroutine job |
|
|
|
47:48.560 --> 47:51.360 |
|
and then take on a job that will last |
|
|
|
47:51.360 --> 47:53.400 |
|
for that person's lifetime. |
|
|
|
47:53.400 --> 47:56.160 |
|
Now, having said that, if you look deeply |
|
|
|
47:56.160 --> 47:58.240 |
|
into Andrew's document, he does cater for that. |
|
|
|
47:58.240 --> 48:03.240 |
|
So I'm not disagreeing with what he's trying to do. |
|
|
|
48:03.280 --> 48:06.360 |
|
But for simplification, sometimes he just says UBI, |
|
|
|
48:06.360 --> 48:08.760 |
|
but simple UBI wouldn't work. |
|
|
|
48:08.760 --> 48:10.600 |
|
And I think you've mentioned elsewhere |
|
|
|
48:10.600 --> 48:15.600 |
|
that the goal isn't necessarily to give people enough money |
|
|
|
48:15.760 --> 48:19.120 |
|
to survive or live or even to prosper. |
|
|
|
48:19.120 --> 48:22.800 |
|
The point is to give them a job that gives them meaning. |
|
|
|
48:22.800 --> 48:25.600 |
|
That meaning is extremely important. |
|
|
|
48:25.600 --> 48:28.600 |
|
That our employment, at least in the United States |
|
|
|
48:28.600 --> 48:31.200 |
|
and perhaps it cares across the world, |
|
|
|
48:31.200 --> 48:34.600 |
|
provides something that's, forgive me for saying, |
|
|
|
48:34.600 --> 48:36.960 |
|
greater than money, it provides meaning. |
|
|
|
48:38.400 --> 48:43.400 |
|
So now what kind of jobs do you think can't be automated? |
|
|
|
48:44.840 --> 48:46.600 |
|
You talk a little bit about creativity |
|
|
|
48:46.600 --> 48:48.200 |
|
and compassion in your book. |
|
|
|
48:48.200 --> 48:50.720 |
|
What aspects do you think it's difficult |
|
|
|
48:50.720 --> 48:52.320 |
|
to automate for an AI system? |
|
|
|
48:52.320 --> 48:57.320 |
|
Because an AI system is currently merely optimizing. |
|
|
|
48:57.360 --> 49:00.120 |
|
It's not able to reason, plan, |
|
|
|
49:00.120 --> 49:02.920 |
|
or think creatively or strategically. |
|
|
|
49:02.920 --> 49:05.320 |
|
It's not able to deal with complex problems. |
|
|
|
49:05.320 --> 49:09.520 |
|
It can't come up with a new problem and solve it. |
|
|
|
49:09.520 --> 49:12.320 |
|
A human needs to find the problem |
|
|
|
49:12.320 --> 49:15.520 |
|
and pose it as an optimization problem, |
|
|
|
49:15.520 --> 49:17.520 |
|
then have the AI work at it. |
|
|
|
49:17.520 --> 49:21.320 |
|
So an AI would have a very hard time |
|
|
|
49:21.320 --> 49:23.320 |
|
discovering a new drug |
|
|
|
49:23.320 --> 49:26.320 |
|
or discovering a new style of painting |
|
|
|
49:27.320 --> 49:30.320 |
|
or dealing with complex tasks |
|
|
|
49:30.320 --> 49:32.320 |
|
such as managing a company |
|
|
|
49:32.320 --> 49:35.320 |
|
that isn't just about optimizing the bottom line, |
|
|
|
49:35.320 --> 49:39.320 |
|
but also about employee satisfaction, corporate brand, |
|
|
|
49:39.320 --> 49:40.320 |
|
and many, many other things. |
|
|
|
49:40.320 --> 49:44.320 |
|
So that is one category of things. |
|
|
|
49:44.320 --> 49:48.320 |
|
And because these things are challenging, creative, complex, |
|
|
|
49:48.320 --> 49:52.320 |
|
doing them creates a higher degree of satisfaction |
|
|
|
49:52.320 --> 49:55.320 |
|
and therefore appealing to our desire for working, |
|
|
|
49:55.320 --> 49:57.320 |
|
which isn't just to make the money, |
|
|
|
49:57.320 --> 49:58.320 |
|
make the ends meet, |
|
|
|
49:58.320 --> 50:00.320 |
|
but also that we've accomplished something |
|
|
|
50:00.320 --> 50:03.320 |
|
that others maybe can't do or can't do as well. |
|
|
|
50:04.320 --> 50:07.320 |
|
Another type of job that is much numerous |
|
|
|
50:07.320 --> 50:09.320 |
|
would be compassionate jobs, |
|
|
|
50:09.320 --> 50:14.320 |
|
jobs that require compassion, empathy, human touch, human trust. |
|
|
|
50:14.320 --> 50:18.320 |
|
AI can't do that because AI is cold, calculating, |
|
|
|
50:18.320 --> 50:22.320 |
|
and even if it can fake that to some extent, |
|
|
|
50:22.320 --> 50:26.320 |
|
it will make errors and that will make it look very silly. |
|
|
|
50:26.320 --> 50:29.320 |
|
And also, I think even if AI did okay, |
|
|
|
50:29.320 --> 50:33.320 |
|
people would want to interact with another person, |
|
|
|
50:33.320 --> 50:38.320 |
|
whether it's for some kind of a service or a teacher or a doctor |
|
|
|
50:38.320 --> 50:41.320 |
|
or a concierge or a masseuse or bartender. |
|
|
|
50:41.320 --> 50:46.320 |
|
There are so many jobs where people just don't want to interact |
|
|
|
50:46.320 --> 50:49.320 |
|
with a cold robot or software. |
|
|
|
50:50.320 --> 50:53.320 |
|
I've had an entrepreneur who built an elderly care robot |
|
|
|
50:53.320 --> 50:58.320 |
|
and they found that the elderly really only use it for customer service. |
|
|
|
50:58.320 --> 51:00.320 |
|
But not to service the product, |
|
|
|
51:00.320 --> 51:05.320 |
|
but they click on customer service and the video of a person comes up |
|
|
|
51:05.320 --> 51:07.320 |
|
and then the person says, |
|
|
|
51:07.320 --> 51:11.320 |
|
how come my daughter didn't call me? Let me show you a picture of her grandkids. |
|
|
|
51:11.320 --> 51:15.320 |
|
So people earn for that, people people interaction. |
|
|
|
51:15.320 --> 51:19.320 |
|
So even if robots improved, people just don't want it. |
|
|
|
51:19.320 --> 51:21.320 |
|
And those jobs are going to be increasing |
|
|
|
51:21.320 --> 51:24.320 |
|
because AI will create a lot of value, |
|
|
|
51:24.320 --> 51:29.320 |
|
$16 trillion to the world in next 11 years according to PWC |
|
|
|
51:29.320 --> 51:34.320 |
|
and that will give people money to enjoy services, |
|
|
|
51:34.320 --> 51:39.320 |
|
whether it's eating a gourmet meal or tourism and traveling |
|
|
|
51:39.320 --> 51:41.320 |
|
or having concierge services. |
|
|
|
51:41.320 --> 51:44.320 |
|
The services revolving around, you know, |
|
|
|
51:44.320 --> 51:47.320 |
|
every dollar of that $16 trillion will be tremendous. |
|
|
|
51:47.320 --> 51:52.320 |
|
It will create more opportunities to service the people who did well |
|
|
|
51:52.320 --> 51:55.320 |
|
through AI with things. |
|
|
|
51:55.320 --> 52:01.320 |
|
But even at the same time, the entire society is very much short |
|
|
|
52:01.320 --> 52:05.320 |
|
in need of many service oriented, compassionate oriented jobs. |
|
|
|
52:05.320 --> 52:10.320 |
|
The best example is probably in healthcare services. |
|
|
|
52:10.320 --> 52:15.320 |
|
There's going to be 2 million new jobs, not counting replacement, |
|
|
|
52:15.320 --> 52:20.320 |
|
just brand new incremental jobs in the next six years in healthcare services. |
|
|
|
52:20.320 --> 52:24.320 |
|
That includes nurses orderly in the hospital, |
|
|
|
52:24.320 --> 52:29.320 |
|
elderly care and also at home care. |
|
|
|
52:29.320 --> 52:31.320 |
|
It's particularly lacking. |
|
|
|
52:31.320 --> 52:34.320 |
|
And those jobs are not likely to be filled. |
|
|
|
52:34.320 --> 52:36.320 |
|
So there's likely to be a shortage. |
|
|
|
52:36.320 --> 52:41.320 |
|
And the reason they're not filled is simply because they don't pay very well |
|
|
|
52:41.320 --> 52:47.320 |
|
and that the social status of these jobs are not very good. |
|
|
|
52:47.320 --> 52:52.320 |
|
So they pay about half as much as a heavy equipment operator, |
|
|
|
52:52.320 --> 52:55.320 |
|
which will be replaced a lot sooner. |
|
|
|
52:55.320 --> 52:59.320 |
|
And they pay probably comparably to someone on the assembly line. |
|
|
|
52:59.320 --> 53:03.320 |
|
And so if we're ignoring all the other issues |
|
|
|
53:03.320 --> 53:07.320 |
|
and just think about satisfaction from one's job, |
|
|
|
53:07.320 --> 53:11.320 |
|
someone repetitively doing the same manual action at an assembly line, |
|
|
|
53:11.320 --> 53:14.320 |
|
that can't create a lot of job satisfaction. |
|
|
|
53:14.320 --> 53:17.320 |
|
But someone taking care of a sick person |
|
|
|
53:17.320 --> 53:21.320 |
|
and getting a hug and thank you from that person and the family, |
|
|
|
53:21.320 --> 53:24.320 |
|
I think is quite satisfying. |
|
|
|
53:24.320 --> 53:28.320 |
|
So if only we could fix the pay for service jobs, |
|
|
|
53:28.320 --> 53:33.320 |
|
there are plenty of jobs that require some training or a lot of training |
|
|
|
53:33.320 --> 53:36.320 |
|
for the people coming off the routine jobs to take. |
|
|
|
53:36.320 --> 53:43.320 |
|
We can easily imagine someone who was maybe a cashier at the grocery store, |
|
|
|
53:43.320 --> 53:49.320 |
|
at stores become automated, learns to become a nurse or at home care. |
|
|
|
53:49.320 --> 53:54.320 |
|
Also, I do want to point out the blue collar jobs are going to stay around a bit longer, |
|
|
|
53:54.320 --> 53:57.320 |
|
some of them quite a bit longer. |
|
|
|
53:57.320 --> 54:01.320 |
|
AI cannot be told, go clean an arbitrary home. |
|
|
|
54:01.320 --> 54:03.320 |
|
That's incredibly hard. |
|
|
|
54:03.320 --> 54:07.320 |
|
Arguably is an L5 level of difficulty. |
|
|
|
54:07.320 --> 54:09.320 |
|
And then AI cannot be a good plumber, |
|
|
|
54:09.320 --> 54:12.320 |
|
because plumber is almost like a mini detective |
|
|
|
54:12.320 --> 54:15.320 |
|
that has to figure out where the leak came from. |
|
|
|
54:15.320 --> 54:22.320 |
|
So yet AI probably can be an assembly line and auto mechanic and so on. |
|
|
|
54:22.320 --> 54:26.320 |
|
So one has to study which blue collar jobs are going away |
|
|
|
54:26.320 --> 54:30.320 |
|
and facilitate retraining for the people to go into the ones that won't go away |
|
|
|
54:30.320 --> 54:32.320 |
|
or maybe even will increase. |
|
|
|
54:32.320 --> 54:39.320 |
|
I mean, it is fascinating that it's easier to build a world champion chess player |
|
|
|
54:39.320 --> 54:41.320 |
|
than it is to build a mediocre plumber. |
|
|
|
54:41.320 --> 54:43.320 |
|
Yes, very true. |
|
|
|
54:43.320 --> 54:47.320 |
|
And to AI, and that goes counterintuitive to a lot of people's understanding |
|
|
|
54:47.320 --> 54:49.320 |
|
of what artificial intelligence is. |
|
|
|
54:49.320 --> 54:53.320 |
|
So it sounds, I mean, you're painting a pretty optimistic picture |
|
|
|
54:53.320 --> 54:56.320 |
|
about retraining, about the number of jobs |
|
|
|
54:56.320 --> 55:01.320 |
|
and actually the meaningful nature of those jobs once we automate repetitive tasks. |
|
|
|
55:01.320 --> 55:07.320 |
|
So overall, are you optimistic about the future |
|
|
|
55:07.320 --> 55:11.320 |
|
where much of the repetitive tasks are automated, |
|
|
|
55:11.320 --> 55:15.320 |
|
that there is a lot of room for humans, for the compassionate, |
|
|
|
55:15.320 --> 55:19.320 |
|
for the creative input that only humans can provide? |
|
|
|
55:19.320 --> 55:23.320 |
|
I am optimistic if we start to take action. |
|
|
|
55:23.320 --> 55:27.320 |
|
If we have no action in the next five years, |
|
|
|
55:27.320 --> 55:33.320 |
|
I think it's going to be hard to deal with the devastating losses that will emerge. |
|
|
|
55:33.320 --> 55:39.320 |
|
So if we start thinking about retraining, maybe with the low hanging fruits, |
|
|
|
55:39.320 --> 55:45.320 |
|
explaining to vocational schools why they should train more plumbers than auto mechanics, |
|
|
|
55:45.320 --> 55:53.320 |
|
maybe starting with some government subsidy for corporations to have more training positions. |
|
|
|
55:53.320 --> 55:57.320 |
|
We start to explain to people why retraining is important. |
|
|
|
55:57.320 --> 56:00.320 |
|
We start to think about what the future of education, |
|
|
|
56:00.320 --> 56:04.320 |
|
how that needs to be tweaked for the era of AI. |
|
|
|
56:04.320 --> 56:06.320 |
|
If we start to make incremental progress, |
|
|
|
56:06.320 --> 56:09.320 |
|
and the greater number of people understand, |
|
|
|
56:09.320 --> 56:12.320 |
|
then there's no reason to think we can't deal with this, |
|
|
|
56:12.320 --> 56:16.320 |
|
because this technological revolution is arguably similar to |
|
|
|
56:16.320 --> 56:20.320 |
|
what electricity, industrial revolutions, and internet brought about. |
|
|
|
56:20.320 --> 56:24.320 |
|
Do you think there's a role for policy, for governments to step in |
|
|
|
56:24.320 --> 56:27.320 |
|
to help with policy to create a better world? |
|
|
|
56:27.320 --> 56:32.320 |
|
Absolutely, and the governments don't have to believe |
|
|
|
56:32.320 --> 56:39.320 |
|
that unemployment will go up, and they don't have to believe automation will be this fast to do something. |
|
|
|
56:39.320 --> 56:42.320 |
|
Revamping vocational school would be one example. |
|
|
|
56:42.320 --> 56:47.320 |
|
Another is if there's a big gap in healthcare service employment, |
|
|
|
56:47.320 --> 56:54.320 |
|
and we know that a country's population is growing older and more longevity living older, |
|
|
|
56:54.320 --> 56:59.320 |
|
because people over 80 require five times as much care as those under 80, |
|
|
|
56:59.320 --> 57:04.320 |
|
then it is a good time to incent training programs for elderly care, |
|
|
|
57:04.320 --> 57:07.320 |
|
to find ways to improve the pay. |
|
|
|
57:07.320 --> 57:13.320 |
|
Maybe one way would be to offer as part of Medicare or the equivalent program |
|
|
|
57:13.320 --> 57:18.320 |
|
for people over 80 to be entitled to a few hours of elderly care at home, |
|
|
|
57:18.320 --> 57:21.320 |
|
and then that might be reimbursable, |
|
|
|
57:21.320 --> 57:28.320 |
|
and that will stimulate the service industry around the policy. |
|
|
|
57:28.320 --> 57:32.320 |
|
Do you have concerns about large entities, |
|
|
|
57:32.320 --> 57:38.320 |
|
whether it's governments or companies, controlling the future of AI development in general? |
|
|
|
57:38.320 --> 57:40.320 |
|
So we talked about companies. |
|
|
|
57:40.320 --> 57:48.320 |
|
Do you have a better sense that governments can better represent the interest of the people |
|
|
|
57:48.320 --> 57:54.320 |
|
than companies, or do you believe companies are better at representing the interest of the people? |
|
|
|
57:54.320 --> 57:56.320 |
|
Or is there no easy answer? |
|
|
|
57:56.320 --> 57:59.320 |
|
I don't think there's an easy answer because it's a double edged sword. |
|
|
|
57:59.320 --> 58:06.320 |
|
The companies and governments can provide better services with more access to data and more access to AI, |
|
|
|
58:06.320 --> 58:13.320 |
|
but that also leads to greater power, which can lead to uncontrollable problems, |
|
|
|
58:13.320 --> 58:17.320 |
|
whether it's monopoly or corruption in the government. |
|
|
|
58:17.320 --> 58:24.320 |
|
So I think one has to be careful to look at how much data that companies and governments have, |
|
|
|
58:24.320 --> 58:29.320 |
|
and some kind of checks and balances would be helpful. |
|
|
|
58:29.320 --> 58:33.320 |
|
So again, I come from Russia. |
|
|
|
58:33.320 --> 58:36.320 |
|
There's something called the Cold War. |
|
|
|
58:36.320 --> 58:40.320 |
|
So let me ask a difficult question here, looking at conflict. |
|
|
|
58:40.320 --> 58:45.320 |
|
Steven Pinker wrote a great book that conflict all over the world is decreasing in general. |
|
|
|
58:45.320 --> 58:51.320 |
|
But do you have a sense that having written the book AI Superpowers, |
|
|
|
58:51.320 --> 58:57.320 |
|
do you see a major international conflict potentially arising between major nations, |
|
|
|
58:57.320 --> 59:02.320 |
|
whatever they are, whether it's Russia, China, European nations, United States, |
|
|
|
59:02.320 --> 59:09.320 |
|
or others in the next 10, 20, 50 years around AI, around the digital space, cyber space? |
|
|
|
59:09.320 --> 59:12.320 |
|
Do you worry about that? |
|
|
|
59:12.320 --> 59:19.320 |
|
Is that something we need to think about and try to alleviate or prevent? |
|
|
|
59:19.320 --> 59:22.320 |
|
I believe in greater engagement. |
|
|
|
59:22.320 --> 59:33.320 |
|
A lot of the worries about more powerful AI are based on an arms race metaphor. |
|
|
|
59:33.320 --> 59:41.320 |
|
And when you extrapolate into military kinds of scenarios, |
|
|
|
59:41.320 --> 59:48.320 |
|
AI can automate autonomous weapons that needs to be controlled somehow. |
|
|
|
59:48.320 --> 59:57.320 |
|
And autonomous decision making can lead to not enough time to fix international crises. |
|
|
|
59:57.320 --> 1:00:02.320 |
|
So I actually believe a Cold War mentality would be very dangerous |
|
|
|
1:00:02.320 --> 1:00:07.320 |
|
because should two countries rely on AI to make certain decisions |
|
|
|
1:00:07.320 --> 1:00:11.320 |
|
and they don't even talk to each other, they do their own scenario planning, |
|
|
|
1:00:11.320 --> 1:00:14.320 |
|
then something could easily go wrong. |
|
|
|
1:00:14.320 --> 1:00:24.320 |
|
I think engagement, interaction, some protocols to avoid inadvertent disasters is actually needed. |
|
|
|
1:00:24.320 --> 1:00:28.320 |
|
So it's natural for each country to want to be the best, |
|
|
|
1:00:28.320 --> 1:00:34.320 |
|
whether it's in nuclear technologies or AI or bio. |
|
|
|
1:00:34.320 --> 1:00:40.320 |
|
But I think it's important to realize if each country has a black box AI |
|
|
|
1:00:40.320 --> 1:00:48.320 |
|
and don't talk to each other, that probably presents greater challenges to humanity |
|
|
|
1:00:48.320 --> 1:00:50.320 |
|
than if they interacted. |
|
|
|
1:00:50.320 --> 1:00:56.320 |
|
I think there can still be competition, but with some degree of protocol for interaction. |
|
|
|
1:00:56.320 --> 1:01:01.320 |
|
Just like when there was a nuclear competition, |
|
|
|
1:01:01.320 --> 1:01:07.320 |
|
there were some protocol for deterrence among US, Russia, and China. |
|
|
|
1:01:07.320 --> 1:01:10.320 |
|
And I think that engagement is needed. |
|
|
|
1:01:10.320 --> 1:01:15.320 |
|
So of course, we're still far from AI presenting that kind of danger. |
|
|
|
1:01:15.320 --> 1:01:22.320 |
|
But what I worry the most about is the level of engagement seems to be coming down. |
|
|
|
1:01:22.320 --> 1:01:25.320 |
|
The level of distrust seems to be going up, |
|
|
|
1:01:25.320 --> 1:01:32.320 |
|
especially from the US towards other large countries such as China and Russia. |
|
|
|
1:01:32.320 --> 1:01:34.320 |
|
Is there a way to make that better? |
|
|
|
1:01:34.320 --> 1:01:40.320 |
|
So that's beautifully put, level of engagement and even just basic trust and communication |
|
|
|
1:01:40.320 --> 1:01:52.320 |
|
as opposed to making artificial enemies out of particular countries. |
|
|
|
1:01:52.320 --> 1:02:01.320 |
|
Do you have a sense how we can make it better, actionable items that as a society we can take on? |
|
|
|
1:02:01.320 --> 1:02:10.320 |
|
I'm not an expert at geopolitics, but I would say that we look pretty foolish as humankind |
|
|
|
1:02:10.320 --> 1:02:19.320 |
|
when we are faced with the opportunity to create $16 trillion for humanity. |
|
|
|
1:02:19.320 --> 1:02:29.320 |
|
And yet we're not solving fundamental problems with parts of the world still in poverty. |
|
|
|
1:02:29.320 --> 1:02:34.320 |
|
And for the first time, we have the resources to overcome poverty and hunger. |
|
|
|
1:02:34.320 --> 1:02:38.320 |
|
We're not using it on that, but we're fueling competition among superpowers. |
|
|
|
1:02:38.320 --> 1:02:41.320 |
|
And that's a very unfortunate thing. |
|
|
|
1:02:41.320 --> 1:02:54.320 |
|
If we become utopian for a moment, imagine a benevolent world government that has this $16 trillion |
|
|
|
1:02:54.320 --> 1:03:02.320 |
|
and maybe some AI to figure out how to use it to deal with diseases and problems and hate and things like that. |
|
|
|
1:03:02.320 --> 1:03:04.320 |
|
World would be a lot better off. |
|
|
|
1:03:04.320 --> 1:03:07.320 |
|
So what is wrong with the current world? |
|
|
|
1:03:07.320 --> 1:03:13.320 |
|
I think the people with more skill than I should think about this. |
|
|
|
1:03:13.320 --> 1:03:19.320 |
|
And then the geopolitics issue with superpower competition is one side of the issue. |
|
|
|
1:03:19.320 --> 1:03:29.320 |
|
There's another side which I worry maybe even more, which is as the $16 trillion all gets made by U.S. and China |
|
|
|
1:03:29.320 --> 1:03:34.320 |
|
and a few of the other developed countries, the poorer country will get nothing |
|
|
|
1:03:34.320 --> 1:03:42.320 |
|
because they don't have technology and the wealth disparity and inequality will increase. |
|
|
|
1:03:42.320 --> 1:03:50.320 |
|
So a poorer country with a large population will not only benefit from the AI boom or other technology booms |
|
|
|
1:03:50.320 --> 1:03:57.320 |
|
but they will have their workers who previously had hoped they could do the China model and do outsource manufacturing |
|
|
|
1:03:57.320 --> 1:04:02.320 |
|
or the India model so they could do the outsource process or call center |
|
|
|
1:04:02.320 --> 1:04:05.320 |
|
while all those jobs are going to be gone in 10 or 15 years. |
|
|
|
1:04:05.320 --> 1:04:14.320 |
|
So the individual citizen may be a net liability, I mean financially speaking, to a poorer country |
|
|
|
1:04:14.320 --> 1:04:19.320 |
|
and not an asset to claw itself out of poverty. |
|
|
|
1:04:19.320 --> 1:04:29.320 |
|
So in that kind of situation, these large countries with not much tech are going to be facing a downward spiral |
|
|
|
1:04:29.320 --> 1:04:37.320 |
|
and it's unclear what could be done and then when we look back and say there's $16 trillion being created |
|
|
|
1:04:37.320 --> 1:04:43.320 |
|
and it's all being kept by U.S. China and other developed countries, it just doesn't feel right. |
|
|
|
1:04:43.320 --> 1:04:50.320 |
|
So I hope people who know about geopolitics can find solutions that's beyond my expertise. |
|
|
|
1:04:50.320 --> 1:04:54.320 |
|
So different countries that we've talked about have different value systems. |
|
|
|
1:04:54.320 --> 1:05:02.320 |
|
If you look at the United States to an almost extreme degree, there is an absolute desire for freedom of speech. |
|
|
|
1:05:02.320 --> 1:05:14.320 |
|
If you look at a country where I was raised, that desire just amongst the people is not as elevated as it is to basically fundamental level |
|
|
|
1:05:14.320 --> 1:05:17.320 |
|
to the essence of what it means to be America, right? |
|
|
|
1:05:17.320 --> 1:05:20.320 |
|
And the same is true with China, there's different value systems. |
|
|
|
1:05:20.320 --> 1:05:30.320 |
|
There is some censorship of internet content that China and Russia and many other countries undertake. |
|
|
|
1:05:30.320 --> 1:05:40.320 |
|
Do you see that having effects on innovation, other aspects of some of the tech stuff, AI development we talked about |
|
|
|
1:05:40.320 --> 1:05:52.320 |
|
and maybe from another angle, do you see that changing in different ways over the next 10 years, 20 years, 50 years as China continues to grow |
|
|
|
1:05:52.320 --> 1:05:55.320 |
|
as it does now in its tech innovation? |
|
|
|
1:05:55.320 --> 1:06:08.320 |
|
There's a common belief that full freedom of speech and expression is correlated with creativity, which is correlated with entrepreneurial success. |
|
|
|
1:06:08.320 --> 1:06:15.320 |
|
I think empirically we have seen that is not true and China has been successful. |
|
|
|
1:06:15.320 --> 1:06:25.320 |
|
That's not to say the fundamental values are not right or not the best, but it's just that perfect correlation isn't there. |
|
|
|
1:06:25.320 --> 1:06:36.320 |
|
It's hard to read the tea leaves on opening up or not in any country and I've not been very good at that in my past predictions. |
|
|
|
1:06:36.320 --> 1:06:46.320 |
|
But I do believe every country shares some fundamental value, a lot of fundamental values for the long term. |
|
|
|
1:06:46.320 --> 1:07:02.320 |
|
So, you know, China is drafting its privacy policy for individual citizens and they don't look that different from the American or European ones. |
|
|
|
1:07:02.320 --> 1:07:13.320 |
|
So, people do want to protect their privacy and have the opportunity to express and I think the fundamental values are there. |
|
|
|
1:07:13.320 --> 1:07:21.320 |
|
The question is in the execution and timing, how soon or when will that start to open up? |
|
|
|
1:07:21.320 --> 1:07:31.320 |
|
So, as long as each government knows, ultimately people want that kind of protection, there should be a plan to move towards that. |
|
|
|
1:07:31.320 --> 1:07:35.320 |
|
As to when or how, again, I'm not an expert. |
|
|
|
1:07:35.320 --> 1:07:38.320 |
|
On the point of privacy to me, it's really interesting. |
|
|
|
1:07:38.320 --> 1:07:44.320 |
|
So, AI needs data to create a personalized awesome experience. |
|
|
|
1:07:44.320 --> 1:07:47.320 |
|
I'm just speaking generally in terms of products. |
|
|
|
1:07:47.320 --> 1:07:53.320 |
|
And then we have currently, depending on the age and depending on the demographics of who we're talking about, |
|
|
|
1:07:53.320 --> 1:07:58.320 |
|
some people are more or less concerned about the amount of data they hand over. |
|
|
|
1:07:58.320 --> 1:08:03.320 |
|
So, in your view, how do we get this balance right? |
|
|
|
1:08:03.320 --> 1:08:09.320 |
|
That we provide an amazing experience to people that use products. |
|
|
|
1:08:09.320 --> 1:08:15.320 |
|
You look at Facebook, you know, the more Facebook knows about you, yes, it's scary to say. |
|
|
|
1:08:15.320 --> 1:08:20.320 |
|
The better it can probably, a better experience it can probably create. |
|
|
|
1:08:20.320 --> 1:08:24.320 |
|
So, in your view, how do we get that balance right? |
|
|
|
1:08:24.320 --> 1:08:38.320 |
|
Yes, I think a lot of people have a misunderstanding that it's okay and possible to just rip all the data out from a provider and give it back to you. |
|
|
|
1:08:38.320 --> 1:08:43.320 |
|
So, you can deny them access to further data and still enjoy the services we have. |
|
|
|
1:08:43.320 --> 1:08:48.320 |
|
If we take back all the data, all the services will give us nonsense. |
|
|
|
1:08:48.320 --> 1:08:57.320 |
|
We'll no longer be able to use products that function well in terms of, you know, right ranking, right products, right user experience. |
|
|
|
1:08:57.320 --> 1:09:04.320 |
|
So, yet I do understand we don't want to permit misuse of the data. |
|
|
|
1:09:04.320 --> 1:09:16.320 |
|
From legal policy standpoint, I think there can be severe punishment for those who have egregious misuse of the data. |
|
|
|
1:09:16.320 --> 1:09:19.320 |
|
That's, I think, a good first step. |
|
|
|
1:09:19.320 --> 1:09:27.320 |
|
Actually, China on this aspect has very strong laws about people who sell or give data to other companies. |
|
|
|
1:09:27.320 --> 1:09:40.320 |
|
And that over the past few years, since that law came into effect, pretty much eradicated the illegal distribution sharing of data. |
|
|
|
1:09:40.320 --> 1:09:52.320 |
|
Additionally, I think giving, I think technology is often a very good way to solve technology misuse. |
|
|
|
1:09:52.320 --> 1:09:58.320 |
|
So, can we come up with new technologies that will let us have our cake and eat it too? |
|
|
|
1:09:58.320 --> 1:10:07.320 |
|
People are looking into homomorphic encryption, which is letting you keep the data, have it encrypted and train encrypted data. |
|
|
|
1:10:07.320 --> 1:10:13.320 |
|
Of course, we haven't solved that one yet, but that kind of direction may be worth pursuing. |
|
|
|
1:10:13.320 --> 1:10:22.320 |
|
Also federated learning, which would allow one hospital to train on its hospitals patient data fully because they have a license for that. |
|
|
|
1:10:22.320 --> 1:10:28.320 |
|
And then hospitals would then share their models, not data, but models to create a supra AI. |
|
|
|
1:10:28.320 --> 1:10:30.320 |
|
And that also maybe has some promise. |
|
|
|
1:10:30.320 --> 1:10:39.320 |
|
So I would want to encourage us to be open minded and think of this as not just the policy binary yes no, |
|
|
|
1:10:39.320 --> 1:10:48.320 |
|
but letting the technologists try to find solutions to let us have our cake and eat it too, or have most of our cake and eat most of it too. |
|
|
|
1:10:48.320 --> 1:10:55.320 |
|
Finally, I think giving each end user a choice is important and having transparency is important. |
|
|
|
1:10:55.320 --> 1:11:04.320 |
|
Also, I think that's universal, but the choice you give to the user should not be at a granular level that the user cannot understand. |
|
|
|
1:11:04.320 --> 1:11:12.320 |
|
GDPR today causes all these pop ups of yes, no, will you give this site this right to use this part of your data? |
|
|
|
1:11:12.320 --> 1:11:20.320 |
|
I don't think any user understands what they're saying yes or no to, and I suspect most are just saying yes because they don't understand it. |
|
|
|
1:11:20.320 --> 1:11:30.320 |
|
So while GDPR in its current implementation has lived up to its promise of transparency and user choice, |
|
|
|
1:11:30.320 --> 1:11:39.320 |
|
it implemented it in such a way that really didn't deliver the spirit of GDPR. |
|
|
|
1:11:39.320 --> 1:11:41.320 |
|
It fit the letter, but not the spirit. |
|
|
|
1:11:41.320 --> 1:11:50.320 |
|
So again, I think we need to think about is there a way to fit the spirit of GDPR by using some kind of technology? |
|
|
|
1:11:50.320 --> 1:11:52.320 |
|
Can we have a slider? |
|
|
|
1:11:52.320 --> 1:12:01.320 |
|
That's an AI trying to figure out how much you want to slide between perfect protection security of your personal data |
|
|
|
1:12:01.320 --> 1:12:07.320 |
|
versus high degree of convenience with some risks of not having full privacy. |
|
|
|
1:12:07.320 --> 1:12:11.320 |
|
Each user should have some preference and that gives you the user choice, |
|
|
|
1:12:11.320 --> 1:12:18.320 |
|
but maybe we should turn the problem on its head and ask can there be an AI algorithm that can customize this |
|
|
|
1:12:18.320 --> 1:12:24.320 |
|
because we can understand the slider, but we sure cannot understand every pop up question. |
|
|
|
1:12:24.320 --> 1:12:30.320 |
|
And I think getting that right requires getting the balance between what we talked about earlier, |
|
|
|
1:12:30.320 --> 1:12:36.320 |
|
which is heart and soul versus profit driven decisions and strategy. |
|
|
|
1:12:36.320 --> 1:12:45.320 |
|
I think from my perspective, the best way to make a lot of money in the long term is to keep your heart and soul intact. |
|
|
|
1:12:45.320 --> 1:12:53.320 |
|
I think getting that slider right in the short term may feel like you'll be sacrificing profit, |
|
|
|
1:12:53.320 --> 1:12:59.320 |
|
but in the long term, you'll be getting user trust and providing a great experience. |
|
|
|
1:12:59.320 --> 1:13:01.320 |
|
Do you share that kind of view in general? |
|
|
|
1:13:01.320 --> 1:13:11.320 |
|
Yes, absolutely. I sure would hope there is a way we can do long term projects that really do the right thing. |
|
|
|
1:13:11.320 --> 1:13:16.320 |
|
I think a lot of people who embrace GDPR, their hearts in the right place. |
|
|
|
1:13:16.320 --> 1:13:20.320 |
|
I think they just need to figure out how to build a solution. |
|
|
|
1:13:20.320 --> 1:13:24.320 |
|
I've heard utopians talk about solutions that get me excited, |
|
|
|
1:13:24.320 --> 1:13:29.320 |
|
but not sure how in the current funding environment they can get started, right? |
|
|
|
1:13:29.320 --> 1:13:37.320 |
|
People talk about, imagine this crowdsourced data collection that we all trust, |
|
|
|
1:13:37.320 --> 1:13:45.320 |
|
and then we have these agents that we ask them to ask the trusted agent. |
|
|
|
1:13:45.320 --> 1:13:48.320 |
|
That agent only, that platform. |
|
|
|
1:13:48.320 --> 1:14:02.320 |
|
A trusted joint platform that we all believe is trustworthy that can give us all the close loop personal suggestions |
|
|
|
1:14:02.320 --> 1:14:07.320 |
|
by the new social network, new search engine, new ecommerce engine |
|
|
|
1:14:07.320 --> 1:14:12.320 |
|
that has access to even more of our data, but not directly but indirectly. |
|
|
|
1:14:12.320 --> 1:14:18.320 |
|
I think that general concept of licensing to some trusted engine |
|
|
|
1:14:18.320 --> 1:14:22.320 |
|
and finding a way to trust that engine seems like a great idea, |
|
|
|
1:14:22.320 --> 1:14:27.320 |
|
but if you think how long it's going to take to implement and tweak and develop it right, |
|
|
|
1:14:27.320 --> 1:14:31.320 |
|
as well as to collect all the trust and the data from the people, |
|
|
|
1:14:31.320 --> 1:14:34.320 |
|
it's beyond the current cycle of venture capital. |
|
|
|
1:14:34.320 --> 1:14:37.320 |
|
How do you do that is a big question. |
|
|
|
1:14:37.320 --> 1:14:44.320 |
|
You've recently had a fight with cancer, stage 4 lymphoma, |
|
|
|
1:14:44.320 --> 1:14:54.320 |
|
and in a sort of deep personal level, what did it feel like in the darker moments to face your own mortality? |
|
|
|
1:14:54.320 --> 1:14:57.320 |
|
Well, I've been the workaholic my whole life, |
|
|
|
1:14:57.320 --> 1:15:04.320 |
|
and I've basically worked 9.96, 9am to 9pm, 6 days a week, roughly. |
|
|
|
1:15:04.320 --> 1:15:10.320 |
|
And I didn't really pay a lot of attention to my family, friends, and people who loved me, |
|
|
|
1:15:10.320 --> 1:15:14.320 |
|
and my life revolved around optimizing for work. |
|
|
|
1:15:14.320 --> 1:15:25.320 |
|
While my work was not routine, my optimization really made my life basically a very mechanical process. |
|
|
|
1:15:25.320 --> 1:15:36.320 |
|
But I got a lot of highs out of it because of accomplishments that I thought were really important and dear and the highest priority to me. |
|
|
|
1:15:36.320 --> 1:15:41.320 |
|
But when I faced mortality and the possible death in matter of months, |
|
|
|
1:15:41.320 --> 1:15:45.320 |
|
I suddenly realized that this really meant nothing to me, |
|
|
|
1:15:45.320 --> 1:15:48.320 |
|
that I didn't feel like working for another minute, |
|
|
|
1:15:48.320 --> 1:15:54.320 |
|
that if I had 6 months left in my life, I would spend it all with my loved ones. |
|
|
|
1:15:54.320 --> 1:16:02.320 |
|
And thanking them, giving them love back, and apologizing to them that I lived my life the wrong way. |
|
|
|
1:16:02.320 --> 1:16:11.320 |
|
So that moment of reckoning caused me to really rethink that why we exist in this world |
|
|
|
1:16:11.320 --> 1:16:22.320 |
|
is something that we might be too much shaped by the society to think that success and accomplishments is why we live. |
|
|
|
1:16:22.320 --> 1:16:29.320 |
|
And while that can get you periodic successes and satisfaction, |
|
|
|
1:16:29.320 --> 1:16:35.320 |
|
it's really in them facing death, you see what's truly important to you. |
|
|
|
1:16:35.320 --> 1:16:41.320 |
|
So as a result of going through the challenges with cancer, |
|
|
|
1:16:41.320 --> 1:16:45.320 |
|
I've resolved to live a more balanced lifestyle. |
|
|
|
1:16:45.320 --> 1:16:48.320 |
|
I'm now in remission, knock on wood, |
|
|
|
1:16:48.320 --> 1:16:52.320 |
|
and I'm spending more time with my family. |
|
|
|
1:16:52.320 --> 1:16:54.320 |
|
My wife travels with me. |
|
|
|
1:16:54.320 --> 1:16:57.320 |
|
When my kids need me, I spend more time with them. |
|
|
|
1:16:57.320 --> 1:17:02.320 |
|
And before, I used to prioritize everything around work. |
|
|
|
1:17:02.320 --> 1:17:05.320 |
|
When I had a little bit of time, I would dole it out to my family. |
|
|
|
1:17:05.320 --> 1:17:09.320 |
|
Now, when my family needs something, really needs something, |
|
|
|
1:17:09.320 --> 1:17:12.320 |
|
I drop everything at work and go to them. |
|
|
|
1:17:12.320 --> 1:17:15.320 |
|
And then in the time remaining, I allocate to work. |
|
|
|
1:17:15.320 --> 1:17:18.320 |
|
But one's family is very understanding. |
|
|
|
1:17:18.320 --> 1:17:22.320 |
|
It's not like they will take 50 hours a week from me. |
|
|
|
1:17:22.320 --> 1:17:26.320 |
|
So I'm actually able to still work pretty hard, |
|
|
|
1:17:26.320 --> 1:17:28.320 |
|
maybe 10 hours less per week. |
|
|
|
1:17:28.320 --> 1:17:35.320 |
|
So I realize the most important thing in my life is really love and the people I love. |
|
|
|
1:17:35.320 --> 1:17:38.320 |
|
And I give that the highest priority. |
|
|
|
1:17:38.320 --> 1:17:40.320 |
|
It isn't the only thing I do. |
|
|
|
1:17:40.320 --> 1:17:45.320 |
|
But when that is needed, I put that at the top priority. |
|
|
|
1:17:45.320 --> 1:17:49.320 |
|
And I feel much better and I feel much more balanced. |
|
|
|
1:17:49.320 --> 1:17:56.320 |
|
And I think this also gives a hint as to a life of routine work, |
|
|
|
1:17:56.320 --> 1:17:58.320 |
|
a life of pursuit of numbers. |
|
|
|
1:17:58.320 --> 1:18:03.320 |
|
While my job was not routine, it wasn't pursuit of numbers. |
|
|
|
1:18:03.320 --> 1:18:05.320 |
|
Pursuit of, can I make more money? |
|
|
|
1:18:05.320 --> 1:18:07.320 |
|
Can I fund more great companies? |
|
|
|
1:18:07.320 --> 1:18:09.320 |
|
Can I raise more money? |
|
|
|
1:18:09.320 --> 1:18:13.320 |
|
Can I make sure our VC is ranked higher and higher every year? |
|
|
|
1:18:13.320 --> 1:18:20.320 |
|
This competitive nature of driving for bigger numbers and better numbers |
|
|
|
1:18:20.320 --> 1:18:27.320 |
|
became an endless pursuit of that's mechanical. |
|
|
|
1:18:27.320 --> 1:18:31.320 |
|
And bigger numbers really didn't make me happier. |
|
|
|
1:18:31.320 --> 1:18:36.320 |
|
And faced with death, I realized bigger numbers really meant nothing. |
|
|
|
1:18:36.320 --> 1:18:42.320 |
|
And what was important is that people who have given their heart and their love to me |
|
|
|
1:18:42.320 --> 1:18:45.320 |
|
deserve for me to do the same. |
|
|
|
1:18:45.320 --> 1:18:52.320 |
|
So there's deep profound truth in that, that everyone should hear and internalize. |
|
|
|
1:18:52.320 --> 1:18:56.320 |
|
And that's really powerful for you to say that. |
|
|
|
1:18:56.320 --> 1:19:02.320 |
|
I have to ask sort of a difficult question here. |
|
|
|
1:19:02.320 --> 1:19:07.320 |
|
So I've competed in sports my whole life, looking historically. |
|
|
|
1:19:07.320 --> 1:19:14.320 |
|
I'd like to challenge some aspect of that a little bit on the point of hard work. |
|
|
|
1:19:14.320 --> 1:19:20.320 |
|
That it feels that there are certain aspects that is the greatest, |
|
|
|
1:19:20.320 --> 1:19:26.320 |
|
the most beautiful aspects of human nature, is the ability to become obsessed, |
|
|
|
1:19:26.320 --> 1:19:33.320 |
|
of becoming extremely passionate to the point where, yes, flaws are revealed |
|
|
|
1:19:33.320 --> 1:19:36.320 |
|
and just giving yourself fully to a task. |
|
|
|
1:19:36.320 --> 1:19:41.320 |
|
That is, in another sense, you mentioned love being important, |
|
|
|
1:19:41.320 --> 1:19:47.320 |
|
but in another sense, this kind of obsession, this pure exhibition of passion and hard work |
|
|
|
1:19:47.320 --> 1:19:50.320 |
|
is truly what it means to be human. |
|
|
|
1:19:50.320 --> 1:19:53.320 |
|
What lessons should we take that's deeper? |
|
|
|
1:19:53.320 --> 1:19:55.320 |
|
Because you've accomplished incredible things. |
|
|
|
1:19:55.320 --> 1:19:57.320 |
|
Like chasing numbers. |
|
|
|
1:19:57.320 --> 1:20:01.320 |
|
But really, there's some incredible work there. |
|
|
|
1:20:01.320 --> 1:20:07.320 |
|
So how do you think about that when you look back in your 20s, your 30s? |
|
|
|
1:20:07.320 --> 1:20:10.320 |
|
What would you do differently? |
|
|
|
1:20:10.320 --> 1:20:16.320 |
|
Would you really take back some of the incredible hard work? |
|
|
|
1:20:16.320 --> 1:20:17.320 |
|
I would. |
|
|
|
1:20:17.320 --> 1:20:20.320 |
|
But it's in percentages, right? |
|
|
|
1:20:20.320 --> 1:20:22.320 |
|
We're both now computer scientists. |
|
|
|
1:20:22.320 --> 1:20:27.320 |
|
So I think when one balances one's life, when one is younger, |
|
|
|
1:20:27.320 --> 1:20:33.320 |
|
you might give a smaller percentage to family, but you would still give them high priority. |
|
|
|
1:20:33.320 --> 1:20:38.320 |
|
And when you get older, you would give a larger percentage to them and still the high priority. |
|
|
|
1:20:38.320 --> 1:20:43.320 |
|
And when you're near retirement, you give most of it to them and the highest priority. |
|
|
|
1:20:43.320 --> 1:20:50.320 |
|
So I think the key point is not that we would work 20 hours less for the whole life |
|
|
|
1:20:50.320 --> 1:20:56.320 |
|
and just spend it aimlessly with the family, but that when the family has a need, |
|
|
|
1:20:56.320 --> 1:21:02.320 |
|
when your wife is having a baby, when your daughter has a birthday, |
|
|
|
1:21:02.320 --> 1:21:07.320 |
|
or when they're depressed, or when they're celebrating something, |
|
|
|
1:21:07.320 --> 1:21:11.320 |
|
or when they have a get together, or when we have family time, |
|
|
|
1:21:11.320 --> 1:21:18.320 |
|
that is important for us to put down our phone and PC and be 100% with them. |
|
|
|
1:21:18.320 --> 1:21:26.320 |
|
And that priority on the things that really matter isn't going to be so taxing |
|
|
|
1:21:26.320 --> 1:21:32.320 |
|
that it would eliminate or even dramatically reduce our accomplishments. |
|
|
|
1:21:32.320 --> 1:21:36.320 |
|
It might have some impact, but it might also have other impact |
|
|
|
1:21:36.320 --> 1:21:39.320 |
|
because if you have a happier family, maybe you fight less. |
|
|
|
1:21:39.320 --> 1:21:45.320 |
|
If you fight less, you don't spend time taking care of all the aftermath of a fight. |
|
|
|
1:21:45.320 --> 1:21:46.320 |
|
That's right. |
|
|
|
1:21:46.320 --> 1:21:48.320 |
|
And I'm sure that it would take more time. |
|
|
|
1:21:48.320 --> 1:21:53.320 |
|
And if it did, I'd be willing to take that reduction. |
|
|
|
1:21:53.320 --> 1:21:56.320 |
|
And it's not a dramatic number, but it's a number |
|
|
|
1:21:56.320 --> 1:22:00.320 |
|
that I think would give me a greater degree of happiness |
|
|
|
1:22:00.320 --> 1:22:03.320 |
|
and knowing that I've done the right thing |
|
|
|
1:22:03.320 --> 1:22:10.320 |
|
and still have plenty of hours to get the success that I want to get. |
|
|
|
1:22:10.320 --> 1:22:14.320 |
|
So given the many successful companies that you've launched |
|
|
|
1:22:14.320 --> 1:22:17.320 |
|
and much success throughout your career, |
|
|
|
1:22:17.320 --> 1:22:25.320 |
|
what advice would you give to young people today looking, |
|
|
|
1:22:25.320 --> 1:22:28.320 |
|
or it doesn't have to be young, but people today looking to launch |
|
|
|
1:22:28.320 --> 1:22:35.320 |
|
and to create the next $1 billion tech startup, or even AI based startup? |
|
|
|
1:22:35.320 --> 1:22:42.320 |
|
I would suggest that people understand technology waves move quickly. |
|
|
|
1:22:42.320 --> 1:22:45.320 |
|
What worked two years ago may not work today. |
|
|
|
1:22:45.320 --> 1:22:49.320 |
|
And that is very much a case in point for AI. |
|
|
|
1:22:49.320 --> 1:22:53.320 |
|
I think two years ago, or maybe three years ago, |
|
|
|
1:22:53.320 --> 1:22:57.320 |
|
you certainly could say I have a couple of super smart PhDs |
|
|
|
1:22:57.320 --> 1:22:59.320 |
|
and we're not sure what we're going to do, |
|
|
|
1:22:59.320 --> 1:23:04.320 |
|
but here's how we're going to start and get funding for a very high valuation. |
|
|
|
1:23:04.320 --> 1:23:11.320 |
|
Those days are over because AI is going from rocket science towards mainstream. |
|
|
|
1:23:11.320 --> 1:23:14.320 |
|
Not yet commodity, but more mainstream. |
|
|
|
1:23:14.320 --> 1:23:20.320 |
|
So first, the creation of any company to eventual capitalist |
|
|
|
1:23:20.320 --> 1:23:25.320 |
|
has to be creation of business value and monetary value. |
|
|
|
1:23:25.320 --> 1:23:29.320 |
|
And when you have a very scarce commodity, |
|
|
|
1:23:29.320 --> 1:23:34.320 |
|
VCs may be willing to accept greater uncertainty. |
|
|
|
1:23:34.320 --> 1:23:40.320 |
|
But now the number of people who have the equivalent of PhD three years ago |
|
|
|
1:23:40.320 --> 1:23:43.320 |
|
because that can be learned more quickly. |
|
|
|
1:23:43.320 --> 1:23:45.320 |
|
Platforms are emerging. |
|
|
|
1:23:45.320 --> 1:23:51.320 |
|
The cost to become an AI engineer is much lower and there are many more AI engineers. |
|
|
|
1:23:51.320 --> 1:23:53.320 |
|
So the market is different. |
|
|
|
1:23:53.320 --> 1:23:57.320 |
|
So I would suggest someone who wants to build an AI company |
|
|
|
1:23:57.320 --> 1:24:01.320 |
|
be thinking about the normal business questions. |
|
|
|
1:24:01.320 --> 1:24:05.320 |
|
What customer cases are you trying to address? |
|
|
|
1:24:05.320 --> 1:24:08.320 |
|
What kind of pain are you trying to address? |
|
|
|
1:24:08.320 --> 1:24:10.320 |
|
How does that translate to value? |
|
|
|
1:24:10.320 --> 1:24:16.320 |
|
How will you extract value and get paid through what channel? |
|
|
|
1:24:16.320 --> 1:24:19.320 |
|
And how much business value will get created? |
|
|
|
1:24:19.320 --> 1:24:26.320 |
|
That today needs to be thought about much earlier up front than it did three years ago. |
|
|
|
1:24:26.320 --> 1:24:30.320 |
|
The scarcity question of AI talent has changed. |
|
|
|
1:24:30.320 --> 1:24:32.320 |
|
The number of AI talent has changed. |
|
|
|
1:24:32.320 --> 1:24:41.320 |
|
So now you need not just AI but also understanding of business customer and the marketplace. |
|
|
|
1:24:41.320 --> 1:24:49.320 |
|
So I also think you should have a more reasonable evaluation expectation |
|
|
|
1:24:49.320 --> 1:24:51.320 |
|
and growth expectation. |
|
|
|
1:24:51.320 --> 1:24:53.320 |
|
There's going to be more competition. |
|
|
|
1:24:53.320 --> 1:25:00.320 |
|
But the good news though is that AI technologies are now more available in open source. |
|
|
|
1:25:00.320 --> 1:25:06.320 |
|
TensorFlow, PyTorch and such tools are much easier to use. |
|
|
|
1:25:06.320 --> 1:25:13.320 |
|
So you should be able to experiment and get results iteratively faster than before. |
|
|
|
1:25:13.320 --> 1:25:18.320 |
|
So take more of a business mindset to this. |
|
|
|
1:25:18.320 --> 1:25:25.320 |
|
Think less of this as a laboratory taken into a company because we've gone beyond that stage. |
|
|
|
1:25:25.320 --> 1:25:31.320 |
|
The only exception is if you truly have a breakthrough in some technology that really no one has, |
|
|
|
1:25:31.320 --> 1:25:34.320 |
|
then the old way still works. |
|
|
|
1:25:34.320 --> 1:25:36.320 |
|
But I think that's harder and harder now. |
|
|
|
1:25:36.320 --> 1:25:44.320 |
|
So I know you believe as many do that we're far from creating an artificial general intelligence system. |
|
|
|
1:25:44.320 --> 1:25:54.320 |
|
But say once we do and you get to ask her one question, what would that question be? |
|
|
|
1:25:54.320 --> 1:26:00.320 |
|
What is it that differentiates you and me? |
|
|
|
1:26:00.320 --> 1:26:05.320 |
|
Beautifully put, Kaifu, thank you so much for your time today. |
|
|
|
1:26:05.320 --> 1:26:26.320 |
|
Thank you. |
|
|
|
|