diff --git "a/vtt/episode_027_small.vtt" "b/vtt/episode_027_small.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_027_small.vtt" @@ -0,0 +1,3944 @@ +WEBVTT + +00:00.000 --> 00:02.960 + The following is a conversation with Kai Fu Li. + +00:02.960 --> 00:06.520 + He's the chairman and CEO of Sinovation Ventures + +00:06.520 --> 00:10.560 + that manages a $2 billion dual currency investment fund + +00:10.560 --> 00:13.160 + with a focus on developing the next generation + +00:13.160 --> 00:15.440 + of Chinese high tech companies. + +00:15.440 --> 00:17.840 + He's the former president of Google China + +00:17.840 --> 00:20.880 + and the founder of what is now called Microsoft Research + +00:20.880 --> 00:24.160 + Asia, an institute that trained many + +00:24.160 --> 00:26.520 + of the artificial intelligence leaders in China, + +00:26.520 --> 00:32.080 + including CTOs or AI execs at Baidu, Tencent, Alibaba, + +00:32.080 --> 00:34.840 + Lenovo, and Huawei. + +00:34.840 --> 00:38.520 + He was named one of the 100 most influential people + +00:38.520 --> 00:40.680 + in the world by Time Magazine. + +00:40.680 --> 00:43.880 + He's the author of seven bestselling books in Chinese + +00:43.880 --> 00:47.080 + and most recently, the New York Times bestseller called + +00:47.080 --> 00:50.600 + AI Superpowers, China, Silicon Valley, + +00:50.600 --> 00:52.760 + and the New World Order. + +00:52.760 --> 00:57.200 + He has unparalleled experience in working across major tech + +00:57.200 --> 01:00.120 + companies and governments on applications of AI. + +01:00.120 --> 01:02.440 + And so he has a unique perspective + +01:02.440 --> 01:05.080 + on global innovation in the future of AI + +01:05.080 --> 01:09.000 + that I think is important to listen to and think about. + +01:09.000 --> 01:11.960 + This is the Artificial Intelligence Podcast. + +01:11.960 --> 01:15.240 + If you enjoy it, subscribe on YouTube and iTunes, + +01:15.240 --> 01:18.880 + support it on Patreon, or simply connect with me on Twitter + +01:18.880 --> 01:21.040 + at Lex Freedman. + +01:21.040 --> 01:26.120 + And now, here's my conversation with Kaifu Li. + +01:26.120 --> 01:29.440 + I immigrated from Russia to US when I was 13. + +01:29.440 --> 01:32.480 + You immigrated to US at about the same age. + +01:32.480 --> 01:35.920 + The Russian people, the American people, the Chinese people, + +01:35.920 --> 01:39.440 + each have a certain soul, a spirit, + +01:39.440 --> 01:42.080 + that permeates throughout the generations. + +01:42.080 --> 01:45.120 + So maybe it's a little bit of a poetic question, + +01:45.120 --> 01:49.240 + but could you describe your sense of what + +01:49.240 --> 01:52.080 + defines the Chinese soul? + +01:52.080 --> 01:56.160 + I think the Chinese soul of people today, right, + +01:56.160 --> 02:02.000 + we're talking about people who have had centuries of burden + +02:02.000 --> 02:05.240 + because of the poverty that the country has gone through + +02:05.240 --> 02:10.560 + and suddenly shined with hope of prosperity + +02:10.560 --> 02:13.440 + in the past 40 years as China opened up + +02:13.440 --> 02:16.440 + and embraced market economy. + +02:16.440 --> 02:20.200 + And undoubtedly, there are two sets of pressures + +02:20.200 --> 02:24.160 + on the people, that of the tradition, + +02:24.160 --> 02:28.040 + that of facing difficult situations, + +02:28.040 --> 02:31.160 + and that of hope of wanting to be the first + +02:31.160 --> 02:33.840 + to become successful and wealthy, + +02:33.840 --> 02:38.360 + so that it's a very strong hunger and strong desire + +02:38.360 --> 02:41.160 + and strong work ethic that drives China forward. + +02:41.160 --> 02:43.960 + And is there roots to not just this generation, + +02:43.960 --> 02:47.880 + but before, that's deeper than just + +02:47.880 --> 02:50.080 + the new economic developments? + +02:50.080 --> 02:52.520 + Is there something that's unique to China + +02:52.520 --> 02:54.960 + that you could speak to that's in the people? + +02:54.960 --> 02:56.000 + Yeah. + +02:56.000 --> 03:00.280 + Well, the Chinese tradition is about excellence, + +03:00.280 --> 03:02.680 + dedication, and results. + +03:02.680 --> 03:07.240 + And the Chinese exams and study subjects in schools + +03:07.240 --> 03:11.080 + have traditionally started from memorizing 10,000 characters, + +03:11.080 --> 03:13.600 + not an easy task to start with. + +03:13.600 --> 03:17.640 + And further by memorizing historic philosophers, + +03:17.640 --> 03:19.000 + literature, poetry. + +03:19.000 --> 03:22.480 + So it really is probably the strongest road + +03:22.480 --> 03:26.920 + learning mechanism created to make sure people had good memory + +03:26.920 --> 03:30.080 + and remembered things extremely well. + +03:30.080 --> 03:33.720 + That, I think, at the same time suppresses + +03:33.720 --> 03:37.360 + the breakthrough innovation. + +03:37.360 --> 03:42.520 + And also enhances the speed execution get results. + +03:42.520 --> 03:47.400 + And that, I think, characterizes the historic basis of China. + +03:47.400 --> 03:49.160 + That's interesting, because there's echoes of that + +03:49.160 --> 03:52.080 + in Russian education as well as rote memorization. + +03:52.080 --> 03:53.800 + So you memorize a lot of poetry. + +03:53.800 --> 03:59.240 + I mean, there's just an emphasis on perfection in all forms + +03:59.240 --> 04:02.240 + that's not conducive to perhaps what you're speaking to, + +04:02.240 --> 04:03.640 + which is creativity. + +04:03.640 --> 04:05.640 + But you think that kind of education + +04:05.640 --> 04:09.040 + holds back the innovative spirit that you + +04:09.040 --> 04:10.960 + might see in the United States? + +04:10.960 --> 04:14.840 + Well, it holds back the breakthrough innovative spirit + +04:14.840 --> 04:16.480 + that we see in the United States. + +04:16.480 --> 04:21.880 + But it does not hold back the valuable execution oriented, + +04:21.880 --> 04:26.320 + result oriented value creating engines, which we see China + +04:26.320 --> 04:27.960 + being very successful. + +04:27.960 --> 04:32.320 + So is there a difference between a Chinese AI engineer + +04:32.320 --> 04:35.600 + today and an American AI engineer perhaps rooted + +04:35.600 --> 04:38.320 + in the culture that we just talked about or the education + +04:38.320 --> 04:41.160 + or the very soul of the people or no? + +04:41.160 --> 04:43.720 + And what would your advice be to each + +04:43.720 --> 04:45.520 + if there's a difference? + +04:45.520 --> 04:47.120 + Well, there's a lot that's similar, + +04:47.120 --> 04:51.240 + because AI is about mastering sciences, + +04:51.240 --> 04:54.880 + about using known technologies and trying new things. + +04:54.880 --> 04:59.760 + But it's also about picking from many parts of possible networks + +04:59.760 --> 05:02.920 + to use and different types of parameters to tune. + +05:02.920 --> 05:05.280 + And that part is somewhat rote. + +05:05.280 --> 05:09.040 + And it is also, as anyone who's built AI products, + +05:09.040 --> 05:12.680 + can tell you a lot about cleansing the data. + +05:12.680 --> 05:15.200 + Because AI runs better with more data. + +05:15.200 --> 05:20.160 + And data is generally unstructured, errorful, + +05:20.160 --> 05:22.360 + and unclean. + +05:22.360 --> 05:26.280 + And the effort to clean the data is immense. + +05:26.280 --> 05:32.280 + So I think the better part of the American AI engineering + +05:32.280 --> 05:36.840 + process is to try new things, to do things people haven't done + +05:36.840 --> 05:41.840 + before, and to use technology to solve most, if not all, + +05:41.840 --> 05:43.480 + problems. + +05:43.480 --> 05:47.160 + So to make the algorithm work despite not so great data, + +05:47.160 --> 05:50.680 + find error tolerant ways to deal with the data. + +05:50.680 --> 05:55.960 + The Chinese way would be to basically enumerate, + +05:55.960 --> 05:58.560 + to the fullest extent, all the possible ways + +05:58.560 --> 06:01.000 + by a lot of machines, try lots of different ways + +06:01.000 --> 06:05.320 + to get it to work, and spend a lot of resources and money + +06:05.320 --> 06:07.720 + and time cleaning up data. + +06:07.720 --> 06:11.880 + That means the AI engineer may be writing data cleansing + +06:11.880 --> 06:15.600 + algorithms, working with thousands of people + +06:15.600 --> 06:19.160 + who label or correct or do things with the data. + +06:19.160 --> 06:21.920 + That is the incredible hard work that + +06:21.920 --> 06:24.040 + might lead to better results. + +06:24.040 --> 06:28.240 + So the Chinese engineer would rely on and ask for more and more + +06:28.240 --> 06:31.120 + data and find ways to cleanse them and make them work + +06:31.120 --> 06:34.200 + in the system, and probably less time thinking + +06:34.200 --> 06:39.320 + about new algorithms that can overcome data or other issues. + +06:39.320 --> 06:40.560 + So where's your intuition? + +06:40.560 --> 06:43.160 + What do you think the biggest impact the next 10 years + +06:43.160 --> 06:43.920 + lies? + +06:43.920 --> 06:47.120 + Is it in some breakthrough algorithms? + +06:47.120 --> 06:53.920 + Or is it in just this at scale rigor, a rigorous approach + +06:53.920 --> 06:57.120 + to data, cleaning data, organizing data + +06:57.120 --> 06:58.440 + onto the same algorithms? + +06:58.440 --> 07:02.600 + What do you think the big impact in the applied world is? + +07:02.600 --> 07:04.560 + Well, if you're really in the company + +07:04.560 --> 07:08.400 + and you have to deliver results, using known techniques + +07:08.400 --> 07:12.240 + and enhancing data seems like the more expedient approach + +07:12.240 --> 07:15.640 + that's very low risk and likely to generate + +07:15.640 --> 07:17.200 + better and better results. + +07:17.200 --> 07:20.520 + And that's why the Chinese approach has done quite well. + +07:20.520 --> 07:24.240 + Now, there are a lot of more challenging startups + +07:24.240 --> 07:28.440 + and problems, such as autonomous vehicles, + +07:28.440 --> 07:32.560 + medical diagnosis, that existing algorithms probably + +07:32.560 --> 07:34.240 + won't solve. + +07:34.240 --> 07:38.680 + And that would put the Chinese approach more challenged + +07:38.680 --> 07:43.720 + and give them more breakthrough innovation approach, more + +07:43.720 --> 07:45.440 + of an edge on those kinds of problems. + +07:45.440 --> 07:47.040 + So let me talk to that a little more. + +07:47.040 --> 07:50.960 + So my intuition, personally, is that data + +07:50.960 --> 07:53.680 + can take us extremely far. + +07:53.680 --> 07:56.480 + So you brought up autonomous vehicles and medical diagnosis. + +07:56.480 --> 08:00.080 + So your intuition is that huge amounts of data + +08:00.080 --> 08:04.000 + might not be able to completely help us solve that problem. + +08:04.000 --> 08:04.600 + Right. + +08:04.600 --> 08:08.080 + So breaking that down further, autonomous vehicle, + +08:08.080 --> 08:10.080 + I think huge amounts of data probably + +08:10.080 --> 08:13.360 + will solve trucks driving on highways, which + +08:13.360 --> 08:15.640 + will deliver significant value. + +08:15.640 --> 08:19.320 + And China will probably lead in that. + +08:19.320 --> 08:24.880 + And full L5 autonomous is likely to require new technologies + +08:24.880 --> 08:26.320 + we don't yet know. + +08:26.320 --> 08:30.320 + And that might require academia and great industrial research, + +08:30.320 --> 08:32.480 + both innovating and working together. + +08:32.480 --> 08:35.360 + And in that case, US has an advantage. + +08:35.360 --> 08:37.040 + So the interesting question there is, + +08:37.040 --> 08:39.280 + I don't know if you're familiar on the autonomous vehicle + +08:39.280 --> 08:43.480 + space and the developments with Tesla and Elon Musk, + +08:43.480 --> 08:49.400 + where they are, in fact, a full steam ahead + +08:49.400 --> 08:53.480 + into this mysterious, complex world of full autonomy, L5, + +08:53.480 --> 08:55.080 + L4, L5. + +08:55.080 --> 08:58.800 + And they're trying to solve that purely with data. + +08:58.800 --> 09:00.800 + So the same kind of thing that you're saying + +09:00.800 --> 09:03.200 + is just for highway, which is what a lot of people + +09:03.200 --> 09:07.200 + share your intuition, they're trying to solve with data. + +09:07.200 --> 09:09.320 + It's just to linger on that moment further. + +09:09.320 --> 09:13.600 + Do you think possible for them to achieve success + +09:13.600 --> 09:17.040 + with simply just a huge amount of this training + +09:17.040 --> 09:20.440 + on edge cases, on difficult cases in urban environments, + +09:20.440 --> 09:22.840 + not just highway and so on? + +09:22.840 --> 09:24.480 + I think they'll be very hard. + +09:24.480 --> 09:27.680 + One could characterize Tesla's approach as kind + +09:27.680 --> 09:31.600 + of a Chinese strength approach, gather all the data you can, + +09:31.600 --> 09:34.000 + and hope that will overcome the problems. + +09:34.000 --> 09:38.480 + But in autonomous driving, clearly a lot of the decisions + +09:38.480 --> 09:41.480 + aren't merely solved by aggregating data + +09:41.480 --> 09:43.520 + and having feedback loop. + +09:43.520 --> 09:48.040 + There are things that are more akin to human thinking. + +09:48.040 --> 09:51.680 + And how would those be integrated and built? + +09:51.680 --> 09:54.000 + There has not yet been a lot of success + +09:54.000 --> 09:57.200 + integrating human intelligence or, you know, + +09:57.200 --> 09:58.800 + colored expert systems, if you will, + +09:58.800 --> 10:02.960 + even though that's a taboo word with the machine learning. + +10:02.960 --> 10:05.600 + And the integration of the two types of thinking + +10:05.600 --> 10:07.840 + hasn't yet been demonstrated. + +10:07.840 --> 10:09.600 + And the question is, how much can you + +10:09.600 --> 10:12.440 + push a purely machine learning approach? + +10:12.440 --> 10:15.480 + And of course, Tesla also has an additional constraint + +10:15.480 --> 10:18.520 + that they don't have all the sensors. + +10:18.520 --> 10:21.120 + I know that they think it's foolish to use LIDARS, + +10:21.120 --> 10:25.920 + but that's clearly a one less, very valuable and reliable + +10:25.920 --> 10:29.200 + source of input that they're foregoing, which + +10:29.200 --> 10:32.440 + may also have consequences. + +10:32.440 --> 10:33.840 + I think the advantage, of course, + +10:33.840 --> 10:37.040 + is capturing data that no one has ever seen before. + +10:37.040 --> 10:41.040 + And in some cases, such as computer vision and speech + +10:41.040 --> 10:44.800 + recognition, I have seen Chinese companies accumulate data + +10:44.800 --> 10:47.320 + that's not seen anywhere in the Western world, + +10:47.320 --> 10:50.200 + and they have delivered superior results. + +10:50.200 --> 10:53.720 + But then speech recognition and object recognition + +10:53.720 --> 10:57.080 + are relatively suitable problems for deep learning + +10:57.080 --> 11:02.440 + and don't have the potentially need for the human intelligence + +11:02.440 --> 11:04.440 + analytical planning elements. + +11:04.440 --> 11:06.400 + And the same on the speech recognition side, + +11:06.400 --> 11:09.440 + your intuition that speech recognition and the machine + +11:09.440 --> 11:11.440 + learning approaches to speech recognition + +11:11.440 --> 11:14.600 + won't take us to a conversational system that + +11:14.600 --> 11:19.160 + can pass the Turing test, which is maybe akin to what + +11:19.160 --> 11:20.040 + driving is. + +11:20.040 --> 11:25.120 + So it needs to have something more than just simply simple + +11:25.120 --> 11:27.480 + language understanding, simple language generation. + +11:27.480 --> 11:32.000 + Roughly right, I would say that based on purely machine + +11:32.000 --> 11:35.160 + learning approaches, it's hard to imagine. + +11:35.160 --> 11:40.520 + It could lead to a full conversational experience + +11:40.520 --> 11:44.600 + across arbitrary domains, which is akin to L5. + +11:44.600 --> 11:46.920 + I'm a little hesitant to use the word Turing test, + +11:46.920 --> 11:50.280 + because the original definition was probably too easy. + +11:50.280 --> 11:52.320 + We probably do that. + +11:52.320 --> 11:55.280 + The spirit of the Turing test is what I was referring to. + +11:55.280 --> 11:56.520 + Of course. + +11:56.520 --> 11:59.400 + So you've had major leadership research positions + +11:59.400 --> 12:01.640 + at Apple, Microsoft, Google. + +12:01.640 --> 12:06.320 + So continuing on the discussion of America, Russia, Chinese soul + +12:06.320 --> 12:10.520 + and culture and so on, what is the culture of Silicon + +12:10.520 --> 12:16.400 + Valley in contrast to China and maybe US broadly? + +12:16.400 --> 12:19.920 + And what is the unique culture of each of these three + +12:19.920 --> 12:22.040 + major companies, in your view? + +12:22.040 --> 12:25.120 + I think in aggregate, Silicon Valley companies, + +12:25.120 --> 12:27.200 + we could probably include Microsoft in that, + +12:27.200 --> 12:29.120 + even though they're not in the Valley, + +12:29.120 --> 12:33.960 + is really dream big and have visionary goals + +12:33.960 --> 12:37.920 + and believe that technology will conquer all + +12:37.920 --> 12:42.240 + and also the self confidence and the self entitlement + +12:42.240 --> 12:45.440 + that whatever they produce, the whole world should use + +12:45.440 --> 12:47.240 + and must use. + +12:47.240 --> 12:54.080 + And those are historically important, I think. + +12:54.080 --> 12:59.120 + Steve Jobs's famous quote that he doesn't do focus groups. + +12:59.120 --> 13:02.360 + He looks in the mirror and asks the person in the mirror, + +13:02.360 --> 13:03.520 + what do you want? + +13:03.520 --> 13:07.000 + And that really is an inspirational comment + +13:07.000 --> 13:10.480 + that says the great company shouldn't just ask users + +13:10.480 --> 13:13.240 + what they want, but develop something + +13:13.240 --> 13:16.200 + that users will know they want when they see it, + +13:16.200 --> 13:18.960 + but they could never come up with themselves. + +13:18.960 --> 13:23.880 + I think that is probably the most exhilarating description + +13:23.880 --> 13:26.560 + of what the essence of Silicon Valley is, + +13:26.560 --> 13:31.840 + that this brilliant idea could cause you to build something + +13:31.840 --> 13:35.520 + that couldn't come out of the focus groups or A.B. tests. + +13:35.520 --> 13:38.040 + And iPhone would be an example of that. + +13:38.040 --> 13:40.560 + No one in the age of BlackBerry would write down + +13:40.560 --> 13:43.720 + they want an iPhone or multi touch, a browser, + +13:43.720 --> 13:44.800 + might be another example. + +13:44.800 --> 13:47.520 + No one would say they want that in the days of FTP, + +13:47.520 --> 13:49.440 + but once they see it, they want it. + +13:49.440 --> 13:55.680 + So I think that is what Silicon Valley is best at. + +13:55.680 --> 13:58.920 + But it also came with a lot of success. + +13:58.920 --> 14:01.960 + These products became global platforms, + +14:01.960 --> 14:05.080 + and there were basically no competitors anywhere. + +14:05.080 --> 14:08.400 + And that has also led to a belief + +14:08.400 --> 14:13.240 + that these are the only things that one should do, + +14:13.240 --> 14:17.960 + that companies should not tread on other companies territory, + +14:17.960 --> 14:24.040 + so that a Groupon and a Yelp and an OpenTable + +14:24.040 --> 14:26.240 + and the Grubhub would each feel, + +14:26.240 --> 14:28.520 + okay, I'm not going to do the other companies business + +14:28.520 --> 14:33.280 + because that would not be the pride of innovating + +14:33.280 --> 14:36.920 + what each of these four companies have innovated. + +14:36.920 --> 14:42.720 + But I think the Chinese approach is do whatever it takes to win. + +14:42.720 --> 14:45.000 + And it's a winner take all market. + +14:45.000 --> 14:47.200 + And in fact, in the internet space, + +14:47.200 --> 14:50.840 + the market leader will get predominantly all the value + +14:50.840 --> 14:53.320 + extracted out of the system. + +14:53.320 --> 14:59.600 + And the system isn't just defined as one narrow category, + +14:59.600 --> 15:01.360 + but gets broader and broader. + +15:01.360 --> 15:07.960 + So it's amazing ambition for success and domination + +15:07.960 --> 15:11.760 + of increasingly larger product categories + +15:11.760 --> 15:15.080 + leading to clear market winner status + +15:15.080 --> 15:19.120 + and the opportunity to extract tremendous value. + +15:19.120 --> 15:25.840 + And that develops a practical, result oriented, + +15:25.840 --> 15:31.520 + ultra ambitious winner take all gladiatorial mentality. + +15:31.520 --> 15:37.400 + And if what it takes is to build what the competitors built, + +15:37.400 --> 15:41.920 + essentially a copycat, that can be done without infringing laws. + +15:41.920 --> 15:46.280 + If what it takes is to satisfy a foreign country's need + +15:46.280 --> 15:48.480 + by forking the code base and building something + +15:48.480 --> 15:51.440 + that looks really ugly and different, they'll do it. + +15:51.440 --> 15:56.280 + So it's contrasted very sharply with the Silicon Valley approach. + +15:56.280 --> 16:00.080 + And I think the flexibility and the speed and execution + +16:00.080 --> 16:01.960 + has helped the Chinese approach. + +16:01.960 --> 16:05.040 + And I think the Silicon Valley approach + +16:05.040 --> 16:10.280 + is potentially challenged if every Chinese entrepreneur is + +16:10.280 --> 16:13.200 + learning from the whole world, US and China, + +16:13.200 --> 16:16.280 + and the American entrepreneurs only look internally + +16:16.280 --> 16:19.600 + and write off China as a copycat. + +16:19.600 --> 16:22.880 + And the second part of your question about the three + +16:22.880 --> 16:23.520 + companies. + +16:23.520 --> 16:26.000 + The unique elements of the three companies, perhaps. + +16:26.000 --> 16:26.840 + Yeah. + +16:26.840 --> 16:33.080 + I think Apple represents, while the user, please the user, + +16:33.080 --> 16:38.520 + and the essence of design and brand, + +16:38.520 --> 16:44.080 + and it's the one company and perhaps the only tech company + +16:44.080 --> 16:49.920 + that draws people with a strong, serious desire + +16:49.920 --> 16:53.560 + for the product and the willingness to pay a premium + +16:53.560 --> 16:57.160 + because of the halo effect of the brand, which + +16:57.160 --> 17:00.960 + came from the attention to detail and great respect + +17:00.960 --> 17:03.360 + for user needs. + +17:03.360 --> 17:09.200 + Microsoft represents a platform approach + +17:09.200 --> 17:14.280 + that builds giant products that become very strong modes + +17:14.280 --> 17:17.680 + that others can't do because it's + +17:17.680 --> 17:21.480 + well architected at the bottom level + +17:21.480 --> 17:26.640 + and the work is efficiently delegated to individuals + +17:26.640 --> 17:30.360 + and then the whole product is built + +17:30.360 --> 17:33.560 + by adding small parts that sum together. + +17:33.560 --> 17:37.760 + So it's probably the most effective high tech assembly + +17:37.760 --> 17:40.480 + line that builds a very difficult product + +17:40.480 --> 17:44.800 + that the whole process of doing that + +17:44.800 --> 17:50.800 + is kind of a differentiation and something competitors + +17:50.800 --> 17:52.480 + can't easily repeat. + +17:52.480 --> 17:54.800 + Are there elements of the Chinese approach + +17:54.800 --> 17:59.280 + in the way Microsoft went about assembling those little pieces + +17:59.280 --> 18:03.920 + and essentially dominating the market for a long time? + +18:03.920 --> 18:05.640 + Or do you see those as distinct? + +18:05.640 --> 18:08.240 + I think there are elements that are the same. + +18:08.240 --> 18:10.440 + I think the three American companies + +18:10.440 --> 18:13.880 + that had or have Chinese characteristics, + +18:13.880 --> 18:16.080 + and obviously as well as American characteristics, + +18:16.080 --> 18:20.400 + are Microsoft, Facebook, and Amazon. + +18:20.400 --> 18:21.720 + Yes, that's right, Amazon. + +18:21.720 --> 18:25.560 + Because these are companies that will tenaciously + +18:25.560 --> 18:31.320 + go after adjacent markets, build up strong product offering, + +18:31.320 --> 18:38.200 + and find ways to extract greater value from a sphere that's + +18:38.200 --> 18:39.960 + ever increasing. + +18:39.960 --> 18:43.520 + And they understand the value of the platforms. + +18:43.520 --> 18:45.600 + So that's the similarity. + +18:45.600 --> 18:53.760 + And then with Google, I think it's a genuinely value oriented + +18:53.760 --> 18:56.960 + company that does have a heart and soul + +18:56.960 --> 18:59.760 + and that wants to do great things for the world + +18:59.760 --> 19:06.040 + by connecting information and that has also + +19:06.040 --> 19:13.280 + very strong technology genes and wants to use technology + +19:13.280 --> 19:19.080 + and has found out of the box ways to use technology + +19:19.080 --> 19:23.680 + to deliver incredible value to the end user. + +19:23.680 --> 19:25.240 + We can look at Google, for example. + +19:25.240 --> 19:28.040 + You mentioned heart and soul. + +19:28.040 --> 19:31.840 + There seems to be an element where Google + +19:31.840 --> 19:34.840 + is after making the world better. + +19:34.840 --> 19:36.520 + There's a more positive view. + +19:36.520 --> 19:38.960 + I mean, they used to have the slogan, don't be evil. + +19:38.960 --> 19:43.120 + And Facebook a little bit more has a negative tend to it, + +19:43.120 --> 19:46.000 + at least in the perception of privacy and so on. + +19:46.000 --> 19:51.280 + Do you have a sense of how these different companies can + +19:51.280 --> 19:53.400 + achieve, because you've talked about how much + +19:53.400 --> 19:55.600 + we can make the world better in all these kinds of ways + +19:55.600 --> 19:59.360 + with AI, what is it about a company that can make, + +19:59.360 --> 20:03.200 + give it a heart and soul, gain the trust of the public, + +20:03.200 --> 20:08.000 + and just actually just not be evil and do good for the world? + +20:08.000 --> 20:09.000 + It's really hard. + +20:09.000 --> 20:13.120 + And I think Google has struggled with that. + +20:13.120 --> 20:15.160 + First, they don't do evil. + +20:15.160 --> 20:18.880 + Mantra is very dangerous, because every employee's + +20:18.880 --> 20:20.800 + definition of evil is different. + +20:20.800 --> 20:23.800 + And that has led to some difficult employee situations + +20:23.800 --> 20:25.240 + for them. + +20:25.240 --> 20:29.520 + So I don't necessarily think that's a good value statement. + +20:29.520 --> 20:31.840 + But just watching the kinds of things + +20:31.840 --> 20:36.440 + Google or its parent company Alphabet does in new areas + +20:36.440 --> 20:40.440 + like health care, like eradicating mosquitoes, + +20:40.440 --> 20:42.360 + things that are really not in the business + +20:42.360 --> 20:45.040 + of a internet tech company, I think + +20:45.040 --> 20:47.200 + that shows that there is a heart and soul + +20:47.200 --> 20:53.920 + and desire to do good and willingness to put in the resources + +20:53.920 --> 20:58.280 + to do something when they see it's good, they will pursue it. + +20:58.280 --> 21:00.640 + That doesn't necessarily mean it has + +21:00.640 --> 21:02.520 + all the trust of the users. + +21:02.520 --> 21:06.400 + I realize while most people would view Facebook + +21:06.400 --> 21:09.760 + as the primary target of their recent unhappiness + +21:09.760 --> 21:12.720 + about Silicon Valley companies, many would put Google + +21:12.720 --> 21:14.080 + in that category. + +21:14.080 --> 21:16.800 + And some have named Google's business practices + +21:16.800 --> 21:19.840 + as predatory also. + +21:19.840 --> 21:24.240 + So it's kind of difficult to have the two parts of a body. + +21:24.240 --> 21:28.080 + The brain wants to do what it's supposed to do for a shareholder, + +21:28.080 --> 21:29.280 + maximize profit. + +21:29.280 --> 21:30.880 + And then the heart and soul wants + +21:30.880 --> 21:36.120 + to do good things that may run against what the brain wants to do. + +21:36.120 --> 21:40.320 + So in this complex balancing that these companies have to do, + +21:40.320 --> 21:44.520 + you've mentioned that you're concerned about a future where + +21:44.520 --> 21:47.360 + too few companies like Google, Facebook, Amazon + +21:47.360 --> 21:51.560 + are controlling our data or are controlling too much + +21:51.560 --> 21:53.360 + of our digital lives. + +21:53.360 --> 21:55.400 + Can you elaborate on this concern? + +21:55.400 --> 21:58.640 + Perhaps do you have a better way forward? + +21:58.640 --> 22:05.000 + I think I'm hardly the most vocal complainer of this. + +22:05.000 --> 22:07.280 + There are a lot louder complainers out there. + +22:07.280 --> 22:11.840 + I do observe that having a lot of data + +22:11.840 --> 22:16.120 + does perpetuate their strength and limits + +22:16.120 --> 22:19.400 + competition in many spaces. + +22:19.400 --> 22:24.200 + But I also believe AI is much broader than the internet space. + +22:24.200 --> 22:26.280 + So the entrepreneurial opportunities + +22:26.280 --> 22:30.480 + still exists in using AI to empower + +22:30.480 --> 22:34.160 + financial, retail, manufacturing, education, + +22:34.160 --> 22:35.480 + applications. + +22:35.480 --> 22:39.800 + So I don't think it's quite a case of full monopolistic dominance + +22:39.800 --> 22:43.960 + that totally stifles innovation. + +22:43.960 --> 22:46.400 + But I do believe in their areas of strength + +22:46.400 --> 22:49.760 + it's hard to dislodge them. + +22:49.760 --> 22:53.280 + I don't know if I have a good solution. + +22:53.280 --> 22:57.160 + Probably the best solution is let the entrepreneurial VC + +22:57.160 --> 23:00.840 + ecosystem work well and find all the places that + +23:00.840 --> 23:04.200 + can create the next Google, the next Facebook. + +23:04.200 --> 23:08.560 + So there will always be increasing number of challengers. + +23:08.560 --> 23:11.360 + In some sense, that has happened a little bit. + +23:11.360 --> 23:15.760 + You see Uber, Airbnb having emerged despite the strength + +23:15.760 --> 23:19.040 + of the big three. + +23:19.040 --> 23:22.400 + And I think China as an environment + +23:22.400 --> 23:25.280 + may be more interesting for the emergence. + +23:25.280 --> 23:28.920 + Because if you look at companies between, let's say, + +23:28.920 --> 23:36.320 + $50 to $300 billion, China has emerged more of such companies + +23:36.320 --> 23:39.880 + than the US in the last three to four years. + +23:39.880 --> 23:42.120 + Because of the larger marketplace, + +23:42.120 --> 23:47.000 + because of the more fearless nature of the entrepreneurs. + +23:47.000 --> 23:50.840 + And the Chinese giants are just as powerful as American ones. + +23:50.840 --> 23:52.920 + Tencent Alibaba are very strong. + +23:52.920 --> 23:57.040 + But Bytes Dance has emerged worth $75 billion. + +23:57.040 --> 24:00.120 + And financial, while it's Alibaba affiliated, + +24:00.120 --> 24:03.920 + it's nevertheless independent and worth $150 billion. + +24:03.920 --> 24:08.280 + And so I do think if we start to extend + +24:08.280 --> 24:12.640 + to traditional businesses, we will see very valuable companies. + +24:12.640 --> 24:18.120 + So it's probably not the case that in five or 10 years, + +24:18.120 --> 24:20.920 + we'll still see the whole world with these five companies + +24:20.920 --> 24:22.680 + having such dominance. + +24:22.680 --> 24:26.040 + So you've mentioned a couple of times + +24:26.040 --> 24:27.840 + this fascinating world of entrepreneurship + +24:27.840 --> 24:31.080 + in China of the fearless nature of the entrepreneurs. + +24:31.080 --> 24:32.640 + So can you maybe talk a little bit + +24:32.640 --> 24:35.520 + about what it takes to be an entrepreneur in China? + +24:35.520 --> 24:38.240 + What are the strategies that are undertaken? + +24:38.240 --> 24:41.120 + What are the ways that you success? + +24:41.120 --> 24:43.960 + What is the dynamic of VCF funding, + +24:43.960 --> 24:46.480 + of the way the government helps companies, and so on? + +24:46.480 --> 24:49.520 + What are the interesting aspects here that are distinct from, + +24:49.520 --> 24:52.880 + that are different from the Silicon Valley world + +24:52.880 --> 24:55.240 + of entrepreneurship? + +24:55.240 --> 24:58.080 + Well, many of the listeners probably + +24:58.080 --> 25:03.000 + still would brand Chinese entrepreneur as copycats. + +25:03.000 --> 25:06.120 + And no doubt, 10 years ago, that would not + +25:06.120 --> 25:09.080 + be an inaccurate description. + +25:09.080 --> 25:12.320 + Back 10 years ago, an entrepreneur probably + +25:12.320 --> 25:14.840 + could not get funding if he or she could not + +25:14.840 --> 25:20.400 + describe what product he or she is copying from the US. + +25:20.400 --> 25:23.520 + The first question is, who has proven this business model, + +25:23.520 --> 25:27.200 + which is a nice way of asking, who are you copying? + +25:27.200 --> 25:29.520 + And that reason is understandable, + +25:29.520 --> 25:34.840 + because China had a much lower internet penetration + +25:34.840 --> 25:40.920 + and didn't have enough indigenous experience + +25:40.920 --> 25:43.200 + to build innovative products. + +25:43.200 --> 25:47.600 + And secondly, internet was emerging. + +25:47.600 --> 25:49.800 + Link startup was the way to do things, + +25:49.800 --> 25:52.920 + building a first minimally viable product, + +25:52.920 --> 25:55.320 + and then expanding was the right way to go. + +25:55.320 --> 25:59.480 + And the American successes have given a shortcut + +25:59.480 --> 26:02.840 + that if you build your minimally viable product based + +26:02.840 --> 26:05.040 + on an American product, it's guaranteed + +26:05.040 --> 26:06.720 + to be a decent starting point. + +26:06.720 --> 26:08.400 + Then you tweak it afterwards. + +26:08.400 --> 26:11.720 + So as long as there are no IP infringement, which, + +26:11.720 --> 26:15.080 + as far as I know, there hasn't been in the mobile and AI + +26:15.080 --> 26:19.360 + spaces, that's a much better shortcut. + +26:19.360 --> 26:23.720 + And I think Silicon Valley would view that as still not + +26:23.720 --> 26:29.200 + very honorable, because that's not your own idea to start with. + +26:29.200 --> 26:32.600 + But you can't really, at the same time, + +26:32.600 --> 26:35.160 + believe every idea must be your own + +26:35.160 --> 26:38.120 + and believe in the link startup methodology, + +26:38.120 --> 26:41.880 + because link startup is intended to try many, many things + +26:41.880 --> 26:44.240 + and then converge when that works. + +26:44.240 --> 26:46.720 + And it's meant to be iterated and changed. + +26:46.720 --> 26:51.240 + So finding a decent starting point without legal violations, + +26:51.240 --> 26:55.520 + there should be nothing morally dishonorable about that. + +26:55.520 --> 26:57.080 + So just a quick pause on that. + +26:57.080 --> 27:01.920 + It's fascinating that that's why is that not honorable, right? + +27:01.920 --> 27:04.680 + It's exactly as you formulated. + +27:04.680 --> 27:08.040 + It seems like a perfect start for business + +27:08.040 --> 27:12.440 + is to take a look at Amazon and say, OK, + +27:12.440 --> 27:14.560 + we'll do exactly what Amazon is doing. + +27:14.560 --> 27:16.800 + Let's start there in this particular market. + +27:16.800 --> 27:20.520 + And then let's out innovate them from that starting point. + +27:20.520 --> 27:22.200 + Yes. Come up with new ways. + +27:22.200 --> 27:26.520 + I mean, is it wrong to be, except the word copycat just + +27:26.520 --> 27:28.800 + sounds bad, but is it wrong to be a copycat? + +27:28.800 --> 27:31.640 + It just seems like a smart strategy. + +27:31.640 --> 27:35.800 + But yes, doesn't have a heroic nature to it + +27:35.800 --> 27:42.280 + that Steve Jobs, Elon Musk, sort of in something completely + +27:42.280 --> 27:43.880 + coming up with something completely new. + +27:43.880 --> 27:45.480 + Yeah, I like the way you describe it. + +27:45.480 --> 27:50.440 + It's a nonheroic, acceptable way to start the company. + +27:50.440 --> 27:52.840 + And maybe more expedient. + +27:52.840 --> 27:58.920 + So that's, I think, a baggage for Silicon Valley, + +27:58.920 --> 28:01.320 + that if it doesn't let go, then it + +28:01.320 --> 28:05.160 + may limit the ultimate ceiling of the company. + +28:05.160 --> 28:07.200 + Take Snapchat as an example. + +28:07.200 --> 28:09.840 + I think Evan's brilliant. + +28:09.840 --> 28:11.480 + He built a great product. + +28:11.480 --> 28:14.160 + But he's very proud that he wants + +28:14.160 --> 28:16.800 + to build his own features, not copy others. + +28:16.800 --> 28:21.000 + While Facebook was more willing to copy his features, + +28:21.000 --> 28:23.440 + and you see what happens in the competition. + +28:23.440 --> 28:27.440 + So I think putting that handcuff on the company + +28:27.440 --> 28:31.560 + would limit its ability to reach the maximum potential. + +28:31.560 --> 28:33.800 + So back to the Chinese environment, + +28:33.800 --> 28:38.400 + copying was merely a way to learn from the American masters. + +28:38.400 --> 28:43.480 + Just like if we learned to play piano or painting, + +28:43.480 --> 28:44.560 + you start by copying. + +28:44.560 --> 28:46.160 + You don't start by innovating when + +28:46.160 --> 28:48.200 + you don't have the basic skill sets. + +28:48.200 --> 28:51.040 + So very amazingly, the Chinese entrepreneurs + +28:51.040 --> 28:56.160 + about six years ago started to branch off + +28:56.160 --> 28:59.520 + with these lean startups built on American ideas + +28:59.520 --> 29:02.280 + to build better products than American products. + +29:02.280 --> 29:04.960 + But they did start from the American idea. + +29:04.960 --> 29:08.600 + And today, WeChat is better than WhatsApp. + +29:08.600 --> 29:10.520 + Weibo is better than Twitter. + +29:10.520 --> 29:12.920 + Zihu is better than Quora and so on. + +29:12.920 --> 29:17.000 + So that, I think, is Chinese entrepreneurs + +29:17.000 --> 29:18.480 + going to step two. + +29:18.480 --> 29:21.760 + And then step three is once these entrepreneurs have + +29:21.760 --> 29:23.720 + done one or two of these companies, + +29:23.720 --> 29:27.400 + they now look at the Chinese market and the opportunities + +29:27.400 --> 29:30.600 + and come up with ideas that didn't exist elsewhere. + +29:30.600 --> 29:36.320 + So products like and financial under which includes Alipay, + +29:36.320 --> 29:42.080 + which is mobile payments, and also the financial products + +29:42.080 --> 29:48.560 + for loans built on that, and also in education, VIP kid, + +29:48.560 --> 29:54.880 + and in social video, social network, TikTok, + +29:54.880 --> 29:58.640 + and in social eCommerce, Pinduoduo, + +29:58.640 --> 30:01.720 + and then in ride sharing, Mobike. + +30:01.720 --> 30:05.640 + These are all Chinese innovative products + +30:05.640 --> 30:08.720 + that now are being copied elsewhere. + +30:08.720 --> 30:13.040 + So an additional interesting observation + +30:13.040 --> 30:16.000 + is some of these products are built on unique Chinese + +30:16.000 --> 30:19.360 + demographics, which may not work in the US, + +30:19.360 --> 30:23.160 + but may work very well in Southeast Asia, Africa, + +30:23.160 --> 30:27.840 + and other developing worlds that are a few years behind China. + +30:27.840 --> 30:31.040 + And a few of these products maybe are universal + +30:31.040 --> 30:33.760 + and are getting traction even in the United States, + +30:33.760 --> 30:35.360 + such as TikTok. + +30:35.360 --> 30:42.080 + So this whole ecosystem is supported by VCs + +30:42.080 --> 30:44.920 + as a virtuous cycle, because a large market + +30:44.920 --> 30:49.400 + with innovative entrepreneurs will draw a lot of money + +30:49.400 --> 30:51.560 + and then invest in these companies. + +30:51.560 --> 30:54.480 + As the market gets larger and larger, + +30:54.480 --> 30:58.400 + China market is easily three, four times larger than the US. + +30:58.400 --> 31:01.120 + They will create greater value and greater returns + +31:01.120 --> 31:05.400 + for the VCs, thereby raising even more money. + +31:05.400 --> 31:10.000 + So at Sinovation Ventures, our first fund was $15 million. + +31:10.000 --> 31:12.040 + Our last fund was $500 million. + +31:12.040 --> 31:16.520 + So it reflects the valuation of the companies + +31:16.520 --> 31:19.840 + and our us going multi stage and things like that. + +31:19.840 --> 31:23.840 + It also has government support, but not + +31:23.840 --> 31:26.080 + in the way most Americans would think of it. + +31:26.080 --> 31:29.520 + The government actually leaves the entrepreneurial space + +31:29.520 --> 31:33.200 + as a private enterprise, so the self regulating. + +31:33.200 --> 31:36.200 + And the government would build infrastructures + +31:36.200 --> 31:39.320 + that would around it to make it work better. + +31:39.320 --> 31:41.960 + For example, the mass entrepreneur mass innovation + +31:41.960 --> 31:44.880 + plan builds 8,000 incubators. + +31:44.880 --> 31:48.360 + So the pipeline is very strong to the VCs + +31:48.360 --> 31:49.680 + for autonomous vehicles. + +31:49.680 --> 31:53.280 + The Chinese government is building smart highways + +31:53.280 --> 31:56.680 + with sensors, smart cities that separate pedestrians + +31:56.680 --> 32:01.560 + from cars that may allow initially an inferior autonomous + +32:01.560 --> 32:05.760 + vehicle company to launch a car without increasing, + +32:05.760 --> 32:11.520 + with lower casualty, because the roads or the city is smart. + +32:11.520 --> 32:13.800 + And the Chinese government at local levels + +32:13.800 --> 32:17.360 + would have these guiding funds acting as LPs, + +32:17.360 --> 32:19.400 + passive LPs to funds. + +32:19.400 --> 32:23.240 + And when the fund makes money, part of the money made + +32:23.240 --> 32:27.280 + is given back to the GPs and potentially other LPs + +32:27.280 --> 32:31.960 + to increase everybody's return at the expense + +32:31.960 --> 32:33.680 + of the government's return. + +32:33.680 --> 32:36.360 + So that's an interesting incentive + +32:36.360 --> 32:41.640 + that entrusts the task of choosing entrepreneurs to VCs + +32:41.640 --> 32:43.800 + who are better at it than the government + +32:43.800 --> 32:46.680 + by letting some of the profits move that way. + +32:46.680 --> 32:48.720 + So this is really fascinating, right? + +32:48.720 --> 32:51.800 + So I look at the Russian government as a case study + +32:51.800 --> 32:54.480 + where, let me put it this way, there + +32:54.480 --> 32:58.520 + is no such government driven, large scale + +32:58.520 --> 33:00.840 + support of entrepreneurship. + +33:00.840 --> 33:04.000 + And probably the same is true in the United States. + +33:04.000 --> 33:07.640 + But the entrepreneurs themselves kind of find a way. + +33:07.640 --> 33:11.680 + So maybe in a form of advice or explanation, + +33:11.680 --> 33:15.560 + how did the Chinese government arrive to be this way, + +33:15.560 --> 33:17.680 + so supportive on entrepreneurship, + +33:17.680 --> 33:21.520 + to be in this particular way so forward thinking + +33:21.520 --> 33:23.120 + at such a large scale? + +33:23.120 --> 33:28.280 + And also perhaps, how can we copy it in other countries? + +33:28.280 --> 33:29.800 + How can we encourage other governments, + +33:29.800 --> 33:31.600 + like even the United States government, + +33:31.600 --> 33:33.760 + to support infrastructure for autonomous vehicles + +33:33.760 --> 33:36.040 + in that same kind of way, perhaps? + +33:36.040 --> 33:36.680 + Yes. + +33:36.680 --> 33:44.440 + So these techniques are the result of several key things, + +33:44.440 --> 33:46.480 + some of which may be learnable, some of which + +33:46.480 --> 33:48.440 + may be very hard. + +33:48.440 --> 33:51.080 + One is just trial and error and watching + +33:51.080 --> 33:52.960 + what everyone else is doing. + +33:52.960 --> 33:54.960 + I think it's important to be humble and not + +33:54.960 --> 33:56.920 + feel like you know all the answers. + +33:56.920 --> 33:59.480 + The guiding funds idea came from Singapore, + +33:59.480 --> 34:01.440 + which came from Israel. + +34:01.440 --> 34:06.080 + And China made a few tweaks and turned it into a, + +34:06.080 --> 34:09.600 + because the Chinese cities and government officials kind + +34:09.600 --> 34:11.320 + of compete with each other. + +34:11.320 --> 34:14.640 + Because they all want to make their city more successful, + +34:14.640 --> 34:20.280 + so they can get the next level in their political career. + +34:20.280 --> 34:22.320 + And it's somewhat competitive. + +34:22.320 --> 34:25.200 + So the central government made it a bit of a competition. + +34:25.200 --> 34:26.840 + Everybody has a budget. + +34:26.840 --> 34:29.840 + They can put it on AI, or they can put it on bio, + +34:29.840 --> 34:32.200 + or they can put it on energy. + +34:32.200 --> 34:35.040 + And then whoever gets the results, the city shines, + +34:35.040 --> 34:38.000 + the people are better off, the mayor gets a promotion. + +34:38.000 --> 34:41.680 + So the tools is kind of almost like an entrepreneurial + +34:41.680 --> 34:44.840 + environment for local governments + +34:44.840 --> 34:47.480 + to see who can do a better job. + +34:47.480 --> 34:52.440 + And also, many of them tried different experiments. + +34:52.440 --> 34:58.440 + Some have given award to very smart researchers, + +34:58.440 --> 35:00.840 + just give them money and hope they'll start a company. + +35:00.840 --> 35:05.840 + Some have given money to academic research labs, + +35:05.840 --> 35:08.440 + maybe government research labs, to see + +35:08.440 --> 35:11.920 + if they can spin off some companies from the science + +35:11.920 --> 35:14.040 + lab or something like that. + +35:14.040 --> 35:17.080 + Some have tried to recruit overseas Chinese + +35:17.080 --> 35:18.960 + to come back and start companies. + +35:18.960 --> 35:20.960 + And they've had mixed results. + +35:20.960 --> 35:23.400 + The one that worked the best was the guiding funds. + +35:23.400 --> 35:25.840 + So it's almost like a lean startup idea + +35:25.840 --> 35:29.160 + where people try different things in what works, sticks, + +35:29.160 --> 35:30.600 + and everybody copies. + +35:30.600 --> 35:32.880 + So now every city has a guiding fund. + +35:32.880 --> 35:35.680 + So that's how that came about. + +35:35.680 --> 35:40.400 + The autonomous vehicle and the massive spending + +35:40.400 --> 35:46.080 + in highways and smart cities, that's a Chinese way. + +35:46.080 --> 35:49.480 + It's about building infrastructure to facilitate. + +35:49.480 --> 35:52.840 + It's a clear division of the government's responsibility + +35:52.840 --> 35:55.400 + from the market. + +35:55.400 --> 36:00.560 + The market should do everything in a private freeway. + +36:00.560 --> 36:02.920 + But there are things the market can't afford to do, + +36:02.920 --> 36:04.520 + like infrastructure. + +36:04.520 --> 36:08.000 + So the government always appropriates + +36:08.000 --> 36:12.000 + large amounts of money for infrastructure building. + +36:12.000 --> 36:16.880 + This happens with not only autonomous vehicle and AI, + +36:16.880 --> 36:20.840 + but happened with the 3G and 4G. + +36:20.840 --> 36:25.320 + You'll find that the Chinese wireless reception + +36:25.320 --> 36:28.760 + is better than the US, because massive spending that + +36:28.760 --> 36:30.720 + tries to cover the whole country. + +36:30.720 --> 36:34.360 + Whereas in the US, it may be a little spotty. + +36:34.360 --> 36:36.160 + It's a government driven, because I think + +36:36.160 --> 36:44.120 + they view the coverage of cell access and 3G, 4G access + +36:44.120 --> 36:47.080 + to be a governmental infrastructure spending, + +36:47.080 --> 36:49.880 + as opposed to capitalistic. + +36:49.880 --> 36:52.160 + So of course, the state or enterprise + +36:52.160 --> 36:55.000 + is also publicly traded, but they also + +36:55.000 --> 36:57.720 + carry a government responsibility + +36:57.720 --> 37:00.240 + to deliver infrastructure to all. + +37:00.240 --> 37:01.880 + So it's a different way of thinking + +37:01.880 --> 37:05.400 + that may be very hard to inject into Western countries + +37:05.400 --> 37:09.280 + to say starting tomorrow, bandwidth infrastructure + +37:09.280 --> 37:13.840 + and highways are going to be governmental spending + +37:13.840 --> 37:16.240 + with some characteristics. + +37:16.240 --> 37:18.240 + What's your sense, and sorry to interrupt, + +37:18.240 --> 37:21.680 + but because it's such a fascinating point, + +37:21.680 --> 37:25.600 + do you think on the autonomous vehicle space + +37:25.600 --> 37:30.120 + it's possible to solve the problem of full autonomy + +37:30.120 --> 37:34.040 + without significant investment in infrastructure? + +37:34.040 --> 37:36.400 + Well, that's really hard to speculate. + +37:36.400 --> 37:38.960 + I think it's not a yes, no question, + +37:38.960 --> 37:41.920 + but how long does it take question? + +37:41.920 --> 37:45.120 + 15 years, 30 years, 45 years. + +37:45.120 --> 37:48.960 + Clearly with infrastructure augmentation, + +37:48.960 --> 37:52.320 + where there's road, the city, or whole city planning, + +37:52.320 --> 37:56.440 + building a new city, I'm sure that will accelerate + +37:56.440 --> 37:59.040 + the day of the L5. + +37:59.040 --> 38:01.520 + I'm not knowledgeable enough, and it's + +38:01.520 --> 38:03.920 + hard to predict even when we're knowledgeable, + +38:03.920 --> 38:07.120 + because a lot of it is speculative. + +38:07.120 --> 38:09.800 + But in the US, I don't think people + +38:09.800 --> 38:13.240 + would consider building a new city the size of Chicago + +38:13.240 --> 38:15.920 + to make it the AI slash autonomous city. + +38:15.920 --> 38:18.840 + There are smaller ones being built, I'm aware of that. + +38:18.840 --> 38:21.280 + But is infrastructure spend really + +38:21.280 --> 38:23.720 + impossible for US or Western countries? + +38:23.720 --> 38:25.680 + I don't think so. + +38:25.680 --> 38:28.920 + The US highway system was built. + +38:28.920 --> 38:31.960 + Was that during President Eisenhower or Kennedy? + +38:31.960 --> 38:33.160 + Eisenhower, yeah. + +38:33.160 --> 38:38.960 + So maybe historians can study how the President Eisenhower + +38:38.960 --> 38:42.960 + get the resources to build this massive infrastructure that + +38:42.960 --> 38:47.560 + surely gave US a tremendous amount of prosperity + +38:47.560 --> 38:50.800 + over the next decade, if not century. + +38:50.800 --> 38:53.240 + If I may comment on that, then, it + +38:53.240 --> 38:54.880 + takes us to artificial intelligence + +38:54.880 --> 38:58.080 + a little bit, because in order to build infrastructure, + +38:58.080 --> 39:00.520 + it creates a lot of jobs. + +39:00.520 --> 39:02.840 + So I'll be actually interested if you + +39:02.840 --> 39:06.120 + would say that you're talking in your book about all kinds + +39:06.120 --> 39:08.960 + of jobs that could and could not be automated. + +39:08.960 --> 39:12.000 + I wonder if building infrastructure + +39:12.000 --> 39:15.720 + is one of the jobs that would not be easily automated, + +39:15.720 --> 39:18.160 + something you can think about, because I think you've mentioned + +39:18.160 --> 39:21.160 + somewhere in a talk, or that there + +39:21.160 --> 39:24.280 + might be, as jobs are being automated, + +39:24.280 --> 39:28.160 + a role for government to create jobs that can't be automated. + +39:28.160 --> 39:31.040 + Yes, I think that's a possibility. + +39:31.040 --> 39:34.280 + Back in the last financial crisis, + +39:34.280 --> 39:40.320 + China put a lot of money to basically give this economy + +39:40.320 --> 39:45.520 + a boost, and a lot of it went into infrastructure building. + +39:45.520 --> 39:49.920 + And I think that's a legitimate way, at the government level, + +39:49.920 --> 39:55.680 + to deal with the employment issues as well as build out + +39:55.680 --> 39:58.960 + the infrastructure, as long as the infrastructures are truly + +39:58.960 --> 40:03.160 + needed, and as long as there is an employment problem, which + +40:03.160 --> 40:04.960 + we don't know. + +40:04.960 --> 40:07.920 + So maybe taking a little step back, + +40:07.920 --> 40:12.840 + if you've been a leader and a researcher in AI + +40:12.840 --> 40:16.200 + for several decades, at least 30 years, + +40:16.200 --> 40:21.040 + so how has AI changed in the West and the East + +40:21.040 --> 40:23.120 + as you've observed, as you've been deep in it + +40:23.120 --> 40:25.120 + over the past 30 years? + +40:25.120 --> 40:28.520 + Well, AI began as the pursuit of understanding + +40:28.520 --> 40:34.160 + human intelligence, and the term itself represents that. + +40:34.160 --> 40:37.680 + But it kind of drifted into the one subarea that + +40:37.680 --> 40:40.880 + worked extremely well, which is machine intelligence. + +40:40.880 --> 40:45.080 + And that's actually more using pattern recognition techniques + +40:45.080 --> 40:51.280 + to basically do incredibly well on a limited domain, + +40:51.280 --> 40:54.840 + large amount of data, but relatively simple kinds + +40:54.840 --> 40:58.720 + of planning, tasks, and not very creative. + +40:58.720 --> 41:02.480 + So we didn't end up building human intelligence. + +41:02.480 --> 41:04.760 + We built a different machine that + +41:04.760 --> 41:08.040 + was a lot better than us, some problems, + +41:08.040 --> 41:11.840 + but nowhere close to us on other problems. + +41:11.840 --> 41:14.200 + So today, I think a lot of people still + +41:14.200 --> 41:18.080 + misunderstand when we say artificial intelligence + +41:18.080 --> 41:20.720 + and what various products can do. + +41:20.720 --> 41:24.160 + People still think it's about replicating human intelligence. + +41:24.160 --> 41:26.160 + But the products out there really + +41:26.160 --> 41:31.680 + are closer to having invented the internet or the spreadsheet + +41:31.680 --> 41:35.360 + or the database and getting broader adoption. + +41:35.360 --> 41:38.400 + And speaking further to the fears, near term fears + +41:38.400 --> 41:41.240 + that people have about AI, so you're commenting + +41:41.240 --> 41:45.680 + on the general intelligence that people + +41:45.680 --> 41:48.040 + in the popular culture from sci fi movies + +41:48.040 --> 41:50.920 + have a sense about AI, but there's practical fears + +41:50.920 --> 41:54.800 + about AI, the kind of narrow AI that you're talking about + +41:54.800 --> 41:57.280 + of automating particular kinds of jobs, + +41:57.280 --> 41:59.400 + and you talk about them in the book. + +41:59.400 --> 42:01.520 + So what are the kinds of jobs in your view + +42:01.520 --> 42:04.840 + that you see in the next five, 10 years beginning + +42:04.840 --> 42:09.240 + to be automated by AI systems algorithms? + +42:09.240 --> 42:13.000 + Yes, this is also maybe a little bit counterintuitive + +42:13.000 --> 42:15.440 + because it's the routine jobs that + +42:15.440 --> 42:18.360 + will be displaced the soonest. + +42:18.360 --> 42:23.120 + And they may not be displaced entirely, maybe 50%, 80% + +42:23.120 --> 42:26.320 + of a job, but when the workload drops by that much, + +42:26.320 --> 42:28.760 + employment will come down. + +42:28.760 --> 42:31.520 + And also another part of misunderstanding + +42:31.520 --> 42:35.720 + is most people think of AI replacing routine jobs, + +42:35.720 --> 42:38.760 + then they think of the assembly line, the workers. + +42:38.760 --> 42:40.960 + Well, that will have some effects, + +42:40.960 --> 42:44.600 + but it's actually the routine white collar workers that's + +42:44.600 --> 42:49.280 + easiest to replace because to replace a white collar worker, + +42:49.280 --> 42:50.720 + you just need software. + +42:50.720 --> 42:53.120 + To replace a blue collar worker, + +42:53.120 --> 42:57.200 + you need robotics, mechanical excellence, + +42:57.200 --> 43:01.880 + and the ability to deal with dexterity, + +43:01.880 --> 43:05.640 + and maybe even unknown environments, very, very difficult. + +43:05.640 --> 43:11.200 + So if we were to categorize the most dangerous white collar + +43:11.200 --> 43:15.600 + jobs, they would be things like back office, + +43:15.600 --> 43:20.800 + people who copy and paste and deal with simple computer + +43:20.800 --> 43:25.560 + programs and data, and maybe paper and OCR, + +43:25.560 --> 43:29.000 + and they don't make strategic decisions, + +43:29.000 --> 43:32.040 + they basically facilitate the process. + +43:32.040 --> 43:34.680 + These software and paper systems don't work, + +43:34.680 --> 43:40.520 + so you have people dealing with new employee orientation, + +43:40.520 --> 43:45.400 + searching for past lawsuits and financial documents, + +43:45.400 --> 43:49.800 + and doing reference check, so basic searching and management + +43:49.800 --> 43:52.800 + of data that's the most in danger of being lost. + +43:52.800 --> 43:56.440 + In addition to the white collar repetitive work, + +43:56.440 --> 43:59.360 + a lot of simple interaction work can also + +43:59.360 --> 44:02.840 + be taken care of, such as tele sales, telemarketing, + +44:02.840 --> 44:07.280 + customer service, as well as many physical jobs + +44:07.280 --> 44:09.880 + that are in the same location and don't + +44:09.880 --> 44:12.240 + require a high degree of dexterity, + +44:12.240 --> 44:17.840 + so fruit picking, dishwashing, assembly line, inspection, + +44:17.840 --> 44:20.360 + our jobs in that category. + +44:20.360 --> 44:25.440 + So altogether, back office is a big part, + +44:25.440 --> 44:29.840 + and the other, the blue collar may be smaller initially, + +44:29.840 --> 44:32.560 + but over time, AI will get better. + +44:32.560 --> 44:36.880 + And when we start to get to over the next 15, 20 years, + +44:36.880 --> 44:39.120 + the ability to actually have the dexterity + +44:39.120 --> 44:42.600 + of doing assembly line, that's a huge chunk of jobs. + +44:42.600 --> 44:44.760 + And when autonomous vehicles start + +44:44.760 --> 44:47.400 + to work initially starting with truck drivers, + +44:47.400 --> 44:49.640 + but eventually to all drivers, that's + +44:49.640 --> 44:52.040 + another huge group of workers. + +44:52.040 --> 44:55.560 + So I see modest numbers in the next five years, + +44:55.560 --> 44:58.080 + but increasing rapidly after that. + +44:58.080 --> 45:01.240 + On the worry of the jobs that are in danger + +45:01.240 --> 45:04.320 + and the gradual loss of jobs, I'm not + +45:04.320 --> 45:06.680 + sure if you're familiar with Andrew Yang. + +45:06.680 --> 45:07.800 + Yes, I am. + +45:07.800 --> 45:10.560 + So there's a candidate for president of the United States + +45:10.560 --> 45:14.960 + whose platform, Andrew Yang, is based around, in part, + +45:14.960 --> 45:17.680 + around job loss due to automation, + +45:17.680 --> 45:21.120 + and also, in addition, the need, perhaps, + +45:21.120 --> 45:26.120 + of universal basic income to support jobs that are folks who + +45:26.120 --> 45:28.560 + lose their job due to automation and so on, + +45:28.560 --> 45:31.960 + and in general, support people under complex, + +45:31.960 --> 45:34.320 + unstable job market. + +45:34.320 --> 45:36.720 + So what are your thoughts about his concerns, + +45:36.720 --> 45:40.000 + him as a candidate, his ideas in general? + +45:40.000 --> 45:44.600 + I think his thinking is generally in the right direction, + +45:44.600 --> 45:48.440 + but his approach as a presidential candidate + +45:48.440 --> 45:52.240 + may be a little bit ahead at the time. + +45:52.240 --> 45:56.080 + I think the displacements will happen, + +45:56.080 --> 45:58.280 + but will they happen soon enough for people + +45:58.280 --> 46:00.480 + to agree to vote for him? + +46:00.480 --> 46:03.760 + The unemployment numbers are not very high yet. + +46:03.760 --> 46:07.600 + And I think he and I have the same challenge. + +46:07.600 --> 46:11.520 + If I want to theoretically convince people this is an issue + +46:11.520 --> 46:13.880 + and he wants to become the president, + +46:13.880 --> 46:17.760 + people have to see how can this be the case when + +46:17.760 --> 46:19.680 + unemployment numbers are low. + +46:19.680 --> 46:21.360 + So that is the challenge. + +46:21.360 --> 46:27.360 + And I think I do agree with him on the displacement issue, + +46:27.360 --> 46:32.280 + on universal basic income, at a very vanilla level. + +46:32.280 --> 46:36.800 + I don't agree with it because I think the main issue + +46:36.800 --> 46:38.320 + is retraining. + +46:38.320 --> 46:43.200 + So people need to be incented not by just giving a monthly + +46:43.200 --> 46:47.160 + $2,000 check or $1,000 check and do whatever they want + +46:47.160 --> 46:50.920 + because they don't have the know how + +46:50.920 --> 46:56.840 + to know what to retrain to go into what type of a job + +46:56.840 --> 46:58.640 + and guidance is needed. + +46:58.640 --> 47:01.720 + And retraining is needed because historically + +47:01.720 --> 47:05.080 + in technology revolutions, when routine jobs were displaced, + +47:05.080 --> 47:06.920 + new routine jobs came up. + +47:06.920 --> 47:09.400 + So there was always room for that. + +47:09.400 --> 47:12.640 + But with AI and automation, the whole point + +47:12.640 --> 47:15.320 + is replacing all routine jobs eventually. + +47:15.320 --> 47:17.840 + So there will be fewer and fewer routine jobs. + +47:17.840 --> 47:22.640 + And AI will create jobs, but it won't create routine jobs + +47:22.640 --> 47:24.840 + because if it creates routine jobs, + +47:24.840 --> 47:26.880 + why wouldn't AI just do it? + +47:26.880 --> 47:30.360 + So therefore, the people who are losing the jobs + +47:30.360 --> 47:32.280 + are losing routine jobs. + +47:32.280 --> 47:35.720 + The jobs that are becoming available are nonroutine jobs. + +47:35.720 --> 47:39.320 + So the social stipend needs to be put in place + +47:39.320 --> 47:42.040 + is for the routine workers who lost their jobs + +47:42.040 --> 47:46.120 + to be retrained maybe in six months, maybe in three years. + +47:46.120 --> 47:48.560 + Takes a while to retrain on the nonroutine job + +47:48.560 --> 47:51.360 + and then take on a job that will last + +47:51.360 --> 47:53.400 + for that person's lifetime. + +47:53.400 --> 47:56.160 + Now, having said that, if you look deeply + +47:56.160 --> 47:58.240 + into Andrew's document, he does cater for that. + +47:58.240 --> 48:03.240 + So I'm not disagreeing with what he's trying to do. + +48:03.280 --> 48:06.360 + But for simplification, sometimes he just says UBI, + +48:06.360 --> 48:08.760 + but simple UBI wouldn't work. + +48:08.760 --> 48:10.600 + And I think you've mentioned elsewhere + +48:10.600 --> 48:15.600 + that the goal isn't necessarily to give people enough money + +48:15.760 --> 48:19.120 + to survive or live or even to prosper. + +48:19.120 --> 48:22.800 + The point is to give them a job that gives them meaning. + +48:22.800 --> 48:25.600 + That meaning is extremely important. + +48:25.600 --> 48:28.600 + That our employment, at least in the United States + +48:28.600 --> 48:31.200 + and perhaps it cares across the world, + +48:31.200 --> 48:34.600 + provides something that's, forgive me for saying, + +48:34.600 --> 48:36.960 + greater than money, it provides meaning. + +48:38.400 --> 48:43.400 + So now what kind of jobs do you think can't be automated? + +48:44.840 --> 48:46.600 + You talk a little bit about creativity + +48:46.600 --> 48:48.200 + and compassion in your book. + +48:48.200 --> 48:50.720 + What aspects do you think it's difficult + +48:50.720 --> 48:52.320 + to automate for an AI system? + +48:52.320 --> 48:57.320 + Because an AI system is currently merely optimizing. + +48:57.360 --> 49:00.120 + It's not able to reason, plan, + +49:00.120 --> 49:02.920 + or think creatively or strategically. + +49:02.920 --> 49:05.320 + It's not able to deal with complex problems. + +49:05.320 --> 49:09.520 + It can't come up with a new problem and solve it. + +49:09.520 --> 49:12.320 + A human needs to find the problem + +49:12.320 --> 49:15.520 + and pose it as an optimization problem, + +49:15.520 --> 49:17.520 + then have the AI work at it. + +49:17.520 --> 49:21.320 + So an AI would have a very hard time + +49:21.320 --> 49:23.320 + discovering a new drug + +49:23.320 --> 49:26.320 + or discovering a new style of painting + +49:27.320 --> 49:30.320 + or dealing with complex tasks + +49:30.320 --> 49:32.320 + such as managing a company + +49:32.320 --> 49:35.320 + that isn't just about optimizing the bottom line, + +49:35.320 --> 49:39.320 + but also about employee satisfaction, corporate brand, + +49:39.320 --> 49:40.320 + and many, many other things. + +49:40.320 --> 49:44.320 + So that is one category of things. + +49:44.320 --> 49:48.320 + And because these things are challenging, creative, complex, + +49:48.320 --> 49:52.320 + doing them creates a higher degree of satisfaction + +49:52.320 --> 49:55.320 + and therefore appealing to our desire for working, + +49:55.320 --> 49:57.320 + which isn't just to make the money, + +49:57.320 --> 49:58.320 + make the ends meet, + +49:58.320 --> 50:00.320 + but also that we've accomplished something + +50:00.320 --> 50:03.320 + that others maybe can't do or can't do as well. + +50:04.320 --> 50:07.320 + Another type of job that is much numerous + +50:07.320 --> 50:09.320 + would be compassionate jobs, + +50:09.320 --> 50:14.320 + jobs that require compassion, empathy, human touch, human trust. + +50:14.320 --> 50:18.320 + AI can't do that because AI is cold, calculating, + +50:18.320 --> 50:22.320 + and even if it can fake that to some extent, + +50:22.320 --> 50:26.320 + it will make errors and that will make it look very silly. + +50:26.320 --> 50:29.320 + And also, I think even if AI did okay, + +50:29.320 --> 50:33.320 + people would want to interact with another person, + +50:33.320 --> 50:38.320 + whether it's for some kind of a service or a teacher or a doctor + +50:38.320 --> 50:41.320 + or a concierge or a masseuse or bartender. + +50:41.320 --> 50:46.320 + There are so many jobs where people just don't want to interact + +50:46.320 --> 50:49.320 + with a cold robot or software. + +50:50.320 --> 50:53.320 + I've had an entrepreneur who built an elderly care robot + +50:53.320 --> 50:58.320 + and they found that the elderly really only use it for customer service. + +50:58.320 --> 51:00.320 + But not to service the product, + +51:00.320 --> 51:05.320 + but they click on customer service and the video of a person comes up + +51:05.320 --> 51:07.320 + and then the person says, + +51:07.320 --> 51:11.320 + how come my daughter didn't call me? Let me show you a picture of her grandkids. + +51:11.320 --> 51:15.320 + So people earn for that, people people interaction. + +51:15.320 --> 51:19.320 + So even if robots improved, people just don't want it. + +51:19.320 --> 51:21.320 + And those jobs are going to be increasing + +51:21.320 --> 51:24.320 + because AI will create a lot of value, + +51:24.320 --> 51:29.320 + $16 trillion to the world in next 11 years according to PWC + +51:29.320 --> 51:34.320 + and that will give people money to enjoy services, + +51:34.320 --> 51:39.320 + whether it's eating a gourmet meal or tourism and traveling + +51:39.320 --> 51:41.320 + or having concierge services. + +51:41.320 --> 51:44.320 + The services revolving around, you know, + +51:44.320 --> 51:47.320 + every dollar of that $16 trillion will be tremendous. + +51:47.320 --> 51:52.320 + It will create more opportunities to service the people who did well + +51:52.320 --> 51:55.320 + through AI with things. + +51:55.320 --> 52:01.320 + But even at the same time, the entire society is very much short + +52:01.320 --> 52:05.320 + in need of many service oriented, compassionate oriented jobs. + +52:05.320 --> 52:10.320 + The best example is probably in healthcare services. + +52:10.320 --> 52:15.320 + There's going to be 2 million new jobs, not counting replacement, + +52:15.320 --> 52:20.320 + just brand new incremental jobs in the next six years in healthcare services. + +52:20.320 --> 52:24.320 + That includes nurses orderly in the hospital, + +52:24.320 --> 52:29.320 + elderly care and also at home care. + +52:29.320 --> 52:31.320 + It's particularly lacking. + +52:31.320 --> 52:34.320 + And those jobs are not likely to be filled. + +52:34.320 --> 52:36.320 + So there's likely to be a shortage. + +52:36.320 --> 52:41.320 + And the reason they're not filled is simply because they don't pay very well + +52:41.320 --> 52:47.320 + and that the social status of these jobs are not very good. + +52:47.320 --> 52:52.320 + So they pay about half as much as a heavy equipment operator, + +52:52.320 --> 52:55.320 + which will be replaced a lot sooner. + +52:55.320 --> 52:59.320 + And they pay probably comparably to someone on the assembly line. + +52:59.320 --> 53:03.320 + And so if we're ignoring all the other issues + +53:03.320 --> 53:07.320 + and just think about satisfaction from one's job, + +53:07.320 --> 53:11.320 + someone repetitively doing the same manual action at an assembly line, + +53:11.320 --> 53:14.320 + that can't create a lot of job satisfaction. + +53:14.320 --> 53:17.320 + But someone taking care of a sick person + +53:17.320 --> 53:21.320 + and getting a hug and thank you from that person and the family, + +53:21.320 --> 53:24.320 + I think is quite satisfying. + +53:24.320 --> 53:28.320 + So if only we could fix the pay for service jobs, + +53:28.320 --> 53:33.320 + there are plenty of jobs that require some training or a lot of training + +53:33.320 --> 53:36.320 + for the people coming off the routine jobs to take. + +53:36.320 --> 53:43.320 + We can easily imagine someone who was maybe a cashier at the grocery store, + +53:43.320 --> 53:49.320 + at stores become automated, learns to become a nurse or at home care. + +53:49.320 --> 53:54.320 + Also, I do want to point out the blue collar jobs are going to stay around a bit longer, + +53:54.320 --> 53:57.320 + some of them quite a bit longer. + +53:57.320 --> 54:01.320 + AI cannot be told, go clean an arbitrary home. + +54:01.320 --> 54:03.320 + That's incredibly hard. + +54:03.320 --> 54:07.320 + Arguably is an L5 level of difficulty. + +54:07.320 --> 54:09.320 + And then AI cannot be a good plumber, + +54:09.320 --> 54:12.320 + because plumber is almost like a mini detective + +54:12.320 --> 54:15.320 + that has to figure out where the leak came from. + +54:15.320 --> 54:22.320 + So yet AI probably can be an assembly line and auto mechanic and so on. + +54:22.320 --> 54:26.320 + So one has to study which blue collar jobs are going away + +54:26.320 --> 54:30.320 + and facilitate retraining for the people to go into the ones that won't go away + +54:30.320 --> 54:32.320 + or maybe even will increase. + +54:32.320 --> 54:39.320 + I mean, it is fascinating that it's easier to build a world champion chess player + +54:39.320 --> 54:41.320 + than it is to build a mediocre plumber. + +54:41.320 --> 54:43.320 + Yes, very true. + +54:43.320 --> 54:47.320 + And to AI, and that goes counterintuitive to a lot of people's understanding + +54:47.320 --> 54:49.320 + of what artificial intelligence is. + +54:49.320 --> 54:53.320 + So it sounds, I mean, you're painting a pretty optimistic picture + +54:53.320 --> 54:56.320 + about retraining, about the number of jobs + +54:56.320 --> 55:01.320 + and actually the meaningful nature of those jobs once we automate repetitive tasks. + +55:01.320 --> 55:07.320 + So overall, are you optimistic about the future + +55:07.320 --> 55:11.320 + where much of the repetitive tasks are automated, + +55:11.320 --> 55:15.320 + that there is a lot of room for humans, for the compassionate, + +55:15.320 --> 55:19.320 + for the creative input that only humans can provide? + +55:19.320 --> 55:23.320 + I am optimistic if we start to take action. + +55:23.320 --> 55:27.320 + If we have no action in the next five years, + +55:27.320 --> 55:33.320 + I think it's going to be hard to deal with the devastating losses that will emerge. + +55:33.320 --> 55:39.320 + So if we start thinking about retraining, maybe with the low hanging fruits, + +55:39.320 --> 55:45.320 + explaining to vocational schools why they should train more plumbers than auto mechanics, + +55:45.320 --> 55:53.320 + maybe starting with some government subsidy for corporations to have more training positions. + +55:53.320 --> 55:57.320 + We start to explain to people why retraining is important. + +55:57.320 --> 56:00.320 + We start to think about what the future of education, + +56:00.320 --> 56:04.320 + how that needs to be tweaked for the era of AI. + +56:04.320 --> 56:06.320 + If we start to make incremental progress, + +56:06.320 --> 56:09.320 + and the greater number of people understand, + +56:09.320 --> 56:12.320 + then there's no reason to think we can't deal with this, + +56:12.320 --> 56:16.320 + because this technological revolution is arguably similar to + +56:16.320 --> 56:20.320 + what electricity, industrial revolutions, and internet brought about. + +56:20.320 --> 56:24.320 + Do you think there's a role for policy, for governments to step in + +56:24.320 --> 56:27.320 + to help with policy to create a better world? + +56:27.320 --> 56:32.320 + Absolutely, and the governments don't have to believe + +56:32.320 --> 56:39.320 + that unemployment will go up, and they don't have to believe automation will be this fast to do something. + +56:39.320 --> 56:42.320 + Revamping vocational school would be one example. + +56:42.320 --> 56:47.320 + Another is if there's a big gap in healthcare service employment, + +56:47.320 --> 56:54.320 + and we know that a country's population is growing older and more longevity living older, + +56:54.320 --> 56:59.320 + because people over 80 require five times as much care as those under 80, + +56:59.320 --> 57:04.320 + then it is a good time to incent training programs for elderly care, + +57:04.320 --> 57:07.320 + to find ways to improve the pay. + +57:07.320 --> 57:13.320 + Maybe one way would be to offer as part of Medicare or the equivalent program + +57:13.320 --> 57:18.320 + for people over 80 to be entitled to a few hours of elderly care at home, + +57:18.320 --> 57:21.320 + and then that might be reimbursable, + +57:21.320 --> 57:28.320 + and that will stimulate the service industry around the policy. + +57:28.320 --> 57:32.320 + Do you have concerns about large entities, + +57:32.320 --> 57:38.320 + whether it's governments or companies, controlling the future of AI development in general? + +57:38.320 --> 57:40.320 + So we talked about companies. + +57:40.320 --> 57:48.320 + Do you have a better sense that governments can better represent the interest of the people + +57:48.320 --> 57:54.320 + than companies, or do you believe companies are better at representing the interest of the people? + +57:54.320 --> 57:56.320 + Or is there no easy answer? + +57:56.320 --> 57:59.320 + I don't think there's an easy answer because it's a double edged sword. + +57:59.320 --> 58:06.320 + The companies and governments can provide better services with more access to data and more access to AI, + +58:06.320 --> 58:13.320 + but that also leads to greater power, which can lead to uncontrollable problems, + +58:13.320 --> 58:17.320 + whether it's monopoly or corruption in the government. + +58:17.320 --> 58:24.320 + So I think one has to be careful to look at how much data that companies and governments have, + +58:24.320 --> 58:29.320 + and some kind of checks and balances would be helpful. + +58:29.320 --> 58:33.320 + So again, I come from Russia. + +58:33.320 --> 58:36.320 + There's something called the Cold War. + +58:36.320 --> 58:40.320 + So let me ask a difficult question here, looking at conflict. + +58:40.320 --> 58:45.320 + Steven Pinker wrote a great book that conflict all over the world is decreasing in general. + +58:45.320 --> 58:51.320 + But do you have a sense that having written the book AI Superpowers, + +58:51.320 --> 58:57.320 + do you see a major international conflict potentially arising between major nations, + +58:57.320 --> 59:02.320 + whatever they are, whether it's Russia, China, European nations, United States, + +59:02.320 --> 59:09.320 + or others in the next 10, 20, 50 years around AI, around the digital space, cyber space? + +59:09.320 --> 59:12.320 + Do you worry about that? + +59:12.320 --> 59:19.320 + Is that something we need to think about and try to alleviate or prevent? + +59:19.320 --> 59:22.320 + I believe in greater engagement. + +59:22.320 --> 59:33.320 + A lot of the worries about more powerful AI are based on an arms race metaphor. + +59:33.320 --> 59:41.320 + And when you extrapolate into military kinds of scenarios, + +59:41.320 --> 59:48.320 + AI can automate autonomous weapons that needs to be controlled somehow. + +59:48.320 --> 59:57.320 + And autonomous decision making can lead to not enough time to fix international crises. + +59:57.320 --> 1:00:02.320 + So I actually believe a Cold War mentality would be very dangerous + +1:00:02.320 --> 1:00:07.320 + because should two countries rely on AI to make certain decisions + +1:00:07.320 --> 1:00:11.320 + and they don't even talk to each other, they do their own scenario planning, + +1:00:11.320 --> 1:00:14.320 + then something could easily go wrong. + +1:00:14.320 --> 1:00:24.320 + I think engagement, interaction, some protocols to avoid inadvertent disasters is actually needed. + +1:00:24.320 --> 1:00:28.320 + So it's natural for each country to want to be the best, + +1:00:28.320 --> 1:00:34.320 + whether it's in nuclear technologies or AI or bio. + +1:00:34.320 --> 1:00:40.320 + But I think it's important to realize if each country has a black box AI + +1:00:40.320 --> 1:00:48.320 + and don't talk to each other, that probably presents greater challenges to humanity + +1:00:48.320 --> 1:00:50.320 + than if they interacted. + +1:00:50.320 --> 1:00:56.320 + I think there can still be competition, but with some degree of protocol for interaction. + +1:00:56.320 --> 1:01:01.320 + Just like when there was a nuclear competition, + +1:01:01.320 --> 1:01:07.320 + there were some protocol for deterrence among US, Russia, and China. + +1:01:07.320 --> 1:01:10.320 + And I think that engagement is needed. + +1:01:10.320 --> 1:01:15.320 + So of course, we're still far from AI presenting that kind of danger. + +1:01:15.320 --> 1:01:22.320 + But what I worry the most about is the level of engagement seems to be coming down. + +1:01:22.320 --> 1:01:25.320 + The level of distrust seems to be going up, + +1:01:25.320 --> 1:01:32.320 + especially from the US towards other large countries such as China and Russia. + +1:01:32.320 --> 1:01:34.320 + Is there a way to make that better? + +1:01:34.320 --> 1:01:40.320 + So that's beautifully put, level of engagement and even just basic trust and communication + +1:01:40.320 --> 1:01:52.320 + as opposed to making artificial enemies out of particular countries. + +1:01:52.320 --> 1:02:01.320 + Do you have a sense how we can make it better, actionable items that as a society we can take on? + +1:02:01.320 --> 1:02:10.320 + I'm not an expert at geopolitics, but I would say that we look pretty foolish as humankind + +1:02:10.320 --> 1:02:19.320 + when we are faced with the opportunity to create $16 trillion for humanity. + +1:02:19.320 --> 1:02:29.320 + And yet we're not solving fundamental problems with parts of the world still in poverty. + +1:02:29.320 --> 1:02:34.320 + And for the first time, we have the resources to overcome poverty and hunger. + +1:02:34.320 --> 1:02:38.320 + We're not using it on that, but we're fueling competition among superpowers. + +1:02:38.320 --> 1:02:41.320 + And that's a very unfortunate thing. + +1:02:41.320 --> 1:02:54.320 + If we become utopian for a moment, imagine a benevolent world government that has this $16 trillion + +1:02:54.320 --> 1:03:02.320 + and maybe some AI to figure out how to use it to deal with diseases and problems and hate and things like that. + +1:03:02.320 --> 1:03:04.320 + World would be a lot better off. + +1:03:04.320 --> 1:03:07.320 + So what is wrong with the current world? + +1:03:07.320 --> 1:03:13.320 + I think the people with more skill than I should think about this. + +1:03:13.320 --> 1:03:19.320 + And then the geopolitics issue with superpower competition is one side of the issue. + +1:03:19.320 --> 1:03:29.320 + There's another side which I worry maybe even more, which is as the $16 trillion all gets made by U.S. and China + +1:03:29.320 --> 1:03:34.320 + and a few of the other developed countries, the poorer country will get nothing + +1:03:34.320 --> 1:03:42.320 + because they don't have technology and the wealth disparity and inequality will increase. + +1:03:42.320 --> 1:03:50.320 + So a poorer country with a large population will not only benefit from the AI boom or other technology booms + +1:03:50.320 --> 1:03:57.320 + but they will have their workers who previously had hoped they could do the China model and do outsource manufacturing + +1:03:57.320 --> 1:04:02.320 + or the India model so they could do the outsource process or call center + +1:04:02.320 --> 1:04:05.320 + while all those jobs are going to be gone in 10 or 15 years. + +1:04:05.320 --> 1:04:14.320 + So the individual citizen may be a net liability, I mean financially speaking, to a poorer country + +1:04:14.320 --> 1:04:19.320 + and not an asset to claw itself out of poverty. + +1:04:19.320 --> 1:04:29.320 + So in that kind of situation, these large countries with not much tech are going to be facing a downward spiral + +1:04:29.320 --> 1:04:37.320 + and it's unclear what could be done and then when we look back and say there's $16 trillion being created + +1:04:37.320 --> 1:04:43.320 + and it's all being kept by U.S. China and other developed countries, it just doesn't feel right. + +1:04:43.320 --> 1:04:50.320 + So I hope people who know about geopolitics can find solutions that's beyond my expertise. + +1:04:50.320 --> 1:04:54.320 + So different countries that we've talked about have different value systems. + +1:04:54.320 --> 1:05:02.320 + If you look at the United States to an almost extreme degree, there is an absolute desire for freedom of speech. + +1:05:02.320 --> 1:05:14.320 + If you look at a country where I was raised, that desire just amongst the people is not as elevated as it is to basically fundamental level + +1:05:14.320 --> 1:05:17.320 + to the essence of what it means to be America, right? + +1:05:17.320 --> 1:05:20.320 + And the same is true with China, there's different value systems. + +1:05:20.320 --> 1:05:30.320 + There is some censorship of internet content that China and Russia and many other countries undertake. + +1:05:30.320 --> 1:05:40.320 + Do you see that having effects on innovation, other aspects of some of the tech stuff, AI development we talked about + +1:05:40.320 --> 1:05:52.320 + and maybe from another angle, do you see that changing in different ways over the next 10 years, 20 years, 50 years as China continues to grow + +1:05:52.320 --> 1:05:55.320 + as it does now in its tech innovation? + +1:05:55.320 --> 1:06:08.320 + There's a common belief that full freedom of speech and expression is correlated with creativity, which is correlated with entrepreneurial success. + +1:06:08.320 --> 1:06:15.320 + I think empirically we have seen that is not true and China has been successful. + +1:06:15.320 --> 1:06:25.320 + That's not to say the fundamental values are not right or not the best, but it's just that perfect correlation isn't there. + +1:06:25.320 --> 1:06:36.320 + It's hard to read the tea leaves on opening up or not in any country and I've not been very good at that in my past predictions. + +1:06:36.320 --> 1:06:46.320 + But I do believe every country shares some fundamental value, a lot of fundamental values for the long term. + +1:06:46.320 --> 1:07:02.320 + So, you know, China is drafting its privacy policy for individual citizens and they don't look that different from the American or European ones. + +1:07:02.320 --> 1:07:13.320 + So, people do want to protect their privacy and have the opportunity to express and I think the fundamental values are there. + +1:07:13.320 --> 1:07:21.320 + The question is in the execution and timing, how soon or when will that start to open up? + +1:07:21.320 --> 1:07:31.320 + So, as long as each government knows, ultimately people want that kind of protection, there should be a plan to move towards that. + +1:07:31.320 --> 1:07:35.320 + As to when or how, again, I'm not an expert. + +1:07:35.320 --> 1:07:38.320 + On the point of privacy to me, it's really interesting. + +1:07:38.320 --> 1:07:44.320 + So, AI needs data to create a personalized awesome experience. + +1:07:44.320 --> 1:07:47.320 + I'm just speaking generally in terms of products. + +1:07:47.320 --> 1:07:53.320 + And then we have currently, depending on the age and depending on the demographics of who we're talking about, + +1:07:53.320 --> 1:07:58.320 + some people are more or less concerned about the amount of data they hand over. + +1:07:58.320 --> 1:08:03.320 + So, in your view, how do we get this balance right? + +1:08:03.320 --> 1:08:09.320 + That we provide an amazing experience to people that use products. + +1:08:09.320 --> 1:08:15.320 + You look at Facebook, you know, the more Facebook knows about you, yes, it's scary to say. + +1:08:15.320 --> 1:08:20.320 + The better it can probably, a better experience it can probably create. + +1:08:20.320 --> 1:08:24.320 + So, in your view, how do we get that balance right? + +1:08:24.320 --> 1:08:38.320 + Yes, I think a lot of people have a misunderstanding that it's okay and possible to just rip all the data out from a provider and give it back to you. + +1:08:38.320 --> 1:08:43.320 + So, you can deny them access to further data and still enjoy the services we have. + +1:08:43.320 --> 1:08:48.320 + If we take back all the data, all the services will give us nonsense. + +1:08:48.320 --> 1:08:57.320 + We'll no longer be able to use products that function well in terms of, you know, right ranking, right products, right user experience. + +1:08:57.320 --> 1:09:04.320 + So, yet I do understand we don't want to permit misuse of the data. + +1:09:04.320 --> 1:09:16.320 + From legal policy standpoint, I think there can be severe punishment for those who have egregious misuse of the data. + +1:09:16.320 --> 1:09:19.320 + That's, I think, a good first step. + +1:09:19.320 --> 1:09:27.320 + Actually, China on this aspect has very strong laws about people who sell or give data to other companies. + +1:09:27.320 --> 1:09:40.320 + And that over the past few years, since that law came into effect, pretty much eradicated the illegal distribution sharing of data. + +1:09:40.320 --> 1:09:52.320 + Additionally, I think giving, I think technology is often a very good way to solve technology misuse. + +1:09:52.320 --> 1:09:58.320 + So, can we come up with new technologies that will let us have our cake and eat it too? + +1:09:58.320 --> 1:10:07.320 + People are looking into homomorphic encryption, which is letting you keep the data, have it encrypted and train encrypted data. + +1:10:07.320 --> 1:10:13.320 + Of course, we haven't solved that one yet, but that kind of direction may be worth pursuing. + +1:10:13.320 --> 1:10:22.320 + Also federated learning, which would allow one hospital to train on its hospitals patient data fully because they have a license for that. + +1:10:22.320 --> 1:10:28.320 + And then hospitals would then share their models, not data, but models to create a supra AI. + +1:10:28.320 --> 1:10:30.320 + And that also maybe has some promise. + +1:10:30.320 --> 1:10:39.320 + So I would want to encourage us to be open minded and think of this as not just the policy binary yes no, + +1:10:39.320 --> 1:10:48.320 + but letting the technologists try to find solutions to let us have our cake and eat it too, or have most of our cake and eat most of it too. + +1:10:48.320 --> 1:10:55.320 + Finally, I think giving each end user a choice is important and having transparency is important. + +1:10:55.320 --> 1:11:04.320 + Also, I think that's universal, but the choice you give to the user should not be at a granular level that the user cannot understand. + +1:11:04.320 --> 1:11:12.320 + GDPR today causes all these pop ups of yes, no, will you give this site this right to use this part of your data? + +1:11:12.320 --> 1:11:20.320 + I don't think any user understands what they're saying yes or no to, and I suspect most are just saying yes because they don't understand it. + +1:11:20.320 --> 1:11:30.320 + So while GDPR in its current implementation has lived up to its promise of transparency and user choice, + +1:11:30.320 --> 1:11:39.320 + it implemented it in such a way that really didn't deliver the spirit of GDPR. + +1:11:39.320 --> 1:11:41.320 + It fit the letter, but not the spirit. + +1:11:41.320 --> 1:11:50.320 + So again, I think we need to think about is there a way to fit the spirit of GDPR by using some kind of technology? + +1:11:50.320 --> 1:11:52.320 + Can we have a slider? + +1:11:52.320 --> 1:12:01.320 + That's an AI trying to figure out how much you want to slide between perfect protection security of your personal data + +1:12:01.320 --> 1:12:07.320 + versus high degree of convenience with some risks of not having full privacy. + +1:12:07.320 --> 1:12:11.320 + Each user should have some preference and that gives you the user choice, + +1:12:11.320 --> 1:12:18.320 + but maybe we should turn the problem on its head and ask can there be an AI algorithm that can customize this + +1:12:18.320 --> 1:12:24.320 + because we can understand the slider, but we sure cannot understand every pop up question. + +1:12:24.320 --> 1:12:30.320 + And I think getting that right requires getting the balance between what we talked about earlier, + +1:12:30.320 --> 1:12:36.320 + which is heart and soul versus profit driven decisions and strategy. + +1:12:36.320 --> 1:12:45.320 + I think from my perspective, the best way to make a lot of money in the long term is to keep your heart and soul intact. + +1:12:45.320 --> 1:12:53.320 + I think getting that slider right in the short term may feel like you'll be sacrificing profit, + +1:12:53.320 --> 1:12:59.320 + but in the long term, you'll be getting user trust and providing a great experience. + +1:12:59.320 --> 1:13:01.320 + Do you share that kind of view in general? + +1:13:01.320 --> 1:13:11.320 + Yes, absolutely. I sure would hope there is a way we can do long term projects that really do the right thing. + +1:13:11.320 --> 1:13:16.320 + I think a lot of people who embrace GDPR, their hearts in the right place. + +1:13:16.320 --> 1:13:20.320 + I think they just need to figure out how to build a solution. + +1:13:20.320 --> 1:13:24.320 + I've heard utopians talk about solutions that get me excited, + +1:13:24.320 --> 1:13:29.320 + but not sure how in the current funding environment they can get started, right? + +1:13:29.320 --> 1:13:37.320 + People talk about, imagine this crowdsourced data collection that we all trust, + +1:13:37.320 --> 1:13:45.320 + and then we have these agents that we ask them to ask the trusted agent. + +1:13:45.320 --> 1:13:48.320 + That agent only, that platform. + +1:13:48.320 --> 1:14:02.320 + A trusted joint platform that we all believe is trustworthy that can give us all the close loop personal suggestions + +1:14:02.320 --> 1:14:07.320 + by the new social network, new search engine, new ecommerce engine + +1:14:07.320 --> 1:14:12.320 + that has access to even more of our data, but not directly but indirectly. + +1:14:12.320 --> 1:14:18.320 + I think that general concept of licensing to some trusted engine + +1:14:18.320 --> 1:14:22.320 + and finding a way to trust that engine seems like a great idea, + +1:14:22.320 --> 1:14:27.320 + but if you think how long it's going to take to implement and tweak and develop it right, + +1:14:27.320 --> 1:14:31.320 + as well as to collect all the trust and the data from the people, + +1:14:31.320 --> 1:14:34.320 + it's beyond the current cycle of venture capital. + +1:14:34.320 --> 1:14:37.320 + How do you do that is a big question. + +1:14:37.320 --> 1:14:44.320 + You've recently had a fight with cancer, stage 4 lymphoma, + +1:14:44.320 --> 1:14:54.320 + and in a sort of deep personal level, what did it feel like in the darker moments to face your own mortality? + +1:14:54.320 --> 1:14:57.320 + Well, I've been the workaholic my whole life, + +1:14:57.320 --> 1:15:04.320 + and I've basically worked 9.96, 9am to 9pm, 6 days a week, roughly. + +1:15:04.320 --> 1:15:10.320 + And I didn't really pay a lot of attention to my family, friends, and people who loved me, + +1:15:10.320 --> 1:15:14.320 + and my life revolved around optimizing for work. + +1:15:14.320 --> 1:15:25.320 + While my work was not routine, my optimization really made my life basically a very mechanical process. + +1:15:25.320 --> 1:15:36.320 + But I got a lot of highs out of it because of accomplishments that I thought were really important and dear and the highest priority to me. + +1:15:36.320 --> 1:15:41.320 + But when I faced mortality and the possible death in matter of months, + +1:15:41.320 --> 1:15:45.320 + I suddenly realized that this really meant nothing to me, + +1:15:45.320 --> 1:15:48.320 + that I didn't feel like working for another minute, + +1:15:48.320 --> 1:15:54.320 + that if I had 6 months left in my life, I would spend it all with my loved ones. + +1:15:54.320 --> 1:16:02.320 + And thanking them, giving them love back, and apologizing to them that I lived my life the wrong way. + +1:16:02.320 --> 1:16:11.320 + So that moment of reckoning caused me to really rethink that why we exist in this world + +1:16:11.320 --> 1:16:22.320 + is something that we might be too much shaped by the society to think that success and accomplishments is why we live. + +1:16:22.320 --> 1:16:29.320 + And while that can get you periodic successes and satisfaction, + +1:16:29.320 --> 1:16:35.320 + it's really in them facing death, you see what's truly important to you. + +1:16:35.320 --> 1:16:41.320 + So as a result of going through the challenges with cancer, + +1:16:41.320 --> 1:16:45.320 + I've resolved to live a more balanced lifestyle. + +1:16:45.320 --> 1:16:48.320 + I'm now in remission, knock on wood, + +1:16:48.320 --> 1:16:52.320 + and I'm spending more time with my family. + +1:16:52.320 --> 1:16:54.320 + My wife travels with me. + +1:16:54.320 --> 1:16:57.320 + When my kids need me, I spend more time with them. + +1:16:57.320 --> 1:17:02.320 + And before, I used to prioritize everything around work. + +1:17:02.320 --> 1:17:05.320 + When I had a little bit of time, I would dole it out to my family. + +1:17:05.320 --> 1:17:09.320 + Now, when my family needs something, really needs something, + +1:17:09.320 --> 1:17:12.320 + I drop everything at work and go to them. + +1:17:12.320 --> 1:17:15.320 + And then in the time remaining, I allocate to work. + +1:17:15.320 --> 1:17:18.320 + But one's family is very understanding. + +1:17:18.320 --> 1:17:22.320 + It's not like they will take 50 hours a week from me. + +1:17:22.320 --> 1:17:26.320 + So I'm actually able to still work pretty hard, + +1:17:26.320 --> 1:17:28.320 + maybe 10 hours less per week. + +1:17:28.320 --> 1:17:35.320 + So I realize the most important thing in my life is really love and the people I love. + +1:17:35.320 --> 1:17:38.320 + And I give that the highest priority. + +1:17:38.320 --> 1:17:40.320 + It isn't the only thing I do. + +1:17:40.320 --> 1:17:45.320 + But when that is needed, I put that at the top priority. + +1:17:45.320 --> 1:17:49.320 + And I feel much better and I feel much more balanced. + +1:17:49.320 --> 1:17:56.320 + And I think this also gives a hint as to a life of routine work, + +1:17:56.320 --> 1:17:58.320 + a life of pursuit of numbers. + +1:17:58.320 --> 1:18:03.320 + While my job was not routine, it wasn't pursuit of numbers. + +1:18:03.320 --> 1:18:05.320 + Pursuit of, can I make more money? + +1:18:05.320 --> 1:18:07.320 + Can I fund more great companies? + +1:18:07.320 --> 1:18:09.320 + Can I raise more money? + +1:18:09.320 --> 1:18:13.320 + Can I make sure our VC is ranked higher and higher every year? + +1:18:13.320 --> 1:18:20.320 + This competitive nature of driving for bigger numbers and better numbers + +1:18:20.320 --> 1:18:27.320 + became an endless pursuit of that's mechanical. + +1:18:27.320 --> 1:18:31.320 + And bigger numbers really didn't make me happier. + +1:18:31.320 --> 1:18:36.320 + And faced with death, I realized bigger numbers really meant nothing. + +1:18:36.320 --> 1:18:42.320 + And what was important is that people who have given their heart and their love to me + +1:18:42.320 --> 1:18:45.320 + deserve for me to do the same. + +1:18:45.320 --> 1:18:52.320 + So there's deep profound truth in that, that everyone should hear and internalize. + +1:18:52.320 --> 1:18:56.320 + And that's really powerful for you to say that. + +1:18:56.320 --> 1:19:02.320 + I have to ask sort of a difficult question here. + +1:19:02.320 --> 1:19:07.320 + So I've competed in sports my whole life, looking historically. + +1:19:07.320 --> 1:19:14.320 + I'd like to challenge some aspect of that a little bit on the point of hard work. + +1:19:14.320 --> 1:19:20.320 + That it feels that there are certain aspects that is the greatest, + +1:19:20.320 --> 1:19:26.320 + the most beautiful aspects of human nature, is the ability to become obsessed, + +1:19:26.320 --> 1:19:33.320 + of becoming extremely passionate to the point where, yes, flaws are revealed + +1:19:33.320 --> 1:19:36.320 + and just giving yourself fully to a task. + +1:19:36.320 --> 1:19:41.320 + That is, in another sense, you mentioned love being important, + +1:19:41.320 --> 1:19:47.320 + but in another sense, this kind of obsession, this pure exhibition of passion and hard work + +1:19:47.320 --> 1:19:50.320 + is truly what it means to be human. + +1:19:50.320 --> 1:19:53.320 + What lessons should we take that's deeper? + +1:19:53.320 --> 1:19:55.320 + Because you've accomplished incredible things. + +1:19:55.320 --> 1:19:57.320 + Like chasing numbers. + +1:19:57.320 --> 1:20:01.320 + But really, there's some incredible work there. + +1:20:01.320 --> 1:20:07.320 + So how do you think about that when you look back in your 20s, your 30s? + +1:20:07.320 --> 1:20:10.320 + What would you do differently? + +1:20:10.320 --> 1:20:16.320 + Would you really take back some of the incredible hard work? + +1:20:16.320 --> 1:20:17.320 + I would. + +1:20:17.320 --> 1:20:20.320 + But it's in percentages, right? + +1:20:20.320 --> 1:20:22.320 + We're both now computer scientists. + +1:20:22.320 --> 1:20:27.320 + So I think when one balances one's life, when one is younger, + +1:20:27.320 --> 1:20:33.320 + you might give a smaller percentage to family, but you would still give them high priority. + +1:20:33.320 --> 1:20:38.320 + And when you get older, you would give a larger percentage to them and still the high priority. + +1:20:38.320 --> 1:20:43.320 + And when you're near retirement, you give most of it to them and the highest priority. + +1:20:43.320 --> 1:20:50.320 + So I think the key point is not that we would work 20 hours less for the whole life + +1:20:50.320 --> 1:20:56.320 + and just spend it aimlessly with the family, but that when the family has a need, + +1:20:56.320 --> 1:21:02.320 + when your wife is having a baby, when your daughter has a birthday, + +1:21:02.320 --> 1:21:07.320 + or when they're depressed, or when they're celebrating something, + +1:21:07.320 --> 1:21:11.320 + or when they have a get together, or when we have family time, + +1:21:11.320 --> 1:21:18.320 + that is important for us to put down our phone and PC and be 100% with them. + +1:21:18.320 --> 1:21:26.320 + And that priority on the things that really matter isn't going to be so taxing + +1:21:26.320 --> 1:21:32.320 + that it would eliminate or even dramatically reduce our accomplishments. + +1:21:32.320 --> 1:21:36.320 + It might have some impact, but it might also have other impact + +1:21:36.320 --> 1:21:39.320 + because if you have a happier family, maybe you fight less. + +1:21:39.320 --> 1:21:45.320 + If you fight less, you don't spend time taking care of all the aftermath of a fight. + +1:21:45.320 --> 1:21:46.320 + That's right. + +1:21:46.320 --> 1:21:48.320 + And I'm sure that it would take more time. + +1:21:48.320 --> 1:21:53.320 + And if it did, I'd be willing to take that reduction. + +1:21:53.320 --> 1:21:56.320 + And it's not a dramatic number, but it's a number + +1:21:56.320 --> 1:22:00.320 + that I think would give me a greater degree of happiness + +1:22:00.320 --> 1:22:03.320 + and knowing that I've done the right thing + +1:22:03.320 --> 1:22:10.320 + and still have plenty of hours to get the success that I want to get. + +1:22:10.320 --> 1:22:14.320 + So given the many successful companies that you've launched + +1:22:14.320 --> 1:22:17.320 + and much success throughout your career, + +1:22:17.320 --> 1:22:25.320 + what advice would you give to young people today looking, + +1:22:25.320 --> 1:22:28.320 + or it doesn't have to be young, but people today looking to launch + +1:22:28.320 --> 1:22:35.320 + and to create the next $1 billion tech startup, or even AI based startup? + +1:22:35.320 --> 1:22:42.320 + I would suggest that people understand technology waves move quickly. + +1:22:42.320 --> 1:22:45.320 + What worked two years ago may not work today. + +1:22:45.320 --> 1:22:49.320 + And that is very much a case in point for AI. + +1:22:49.320 --> 1:22:53.320 + I think two years ago, or maybe three years ago, + +1:22:53.320 --> 1:22:57.320 + you certainly could say I have a couple of super smart PhDs + +1:22:57.320 --> 1:22:59.320 + and we're not sure what we're going to do, + +1:22:59.320 --> 1:23:04.320 + but here's how we're going to start and get funding for a very high valuation. + +1:23:04.320 --> 1:23:11.320 + Those days are over because AI is going from rocket science towards mainstream. + +1:23:11.320 --> 1:23:14.320 + Not yet commodity, but more mainstream. + +1:23:14.320 --> 1:23:20.320 + So first, the creation of any company to eventual capitalist + +1:23:20.320 --> 1:23:25.320 + has to be creation of business value and monetary value. + +1:23:25.320 --> 1:23:29.320 + And when you have a very scarce commodity, + +1:23:29.320 --> 1:23:34.320 + VCs may be willing to accept greater uncertainty. + +1:23:34.320 --> 1:23:40.320 + But now the number of people who have the equivalent of PhD three years ago + +1:23:40.320 --> 1:23:43.320 + because that can be learned more quickly. + +1:23:43.320 --> 1:23:45.320 + Platforms are emerging. + +1:23:45.320 --> 1:23:51.320 + The cost to become an AI engineer is much lower and there are many more AI engineers. + +1:23:51.320 --> 1:23:53.320 + So the market is different. + +1:23:53.320 --> 1:23:57.320 + So I would suggest someone who wants to build an AI company + +1:23:57.320 --> 1:24:01.320 + be thinking about the normal business questions. + +1:24:01.320 --> 1:24:05.320 + What customer cases are you trying to address? + +1:24:05.320 --> 1:24:08.320 + What kind of pain are you trying to address? + +1:24:08.320 --> 1:24:10.320 + How does that translate to value? + +1:24:10.320 --> 1:24:16.320 + How will you extract value and get paid through what channel? + +1:24:16.320 --> 1:24:19.320 + And how much business value will get created? + +1:24:19.320 --> 1:24:26.320 + That today needs to be thought about much earlier up front than it did three years ago. + +1:24:26.320 --> 1:24:30.320 + The scarcity question of AI talent has changed. + +1:24:30.320 --> 1:24:32.320 + The number of AI talent has changed. + +1:24:32.320 --> 1:24:41.320 + So now you need not just AI but also understanding of business customer and the marketplace. + +1:24:41.320 --> 1:24:49.320 + So I also think you should have a more reasonable evaluation expectation + +1:24:49.320 --> 1:24:51.320 + and growth expectation. + +1:24:51.320 --> 1:24:53.320 + There's going to be more competition. + +1:24:53.320 --> 1:25:00.320 + But the good news though is that AI technologies are now more available in open source. + +1:25:00.320 --> 1:25:06.320 + TensorFlow, PyTorch and such tools are much easier to use. + +1:25:06.320 --> 1:25:13.320 + So you should be able to experiment and get results iteratively faster than before. + +1:25:13.320 --> 1:25:18.320 + So take more of a business mindset to this. + +1:25:18.320 --> 1:25:25.320 + Think less of this as a laboratory taken into a company because we've gone beyond that stage. + +1:25:25.320 --> 1:25:31.320 + The only exception is if you truly have a breakthrough in some technology that really no one has, + +1:25:31.320 --> 1:25:34.320 + then the old way still works. + +1:25:34.320 --> 1:25:36.320 + But I think that's harder and harder now. + +1:25:36.320 --> 1:25:44.320 + So I know you believe as many do that we're far from creating an artificial general intelligence system. + +1:25:44.320 --> 1:25:54.320 + But say once we do and you get to ask her one question, what would that question be? + +1:25:54.320 --> 1:26:00.320 + What is it that differentiates you and me? + +1:26:00.320 --> 1:26:05.320 + Beautifully put, Kaifu, thank you so much for your time today. + +1:26:05.320 --> 1:26:26.320 + Thank you. +