diff --git "a/vtt/episode_040_large.vtt" "b/vtt/episode_040_large.vtt" new file mode 100644--- /dev/null +++ "b/vtt/episode_040_large.vtt" @@ -0,0 +1,4865 @@ +WEBVTT + +00:00.000 --> 00:03.220 + The following is a conversation with Regina Barzilay. + +00:03.220 --> 00:06.700 + She's a professor at MIT and a world class researcher + +00:06.700 --> 00:08.340 + in natural language processing + +00:08.340 --> 00:12.460 + and applications of deep learning to chemistry and oncology + +00:12.460 --> 00:15.340 + or the use of deep learning for early diagnosis, + +00:15.340 --> 00:18.300 + prevention and treatment of cancer. + +00:18.300 --> 00:21.020 + She has also been recognized for teaching + +00:21.020 --> 00:24.700 + of several successful AI related courses at MIT, + +00:24.700 --> 00:26.840 + including the popular Introduction + +00:26.840 --> 00:28.920 + to Machine Learning course. + +00:28.920 --> 00:32.160 + This is the Artificial Intelligence podcast. + +00:32.160 --> 00:34.560 + If you enjoy it, subscribe on YouTube, + +00:34.560 --> 00:37.840 + give it five stars on iTunes, support it on Patreon + +00:37.840 --> 00:39.840 + or simply connect with me on Twitter + +00:39.840 --> 00:43.760 + at Lex Friedman spelled F R I D M A N. + +00:43.760 --> 00:47.760 + And now here's my conversation with Regina Barzilay. + +00:48.840 --> 00:50.320 + In an interview you've mentioned + +00:50.320 --> 00:51.960 + that if there's one course you would take, + +00:51.960 --> 00:54.600 + it would be a literature course with a friend of yours + +00:54.600 --> 00:56.360 + that a friend of yours teaches. + +00:56.360 --> 00:59.160 + Just out of curiosity, because I couldn't find anything + +00:59.160 --> 01:04.160 + on it, are there books or ideas that had profound impact + +01:04.400 --> 01:07.200 + on your life journey, books and ideas perhaps + +01:07.200 --> 01:10.800 + outside of computer science and the technical fields? + +01:11.780 --> 01:14.680 + I think because I'm spending a lot of my time at MIT + +01:14.680 --> 01:18.280 + and previously in other institutions where I was a student, + +01:18.280 --> 01:21.040 + I have limited ability to interact with people. + +01:21.040 --> 01:22.640 + So a lot of what I know about the world + +01:22.640 --> 01:24.220 + actually comes from books. + +01:24.220 --> 01:27.240 + And there were quite a number of books + +01:27.240 --> 01:31.380 + that had profound impact on me and how I view the world. + +01:31.380 --> 01:35.820 + Let me just give you one example of such a book. + +01:35.820 --> 01:39.660 + I've maybe a year ago read a book + +01:39.660 --> 01:42.500 + called The Emperor of All Melodies. + +01:42.500 --> 01:45.740 + It's a book about, it's kind of a history of science book + +01:45.740 --> 01:50.740 + on how the treatments and drugs for cancer were developed. + +01:50.740 --> 01:54.580 + And that book, despite the fact that I am in the business + +01:54.580 --> 01:59.580 + of science, really opened my eyes on how imprecise + +01:59.780 --> 02:03.060 + and imperfect the discovery process is + +02:03.060 --> 02:05.820 + and how imperfect our current solutions + +02:06.980 --> 02:11.060 + and what makes science succeed and be implemented. + +02:11.060 --> 02:14.100 + And sometimes it's actually not the strengths of the idea, + +02:14.100 --> 02:17.420 + but devotion of the person who wants to see it implemented. + +02:17.420 --> 02:19.780 + So this is one of the books that, you know, + +02:19.780 --> 02:22.300 + at least for the last year, quite changed the way + +02:22.300 --> 02:24.940 + I'm thinking about scientific process + +02:24.940 --> 02:26.700 + just from the historical perspective + +02:26.700 --> 02:31.700 + and what do I need to do to make my ideas really implemented. + +02:33.460 --> 02:36.060 + Let me give you an example of a book + +02:36.060 --> 02:39.580 + which is not kind of, which is a fiction book. + +02:40.620 --> 02:43.100 + It's a book called Americana. + +02:44.420 --> 02:48.780 + And this is a book about a young female student + +02:48.780 --> 02:53.260 + who comes from Africa to study in the United States. + +02:53.260 --> 02:57.740 + And it describes her past, you know, within her studies + +02:57.740 --> 03:02.020 + and her life transformation that, you know, + +03:02.020 --> 03:06.540 + in a new country and kind of adaptation to a new culture. + +03:06.540 --> 03:11.220 + And when I read this book, I saw myself + +03:11.220 --> 03:13.540 + in many different points of it, + +03:13.540 --> 03:18.540 + but it also kind of gave me the lens on different events. + +03:20.140 --> 03:22.060 + And some of it that I never actually paid attention. + +03:22.060 --> 03:24.700 + One of the funny stories in this book + +03:24.700 --> 03:29.700 + is how she arrives to her new college + +03:30.420 --> 03:32.900 + and she starts speaking in English + +03:32.900 --> 03:35.700 + and she had this beautiful British accent + +03:35.700 --> 03:39.860 + because that's how she was educated in her country. + +03:39.860 --> 03:40.980 + This is not my case. + +03:40.980 --> 03:45.460 + And then she notices that the person who talks to her, + +03:45.460 --> 03:47.220 + you know, talks to her in a very funny way, + +03:47.220 --> 03:48.340 + in a very slow way. + +03:48.340 --> 03:51.460 + And she's thinking that this woman is disabled + +03:51.460 --> 03:54.500 + and she's also trying to kind of to accommodate her. + +03:54.500 --> 03:56.700 + And then after a while, when she finishes her discussion + +03:56.700 --> 03:58.580 + with this officer from her college, + +03:59.860 --> 04:02.100 + she sees how she interacts with the other students, + +04:02.100 --> 04:03.020 + with American students. + +04:03.020 --> 04:08.020 + And she discovers that actually she talked to her this way + +04:08.020 --> 04:11.020 + because she saw that she doesn't understand English. + +04:11.020 --> 04:14.180 + And I thought, wow, this is a funny experience. + +04:14.180 --> 04:16.940 + And literally within few weeks, + +04:16.940 --> 04:20.820 + I went to LA to a conference + +04:20.820 --> 04:23.180 + and I asked somebody in the airport, + +04:23.180 --> 04:25.580 + you know, how to find like a cab or something. + +04:25.580 --> 04:28.380 + And then I noticed that this person is talking + +04:28.380 --> 04:29.220 + in a very strange way. + +04:29.220 --> 04:31.100 + And my first thought was that this person + +04:31.100 --> 04:34.500 + have some, you know, pronunciation issues or something. + +04:34.500 --> 04:36.060 + And I'm trying to talk very slowly to him + +04:36.060 --> 04:38.580 + and I was with another professor, Ernst Frankel. + +04:38.580 --> 04:42.180 + And he's like laughing because it's funny + +04:42.180 --> 04:44.860 + that I don't get that the guy is talking in this way + +04:44.860 --> 04:46.060 + because he thinks that I cannot speak. + +04:46.060 --> 04:49.100 + So it was really kind of mirroring experience. + +04:49.100 --> 04:53.300 + And it led me think a lot about my own experiences + +04:53.300 --> 04:56.060 + moving, you know, from different countries. + +04:56.060 --> 04:59.300 + So I think that books play a big role + +04:59.300 --> 05:01.780 + in my understanding of the world. + +05:01.780 --> 05:06.420 + On the science question, you mentioned that + +05:06.420 --> 05:09.780 + it made you discover that personalities of human beings + +05:09.780 --> 05:12.420 + are more important than perhaps ideas. + +05:12.420 --> 05:13.660 + Is that what I heard? + +05:13.660 --> 05:15.740 + It's not necessarily that they are more important + +05:15.740 --> 05:19.180 + than ideas, but I think that ideas on their own + +05:19.180 --> 05:20.460 + are not sufficient. + +05:20.460 --> 05:24.660 + And many times, at least at the local horizon, + +05:24.660 --> 05:29.140 + it's the personalities and their devotion to their ideas + +05:29.140 --> 05:32.980 + is really that locally changes the landscape. + +05:32.980 --> 05:37.500 + Now, if you're looking at AI, like let's say 30 years ago, + +05:37.500 --> 05:39.180 + you know, dark ages of AI or whatever, + +05:39.180 --> 05:42.420 + what is symbolic times, you can use any word. + +05:42.420 --> 05:44.660 + You know, there were some people, + +05:44.660 --> 05:46.620 + now we're looking at a lot of that work + +05:46.620 --> 05:48.780 + and we're kind of thinking this was not really + +05:48.780 --> 05:52.220 + maybe a relevant work, but you can see that some people + +05:52.220 --> 05:54.900 + managed to take it and to make it so shiny + +05:54.900 --> 05:59.260 + and dominate the academic world + +05:59.260 --> 06:02.380 + and make it to be the standard. + +06:02.380 --> 06:05.180 + If you look at the area of natural language processing, + +06:06.420 --> 06:09.140 + it is well known fact that the reason that statistics + +06:09.140 --> 06:13.980 + in NLP took such a long time to become mainstream + +06:13.980 --> 06:16.860 + because there were quite a number of personalities + +06:16.860 --> 06:18.460 + which didn't believe in this idea + +06:18.460 --> 06:22.060 + and didn't stop research progress in this area. + +06:22.060 --> 06:25.900 + So I do not think that, you know, + +06:25.900 --> 06:28.940 + kind of asymptotically maybe personalities matters, + +06:28.940 --> 06:33.940 + but I think locally it does make quite a bit of impact + +06:33.940 --> 06:36.900 + and it's generally, you know, + +06:36.900 --> 06:41.340 + speeds up the rate of adoption of the new ideas. + +06:41.340 --> 06:43.500 + Yeah, and the other interesting question + +06:43.500 --> 06:46.540 + is in the early days of particular discipline, + +06:46.540 --> 06:50.460 + I think you mentioned in that book + +06:50.460 --> 06:52.340 + is ultimately a book of cancer. + +06:52.340 --> 06:55.100 + It's called The Emperor of All Melodies. + +06:55.100 --> 06:58.580 + Yeah, and those melodies included the trying to, + +06:58.580 --> 07:00.740 + the medicine, was it centered around? + +07:00.740 --> 07:04.900 + So it was actually centered on, you know, + +07:04.900 --> 07:07.180 + how people thought of curing cancer. + +07:07.180 --> 07:10.660 + Like for me, it was really a discovery how people, + +07:10.660 --> 07:14.140 + what was the science of chemistry behind drug development + +07:14.140 --> 07:17.220 + that it actually grew up out of dying, + +07:17.220 --> 07:19.780 + like coloring industry that people + +07:19.780 --> 07:23.780 + who developed chemistry in 19th century in Germany + +07:23.780 --> 07:28.140 + and Britain to do, you know, the really new dyes. + +07:28.140 --> 07:30.180 + They looked at the molecules and identified + +07:30.180 --> 07:32.140 + that they do certain things to cells. + +07:32.140 --> 07:34.500 + And from there, the process started. + +07:34.500 --> 07:35.740 + And, you know, like historically saying, + +07:35.740 --> 07:36.900 + yeah, this is fascinating + +07:36.900 --> 07:38.700 + that they managed to make the connection + +07:38.700 --> 07:42.300 + and look under the microscope and do all this discovery. + +07:42.300 --> 07:44.340 + But as you continue reading about it + +07:44.340 --> 07:48.780 + and you read about how chemotherapy drugs + +07:48.780 --> 07:50.500 + which were developed in Boston, + +07:50.500 --> 07:52.500 + and some of them were developed. + +07:52.500 --> 07:57.500 + And Farber, Dr. Farber from Dana Farber, + +07:57.500 --> 08:00.460 + you know, how the experiments were done + +08:00.460 --> 08:03.340 + that, you know, there was some miscalculation, + +08:03.340 --> 08:04.540 + let's put it this way. + +08:04.540 --> 08:06.740 + And they tried it on the patients and they just, + +08:06.740 --> 08:09.980 + and those were children with leukemia and they died. + +08:09.980 --> 08:11.660 + And then they tried another modification. + +08:11.660 --> 08:15.020 + You look at the process, how imperfect is this process? + +08:15.020 --> 08:17.500 + And, you know, like, if we're again looking back + +08:17.500 --> 08:19.180 + like 60 years ago, 70 years ago, + +08:19.180 --> 08:20.780 + you can kind of understand it. + +08:20.780 --> 08:23.020 + But some of the stories in this book + +08:23.020 --> 08:24.620 + which were really shocking to me + +08:24.620 --> 08:27.980 + were really happening, you know, maybe decades ago. + +08:27.980 --> 08:30.660 + And we still don't have a vehicle + +08:30.660 --> 08:35.100 + to do it much more fast and effective and, you know, + +08:35.100 --> 08:38.220 + scientific the way I'm thinking computer science scientific. + +08:38.220 --> 08:40.420 + So from the perspective of computer science, + +08:40.420 --> 08:43.780 + you've gotten a chance to work the application to cancer + +08:43.780 --> 08:44.860 + and to medicine in general. + +08:44.860 --> 08:48.420 + From a perspective of an engineer and a computer scientist, + +08:48.420 --> 08:51.780 + how far along are we from understanding the human body, + +08:51.780 --> 08:55.140 + biology of being able to manipulate it + +08:55.140 --> 08:57.940 + in a way we can cure some of the maladies, + +08:57.940 --> 08:59.740 + some of the diseases? + +08:59.740 --> 09:02.220 + So this is very interesting question. + +09:03.460 --> 09:06.020 + And if you're thinking as a computer scientist + +09:06.020 --> 09:09.820 + about this problem, I think one of the reasons + +09:09.820 --> 09:11.900 + that we succeeded in the areas + +09:11.900 --> 09:13.980 + we as a computer scientist succeeded + +09:13.980 --> 09:16.260 + is because we don't have, + +09:16.260 --> 09:18.980 + we are not trying to understand in some ways. + +09:18.980 --> 09:22.260 + Like if you're thinking about like eCommerce, Amazon, + +09:22.260 --> 09:24.220 + Amazon doesn't really understand you. + +09:24.220 --> 09:27.700 + And that's why it recommends you certain books + +09:27.700 --> 09:29.580 + or certain products, correct? + +09:30.660 --> 09:34.660 + And, you know, traditionally when people + +09:34.660 --> 09:36.380 + were thinking about marketing, you know, + +09:36.380 --> 09:39.780 + they divided the population to different kind of subgroups, + +09:39.780 --> 09:41.740 + identify the features of this subgroup + +09:41.740 --> 09:43.140 + and come up with a strategy + +09:43.140 --> 09:45.580 + which is specific to that subgroup. + +09:45.580 --> 09:47.340 + If you're looking about recommendation system, + +09:47.340 --> 09:50.580 + they're not claiming that they're understanding somebody, + +09:50.580 --> 09:52.700 + they're just managing to, + +09:52.700 --> 09:54.780 + from the patterns of your behavior + +09:54.780 --> 09:57.540 + to recommend you a product. + +09:57.540 --> 09:59.580 + Now, if you look at the traditional biology, + +09:59.580 --> 10:03.180 + and obviously I wouldn't say that I + +10:03.180 --> 10:06.180 + at any way, you know, educated in this field, + +10:06.180 --> 10:09.300 + but you know what I see, there's really a lot of emphasis + +10:09.300 --> 10:10.660 + on mechanistic understanding. + +10:10.660 --> 10:12.540 + And it was very surprising to me + +10:12.540 --> 10:13.820 + coming from computer science, + +10:13.820 --> 10:17.580 + how much emphasis is on this understanding. + +10:17.580 --> 10:20.740 + And given the complexity of the system, + +10:20.740 --> 10:23.220 + maybe the deterministic full understanding + +10:23.220 --> 10:27.380 + of this process is, you know, beyond our capacity. + +10:27.380 --> 10:29.460 + And the same ways in computer science + +10:29.460 --> 10:31.540 + when we're doing recognition, when you do recommendation + +10:31.540 --> 10:32.780 + and many other areas, + +10:32.780 --> 10:35.940 + it's just probabilistic matching process. + +10:35.940 --> 10:40.100 + And in some way, maybe in certain cases, + +10:40.100 --> 10:42.940 + we shouldn't even attempt to understand + +10:42.940 --> 10:45.780 + or we can attempt to understand, but in parallel, + +10:45.780 --> 10:48.060 + we can actually do this kind of matchings + +10:48.060 --> 10:51.060 + that would help us to find key role + +10:51.060 --> 10:54.100 + to do early diagnostics and so on. + +10:54.100 --> 10:55.860 + And I know that in these communities, + +10:55.860 --> 10:59.060 + it's really important to understand, + +10:59.060 --> 11:00.700 + but I'm sometimes wondering, you know, + +11:00.700 --> 11:02.940 + what exactly does it mean to understand here? + +11:02.940 --> 11:05.500 + Well, there's stuff that works and, + +11:05.500 --> 11:07.620 + but that can be, like you said, + +11:07.620 --> 11:10.340 + separate from this deep human desire + +11:10.340 --> 11:12.700 + to uncover the mysteries of the universe, + +11:12.700 --> 11:16.140 + of science, of the way the body works, + +11:16.140 --> 11:17.620 + the way the mind works. + +11:17.620 --> 11:19.540 + It's the dream of symbolic AI, + +11:19.540 --> 11:24.540 + of being able to reduce human knowledge into logic + +11:25.220 --> 11:26.900 + and be able to play with that logic + +11:26.900 --> 11:28.700 + in a way that's very explainable + +11:28.700 --> 11:30.300 + and understandable for us humans. + +11:30.300 --> 11:31.780 + I mean, that's a beautiful dream. + +11:31.780 --> 11:34.860 + So I understand it, but it seems that + +11:34.860 --> 11:37.900 + what seems to work today and we'll talk about it more + +11:37.900 --> 11:40.780 + is as much as possible, reduce stuff into data, + +11:40.780 --> 11:43.900 + reduce whatever problem you're interested in to data + +11:43.900 --> 11:47.060 + and try to apply statistical methods, + +11:47.060 --> 11:49.100 + apply machine learning to that. + +11:49.100 --> 11:51.140 + On a personal note, + +11:51.140 --> 11:54.140 + you were diagnosed with breast cancer in 2014. + +11:55.380 --> 11:58.420 + What did facing your mortality make you think about? + +11:58.420 --> 12:00.260 + How did it change you? + +12:00.260 --> 12:01.860 + You know, this is a great question + +12:01.860 --> 12:03.820 + and I think that I was interviewed many times + +12:03.820 --> 12:05.740 + and nobody actually asked me this question. + +12:05.740 --> 12:09.700 + I think I was 43 at a time. + +12:09.700 --> 12:12.860 + And the first time I realized in my life that I may die + +12:12.860 --> 12:14.460 + and I never thought about it before. + +12:14.460 --> 12:17.260 + And there was a long time since you're diagnosed + +12:17.260 --> 12:18.580 + until you actually know what you have + +12:18.580 --> 12:20.180 + and how severe is your disease. + +12:20.180 --> 12:23.500 + For me, it was like maybe two and a half months. + +12:23.500 --> 12:28.340 + And I didn't know where I am during this time + +12:28.340 --> 12:30.660 + because I was getting different tests + +12:30.660 --> 12:33.380 + and one would say it's bad and I would say, no, it is not. + +12:33.380 --> 12:34.900 + So until I knew where I am, + +12:34.900 --> 12:36.300 + I really was thinking about + +12:36.300 --> 12:38.220 + all these different possible outcomes. + +12:38.220 --> 12:39.700 + Were you imagining the worst + +12:39.700 --> 12:41.940 + or were you trying to be optimistic or? + +12:41.940 --> 12:43.540 + It would be really, + +12:43.540 --> 12:47.340 + I don't remember what was my thinking. + +12:47.340 --> 12:51.100 + It was really a mixture with many components at the time + +12:51.100 --> 12:54.100 + speaking in our terms. + +12:54.100 --> 12:59.100 + And one thing that I remember, + +12:59.340 --> 13:01.500 + and every test comes and then you're saying, + +13:01.500 --> 13:03.300 + oh, it could be this or it may not be this. + +13:03.300 --> 13:04.700 + And you're hopeful and then you're desperate. + +13:04.700 --> 13:07.660 + So it's like, there is a whole slew of emotions + +13:07.660 --> 13:08.700 + that goes through you. + +13:09.820 --> 13:14.820 + But what I remember is that when I came back to MIT, + +13:15.100 --> 13:17.780 + I was kind of going the whole time through the treatment + +13:17.780 --> 13:19.780 + to MIT, but my brain was not really there. + +13:19.780 --> 13:21.820 + But when I came back, really finished my treatment + +13:21.820 --> 13:23.860 + and I was here teaching and everything, + +13:24.900 --> 13:27.060 + I look back at what my group was doing, + +13:27.060 --> 13:28.820 + what other groups was doing. + +13:28.820 --> 13:30.820 + And I saw these trivialities. + +13:30.820 --> 13:33.260 + It's like people are building their careers + +13:33.260 --> 13:36.900 + on improving some parts around two or 3% or whatever. + +13:36.900 --> 13:38.380 + I was, it's like, seriously, + +13:38.380 --> 13:40.740 + I did a work on how to decipher ugaritic, + +13:40.740 --> 13:42.860 + like a language that nobody speak and whatever, + +13:42.860 --> 13:46.140 + like what is significance? + +13:46.140 --> 13:49.020 + When all of a sudden, I walked out of MIT, + +13:49.020 --> 13:51.860 + which is when people really do care + +13:51.860 --> 13:54.500 + what happened to your ICLR paper, + +13:54.500 --> 13:57.900 + what is your next publication to ACL, + +13:57.900 --> 14:01.860 + to the world where people, you see a lot of suffering + +14:01.860 --> 14:04.900 + that I'm kind of totally shielded on it on daily basis. + +14:04.900 --> 14:07.460 + And it's like the first time I've seen like real life + +14:07.460 --> 14:08.660 + and real suffering. + +14:09.700 --> 14:13.260 + And I was thinking, why are we trying to improve the parser + +14:13.260 --> 14:18.260 + or deal with trivialities when we have capacity + +14:18.340 --> 14:20.700 + to really make a change? + +14:20.700 --> 14:24.620 + And it was really challenging to me because on one hand, + +14:24.620 --> 14:27.420 + I have my graduate students really want to do their papers + +14:27.420 --> 14:29.860 + and their work, and they want to continue to do + +14:29.860 --> 14:31.900 + what they were doing, which was great. + +14:31.900 --> 14:36.300 + And then it was me who really kind of reevaluated + +14:36.300 --> 14:37.460 + what is the importance. + +14:37.460 --> 14:40.260 + And also at that point, because I had to take some break, + +14:42.500 --> 14:47.500 + I look back into like my years in science + +14:47.740 --> 14:50.460 + and I was thinking, like 10 years ago, + +14:50.460 --> 14:52.940 + this was the biggest thing, I don't know, topic models. + +14:52.940 --> 14:55.340 + We have like millions of papers on topic models + +14:55.340 --> 14:56.500 + and variation of topics models. + +14:56.500 --> 14:58.580 + Now it's totally like irrelevant. + +14:58.580 --> 15:02.460 + And you start looking at this, what do you perceive + +15:02.460 --> 15:04.500 + as important at different point of time + +15:04.500 --> 15:08.900 + and how it fades over time. + +15:08.900 --> 15:12.980 + And since we have a limited time, + +15:12.980 --> 15:14.900 + all of us have limited time on us, + +15:14.900 --> 15:18.380 + it's really important to prioritize things + +15:18.380 --> 15:20.540 + that really matter to you, maybe matter to you + +15:20.540 --> 15:22.020 + at that particular point. + +15:22.020 --> 15:24.380 + But it's important to take some time + +15:24.380 --> 15:26.940 + and understand what matters to you, + +15:26.940 --> 15:28.860 + which may not necessarily be the same + +15:28.860 --> 15:31.700 + as what matters to the rest of your scientific community + +15:31.700 --> 15:34.580 + and pursue that vision. + +15:34.580 --> 15:38.460 + So that moment, did it make you cognizant? + +15:38.460 --> 15:42.500 + You mentioned suffering of just the general amount + +15:42.500 --> 15:44.340 + of suffering in the world. + +15:44.340 --> 15:45.620 + Is that what you're referring to? + +15:45.620 --> 15:47.420 + So as opposed to topic models + +15:47.420 --> 15:50.780 + and specific detailed problems in NLP, + +15:50.780 --> 15:54.460 + did you start to think about other people + +15:54.460 --> 15:56.940 + who have been diagnosed with cancer? + +15:56.940 --> 16:00.020 + Is that the way you started to see the world perhaps? + +16:00.020 --> 16:00.860 + Oh, absolutely. + +16:00.860 --> 16:04.980 + And it actually creates, because like, for instance, + +16:04.980 --> 16:05.820 + there is parts of the treatment + +16:05.820 --> 16:08.500 + where you need to go to the hospital every day + +16:08.500 --> 16:11.620 + and you see the community of people that you see + +16:11.620 --> 16:16.100 + and many of them are much worse than I was at a time. + +16:16.100 --> 16:20.460 + And you all of a sudden see it all. + +16:20.460 --> 16:23.940 + And people who are happier someday + +16:23.940 --> 16:25.300 + just because they feel better. + +16:25.300 --> 16:28.500 + And for people who are in our normal realm, + +16:28.500 --> 16:30.820 + you take it totally for granted that you feel well, + +16:30.820 --> 16:32.940 + that if you decide to go running, you can go running + +16:32.940 --> 16:35.900 + and you're pretty much free + +16:35.900 --> 16:37.620 + to do whatever you want with your body. + +16:37.620 --> 16:40.180 + Like I saw like a community, + +16:40.180 --> 16:42.820 + my community became those people. + +16:42.820 --> 16:47.460 + And I remember one of my friends, Dina Katabi, + +16:47.460 --> 16:50.420 + took me to Prudential to buy me a gift for my birthday. + +16:50.420 --> 16:52.340 + And it was like the first time in months + +16:52.340 --> 16:54.980 + that I went to kind of to see other people. + +16:54.980 --> 16:58.180 + And I was like, wow, first of all, these people, + +16:58.180 --> 16:59.820 + they are happy and they're laughing + +16:59.820 --> 17:02.620 + and they're very different from these other my people. + +17:02.620 --> 17:04.620 + And second of thing, I think it's totally crazy. + +17:04.620 --> 17:06.620 + They're like laughing and wasting their money + +17:06.620 --> 17:08.420 + on some stupid gifts. + +17:08.420 --> 17:12.540 + And they may die. + +17:12.540 --> 17:15.940 + They already may have cancer and they don't understand it. + +17:15.940 --> 17:20.060 + So you can really see how the mind changes + +17:20.060 --> 17:22.340 + that you can see that, + +17:22.340 --> 17:23.180 + before that you can ask, + +17:23.180 --> 17:24.380 + didn't you know that you're gonna die? + +17:24.380 --> 17:28.340 + Of course I knew, but it was a kind of a theoretical notion. + +17:28.340 --> 17:31.060 + It wasn't something which was concrete. + +17:31.060 --> 17:33.900 + And at that point, when you really see it + +17:33.900 --> 17:38.060 + and see how little means sometimes the system has + +17:38.060 --> 17:41.740 + to have them, you really feel that we need to take a lot + +17:41.740 --> 17:45.420 + of our brilliance that we have here at MIT + +17:45.420 --> 17:48.020 + and translate it into something useful. + +17:48.020 --> 17:50.540 + Yeah, and you still couldn't have a lot of definitions, + +17:50.540 --> 17:53.620 + but of course, alleviating, suffering, alleviating, + +17:53.620 --> 17:57.460 + trying to cure cancer is a beautiful mission. + +17:57.460 --> 18:01.940 + So I of course know theoretically the notion of cancer, + +18:01.940 --> 18:06.940 + but just reading more and more about it's 1.7 million + +18:07.100 --> 18:09.860 + new cancer cases in the United States every year, + +18:09.860 --> 18:13.460 + 600,000 cancer related deaths every year. + +18:13.460 --> 18:18.460 + So this has a huge impact, United States globally. + +18:19.340 --> 18:24.340 + When broadly, before we talk about how machine learning, + +18:24.340 --> 18:27.180 + how MIT can help, + +18:27.180 --> 18:32.100 + when do you think we as a civilization will cure cancer? + +18:32.100 --> 18:34.980 + How hard of a problem is it from everything you've learned + +18:34.980 --> 18:35.940 + from it recently? + +18:37.260 --> 18:39.300 + I cannot really assess it. + +18:39.300 --> 18:42.100 + What I do believe will happen with the advancement + +18:42.100 --> 18:45.940 + in machine learning is that a lot of types of cancer + +18:45.940 --> 18:48.500 + we will be able to predict way early + +18:48.500 --> 18:53.420 + and more effectively utilize existing treatments. + +18:53.420 --> 18:57.540 + I think, I hope at least that with all the advancements + +18:57.540 --> 19:01.180 + in AI and drug discovery, we would be able + +19:01.180 --> 19:04.700 + to much faster find relevant molecules. + +19:04.700 --> 19:08.220 + What I'm not sure about is how long it will take + +19:08.220 --> 19:11.940 + the medical establishment and regulatory bodies + +19:11.940 --> 19:14.780 + to kind of catch up and to implement it. + +19:14.780 --> 19:17.420 + And I think this is a very big piece of puzzle + +19:17.420 --> 19:20.420 + that is currently not addressed. + +19:20.420 --> 19:21.780 + That's the really interesting question. + +19:21.780 --> 19:25.460 + So first a small detail that I think the answer is yes, + +19:25.460 --> 19:30.460 + but is cancer one of the diseases that when detected earlier + +19:33.700 --> 19:37.820 + that's a significantly improves the outcomes? + +19:37.820 --> 19:41.020 + So like, cause we will talk about there's the cure + +19:41.020 --> 19:43.020 + and then there is detection. + +19:43.020 --> 19:45.180 + And I think where machine learning can really help + +19:45.180 --> 19:46.660 + is earlier detection. + +19:46.660 --> 19:48.580 + So does detection help? + +19:48.580 --> 19:49.660 + Detection is crucial. + +19:49.660 --> 19:53.940 + For instance, the vast majority of pancreatic cancer patients + +19:53.940 --> 19:57.300 + are detected at the stage that they are incurable. + +19:57.300 --> 20:02.300 + That's why they have such a terrible survival rate. + +20:03.740 --> 20:07.300 + It's like just few percent over five years. + +20:07.300 --> 20:09.820 + It's pretty much today the sentence. + +20:09.820 --> 20:13.620 + But if you can discover this disease early, + +20:14.500 --> 20:16.740 + there are mechanisms to treat it. + +20:16.740 --> 20:20.740 + And in fact, I know a number of people who were diagnosed + +20:20.740 --> 20:23.580 + and saved just because they had food poisoning. + +20:23.580 --> 20:25.020 + They had terrible food poisoning. + +20:25.020 --> 20:28.540 + They went to ER, they got scan. + +20:28.540 --> 20:30.660 + There were early signs on the scan + +20:30.660 --> 20:33.540 + and that would save their lives. + +20:33.540 --> 20:35.820 + But this wasn't really an accidental case. + +20:35.820 --> 20:40.820 + So as we become better, we would be able to help + +20:41.260 --> 20:46.260 + to many more people that are likely to develop diseases. + +20:46.540 --> 20:51.020 + And I just want to say that as I got more into this field, + +20:51.020 --> 20:53.620 + I realized that cancer is of course terrible disease, + +20:53.620 --> 20:56.700 + but there are really the whole slew of terrible diseases + +20:56.700 --> 21:00.820 + out there like neurodegenerative diseases and others. + +21:01.660 --> 21:04.580 + So we, of course, a lot of us are fixated on cancer + +21:04.580 --> 21:06.420 + because it's so prevalent in our society. + +21:06.420 --> 21:08.540 + And you see these people where there are a lot of patients + +21:08.540 --> 21:10.340 + with neurodegenerative diseases + +21:10.340 --> 21:12.540 + and the kind of aging diseases + +21:12.540 --> 21:17.100 + that we still don't have a good solution for. + +21:17.100 --> 21:22.100 + And I felt as a computer scientist, + +21:22.860 --> 21:25.460 + we kind of decided that it's other people's job + +21:25.460 --> 21:29.340 + to treat these diseases because it's like traditionally + +21:29.340 --> 21:32.420 + people in biology or in chemistry or MDs + +21:32.420 --> 21:35.340 + are the ones who are thinking about it. + +21:35.340 --> 21:37.420 + And after kind of start paying attention, + +21:37.420 --> 21:40.340 + I think that it's really a wrong assumption + +21:40.340 --> 21:42.940 + and we all need to join the battle. + +21:42.940 --> 21:46.460 + So how it seems like in cancer specifically + +21:46.460 --> 21:49.140 + that there's a lot of ways that machine learning can help. + +21:49.140 --> 21:51.860 + So what's the role of machine learning + +21:51.860 --> 21:54.100 + in the diagnosis of cancer? + +21:55.260 --> 21:58.700 + So for many cancers today, we really don't know + +21:58.700 --> 22:03.460 + what is your likelihood to get cancer. + +22:03.460 --> 22:06.300 + And for the vast majority of patients, + +22:06.300 --> 22:07.940 + especially on the younger patients, + +22:07.940 --> 22:09.580 + it really comes as a surprise. + +22:09.580 --> 22:11.140 + Like for instance, for breast cancer, + +22:11.140 --> 22:13.860 + 80% of the patients are first in their families, + +22:13.860 --> 22:15.380 + it's like me. + +22:15.380 --> 22:18.460 + And I never saw that I had any increased risk + +22:18.460 --> 22:20.820 + because nobody had it in my family. + +22:20.820 --> 22:22.300 + And for some reason in my head, + +22:22.300 --> 22:24.820 + it was kind of inherited disease. + +22:26.580 --> 22:28.380 + But even if I would pay attention, + +22:28.380 --> 22:32.420 + the very simplistic statistical models + +22:32.420 --> 22:34.540 + that are currently used in clinical practice, + +22:34.540 --> 22:37.460 + they really don't give you an answer, so you don't know. + +22:37.460 --> 22:40.380 + And the same true for pancreatic cancer, + +22:40.380 --> 22:45.380 + the same true for non smoking lung cancer and many others. + +22:45.380 --> 22:47.340 + So what machine learning can do here + +22:47.340 --> 22:51.620 + is utilize all this data to tell us early + +22:51.620 --> 22:53.140 + who is likely to be susceptible + +22:53.140 --> 22:55.980 + and using all the information that is already there, + +22:55.980 --> 22:59.980 + be it imaging, be it your other tests, + +22:59.980 --> 23:04.860 + and eventually liquid biopsies and others, + +23:04.860 --> 23:08.180 + where the signal itself is not sufficiently strong + +23:08.180 --> 23:11.300 + for human eye to do good discrimination + +23:11.300 --> 23:12.940 + because the signal may be weak, + +23:12.940 --> 23:15.620 + but by combining many sources, + +23:15.620 --> 23:18.100 + machine which is trained on large volumes of data + +23:18.100 --> 23:20.700 + can really detect it early. + +23:20.700 --> 23:22.500 + And that's what we've seen with breast cancer + +23:22.500 --> 23:25.900 + and people are reporting it in other diseases as well. + +23:25.900 --> 23:28.260 + That really boils down to data, right? + +23:28.260 --> 23:30.980 + And in the different kinds of sources of data. + +23:30.980 --> 23:33.740 + And you mentioned regulatory challenges. + +23:33.740 --> 23:35.180 + So what are the challenges + +23:35.180 --> 23:39.260 + in gathering large data sets in this space? + +23:40.860 --> 23:42.660 + Again, another great question. + +23:42.660 --> 23:45.500 + So it took me after I decided that I want to work on it + +23:45.500 --> 23:48.740 + two years to get access to data. + +23:48.740 --> 23:50.580 + Any data, like any significant data set? + +23:50.580 --> 23:53.580 + Any significant amount, like right now in this country, + +23:53.580 --> 23:57.060 + there is no publicly available data set + +23:57.060 --> 23:58.820 + of modern mammograms that you can just go + +23:58.820 --> 24:01.860 + on your computer, sign a document and get it. + +24:01.860 --> 24:03.180 + It just doesn't exist. + +24:03.180 --> 24:06.860 + I mean, obviously every hospital has its own collection + +24:06.860 --> 24:07.700 + of mammograms. + +24:07.700 --> 24:11.300 + There are data that came out of clinical trials. + +24:11.300 --> 24:13.220 + What we're talking about here is a computer scientist + +24:13.220 --> 24:17.140 + who just wants to run his or her model + +24:17.140 --> 24:19.060 + and see how it works. + +24:19.060 --> 24:22.900 + This data, like ImageNet, doesn't exist. + +24:22.900 --> 24:27.900 + And there is a set which is called like Florida data set + +24:28.620 --> 24:30.860 + which is a film mammogram from 90s + +24:30.860 --> 24:32.420 + which is totally not representative + +24:32.420 --> 24:33.860 + of the current developments. + +24:33.860 --> 24:35.780 + Whatever you're learning on them doesn't scale up. + +24:35.780 --> 24:39.300 + This is the only resource that is available. + +24:39.300 --> 24:42.780 + And today there are many agencies + +24:42.780 --> 24:44.460 + that govern access to data. + +24:44.460 --> 24:46.300 + Like the hospital holds your data + +24:46.300 --> 24:49.260 + and the hospital decides whether they would give it + +24:49.260 --> 24:52.340 + to the researcher to work with this data or not. + +24:52.340 --> 24:54.180 + Individual hospital? + +24:54.180 --> 24:55.020 + Yeah. + +24:55.020 --> 24:57.220 + I mean, the hospital may, you know, + +24:57.220 --> 24:59.220 + assuming that you're doing research collaboration, + +24:59.220 --> 25:01.980 + you can submit, you know, + +25:01.980 --> 25:05.060 + there is a proper approval process guided by RB + +25:05.060 --> 25:07.820 + and if you go through all the processes, + +25:07.820 --> 25:10.140 + you can eventually get access to the data. + +25:10.140 --> 25:13.540 + But if you yourself know our OEI community, + +25:13.540 --> 25:16.100 + there are not that many people who actually ever got access + +25:16.100 --> 25:20.260 + to data because it's very challenging process. + +25:20.260 --> 25:22.780 + And sorry, just in a quick comment, + +25:22.780 --> 25:25.780 + MGH or any kind of hospital, + +25:25.780 --> 25:28.100 + are they scanning the data? + +25:28.100 --> 25:29.740 + Are they digitally storing it? + +25:29.740 --> 25:31.580 + Oh, it is already digitally stored. + +25:31.580 --> 25:34.180 + You don't need to do any extra processing steps. + +25:34.180 --> 25:38.340 + It's already there in the right format is that right now + +25:38.340 --> 25:41.180 + there are a lot of issues that govern access to the data + +25:41.180 --> 25:46.180 + because the hospital is legally responsible for the data. + +25:46.180 --> 25:51.020 + And, you know, they have a lot to lose + +25:51.020 --> 25:53.140 + if they give the data to the wrong person, + +25:53.140 --> 25:56.460 + but they may not have a lot to gain if they give it + +25:56.460 --> 26:00.580 + as a hospital, as a legal entity has given it to you. + +26:00.580 --> 26:02.740 + And the way, you know, what I would imagine + +26:02.740 --> 26:05.220 + happening in the future is the same thing that happens + +26:05.220 --> 26:06.780 + when you're getting your driving license, + +26:06.780 --> 26:09.820 + you can decide whether you want to donate your organs. + +26:09.820 --> 26:13.100 + You can imagine that whenever a person goes to the hospital, + +26:13.100 --> 26:17.540 + they, it should be easy for them to donate their data + +26:17.540 --> 26:19.420 + for research and it can be different kind of, + +26:19.420 --> 26:22.420 + do they only give you a test results or only mammogram + +26:22.420 --> 26:25.900 + or only imaging data or the whole medical record? + +26:27.060 --> 26:28.980 + Because at the end, + +26:30.540 --> 26:33.860 + we all will benefit from all this insights. + +26:33.860 --> 26:36.060 + And it's not like you say, I want to keep my data private, + +26:36.060 --> 26:38.780 + but I would really love to get it from other people + +26:38.780 --> 26:40.740 + because other people are thinking the same way. + +26:40.740 --> 26:45.740 + So if there is a mechanism to do this donation + +26:45.740 --> 26:48.020 + and the patient has an ability to say + +26:48.020 --> 26:50.820 + how they want to use their data for research, + +26:50.820 --> 26:54.100 + it would be really a game changer. + +26:54.100 --> 26:56.460 + People, when they think about this problem, + +26:56.460 --> 26:58.460 + there's a, it depends on the population, + +26:58.460 --> 27:00.140 + depends on the demographics, + +27:00.140 --> 27:03.420 + but there's some privacy concerns generally, + +27:03.420 --> 27:05.860 + not just medical data, just any kind of data. + +27:05.860 --> 27:09.620 + It's what you said, my data, it should belong kind of to me. + +27:09.620 --> 27:11.660 + I'm worried how it's going to be misused. + +27:12.540 --> 27:15.620 + How do we alleviate those concerns? + +27:17.100 --> 27:19.460 + Because that seems like a problem that needs to be, + +27:19.460 --> 27:22.980 + that problem of trust, of transparency needs to be solved + +27:22.980 --> 27:27.260 + before we build large data sets that help detect cancer, + +27:27.260 --> 27:30.180 + help save those very people in the future. + +27:30.180 --> 27:31.940 + So I think there are two things that could be done. + +27:31.940 --> 27:34.460 + There is a technical solutions + +27:34.460 --> 27:38.220 + and there are societal solutions. + +27:38.220 --> 27:40.180 + So on the technical end, + +27:41.460 --> 27:46.460 + we today have ability to improve disambiguation. + +27:48.140 --> 27:49.740 + Like, for instance, for imaging, + +27:49.740 --> 27:54.740 + it's, you know, for imaging, you can do it pretty well. + +27:55.620 --> 27:56.780 + What's disambiguation? + +27:56.780 --> 27:58.540 + And disambiguation, sorry, disambiguation, + +27:58.540 --> 27:59.860 + removing the identification, + +27:59.860 --> 28:02.220 + removing the names of the people. + +28:02.220 --> 28:04.820 + There are other data, like if it is a raw tax, + +28:04.820 --> 28:08.180 + you cannot really achieve 99.9%, + +28:08.180 --> 28:10.060 + but there are all these techniques + +28:10.060 --> 28:12.460 + that actually some of them are developed at MIT, + +28:12.460 --> 28:15.460 + how you can do learning on the encoded data + +28:15.460 --> 28:17.420 + where you locally encode the image, + +28:17.420 --> 28:22.420 + you train a network which only works on the encoded images + +28:22.420 --> 28:24.940 + and then you send the outcome back to the hospital + +28:24.940 --> 28:26.580 + and you can open it up. + +28:26.580 --> 28:28.020 + So those are the technical solutions. + +28:28.020 --> 28:30.660 + There are a lot of people who are working in this space + +28:30.660 --> 28:33.780 + where the learning happens in the encoded form. + +28:33.780 --> 28:36.180 + We are still early, + +28:36.180 --> 28:39.260 + but this is an interesting research area + +28:39.260 --> 28:41.900 + where I think we'll make more progress. + +28:43.340 --> 28:45.620 + There is a lot of work in natural language processing + +28:45.620 --> 28:48.620 + community how to do the identification better. + +28:50.380 --> 28:54.020 + But even today, there are already a lot of data + +28:54.020 --> 28:55.900 + which can be deidentified perfectly, + +28:55.900 --> 28:58.780 + like your test data, for instance, correct, + +28:58.780 --> 29:00.980 + where you can just, you know the name of the patient, + +29:00.980 --> 29:04.300 + you just want to extract the part with the numbers. + +29:04.300 --> 29:07.460 + The big problem here is again, + +29:08.420 --> 29:10.420 + hospitals don't see much incentive + +29:10.420 --> 29:12.660 + to give this data away on one hand + +29:12.660 --> 29:14.220 + and then there is general concern. + +29:14.220 --> 29:17.700 + Now, when I'm talking about societal benefits + +29:17.700 --> 29:19.660 + and about the education, + +29:19.660 --> 29:24.340 + the public needs to understand that I think + +29:25.700 --> 29:29.420 + that there are situation and I still remember myself + +29:29.420 --> 29:33.380 + when I really needed an answer, I had to make a choice. + +29:33.380 --> 29:35.220 + There was no information to make a choice, + +29:35.220 --> 29:36.660 + you're just guessing. + +29:36.660 --> 29:41.060 + And at that moment you feel that your life is at the stake, + +29:41.060 --> 29:44.820 + but you just don't have information to make the choice. + +29:44.820 --> 29:48.740 + And many times when I give talks, + +29:48.740 --> 29:51.300 + I get emails from women who say, + +29:51.300 --> 29:52.820 + you know, I'm in this situation, + +29:52.820 --> 29:55.940 + can you please run statistic and see what are the outcomes? + +29:57.100 --> 30:01.300 + We get almost every week a mammogram that comes by mail + +30:01.300 --> 30:03.460 + to my office at MIT, I'm serious. + +30:04.380 --> 30:07.860 + That people ask to run because they need to make + +30:07.860 --> 30:10.020 + life changing decisions. + +30:10.020 --> 30:12.980 + And of course, I'm not planning to open a clinic here, + +30:12.980 --> 30:16.660 + but we do run and give them the results for their doctors. + +30:16.660 --> 30:20.100 + But the point that I'm trying to make, + +30:20.100 --> 30:23.780 + that we all at some point or our loved ones + +30:23.780 --> 30:26.620 + will be in the situation where you need information + +30:26.620 --> 30:28.860 + to make the best choice. + +30:28.860 --> 30:31.860 + And if this information is not available, + +30:31.860 --> 30:35.100 + you would feel vulnerable and unprotected. + +30:35.100 --> 30:37.860 + And then the question is, you know, what do I care more? + +30:37.860 --> 30:40.380 + Because at the end, everything is a trade off, correct? + +30:40.380 --> 30:41.700 + Yeah, exactly. + +30:41.700 --> 30:45.580 + Just out of curiosity, it seems like one possible solution, + +30:45.580 --> 30:47.420 + I'd like to see what you think of it, + +30:49.340 --> 30:50.660 + based on what you just said, + +30:50.660 --> 30:52.500 + based on wanting to know answers + +30:52.500 --> 30:55.060 + for when you're yourself in that situation. + +30:55.060 --> 30:58.420 + Is it possible for patients to own their data + +30:58.420 --> 31:01.020 + as opposed to hospitals owning their data? + +31:01.020 --> 31:04.100 + Of course, theoretically, I guess patients own their data, + +31:04.100 --> 31:06.620 + but can you walk out there with a USB stick + +31:07.580 --> 31:10.620 + containing everything or upload it to the cloud? + +31:10.620 --> 31:14.500 + Where a company, you know, I remember Microsoft + +31:14.500 --> 31:17.820 + had a service, like I try, I was really excited about + +31:17.820 --> 31:19.260 + and Google Health was there. + +31:19.260 --> 31:21.900 + I tried to give, I was excited about it. + +31:21.900 --> 31:24.780 + Basically companies helping you upload your data + +31:24.780 --> 31:27.940 + to the cloud so that you can move from hospital to hospital + +31:27.940 --> 31:29.260 + from doctor to doctor. + +31:29.260 --> 31:32.700 + Do you see a promise of that kind of possibility? + +31:32.700 --> 31:34.660 + I absolutely think this is, you know, + +31:34.660 --> 31:38.180 + the right way to exchange the data. + +31:38.180 --> 31:41.700 + I don't know now who's the biggest player in this field, + +31:41.700 --> 31:45.940 + but I can clearly see that even for totally selfish + +31:45.940 --> 31:49.300 + health reasons, when you are going to a new facility + +31:49.300 --> 31:52.620 + and many of us are sent to some specialized treatment, + +31:52.620 --> 31:55.740 + they don't easily have access to your data. + +31:55.740 --> 31:59.420 + And today, you know, we might want to send this mammogram, + +31:59.420 --> 32:01.780 + need to go to the hospital, find some small office + +32:01.780 --> 32:04.820 + which gives them the CD and they ship as a CD. + +32:04.820 --> 32:08.340 + So you can imagine we're looking at kind of decades old + +32:08.340 --> 32:10.100 + mechanism of data exchange. + +32:11.340 --> 32:15.620 + So I definitely think this is an area where hopefully + +32:15.620 --> 32:20.380 + all the right regulatory and technical forces will align + +32:20.380 --> 32:23.220 + and we will see it actually implemented. + +32:23.220 --> 32:27.500 + It's sad because unfortunately, and I need to research + +32:27.500 --> 32:30.620 + why that happened, but I'm pretty sure Google Health + +32:30.620 --> 32:32.940 + and Microsoft Health Vault or whatever it's called + +32:32.940 --> 32:36.100 + both closed down, which means that there was + +32:36.100 --> 32:39.100 + either regulatory pressure or there's not a business case + +32:39.100 --> 32:41.820 + or there's challenges from hospitals, + +32:41.820 --> 32:43.260 + which is very disappointing. + +32:43.260 --> 32:46.500 + So when you say you don't know what the biggest players are, + +32:46.500 --> 32:50.540 + the two biggest that I was aware of closed their doors. + +32:50.540 --> 32:53.140 + So I'm hoping, I'd love to see why + +32:53.140 --> 32:54.780 + and I'd love to see who else can come up. + +32:54.780 --> 32:59.620 + It seems like one of those Elon Musk style problems + +32:59.620 --> 33:01.300 + that are obvious needs to be solved + +33:01.300 --> 33:02.980 + and somebody needs to step up and actually do + +33:02.980 --> 33:07.540 + this large scale data collection. + +33:07.540 --> 33:09.620 + So I know there is an initiative in Massachusetts, + +33:09.620 --> 33:11.740 + I think, which you led by the governor + +33:11.740 --> 33:15.460 + to try to create this kind of health exchange system + +33:15.460 --> 33:17.860 + where at least to help people who kind of when you show up + +33:17.860 --> 33:20.220 + in emergency room and there is no information + +33:20.220 --> 33:23.540 + about what are your allergies and other things. + +33:23.540 --> 33:26.140 + So I don't know how far it will go. + +33:26.140 --> 33:28.180 + But another thing that you said + +33:28.180 --> 33:30.780 + and I find it very interesting is actually + +33:30.780 --> 33:33.780 + who are the successful players in this space + +33:33.780 --> 33:37.260 + and the whole implementation, how does it go? + +33:37.260 --> 33:40.300 + To me, it is from the anthropological perspective, + +33:40.300 --> 33:44.660 + it's more fascinating that AI that today goes in healthcare, + +33:44.660 --> 33:50.380 + we've seen so many attempts and so very little successes. + +33:50.380 --> 33:54.220 + And it's interesting to understand that I've by no means + +33:54.220 --> 33:56.700 + have knowledge to assess it, + +33:56.700 --> 33:59.620 + why we are in the position where we are. + +33:59.620 --> 34:02.940 + Yeah, it's interesting because data is really fuel + +34:02.940 --> 34:04.980 + for a lot of successful applications. + +34:04.980 --> 34:08.500 + And when that data acquires regulatory approval, + +34:08.500 --> 34:12.940 + like the FDA or any kind of approval, + +34:12.940 --> 34:15.740 + it seems that the computer scientists + +34:15.740 --> 34:17.460 + are not quite there yet in being able + +34:17.460 --> 34:18.900 + to play the regulatory game, + +34:18.900 --> 34:21.220 + understanding the fundamentals of it. + +34:21.220 --> 34:26.500 + I think that in many cases when even people do have data, + +34:26.500 --> 34:31.300 + we still don't know what exactly do you need to demonstrate + +34:31.300 --> 34:33.860 + to change the standard of care. + +34:35.500 --> 34:37.180 + Like let me give you an example + +34:37.180 --> 34:41.100 + related to my breast cancer research. + +34:41.100 --> 34:45.500 + So in traditional breast cancer risk assessment, + +34:45.500 --> 34:47.140 + there is something called density, + +34:47.140 --> 34:50.500 + which determines the likelihood of a woman to get cancer. + +34:50.500 --> 34:51.700 + And this pretty much says, + +34:51.700 --> 34:54.220 + how much white do you see on the mammogram? + +34:54.220 --> 34:58.980 + The whiter it is, the more likely the tissue is dense. + +34:58.980 --> 35:03.660 + And the idea behind density, it's not a bad idea. + +35:03.660 --> 35:08.100 + In 1967, a radiologist called Wolf decided to look back + +35:08.100 --> 35:09.780 + at women who were diagnosed + +35:09.780 --> 35:12.420 + and see what is special in their images. + +35:12.420 --> 35:14.700 + Can we look back and say that they're likely to develop? + +35:14.700 --> 35:16.180 + So he come up with some patterns. + +35:16.180 --> 35:20.660 + And it was the best that his human eye can identify. + +35:20.660 --> 35:22.060 + Then it was kind of formalized + +35:22.060 --> 35:24.220 + and coded into four categories. + +35:24.220 --> 35:26.940 + And that's what we are using today. + +35:26.940 --> 35:31.020 + And today this density assessment + +35:31.020 --> 35:34.620 + is actually a federal law from 2019, + +35:34.620 --> 35:36.180 + approved by President Trump + +35:36.180 --> 35:40.100 + and for the previous FDA commissioner, + +35:40.100 --> 35:43.620 + where women are supposed to be advised by their providers + +35:43.620 --> 35:45.100 + if they have high density, + +35:45.100 --> 35:47.260 + putting them into higher risk category. + +35:47.260 --> 35:49.460 + And in some states, + +35:49.460 --> 35:51.260 + you can actually get supplementary screening + +35:51.260 --> 35:53.700 + paid by your insurance because you're in this category. + +35:53.700 --> 35:56.780 + Now you can say, how much science do we have behind it? + +35:56.780 --> 36:00.820 + Whatever, biological science or epidemiological evidence. + +36:00.820 --> 36:05.140 + So it turns out that between 40 and 50% of women + +36:05.140 --> 36:06.660 + have dense breasts. + +36:06.660 --> 36:11.140 + So about 40% of patients are coming out of their screening + +36:11.140 --> 36:15.020 + and somebody tells them, you are in high risk. + +36:15.020 --> 36:16.860 + Now, what exactly does it mean + +36:16.860 --> 36:19.620 + if you as half of the population in high risk? + +36:19.620 --> 36:22.060 + It's from saying, maybe I'm not, + +36:22.060 --> 36:23.700 + or what do I really need to do with it? + +36:23.700 --> 36:27.220 + Because the system doesn't provide me + +36:27.220 --> 36:28.340 + a lot of the solutions + +36:28.340 --> 36:30.140 + because there are so many people like me, + +36:30.140 --> 36:34.620 + we cannot really provide very expensive solutions for them. + +36:34.620 --> 36:38.740 + And the reason this whole density became this big deal, + +36:38.740 --> 36:40.820 + it's actually advocated by the patients + +36:40.820 --> 36:42.500 + who felt very unprotected + +36:42.500 --> 36:44.900 + because many women went and did the mammograms + +36:44.900 --> 36:46.260 + which were normal. + +36:46.260 --> 36:49.460 + And then it turns out that they already had cancer, + +36:49.460 --> 36:50.580 + quite developed cancer. + +36:50.580 --> 36:54.420 + So they didn't have a way to know who is really at risk + +36:54.420 --> 36:56.300 + and what is the likelihood that when the doctor tells you, + +36:56.300 --> 36:58.060 + you're okay, you are not okay. + +36:58.060 --> 37:02.140 + So at the time, and it was 15 years ago, + +37:02.140 --> 37:06.820 + this maybe was the best piece of science that we had. + +37:06.820 --> 37:11.820 + And it took quite 15, 16 years to make it federal law. + +37:12.180 --> 37:15.660 + But now this is a standard. + +37:15.660 --> 37:17.620 + Now with a deep learning model, + +37:17.620 --> 37:19.660 + we can so much more accurately predict + +37:19.660 --> 37:21.580 + who is gonna develop breast cancer + +37:21.580 --> 37:23.700 + just because you're trained on a logical thing. + +37:23.700 --> 37:26.060 + And instead of describing how much white + +37:26.060 --> 37:27.380 + and what kind of white machine + +37:27.380 --> 37:30.140 + can systematically identify the patterns, + +37:30.140 --> 37:32.780 + which was the original idea behind the thought + +37:32.780 --> 37:33.700 + of the cardiologist, + +37:33.700 --> 37:35.740 + machines can do it much more systematically + +37:35.740 --> 37:38.260 + and predict the risk when you're training the machine + +37:38.260 --> 37:42.140 + to look at the image and to say the risk in one to five years. + +37:42.140 --> 37:45.060 + Now you can ask me how long it will take + +37:45.060 --> 37:46.460 + to substitute this density, + +37:46.460 --> 37:48.620 + which is broadly used across the country + +37:48.620 --> 37:53.620 + and really is not helping to bring this new models. + +37:54.380 --> 37:56.700 + And I would say it's not a matter of the algorithm. + +37:56.700 --> 37:58.780 + Algorithms use already orders of magnitude better + +37:58.780 --> 38:00.460 + than what is currently in practice. + +38:00.460 --> 38:02.500 + I think it's really the question, + +38:02.500 --> 38:04.380 + who do you need to convince? + +38:04.380 --> 38:07.460 + How many hospitals do you need to run the experiment? + +38:07.460 --> 38:11.500 + What, you know, all this mechanism of adoption + +38:11.500 --> 38:15.180 + and how do you explain to patients + +38:15.180 --> 38:17.580 + and to women across the country + +38:17.580 --> 38:20.460 + that this is really a better measure? + +38:20.460 --> 38:22.740 + And again, I don't think it's an AI question. + +38:22.740 --> 38:25.940 + We can work more and make the algorithm even better, + +38:25.940 --> 38:29.300 + but I don't think that this is the current, you know, + +38:29.300 --> 38:32.060 + the barrier, the barrier is really this other piece + +38:32.060 --> 38:35.260 + that for some reason is not really explored. + +38:35.260 --> 38:36.860 + It's like anthropological piece. + +38:36.860 --> 38:39.860 + And coming back to your question about books, + +38:39.860 --> 38:42.980 + there is a book that I'm reading. + +38:42.980 --> 38:47.980 + It's called American Sickness by Elizabeth Rosenthal. + +38:48.260 --> 38:51.580 + And I got this book from my clinical collaborator, + +38:51.580 --> 38:53.100 + Dr. Connie Lehman. + +38:53.100 --> 38:54.820 + And I said, I know everything that I need to know + +38:54.820 --> 38:56.020 + about American health system, + +38:56.020 --> 38:59.220 + but you know, every page doesn't fail to surprise me. + +38:59.220 --> 39:03.140 + And I think there is a lot of interesting + +39:03.140 --> 39:06.860 + and really deep lessons for people like us + +39:06.860 --> 39:09.660 + from computer science who are coming into this field + +39:09.660 --> 39:13.660 + to really understand how complex is the system of incentives + +39:13.660 --> 39:17.660 + in the system to understand how you really need to play + +39:17.660 --> 39:18.780 + to drive adoption. + +39:19.740 --> 39:21.180 + You just said it's complex, + +39:21.180 --> 39:23.980 + but if we're trying to simplify it, + +39:23.980 --> 39:27.380 + who do you think most likely would be successful + +39:27.380 --> 39:29.540 + if we push on this group of people? + +39:29.540 --> 39:30.780 + Is it the doctors? + +39:30.780 --> 39:31.820 + Is it the hospitals? + +39:31.820 --> 39:34.300 + Is it the governments or policymakers? + +39:34.300 --> 39:37.380 + Is it the individual patients, consumers? + +39:38.860 --> 39:43.860 + Who needs to be inspired to most likely lead to adoption? + +39:45.180 --> 39:47.100 + Or is there no simple answer? + +39:47.100 --> 39:48.260 + There's no simple answer, + +39:48.260 --> 39:51.980 + but I think there is a lot of good people in medical system + +39:51.980 --> 39:55.180 + who do want to make a change. + +39:56.460 --> 40:01.460 + And I think a lot of power will come from us as consumers + +40:01.540 --> 40:04.260 + because we all are consumers or future consumers + +40:04.260 --> 40:06.500 + of healthcare services. + +40:06.500 --> 40:11.500 + And I think we can do so much more + +40:12.060 --> 40:15.500 + in explaining the potential and not in the hype terms + +40:15.500 --> 40:17.900 + and not saying that we now killed all Alzheimer + +40:17.900 --> 40:20.500 + and I'm really sick of reading this kind of articles + +40:20.500 --> 40:22.100 + which make these claims, + +40:22.100 --> 40:24.780 + but really to show with some examples + +40:24.780 --> 40:29.060 + what this implementation does and how it changes the care. + +40:29.060 --> 40:30.020 + Because I can't imagine, + +40:30.020 --> 40:33.220 + it doesn't matter what kind of politician it is, + +40:33.220 --> 40:35.220 + we all are susceptible to these diseases. + +40:35.220 --> 40:37.740 + There is no one who is free. + +40:37.740 --> 40:41.060 + And eventually, we all are humans + +40:41.060 --> 40:44.860 + and we're looking for a way to alleviate the suffering. + +40:44.860 --> 40:47.260 + And this is one possible way + +40:47.260 --> 40:49.300 + where we currently are under utilizing, + +40:49.300 --> 40:50.940 + which I think can help. + +40:51.860 --> 40:55.100 + So it sounds like the biggest problems are outside of AI + +40:55.100 --> 40:57.980 + in terms of the biggest impact at this point. + +40:57.980 --> 41:00.420 + But are there any open problems + +41:00.420 --> 41:03.780 + in the application of ML to oncology in general? + +41:03.780 --> 41:07.540 + So improving the detection or any other creative methods, + +41:07.540 --> 41:09.620 + whether it's on the detection segmentations + +41:09.620 --> 41:11.780 + or the vision perception side + +41:11.780 --> 41:16.260 + or some other clever of inference? + +41:16.260 --> 41:19.620 + Yeah, what in general in your view are the open problems + +41:19.620 --> 41:20.460 + in this space? + +41:20.460 --> 41:22.460 + Yeah, I just want to mention that beside detection, + +41:22.460 --> 41:24.820 + not the area where I am kind of quite active + +41:24.820 --> 41:28.580 + and I think it's really an increasingly important area + +41:28.580 --> 41:30.940 + in healthcare is drug design. + +41:32.260 --> 41:33.100 + Absolutely. + +41:33.100 --> 41:36.900 + Because it's fine if you detect something early, + +41:36.900 --> 41:41.100 + but you still need to get drugs + +41:41.100 --> 41:43.860 + and new drugs for these conditions. + +41:43.860 --> 41:46.740 + And today, all of the drug design, + +41:46.740 --> 41:48.300 + ML is non existent there. + +41:48.300 --> 41:52.980 + We don't have any drug that was developed by the ML model + +41:52.980 --> 41:54.900 + or even not developed, + +41:54.900 --> 41:57.060 + but at least even knew that ML model + +41:57.060 --> 41:59.260 + plays some significant role. + +41:59.260 --> 42:03.300 + I think this area with all the new ability + +42:03.300 --> 42:05.780 + to generate molecules with desired properties + +42:05.780 --> 42:10.780 + to do in silica screening is really a big open area. + +42:11.460 --> 42:12.740 + To be totally honest with you, + +42:12.740 --> 42:14.900 + when we are doing diagnostics and imaging, + +42:14.900 --> 42:17.260 + primarily taking the ideas that were developed + +42:17.260 --> 42:20.460 + for other areas and you applying them with some adaptation, + +42:20.460 --> 42:25.460 + the area of drug design is really technically interesting + +42:26.820 --> 42:27.980 + and exciting area. + +42:27.980 --> 42:30.380 + You need to work a lot with graphs + +42:30.380 --> 42:34.580 + and capture various 3D properties. + +42:34.580 --> 42:37.420 + There are lots and lots of opportunities + +42:37.420 --> 42:39.820 + to be technically creative. + +42:39.820 --> 42:44.820 + And I think there are a lot of open questions in this area. + +42:46.820 --> 42:48.820 + We're already getting a lot of successes + +42:48.820 --> 42:52.700 + even with kind of the first generation of these models, + +42:52.700 --> 42:56.500 + but there is much more new creative things that you can do. + +42:56.500 --> 42:59.260 + And what's very nice to see is that actually + +42:59.260 --> 43:04.180 + the more powerful, the more interesting models + +43:04.180 --> 43:05.460 + actually do do better. + +43:05.460 --> 43:10.460 + So there is a place to innovate in machine learning + +43:11.300 --> 43:12.540 + in this area. + +43:13.900 --> 43:16.820 + And some of these techniques are really unique to, + +43:16.820 --> 43:19.620 + let's say, to graph generation and other things. + +43:19.620 --> 43:20.820 + So... + +43:20.820 --> 43:23.980 + What, just to interrupt really quick, I'm sorry, + +43:23.980 --> 43:28.980 + graph generation or graphs, drug discovery in general, + +43:30.620 --> 43:31.940 + how do you discover a drug? + +43:31.940 --> 43:33.340 + Is this chemistry? + +43:33.340 --> 43:37.500 + Is this trying to predict different chemical reactions? + +43:37.500 --> 43:39.660 + Or is it some kind of... + +43:39.660 --> 43:42.100 + What do graphs even represent in this space? + +43:42.100 --> 43:43.980 + Oh, sorry, sorry. + +43:43.980 --> 43:45.340 + And what's a drug? + +43:45.340 --> 43:47.140 + Okay, so let's say you're thinking + +43:47.140 --> 43:48.540 + there are many different types of drugs, + +43:48.540 --> 43:50.580 + but let's say you're gonna talk about small molecules + +43:50.580 --> 43:52.860 + because I think today the majority of drugs + +43:52.860 --> 43:53.700 + are small molecules. + +43:53.700 --> 43:55.020 + So small molecule is a graph. + +43:55.020 --> 43:59.180 + The molecule is just where the node in the graph + +43:59.180 --> 44:01.500 + is an atom and then you have the bonds. + +44:01.500 --> 44:03.220 + So it's really a graph representation. + +44:03.220 --> 44:05.540 + If you look at it in 2D, correct, + +44:05.540 --> 44:07.460 + you can do it 3D, but let's say, + +44:07.460 --> 44:09.540 + let's keep it simple and stick in 2D. + +44:11.500 --> 44:14.740 + So pretty much my understanding today, + +44:14.740 --> 44:18.620 + how it is done at scale in the companies, + +44:18.620 --> 44:20.220 + without machine learning, + +44:20.220 --> 44:22.100 + you have high throughput screening. + +44:22.100 --> 44:23.740 + So you know that you are interested + +44:23.740 --> 44:26.540 + to get certain biological activity of the compound. + +44:26.540 --> 44:28.860 + So you scan a lot of compounds, + +44:28.860 --> 44:30.700 + like maybe hundreds of thousands, + +44:30.700 --> 44:32.980 + some really big number of compounds. + +44:32.980 --> 44:36.060 + You identify some compounds which have the right activity + +44:36.060 --> 44:39.220 + and then at this point, the chemists come + +44:39.220 --> 44:43.220 + and they're trying to now to optimize + +44:43.220 --> 44:45.340 + this original heat to different properties + +44:45.340 --> 44:47.180 + that you want it to be maybe soluble, + +44:47.180 --> 44:49.060 + you want it to decrease toxicity, + +44:49.060 --> 44:51.620 + you want it to decrease the side effects. + +44:51.620 --> 44:54.020 + Are those, sorry again to interrupt, + +44:54.020 --> 44:55.500 + can that be done in simulation + +44:55.500 --> 44:57.700 + or just by looking at the molecules + +44:57.700 --> 44:59.820 + or do you need to actually run reactions + +44:59.820 --> 45:02.460 + in real labs with lab coats and stuff? + +45:02.460 --> 45:04.020 + So when you do high throughput screening, + +45:04.020 --> 45:06.100 + you really do screening. + +45:06.100 --> 45:07.020 + It's in the lab. + +45:07.020 --> 45:09.140 + It's really the lab screening. + +45:09.140 --> 45:10.980 + You screen the molecules, correct? + +45:10.980 --> 45:12.580 + I don't know what screening is. + +45:12.580 --> 45:15.060 + The screening is just check them for certain property. + +45:15.060 --> 45:17.260 + Like in the physical space, in the physical world, + +45:17.260 --> 45:18.740 + like actually there's a machine probably + +45:18.740 --> 45:21.420 + that's actually running the reaction. + +45:21.420 --> 45:22.900 + Actually running the reactions, yeah. + +45:22.900 --> 45:25.420 + So there is a process where you can run + +45:25.420 --> 45:26.660 + and that's why it's called high throughput + +45:26.660 --> 45:29.580 + that it become cheaper and faster + +45:29.580 --> 45:33.820 + to do it on very big number of molecules. + +45:33.820 --> 45:35.820 + You run the screening, + +45:35.820 --> 45:40.300 + you identify potential good starts + +45:40.300 --> 45:42.340 + and then when the chemists come in + +45:42.340 --> 45:44.060 + who have done it many times + +45:44.060 --> 45:46.180 + and then they can try to look at it and say, + +45:46.180 --> 45:48.260 + how can you change the molecule + +45:48.260 --> 45:51.780 + to get the desired profile + +45:51.780 --> 45:53.460 + in terms of all other properties? + +45:53.460 --> 45:56.500 + So maybe how do I make it more bioactive and so on? + +45:56.500 --> 45:59.460 + And there the creativity of the chemists + +45:59.460 --> 46:03.980 + really is the one that determines the success + +46:03.980 --> 46:07.460 + of this design because again, + +46:07.460 --> 46:09.300 + they have a lot of domain knowledge + +46:09.300 --> 46:12.900 + of what works, how do you decrease the CCD and so on + +46:12.900 --> 46:15.020 + and that's what they do. + +46:15.020 --> 46:17.860 + So all the drugs that are currently + +46:17.860 --> 46:20.220 + in the FDA approved drugs + +46:20.220 --> 46:22.140 + or even drugs that are in clinical trials, + +46:22.140 --> 46:27.100 + they are designed using these domain experts + +46:27.100 --> 46:30.060 + which goes through this combinatorial space + +46:30.060 --> 46:31.940 + of molecules or graphs or whatever + +46:31.940 --> 46:35.140 + and find the right one or adjust it to be the right ones. + +46:35.140 --> 46:38.060 + It sounds like the breast density heuristic + +46:38.060 --> 46:40.460 + from 67 to the same echoes. + +46:40.460 --> 46:41.820 + It's not necessarily that. + +46:41.820 --> 46:45.380 + It's really driven by deep understanding. + +46:45.380 --> 46:46.820 + It's not like they just observe it. + +46:46.820 --> 46:48.540 + I mean, they do deeply understand chemistry + +46:48.540 --> 46:50.460 + and they do understand how different groups + +46:50.460 --> 46:53.140 + and how does it changes the properties. + +46:53.140 --> 46:56.660 + So there is a lot of science that gets into it + +46:56.660 --> 46:58.740 + and a lot of kind of simulation, + +46:58.740 --> 47:00.940 + how do you want it to behave? + +47:01.900 --> 47:03.900 + It's very, very complex. + +47:03.900 --> 47:06.140 + So they're quite effective at this design, obviously. + +47:06.140 --> 47:08.420 + Now effective, yeah, we have drugs. + +47:08.420 --> 47:10.780 + Like depending on how do you measure effective, + +47:10.780 --> 47:13.940 + if you measure it in terms of cost, it's prohibitive. + +47:13.940 --> 47:15.820 + If you measure it in terms of times, + +47:15.820 --> 47:18.420 + we have lots of diseases for which we don't have any drugs + +47:18.420 --> 47:20.060 + and we don't even know how to approach + +47:20.060 --> 47:23.460 + and don't need to mention few drugs + +47:23.460 --> 47:27.140 + or neurodegenerative disease drugs that fail. + +47:27.140 --> 47:32.140 + So there are lots of trials that fail in later stages, + +47:32.180 --> 47:35.180 + which is really catastrophic from the financial perspective. + +47:35.180 --> 47:39.540 + So is it the effective, the most effective mechanism? + +47:39.540 --> 47:42.700 + Absolutely no, but this is the only one that currently works. + +47:44.300 --> 47:47.900 + And I was closely interacting + +47:47.900 --> 47:49.260 + with people in pharmaceutical industry. + +47:49.260 --> 47:51.340 + I was really fascinated on how sharp + +47:51.340 --> 47:55.260 + and what a deep understanding of the domain do they have. + +47:55.260 --> 47:57.020 + It's not observation driven. + +47:57.020 --> 48:00.220 + There is really a lot of science behind what they do. + +48:00.220 --> 48:02.300 + But if you ask me, can machine learning change it, + +48:02.300 --> 48:05.300 + I firmly believe yes, + +48:05.300 --> 48:07.860 + because even the most experienced chemists + +48:07.860 --> 48:11.100 + cannot hold in their memory and understanding + +48:11.100 --> 48:12.500 + everything that you can learn + +48:12.500 --> 48:15.420 + from millions of molecules and reactions. + +48:17.220 --> 48:19.900 + And the space of graphs is a totally new space. + +48:19.900 --> 48:22.060 + I mean, it's a really interesting space + +48:22.060 --> 48:23.980 + for machine learning to explore, graph generation. + +48:23.980 --> 48:26.260 + Yeah, so there are a lot of things that you can do here. + +48:26.260 --> 48:28.740 + So we do a lot of work. + +48:28.740 --> 48:31.620 + So the first tool that we started with + +48:31.620 --> 48:36.300 + was the tool that can predict properties of the molecules. + +48:36.300 --> 48:39.420 + So you can just give the molecule and the property. + +48:39.420 --> 48:41.340 + It can be by activity property, + +48:41.340 --> 48:44.300 + or it can be some other property. + +48:44.300 --> 48:46.460 + And you train the molecules + +48:46.460 --> 48:50.020 + and you can now take a new molecule + +48:50.020 --> 48:52.180 + and predict this property. + +48:52.180 --> 48:54.860 + Now, when people started working in this area, + +48:54.860 --> 48:55.980 + it is something very simple. + +48:55.980 --> 48:58.580 + They do kind of existing fingerprints, + +48:58.580 --> 49:00.740 + which is kind of handcrafted features of the molecule. + +49:00.740 --> 49:02.980 + When you break the graph to substructures + +49:02.980 --> 49:05.980 + and then you run it in a feed forward neural network. + +49:05.980 --> 49:08.500 + And what was interesting to see that clearly, + +49:08.500 --> 49:11.020 + this was not the most effective way to proceed. + +49:11.020 --> 49:14.060 + And you need to have much more complex models + +49:14.060 --> 49:16.300 + that can induce a representation, + +49:16.300 --> 49:19.220 + which can translate this graph into the embeddings + +49:19.220 --> 49:21.300 + and do these predictions. + +49:21.300 --> 49:23.220 + So this is one direction. + +49:23.220 --> 49:25.260 + Then another direction, which is kind of related + +49:25.260 --> 49:29.180 + is not only to stop by looking at the embedding itself, + +49:29.180 --> 49:32.780 + but actually modify it to produce better molecules. + +49:32.780 --> 49:36.020 + So you can think about it as machine translation + +49:36.020 --> 49:38.140 + that you can start with a molecule + +49:38.140 --> 49:40.580 + and then there is an improved version of molecule. + +49:40.580 --> 49:42.860 + And you can again, with encoder translate it + +49:42.860 --> 49:45.380 + into the hidden space and then learn how to modify it + +49:45.380 --> 49:49.340 + to improve the in some ways version of the molecules. + +49:49.340 --> 49:52.620 + So that's, it's kind of really exciting. + +49:52.620 --> 49:54.740 + We already have seen that the property prediction + +49:54.740 --> 49:56.140 + works pretty well. + +49:56.140 --> 49:59.780 + And now we are generating molecules + +49:59.780 --> 50:01.820 + and there is actually labs + +50:01.820 --> 50:04.180 + which are manufacturing this molecule. + +50:04.180 --> 50:06.340 + So we'll see where it will get us. + +50:06.340 --> 50:07.780 + Okay, that's really exciting. + +50:07.780 --> 50:08.860 + There's a lot of promise. + +50:08.860 --> 50:11.820 + Speaking of machine translation and embeddings, + +50:11.820 --> 50:15.580 + I think you have done a lot of really great research + +50:15.580 --> 50:17.540 + in NLP, natural language processing. + +50:19.260 --> 50:21.540 + Can you tell me your journey through NLP? + +50:21.540 --> 50:25.100 + What ideas, problems, approaches were you working on? + +50:25.100 --> 50:28.180 + Were you fascinated with, did you explore + +50:28.180 --> 50:33.180 + before this magic of deep learning reemerged and after? + +50:34.020 --> 50:37.180 + So when I started my work in NLP, it was in 97. + +50:38.180 --> 50:39.460 + This was very interesting time. + +50:39.460 --> 50:42.620 + It was exactly the time that I came to ACL. + +50:43.500 --> 50:46.140 + And at the time I could barely understand English, + +50:46.140 --> 50:48.500 + but it was exactly like the transition point + +50:48.500 --> 50:53.500 + because half of the papers were really rule based approaches + +50:53.500 --> 50:56.180 + where people took more kind of heavy linguistic approaches + +50:56.180 --> 51:00.060 + for small domains and try to build up from there. + +51:00.060 --> 51:02.220 + And then there were the first generation of papers + +51:02.220 --> 51:04.500 + which were corpus based papers. + +51:04.500 --> 51:06.420 + And they were very simple in our terms + +51:06.420 --> 51:07.900 + when you collect some statistics + +51:07.900 --> 51:10.020 + and do prediction based on them. + +51:10.020 --> 51:13.100 + And I found it really fascinating that one community + +51:13.100 --> 51:18.100 + can think so very differently about the problem. + +51:19.220 --> 51:22.820 + And I remember my first paper that I wrote, + +51:22.820 --> 51:24.460 + it didn't have a single formula. + +51:24.460 --> 51:25.740 + It didn't have evaluation. + +51:25.740 --> 51:28.340 + It just had examples of outputs. + +51:28.340 --> 51:32.020 + And this was a standard of the field at the time. + +51:32.020 --> 51:35.860 + In some ways, I mean, people maybe just started emphasizing + +51:35.860 --> 51:38.940 + the empirical evaluation, but for many applications + +51:38.940 --> 51:42.780 + like summarization, you just show some examples of outputs. + +51:42.780 --> 51:45.460 + And then increasingly you can see that how + +51:45.460 --> 51:48.300 + the statistical approaches dominated the field + +51:48.300 --> 51:52.100 + and we've seen increased performance + +51:52.100 --> 51:56.020 + across many basic tasks. + +51:56.020 --> 52:00.420 + The sad part of the story maybe that if you look again + +52:00.420 --> 52:05.100 + through this journey, we see that the role of linguistics + +52:05.100 --> 52:07.460 + in some ways greatly diminishes. + +52:07.460 --> 52:11.580 + And I think that you really need to look + +52:11.580 --> 52:14.540 + through the whole proceeding to find one or two papers + +52:14.540 --> 52:17.260 + which make some interesting linguistic references. + +52:17.260 --> 52:18.100 + It's really big. + +52:18.100 --> 52:18.920 + Today, yeah. + +52:18.920 --> 52:19.760 + Today, today. + +52:19.760 --> 52:20.600 + This was definitely one of the. + +52:20.600 --> 52:23.140 + Things like syntactic trees, just even basically + +52:23.140 --> 52:26.180 + against our conversation about human understanding + +52:26.180 --> 52:30.300 + of language, which I guess what linguistics would be + +52:30.300 --> 52:34.300 + structured, hierarchical representing language + +52:34.300 --> 52:37.140 + in a way that's human explainable, understandable + +52:37.140 --> 52:39.500 + is missing today. + +52:39.500 --> 52:42.380 + I don't know if it is, what is explainable + +52:42.380 --> 52:43.620 + and understandable. + +52:43.620 --> 52:47.360 + In the end, we perform functions and it's okay + +52:47.360 --> 52:50.140 + to have machine which performs a function. + +52:50.140 --> 52:53.200 + Like when you're thinking about your calculator, correct? + +52:53.200 --> 52:56.100 + Your calculator can do calculation very different + +52:56.100 --> 52:57.620 + from you would do the calculation, + +52:57.620 --> 52:58.860 + but it's very effective in it. + +52:58.860 --> 53:02.560 + And this is fine if we can achieve certain tasks + +53:02.560 --> 53:05.760 + with high accuracy, doesn't necessarily mean + +53:05.760 --> 53:09.300 + that it has to understand it the same way as we understand. + +53:09.300 --> 53:11.260 + In some ways, it's even naive to request + +53:11.260 --> 53:14.940 + because you have so many other sources of information + +53:14.940 --> 53:17.900 + that are absent when you are training your system. + +53:17.900 --> 53:19.220 + So it's okay. + +53:19.220 --> 53:20.060 + Is it delivered? + +53:20.060 --> 53:21.500 + And I would tell you one application + +53:21.500 --> 53:22.780 + that is really fascinating. + +53:22.780 --> 53:25.060 + In 97, when it came to ACL, there were some papers + +53:25.060 --> 53:25.900 + on machine translation. + +53:25.900 --> 53:27.440 + They were like primitive. + +53:27.440 --> 53:31.060 + Like people were trying really, really simple. + +53:31.060 --> 53:34.260 + And the feeling, my feeling was that, you know, + +53:34.260 --> 53:36.260 + to make real machine translation system, + +53:36.260 --> 53:39.580 + it's like to fly at the moon and build a house there + +53:39.580 --> 53:41.580 + and the garden and live happily ever after. + +53:41.580 --> 53:42.600 + I mean, it's like impossible. + +53:42.600 --> 53:46.740 + I never could imagine that within, you know, 10 years, + +53:46.740 --> 53:48.540 + we would already see the system working. + +53:48.540 --> 53:51.420 + And now, you know, nobody is even surprised + +53:51.420 --> 53:54.420 + to utilize the system on daily basis. + +53:54.420 --> 53:56.220 + So this was like a huge, huge progress, + +53:56.220 --> 53:57.860 + saying that people for very long time + +53:57.860 --> 54:00.820 + tried to solve using other mechanisms. + +54:00.820 --> 54:03.220 + And they were unable to solve it. + +54:03.220 --> 54:06.140 + That's why coming back to your question about biology, + +54:06.140 --> 54:10.800 + that, you know, in linguistics, people try to go this way + +54:10.800 --> 54:13.500 + and try to write the syntactic trees + +54:13.500 --> 54:17.500 + and try to abstract it and to find the right representation. + +54:17.500 --> 54:22.240 + And, you know, they couldn't get very far + +54:22.240 --> 54:26.580 + with this understanding while these models using, + +54:26.580 --> 54:29.640 + you know, other sources actually capable + +54:29.640 --> 54:31.680 + to make a lot of progress. + +54:31.680 --> 54:33.960 + Now, I'm not naive to think + +54:33.960 --> 54:36.780 + that we are in this paradise space in NLP. + +54:36.780 --> 54:38.580 + And sure as you know, + +54:38.580 --> 54:40.860 + that when we slightly change the domain + +54:40.860 --> 54:42.620 + and when we decrease the amount of training, + +54:42.620 --> 54:44.740 + it can do like really bizarre and funny thing. + +54:44.740 --> 54:46.500 + But I think it's just a matter + +54:46.500 --> 54:48.540 + of improving generalization capacity, + +54:48.540 --> 54:51.500 + which is just a technical question. + +54:51.500 --> 54:54.340 + Wow, so that's the question. + +54:54.340 --> 54:57.720 + How much of language understanding can be solved + +54:57.720 --> 54:59.180 + with deep neural networks? + +54:59.180 --> 55:03.740 + In your intuition, I mean, it's unknown, I suppose. + +55:03.740 --> 55:07.660 + But as we start to creep towards romantic notions + +55:07.660 --> 55:10.620 + of the spirit of the Turing test + +55:10.620 --> 55:14.220 + and conversation and dialogue + +55:14.220 --> 55:18.340 + and something that maybe to me or to us, + +55:18.340 --> 55:21.620 + so the humans feels like it needs real understanding. + +55:21.620 --> 55:23.500 + How much can that be achieved + +55:23.500 --> 55:27.180 + with these neural networks or statistical methods? + +55:27.180 --> 55:32.180 + So I guess I am very much driven by the outcomes. + +55:33.340 --> 55:35.420 + Can we achieve the performance + +55:35.420 --> 55:40.420 + which would be satisfactory for us for different tasks? + +55:40.700 --> 55:43.020 + Now, if you again look at machine translation system, + +55:43.020 --> 55:46.020 + which are trained on large amounts of data, + +55:46.020 --> 55:48.780 + they really can do a remarkable job + +55:48.780 --> 55:51.300 + relatively to where they've been a few years ago. + +55:51.300 --> 55:54.620 + And if you project into the future, + +55:54.620 --> 55:59.380 + if it will be the same speed of improvement, you know, + +55:59.380 --> 56:00.220 + this is great. + +56:00.220 --> 56:01.060 + Now, does it bother me + +56:01.060 --> 56:04.860 + that it's not doing the same translation as we are doing? + +56:04.860 --> 56:06.620 + Now, if you go to cognitive science, + +56:06.620 --> 56:09.460 + we still don't really understand what we are doing. + +56:10.460 --> 56:11.860 + I mean, there are a lot of theories + +56:11.860 --> 56:13.840 + and there's obviously a lot of progress and studying, + +56:13.840 --> 56:17.540 + but our understanding what exactly goes on in our brains + +56:17.540 --> 56:21.020 + when we process language is still not crystal clear + +56:21.020 --> 56:25.460 + and precise that we can translate it into machines. + +56:25.460 --> 56:29.220 + What does bother me is that, you know, + +56:29.220 --> 56:31.700 + again, that machines can be extremely brittle + +56:31.700 --> 56:33.980 + when you go out of your comfort zone + +56:33.980 --> 56:36.060 + of when there is a distributional shift + +56:36.060 --> 56:37.300 + between training and testing. + +56:37.300 --> 56:39.020 + And it have been years and years, + +56:39.020 --> 56:41.320 + every year when I teach an LP class, + +56:41.320 --> 56:43.560 + now show them some examples of translation + +56:43.560 --> 56:47.300 + from some newspaper in Hebrew or whatever, it was perfect. + +56:47.300 --> 56:51.300 + And then I have a recipe that Tomi Yakel's system + +56:51.300 --> 56:53.900 + sent me a while ago and it was written in Finnish + +56:53.900 --> 56:55.720 + of Karelian pies. + +56:55.720 --> 56:59.280 + And it's just a terrible translation. + +56:59.280 --> 57:01.460 + You cannot understand anything what it does. + +57:01.460 --> 57:04.180 + It's not like some syntactic mistakes, it's just terrible. + +57:04.180 --> 57:07.020 + And year after year, I tried and will translate + +57:07.020 --> 57:08.980 + and year after year, it does this terrible work + +57:08.980 --> 57:10.980 + because I guess, you know, the recipes + +57:10.980 --> 57:14.580 + are not a big part of their training repertoire. + +57:14.580 --> 57:19.020 + So, but in terms of outcomes, that's a really clean, + +57:19.020 --> 57:20.240 + good way to look at it. + +57:21.100 --> 57:23.140 + I guess the question I was asking is, + +57:24.060 --> 57:27.700 + do you think, imagine a future, + +57:27.700 --> 57:30.540 + do you think the current approaches can pass + +57:30.540 --> 57:32.460 + the Turing test in the way, + +57:34.700 --> 57:37.060 + in the best possible formulation of the Turing test? + +57:37.060 --> 57:39.460 + Which is, would you wanna have a conversation + +57:39.460 --> 57:42.340 + with a neural network for an hour? + +57:42.340 --> 57:45.820 + Oh God, no, no, there are not that many people + +57:45.820 --> 57:48.380 + that I would want to talk for an hour, but. + +57:48.380 --> 57:51.500 + There are some people in this world, alive or not, + +57:51.500 --> 57:53.260 + that you would like to talk to for an hour. + +57:53.260 --> 57:56.700 + Could a neural network achieve that outcome? + +57:56.700 --> 57:58.860 + So I think it would be really hard to create + +57:58.860 --> 58:02.300 + a successful training set, which would enable it + +58:02.300 --> 58:04.980 + to have a conversation, a contextual conversation + +58:04.980 --> 58:05.820 + for an hour. + +58:05.820 --> 58:08.140 + Do you think it's a problem of data, perhaps? + +58:08.140 --> 58:09.940 + I think in some ways it's not a problem of data, + +58:09.940 --> 58:13.620 + it's a problem both of data and the problem of + +58:13.620 --> 58:15.780 + the way we're training our systems, + +58:15.780 --> 58:18.060 + their ability to truly, to generalize, + +58:18.060 --> 58:19.300 + to be very compositional. + +58:19.300 --> 58:23.220 + In some ways it's limited in the current capacity, + +58:23.220 --> 58:27.980 + at least we can translate well, + +58:27.980 --> 58:32.540 + we can find information well, we can extract information. + +58:32.540 --> 58:35.180 + So there are many capacities in which it's doing very well. + +58:35.180 --> 58:38.000 + And you can ask me, would you trust the machine + +58:38.000 --> 58:39.820 + to translate for you and use it as a source? + +58:39.820 --> 58:42.580 + I would say absolutely, especially if we're talking about + +58:42.580 --> 58:45.660 + newspaper data or other data which is in the realm + +58:45.660 --> 58:47.900 + of its own training set, I would say yes. + +58:48.900 --> 58:52.900 + But having conversations with the machine, + +58:52.900 --> 58:56.460 + it's not something that I would choose to do. + +58:56.460 --> 58:59.420 + But I would tell you something, talking about Turing tests + +58:59.420 --> 59:02.940 + and about all this kind of ELISA conversations, + +59:02.940 --> 59:05.540 + I remember visiting Tencent in China + +59:05.540 --> 59:07.620 + and they have this chat board and they claim + +59:07.620 --> 59:10.780 + there is really humongous amount of the local population + +59:10.780 --> 59:12.940 + which for hours talks to the chat board. + +59:12.940 --> 59:15.340 + To me it was, I cannot believe it, + +59:15.340 --> 59:18.000 + but apparently it's documented that there are some people + +59:18.000 --> 59:20.760 + who enjoy this conversation. + +59:20.760 --> 59:24.540 + And it brought to me another MIT story + +59:24.540 --> 59:26.980 + about ELISA and Weisenbaum. + +59:26.980 --> 59:29.340 + I don't know if you're familiar with the story. + +59:29.340 --> 59:31.020 + So Weisenbaum was a professor at MIT + +59:31.020 --> 59:32.580 + and when he developed this ELISA, + +59:32.580 --> 59:34.620 + which was just doing string matching, + +59:34.620 --> 59:38.540 + very trivial, like restating of what you said + +59:38.540 --> 59:41.260 + with very few rules, no syntax. + +59:41.260 --> 59:43.740 + Apparently there were secretaries at MIT + +59:43.740 --> 59:48.180 + that would sit for hours and converse with this trivial thing + +59:48.180 --> 59:50.180 + and at the time there was no beautiful interfaces + +59:50.180 --> 59:51.820 + so you actually need to go through the pain + +59:51.820 --> 59:53.540 + of communicating. + +59:53.540 --> 59:56.940 + And Weisenbaum himself was so horrified by this phenomenon + +59:56.940 --> 59:59.300 + that people can believe enough to the machine + +59:59.300 --> 1:00:00.820 + that you just need to give them the hint + +1:00:00.820 --> 1:00:03.940 + that machine understands you and you can complete the rest + +1:00:03.940 --> 1:00:05.420 + that he kind of stopped this research + +1:00:05.420 --> 1:00:08.660 + and went into kind of trying to understand + +1:00:08.660 --> 1:00:11.480 + what this artificial intelligence can do to our brains. + +1:00:12.740 --> 1:00:14.380 + So my point is, you know, + +1:00:14.380 --> 1:00:19.300 + how much, it's not how good is the technology, + +1:00:19.300 --> 1:00:22.620 + it's how ready we are to believe + +1:00:22.620 --> 1:00:25.580 + that it delivers the goods that we are trying to get. + +1:00:25.580 --> 1:00:27.200 + That's a really beautiful way to put it. + +1:00:27.200 --> 1:00:29.800 + I, by the way, I'm not horrified by that possibility, + +1:00:29.800 --> 1:00:33.140 + but inspired by it because, + +1:00:33.140 --> 1:00:35.920 + I mean, human connection, + +1:00:35.920 --> 1:00:38.220 + whether it's through language or through love, + +1:00:39.860 --> 1:00:44.860 + it seems like it's very amenable to machine learning + +1:00:44.900 --> 1:00:49.340 + and the rest is just challenges of psychology. + +1:00:49.340 --> 1:00:52.460 + Like you said, the secretaries who enjoy spending hours. + +1:00:52.460 --> 1:00:55.020 + I would say I would describe most of our lives + +1:00:55.020 --> 1:00:58.020 + as enjoying spending hours with those we love + +1:00:58.020 --> 1:01:00.820 + for very silly reasons. + +1:01:00.820 --> 1:01:02.780 + All we're doing is keyword matching as well. + +1:01:02.780 --> 1:01:05.100 + So I'm not sure how much intelligence + +1:01:05.100 --> 1:01:08.140 + we exhibit to each other with the people we love + +1:01:08.140 --> 1:01:09.820 + that we're close with. + +1:01:09.820 --> 1:01:12.660 + So it's a very interesting point + +1:01:12.660 --> 1:01:16.020 + of what it means to pass the Turing test with language. + +1:01:16.020 --> 1:01:16.860 + I think you're right. + +1:01:16.860 --> 1:01:18.220 + In terms of conversation, + +1:01:18.220 --> 1:01:20.180 + I think machine translation + +1:01:21.420 --> 1:01:24.420 + has very clear performance and improvement, right? + +1:01:24.420 --> 1:01:28.020 + What it means to have a fulfilling conversation + +1:01:28.020 --> 1:01:32.660 + is very person dependent and context dependent + +1:01:32.660 --> 1:01:33.580 + and so on. + +1:01:33.580 --> 1:01:36.340 + That's, yeah, it's very well put. + +1:01:36.340 --> 1:01:40.740 + But in your view, what's a benchmark in natural language, + +1:01:40.740 --> 1:01:43.640 + a test that's just out of reach right now, + +1:01:43.640 --> 1:01:46.020 + but we might be able to, that's exciting. + +1:01:46.020 --> 1:01:49.100 + Is it in perfecting machine translation + +1:01:49.100 --> 1:01:51.900 + or is there other, is it summarization? + +1:01:51.900 --> 1:01:52.740 + What's out there just out of reach? + +1:01:52.740 --> 1:01:55.820 + I think it goes across specific application. + +1:01:55.820 --> 1:01:59.500 + It's more about the ability to learn from few examples + +1:01:59.500 --> 1:02:03.300 + for real, what we call few short learning and all these cases + +1:02:03.300 --> 1:02:05.940 + because the way we publish these papers today, + +1:02:05.940 --> 1:02:09.900 + we say, if we have like naively, we get 55, + +1:02:09.900 --> 1:02:12.500 + but now we had a few example and we can move to 65. + +1:02:12.500 --> 1:02:13.540 + None of these methods + +1:02:13.540 --> 1:02:15.980 + actually are realistically doing anything useful. + +1:02:15.980 --> 1:02:18.540 + You cannot use them today. + +1:02:18.540 --> 1:02:23.540 + And the ability to be able to generalize and to move + +1:02:25.460 --> 1:02:28.940 + or to be autonomous in finding the data + +1:02:28.940 --> 1:02:30.300 + that you need to learn, + +1:02:31.340 --> 1:02:34.280 + to be able to perfect new tasks or new language, + +1:02:35.300 --> 1:02:38.100 + this is an area where I think we really need + +1:02:39.200 --> 1:02:43.020 + to move forward to and we are not yet there. + +1:02:43.020 --> 1:02:45.060 + Are you at all excited, + +1:02:45.060 --> 1:02:46.540 + curious by the possibility + +1:02:46.540 --> 1:02:48.520 + of creating human level intelligence? + +1:02:49.900 --> 1:02:52.540 + Is this, cause you've been very in your discussion. + +1:02:52.540 --> 1:02:54.340 + So if we look at oncology, + +1:02:54.340 --> 1:02:58.100 + you're trying to use machine learning to help the world + +1:02:58.100 --> 1:02:59.700 + in terms of alleviating suffering. + +1:02:59.700 --> 1:03:02.340 + If you look at natural language processing, + +1:03:02.340 --> 1:03:05.300 + you're focused on the outcomes of improving practical things + +1:03:05.300 --> 1:03:06.820 + like machine translation. + +1:03:06.820 --> 1:03:09.880 + But human level intelligence is a thing + +1:03:09.880 --> 1:03:13.800 + that our civilization has dreamed about creating, + +1:03:13.800 --> 1:03:15.740 + super human level intelligence. + +1:03:15.740 --> 1:03:16.940 + Do you think about this? + +1:03:16.940 --> 1:03:19.040 + Do you think it's at all within our reach? + +1:03:20.380 --> 1:03:22.660 + So as you said yourself, Elie, + +1:03:22.660 --> 1:03:26.140 + talking about how do you perceive + +1:03:26.140 --> 1:03:28.980 + our communications with each other, + +1:03:28.980 --> 1:03:31.940 + that we're matching keywords and certain behaviors + +1:03:31.940 --> 1:03:33.020 + and so on. + +1:03:33.020 --> 1:03:36.860 + So at the end, whenever one assesses, + +1:03:36.860 --> 1:03:38.680 + let's say relations with another person, + +1:03:38.680 --> 1:03:41.460 + you have separate kind of measurements and outcomes + +1:03:41.460 --> 1:03:43.620 + inside your head that determine + +1:03:43.620 --> 1:03:45.860 + what is the status of the relation. + +1:03:45.860 --> 1:03:48.580 + So one way, this is this classical level, + +1:03:48.580 --> 1:03:49.600 + what is the intelligence? + +1:03:49.600 --> 1:03:51.860 + Is it the fact that now we are gonna do the same way + +1:03:51.860 --> 1:03:52.940 + as human is doing, + +1:03:52.940 --> 1:03:55.500 + when we don't even understand what the human is doing? + +1:03:55.500 --> 1:03:59.100 + Or we now have an ability to deliver these outcomes, + +1:03:59.100 --> 1:04:01.300 + but not in one area, not in NLP, + +1:04:01.300 --> 1:04:03.940 + not just to translate or just to answer questions, + +1:04:03.940 --> 1:04:05.380 + but across many, many areas + +1:04:05.380 --> 1:04:08.100 + that we can achieve the functionalities + +1:04:08.100 --> 1:04:11.060 + that humans can achieve with their ability to learn + +1:04:11.060 --> 1:04:12.380 + and do other things. + +1:04:12.380 --> 1:04:15.500 + I think this is, and this we can actually measure + +1:04:15.500 --> 1:04:17.560 + how far we are. + +1:04:17.560 --> 1:04:21.580 + And that's what makes me excited that we, + +1:04:21.580 --> 1:04:23.780 + in my lifetime, at least so far what we've seen, + +1:04:23.780 --> 1:04:25.840 + it's like tremendous progress + +1:04:25.840 --> 1:04:28.700 + across these different functionalities. + +1:04:28.700 --> 1:04:32.260 + And I think it will be really exciting + +1:04:32.260 --> 1:04:35.540 + to see where we will be. + +1:04:35.540 --> 1:04:39.300 + And again, one way to think about it, + +1:04:39.300 --> 1:04:41.820 + there are machines which are improving their functionality. + +1:04:41.820 --> 1:04:44.940 + Another one is to think about us with our brains, + +1:04:44.940 --> 1:04:46.420 + which are imperfect, + +1:04:46.420 --> 1:04:51.420 + how they can be accelerated by this technology + +1:04:51.420 --> 1:04:55.900 + as it becomes stronger and stronger. + +1:04:55.900 --> 1:04:57.260 + Coming back to another book + +1:04:57.260 --> 1:05:01.060 + that I love, Flowers for Algernon. + +1:05:01.060 --> 1:05:02.100 + Have you read this book? + +1:05:02.100 --> 1:05:02.940 + Yes. + +1:05:02.940 --> 1:05:05.700 + So there is this point that the patient gets + +1:05:05.700 --> 1:05:07.980 + this miracle cure, which changes his brain. + +1:05:07.980 --> 1:05:11.020 + And all of a sudden they see life in a different way + +1:05:11.020 --> 1:05:13.300 + and can do certain things better, + +1:05:13.300 --> 1:05:14.860 + but certain things much worse. + +1:05:14.860 --> 1:05:19.860 + So you can imagine this kind of computer augmented cognition + +1:05:22.400 --> 1:05:24.800 + where it can bring you that now in the same way + +1:05:24.800 --> 1:05:28.120 + as the cars enable us to get to places + +1:05:28.120 --> 1:05:30.080 + where we've never been before, + +1:05:30.080 --> 1:05:31.640 + can we think differently? + +1:05:31.640 --> 1:05:33.600 + Can we think faster? + +1:05:33.600 --> 1:05:36.680 + And we already see a lot of it happening + +1:05:36.680 --> 1:05:38.260 + in how it impacts us, + +1:05:38.260 --> 1:05:42.200 + but I think we have a long way to go there. + +1:05:42.200 --> 1:05:45.040 + So that's sort of artificial intelligence + +1:05:45.040 --> 1:05:47.280 + and technology affecting our, + +1:05:47.280 --> 1:05:50.440 + augmenting our intelligence as humans. + +1:05:50.440 --> 1:05:55.440 + Yesterday, a company called Neuralink announced, + +1:05:55.520 --> 1:05:56.800 + they did this whole demonstration. + +1:05:56.800 --> 1:05:57.980 + I don't know if you saw it. + +1:05:57.980 --> 1:06:01.000 + It's, they demonstrated brain computer, + +1:06:01.000 --> 1:06:02.680 + brain machine interface, + +1:06:02.680 --> 1:06:06.360 + where there's like a sewing machine for the brain. + +1:06:06.360 --> 1:06:11.120 + Do you, you know, a lot of that is quite out there + +1:06:11.120 --> 1:06:14.040 + in terms of things that some people would say + +1:06:14.040 --> 1:06:16.340 + are impossible, but they're dreamers + +1:06:16.340 --> 1:06:18.080 + and want to engineer systems like that. + +1:06:18.080 --> 1:06:20.360 + Do you see, based on what you just said, + +1:06:20.360 --> 1:06:23.820 + a hope for that more direct interaction with the brain? + +1:06:25.120 --> 1:06:27.040 + I think there are different ways. + +1:06:27.040 --> 1:06:29.000 + One is a direct interaction with the brain. + +1:06:29.000 --> 1:06:30.900 + And again, there are lots of companies + +1:06:30.900 --> 1:06:32.280 + that work in this space + +1:06:32.280 --> 1:06:35.080 + and I think there will be a lot of developments. + +1:06:35.080 --> 1:06:36.600 + But I'm just thinking that many times + +1:06:36.600 --> 1:06:39.080 + we are not aware of our feelings, + +1:06:39.080 --> 1:06:41.400 + of motivation, what drives us. + +1:06:41.400 --> 1:06:44.200 + Like, let me give you a trivial example, our attention. + +1:06:45.520 --> 1:06:47.260 + There are a lot of studies that demonstrate + +1:06:47.260 --> 1:06:49.200 + that it takes a while to a person to understand + +1:06:49.200 --> 1:06:51.080 + that they are not attentive anymore. + +1:06:51.080 --> 1:06:52.160 + And we know that there are people + +1:06:52.160 --> 1:06:54.520 + who really have strong capacity to hold attention. + +1:06:54.520 --> 1:06:57.080 + There are other end of the spectrum people with ADD + +1:06:57.080 --> 1:06:58.800 + and other issues that they have problem + +1:06:58.800 --> 1:07:00.760 + to regulate their attention. + +1:07:00.760 --> 1:07:03.520 + Imagine to yourself that you have like a cognitive aid + +1:07:03.520 --> 1:07:06.280 + that just alerts you based on your gaze, + +1:07:06.280 --> 1:07:09.280 + that your attention is now not on what you are doing. + +1:07:09.280 --> 1:07:10.560 + And instead of writing a paper, + +1:07:10.560 --> 1:07:12.760 + you're now dreaming of what you're gonna do in the evening. + +1:07:12.760 --> 1:07:16.360 + So even this kind of simple measurement things, + +1:07:16.360 --> 1:07:17.840 + how they can change us. + +1:07:17.840 --> 1:07:22.400 + And I see it even in simple ways with myself. + +1:07:22.400 --> 1:07:26.480 + I have my zone app that I got in MIT gym. + +1:07:26.480 --> 1:07:28.800 + It kind of records, you know, how much did you run + +1:07:28.800 --> 1:07:29.800 + and you have some points + +1:07:29.800 --> 1:07:32.880 + and you can get some status, whatever. + +1:07:32.880 --> 1:07:35.840 + Like, I said, what is this ridiculous thing? + +1:07:35.840 --> 1:07:38.800 + Who would ever care about some status in some app? + +1:07:38.800 --> 1:07:39.640 + Guess what? + +1:07:39.640 --> 1:07:41.560 + So to maintain the status, + +1:07:41.560 --> 1:07:44.640 + you have to do set a number of points every month. + +1:07:44.640 --> 1:07:48.040 + And not only is that I do it every single month + +1:07:48.040 --> 1:07:50.560 + for the last 18 months, + +1:07:50.560 --> 1:07:54.160 + it went to the point that I was injured. + +1:07:54.160 --> 1:07:56.160 + And when I could run again, + +1:07:58.120 --> 1:08:02.560 + in two days, I did like some humongous amount of running + +1:08:02.560 --> 1:08:04.080 + just to complete the points. + +1:08:04.080 --> 1:08:05.920 + It was like really not safe. + +1:08:05.920 --> 1:08:08.440 + It was like, I'm not gonna lose my status + +1:08:08.440 --> 1:08:10.240 + because I want to get there. + +1:08:10.240 --> 1:08:13.320 + So you can already see that this direct measurement + +1:08:13.320 --> 1:08:15.160 + and the feedback is, you know, + +1:08:15.160 --> 1:08:16.320 + we're looking at video games + +1:08:16.320 --> 1:08:18.720 + and see why, you know, the addiction aspect of it, + +1:08:18.720 --> 1:08:21.200 + but you can imagine that the same idea can be expanded + +1:08:21.200 --> 1:08:23.640 + to many other areas of our life. + +1:08:23.640 --> 1:08:25.960 + When we really can get feedback + +1:08:25.960 --> 1:08:28.480 + and imagine in your case in relations, + +1:08:29.880 --> 1:08:31.240 + when we are doing keyword matching, + +1:08:31.240 --> 1:08:36.120 + imagine that the person who is generating the keywords, + +1:08:36.120 --> 1:08:37.720 + that person gets direct feedback + +1:08:37.720 --> 1:08:39.560 + before the whole thing explodes. + +1:08:39.560 --> 1:08:42.000 + Is it maybe at this happy point, + +1:08:42.000 --> 1:08:44.000 + we are going in the wrong direction. + +1:08:44.000 --> 1:08:48.040 + Maybe it will be really a behavior modifying moment. + +1:08:48.040 --> 1:08:51.360 + So yeah, it's a relationship management too. + +1:08:51.360 --> 1:08:54.200 + So yeah, that's a fascinating whole area + +1:08:54.200 --> 1:08:56.120 + of psychology actually as well, + +1:08:56.120 --> 1:08:58.240 + of seeing how our behavior has changed + +1:08:58.240 --> 1:09:01.840 + with basically all human relations now have + +1:09:01.840 --> 1:09:06.200 + other nonhuman entities helping us out. + +1:09:06.200 --> 1:09:09.440 + So you teach a large, + +1:09:09.440 --> 1:09:12.600 + a huge machine learning course here at MIT. + +1:09:14.000 --> 1:09:15.360 + I can ask you a million questions, + +1:09:15.360 --> 1:09:17.560 + but you've seen a lot of students. + +1:09:17.560 --> 1:09:20.920 + What ideas do students struggle with the most + +1:09:20.920 --> 1:09:23.920 + as they first enter this world of machine learning? + +1:09:23.920 --> 1:09:26.520 + Actually, this year was the first time + +1:09:26.520 --> 1:09:28.480 + I started teaching a small machine learning class. + +1:09:28.480 --> 1:09:31.160 + And it came as a result of what I saw + +1:09:31.160 --> 1:09:34.640 + in my big machine learning class that Tomi Yakel and I built + +1:09:34.640 --> 1:09:36.640 + maybe six years ago. + +1:09:38.040 --> 1:09:40.360 + What we've seen that as this area become more + +1:09:40.360 --> 1:09:43.440 + and more popular, more and more people at MIT + +1:09:43.440 --> 1:09:45.360 + want to take this class. + +1:09:45.360 --> 1:09:48.320 + And while we designed it for computer science majors, + +1:09:48.320 --> 1:09:50.760 + there were a lot of people who really are interested + +1:09:50.760 --> 1:09:52.600 + to learn it, but unfortunately, + +1:09:52.600 --> 1:09:55.720 + their background was not enabling them + +1:09:55.720 --> 1:09:57.200 + to do well in the class. + +1:09:57.200 --> 1:09:59.360 + And many of them associated machine learning + +1:09:59.360 --> 1:10:01.360 + with the word struggle and failure, + +1:10:02.480 --> 1:10:04.640 + primarily for non majors. + +1:10:04.640 --> 1:10:06.840 + And that's why we actually started a new class + +1:10:06.840 --> 1:10:10.800 + which we call machine learning from algorithms to modeling, + +1:10:10.800 --> 1:10:15.000 + which emphasizes more the modeling aspects of it + +1:10:15.000 --> 1:10:20.000 + and focuses on, it has majors and non majors. + +1:10:20.000 --> 1:10:23.480 + So we kind of try to extract the relevant parts + +1:10:23.480 --> 1:10:25.560 + and make it more accessible, + +1:10:25.560 --> 1:10:27.800 + because the fact that we're teaching 20 classifiers + +1:10:27.800 --> 1:10:29.240 + in standard machine learning class, + +1:10:29.240 --> 1:10:32.200 + it's really a big question to really need it. + +1:10:32.200 --> 1:10:34.520 + But it was interesting to see this + +1:10:34.520 --> 1:10:36.480 + from first generation of students, + +1:10:36.480 --> 1:10:39.080 + when they came back from their internships + +1:10:39.080 --> 1:10:42.320 + and from their jobs, + +1:10:42.320 --> 1:10:45.560 + what different and exciting things they can do. + +1:10:45.560 --> 1:10:47.600 + I would never think that you can even apply + +1:10:47.600 --> 1:10:50.800 + machine learning to, some of them are like matching, + +1:10:50.800 --> 1:10:53.480 + the relations and other things like variety. + +1:10:53.480 --> 1:10:56.080 + Everything is amenable as the machine learning. + +1:10:56.080 --> 1:10:58.320 + That actually brings up an interesting point + +1:10:58.320 --> 1:11:00.680 + of computer science in general. + +1:11:00.680 --> 1:11:03.520 + It almost seems, maybe I'm crazy, + +1:11:03.520 --> 1:11:06.520 + but it almost seems like everybody needs to learn + +1:11:06.520 --> 1:11:08.160 + how to program these days. + +1:11:08.160 --> 1:11:11.400 + If you're 20 years old, or if you're starting school, + +1:11:11.400 --> 1:11:14.200 + even if you're an English major, + +1:11:14.200 --> 1:11:19.200 + it seems like programming unlocks so much possibility + +1:11:20.480 --> 1:11:21.880 + in this world. + +1:11:21.880 --> 1:11:25.000 + So when you interacted with those non majors, + +1:11:25.000 --> 1:11:30.000 + is there skills that they were simply lacking at the time + +1:11:30.280 --> 1:11:33.000 + that you wish they had and that they learned + +1:11:33.000 --> 1:11:34.680 + in high school and so on? + +1:11:34.680 --> 1:11:37.520 + Like how should education change + +1:11:37.520 --> 1:11:41.320 + in this computerized world that we live in? + +1:11:41.320 --> 1:11:44.320 + I think because I knew that there is a Python component + +1:11:44.320 --> 1:11:47.000 + in the class, their Python skills were okay + +1:11:47.000 --> 1:11:49.160 + and the class isn't really heavy on programming. + +1:11:49.160 --> 1:11:52.400 + They primarily kind of add parts to the programs. + +1:11:52.400 --> 1:11:55.440 + I think it was more of the mathematical barriers + +1:11:55.440 --> 1:11:58.200 + and the class, again, with the design on the majors + +1:11:58.200 --> 1:12:01.200 + was using the notation, like big O for complexity + +1:12:01.200 --> 1:12:04.520 + and others, people who come from different backgrounds + +1:12:04.520 --> 1:12:05.800 + just don't have it in the lexical, + +1:12:05.800 --> 1:12:09.120 + so necessarily very challenging notion, + +1:12:09.120 --> 1:12:11.480 + but they were just not aware. + +1:12:12.360 --> 1:12:16.240 + So I think that kind of linear algebra and probability, + +1:12:16.240 --> 1:12:19.120 + the basics, the calculus, multivariate calculus, + +1:12:19.120 --> 1:12:20.840 + things that can help. + +1:12:20.840 --> 1:12:23.520 + What advice would you give to students + +1:12:23.520 --> 1:12:25.280 + interested in machine learning, + +1:12:25.280 --> 1:12:29.240 + interested, you've talked about detecting, + +1:12:29.240 --> 1:12:31.360 + curing cancer, drug design, + +1:12:31.360 --> 1:12:34.520 + if they want to get into that field, what should they do? + +1:12:36.320 --> 1:12:39.040 + Get into it and succeed as researchers + +1:12:39.040 --> 1:12:42.080 + and entrepreneurs. + +1:12:43.320 --> 1:12:45.240 + The first good piece of news is that right now + +1:12:45.240 --> 1:12:47.400 + there are lots of resources + +1:12:47.400 --> 1:12:50.160 + that are created at different levels + +1:12:50.160 --> 1:12:54.800 + and you can find online in your school classes + +1:12:54.800 --> 1:12:57.560 + which are more mathematical, more applied and so on. + +1:12:57.560 --> 1:13:01.320 + So you can find a kind of a preacher + +1:13:01.320 --> 1:13:02.760 + which preaches in your own language + +1:13:02.760 --> 1:13:04.520 + where you can enter the field + +1:13:04.520 --> 1:13:06.720 + and you can make many different types of contribution + +1:13:06.720 --> 1:13:09.640 + depending of what is your strengths. + +1:13:10.760 --> 1:13:13.720 + And the second point, I think it's really important + +1:13:13.720 --> 1:13:18.160 + to find some area which you really care about + +1:13:18.160 --> 1:13:20.240 + and it can motivate your learning + +1:13:20.240 --> 1:13:22.640 + and it can be for somebody curing cancer + +1:13:22.640 --> 1:13:25.360 + or doing self driving cars or whatever, + +1:13:25.360 --> 1:13:29.680 + but to find an area where there is data + +1:13:29.680 --> 1:13:31.320 + where you believe there are strong patterns + +1:13:31.320 --> 1:13:33.600 + and we should be doing it and we're still not doing it + +1:13:33.600 --> 1:13:35.280 + or you can do it better + +1:13:35.280 --> 1:13:39.680 + and just start there and see where it can bring you. + +1:13:40.800 --> 1:13:45.600 + So you've been very successful in many directions in life, + +1:13:46.480 --> 1:13:48.840 + but you also mentioned Flowers of Argonon. + +1:13:51.200 --> 1:13:53.840 + And I think I've read or listened to you mention somewhere + +1:13:53.840 --> 1:13:55.360 + that researchers often get lost + +1:13:55.360 --> 1:13:56.720 + in the details of their work. + +1:13:56.720 --> 1:14:00.240 + This is per our original discussion with cancer and so on + +1:14:00.240 --> 1:14:02.200 + and don't look at the bigger picture, + +1:14:02.200 --> 1:14:05.320 + bigger questions of meaning and so on. + +1:14:05.320 --> 1:14:07.440 + So let me ask you the impossible question + +1:14:08.640 --> 1:14:11.560 + of what's the meaning of this thing, + +1:14:11.560 --> 1:14:16.560 + of life, of your life, of research. + +1:14:16.720 --> 1:14:21.440 + Why do you think we descendant of great apes + +1:14:21.440 --> 1:14:24.480 + are here on this spinning ball? + +1:14:26.800 --> 1:14:30.320 + You know, I don't think that I have really a global answer. + +1:14:30.320 --> 1:14:32.800 + You know, maybe that's why I didn't go to humanities + +1:14:33.760 --> 1:14:36.480 + and I didn't take humanities classes in my undergrad. + +1:14:39.480 --> 1:14:43.560 + But the way I'm thinking about it, + +1:14:43.560 --> 1:14:48.200 + each one of us inside of them have their own set of, + +1:14:48.200 --> 1:14:51.120 + you know, things that we believe are important. + +1:14:51.120 --> 1:14:53.360 + And it just happens that we are busy + +1:14:53.360 --> 1:14:56.240 + with achieving various goals, busy listening to others + +1:14:56.240 --> 1:15:00.960 + and to kind of try to conform and to be part of the crowd, + +1:15:00.960 --> 1:15:03.680 + that we don't listen to that part. + +1:15:04.600 --> 1:15:09.600 + And, you know, we all should find some time to understand + +1:15:09.600 --> 1:15:11.840 + what is our own individual missions. + +1:15:11.840 --> 1:15:14.080 + And we may have very different missions + +1:15:14.080 --> 1:15:18.200 + and to make sure that while we are running 10,000 things, + +1:15:18.200 --> 1:15:21.920 + we are not, you know, missing out + +1:15:21.920 --> 1:15:26.800 + and we're putting all the resources to satisfy + +1:15:26.800 --> 1:15:28.440 + our own mission. + +1:15:28.440 --> 1:15:32.400 + And if I look over my time, when I was younger, + +1:15:32.400 --> 1:15:35.000 + most of these missions, you know, + +1:15:35.000 --> 1:15:38.600 + I was primarily driven by the external stimulus, + +1:15:38.600 --> 1:15:41.520 + you know, to achieve this or to be that. + +1:15:41.520 --> 1:15:46.520 + And now a lot of what I do is driven by really thinking + +1:15:47.640 --> 1:15:51.360 + what is important for me to achieve independently + +1:15:51.360 --> 1:15:55.160 + of the external recognition. + +1:15:55.160 --> 1:16:00.080 + And, you know, I don't mind to be viewed in certain ways. + +1:16:01.400 --> 1:16:05.760 + The most important thing for me is to be true to myself, + +1:16:05.760 --> 1:16:07.520 + to what I think is right. + +1:16:07.520 --> 1:16:08.680 + How long did it take? + +1:16:08.680 --> 1:16:13.240 + How hard was it to find the you that you have to be true to? + +1:16:14.160 --> 1:16:15.520 + So it takes time. + +1:16:15.520 --> 1:16:17.760 + And even now, sometimes, you know, + +1:16:17.760 --> 1:16:20.880 + the vanity and the triviality can take, you know. + +1:16:20.880 --> 1:16:22.560 + At MIT. + +1:16:22.560 --> 1:16:25.080 + Yeah, it can everywhere, you know, + +1:16:25.080 --> 1:16:26.960 + it's just the vanity at MIT is different, + +1:16:26.960 --> 1:16:28.160 + the vanity in different places, + +1:16:28.160 --> 1:16:30.920 + but we all have our piece of vanity. + +1:16:30.920 --> 1:16:35.920 + But I think actually for me, many times the place + +1:16:38.720 --> 1:16:43.720 + to get back to it is, you know, when I'm alone + +1:16:43.800 --> 1:16:45.800 + and also when I read. + +1:16:45.800 --> 1:16:47.760 + And I think by selecting the right books, + +1:16:47.760 --> 1:16:52.760 + you can get the right questions and learn from what you read. + +1:16:54.880 --> 1:16:58.080 + So, but again, it's not perfect. + +1:16:58.080 --> 1:17:02.040 + Like vanity sometimes dominates. + +1:17:02.040 --> 1:17:04.800 + Well, that's a beautiful way to end. + +1:17:04.800 --> 1:17:06.400 + Thank you so much for talking today. + +1:17:06.400 --> 1:17:07.240 + Thank you. + +1:17:07.240 --> 1:17:08.080 + That was fun. + +1:17:08.080 --> 1:17:28.080 + That was fun. +