WEBVTT 00:00.000 --> 00:03.220 The following is a conversation with Regina Barzilay. 00:03.220 --> 00:06.700 She's a professor at MIT and a world class researcher 00:06.700 --> 00:08.340 in natural language processing 00:08.340 --> 00:12.460 and applications of deep learning to chemistry and oncology 00:12.460 --> 00:15.340 or the use of deep learning for early diagnosis, 00:15.340 --> 00:18.300 prevention and treatment of cancer. 00:18.300 --> 00:21.020 She has also been recognized for teaching 00:21.020 --> 00:24.700 of several successful AI related courses at MIT, 00:24.700 --> 00:26.840 including the popular Introduction 00:26.840 --> 00:28.920 to Machine Learning course. 00:28.920 --> 00:32.160 This is the Artificial Intelligence podcast. 00:32.160 --> 00:34.560 If you enjoy it, subscribe on YouTube, 00:34.560 --> 00:37.840 give it five stars on iTunes, support it on Patreon 00:37.840 --> 00:39.840 or simply connect with me on Twitter 00:39.840 --> 00:43.760 at Lex Friedman spelled F R I D M A N. 00:43.760 --> 00:47.760 And now here's my conversation with Regina Barzilay. 00:48.840 --> 00:50.320 In an interview you've mentioned 00:50.320 --> 00:51.960 that if there's one course you would take, 00:51.960 --> 00:54.600 it would be a literature course with a friend of yours 00:54.600 --> 00:56.360 that a friend of yours teaches. 00:56.360 --> 00:59.160 Just out of curiosity, because I couldn't find anything 00:59.160 --> 01:04.160 on it, are there books or ideas that had profound impact 01:04.400 --> 01:07.200 on your life journey, books and ideas perhaps 01:07.200 --> 01:10.800 outside of computer science and the technical fields? 01:11.780 --> 01:14.680 I think because I'm spending a lot of my time at MIT 01:14.680 --> 01:18.280 and previously in other institutions where I was a student, 01:18.280 --> 01:21.040 I have limited ability to interact with people. 01:21.040 --> 01:22.640 So a lot of what I know about the world 01:22.640 --> 01:24.220 actually comes from books. 01:24.220 --> 01:27.240 And there were quite a number of books 01:27.240 --> 01:31.380 that had profound impact on me and how I view the world. 01:31.380 --> 01:35.820 Let me just give you one example of such a book. 01:35.820 --> 01:39.660 I've maybe a year ago read a book 01:39.660 --> 01:42.500 called The Emperor of All Melodies. 01:42.500 --> 01:45.740 It's a book about, it's kind of a history of science book 01:45.740 --> 01:50.740 on how the treatments and drugs for cancer were developed. 01:50.740 --> 01:54.580 And that book, despite the fact that I am in the business 01:54.580 --> 01:59.580 of science, really opened my eyes on how imprecise 01:59.780 --> 02:03.060 and imperfect the discovery process is 02:03.060 --> 02:05.820 and how imperfect our current solutions 02:06.980 --> 02:11.060 and what makes science succeed and be implemented. 02:11.060 --> 02:14.100 And sometimes it's actually not the strengths of the idea, 02:14.100 --> 02:17.420 but devotion of the person who wants to see it implemented. 02:17.420 --> 02:19.780 So this is one of the books that, you know, 02:19.780 --> 02:22.300 at least for the last year, quite changed the way 02:22.300 --> 02:24.940 I'm thinking about scientific process 02:24.940 --> 02:26.700 just from the historical perspective 02:26.700 --> 02:31.700 and what do I need to do to make my ideas really implemented. 02:33.460 --> 02:36.060 Let me give you an example of a book 02:36.060 --> 02:39.580 which is not kind of, which is a fiction book. 02:40.620 --> 02:43.100 It's a book called Americana. 02:44.420 --> 02:48.780 And this is a book about a young female student 02:48.780 --> 02:53.260 who comes from Africa to study in the United States. 02:53.260 --> 02:57.740 And it describes her past, you know, within her studies 02:57.740 --> 03:02.020 and her life transformation that, you know, 03:02.020 --> 03:06.540 in a new country and kind of adaptation to a new culture. 03:06.540 --> 03:11.220 And when I read this book, I saw myself 03:11.220 --> 03:13.540 in many different points of it, 03:13.540 --> 03:18.540 but it also kind of gave me the lens on different events. 03:20.140 --> 03:22.060 And some of it that I never actually paid attention. 03:22.060 --> 03:24.700 One of the funny stories in this book 03:24.700 --> 03:29.700 is how she arrives to her new college 03:30.420 --> 03:32.900 and she starts speaking in English 03:32.900 --> 03:35.700 and she had this beautiful British accent 03:35.700 --> 03:39.860 because that's how she was educated in her country. 03:39.860 --> 03:40.980 This is not my case. 03:40.980 --> 03:45.460 And then she notices that the person who talks to her, 03:45.460 --> 03:47.220 you know, talks to her in a very funny way, 03:47.220 --> 03:48.340 in a very slow way. 03:48.340 --> 03:51.460 And she's thinking that this woman is disabled 03:51.460 --> 03:54.500 and she's also trying to kind of to accommodate her. 03:54.500 --> 03:56.700 And then after a while, when she finishes her discussion 03:56.700 --> 03:58.580 with this officer from her college, 03:59.860 --> 04:02.100 she sees how she interacts with the other students, 04:02.100 --> 04:03.020 with American students. 04:03.020 --> 04:08.020 And she discovers that actually she talked to her this way 04:08.020 --> 04:11.020 because she saw that she doesn't understand English. 04:11.020 --> 04:14.180 And I thought, wow, this is a funny experience. 04:14.180 --> 04:16.940 And literally within few weeks, 04:16.940 --> 04:20.820 I went to LA to a conference 04:20.820 --> 04:23.180 and I asked somebody in the airport, 04:23.180 --> 04:25.580 you know, how to find like a cab or something. 04:25.580 --> 04:28.380 And then I noticed that this person is talking 04:28.380 --> 04:29.220 in a very strange way. 04:29.220 --> 04:31.100 And my first thought was that this person 04:31.100 --> 04:34.500 have some, you know, pronunciation issues or something. 04:34.500 --> 04:36.060 And I'm trying to talk very slowly to him 04:36.060 --> 04:38.580 and I was with another professor, Ernst Frankel. 04:38.580 --> 04:42.180 And he's like laughing because it's funny 04:42.180 --> 04:44.860 that I don't get that the guy is talking in this way 04:44.860 --> 04:46.060 because he thinks that I cannot speak. 04:46.060 --> 04:49.100 So it was really kind of mirroring experience. 04:49.100 --> 04:53.300 And it led me think a lot about my own experiences 04:53.300 --> 04:56.060 moving, you know, from different countries. 04:56.060 --> 04:59.300 So I think that books play a big role 04:59.300 --> 05:01.780 in my understanding of the world. 05:01.780 --> 05:06.420 On the science question, you mentioned that 05:06.420 --> 05:09.780 it made you discover that personalities of human beings 05:09.780 --> 05:12.420 are more important than perhaps ideas. 05:12.420 --> 05:13.660 Is that what I heard? 05:13.660 --> 05:15.740 It's not necessarily that they are more important 05:15.740 --> 05:19.180 than ideas, but I think that ideas on their own 05:19.180 --> 05:20.460 are not sufficient. 05:20.460 --> 05:24.660 And many times, at least at the local horizon, 05:24.660 --> 05:29.140 it's the personalities and their devotion to their ideas 05:29.140 --> 05:32.980 is really that locally changes the landscape. 05:32.980 --> 05:37.500 Now, if you're looking at AI, like let's say 30 years ago, 05:37.500 --> 05:39.180 you know, dark ages of AI or whatever, 05:39.180 --> 05:42.420 what is symbolic times, you can use any word. 05:42.420 --> 05:44.660 You know, there were some people, 05:44.660 --> 05:46.620 now we're looking at a lot of that work 05:46.620 --> 05:48.780 and we're kind of thinking this was not really 05:48.780 --> 05:52.220 maybe a relevant work, but you can see that some people 05:52.220 --> 05:54.900 managed to take it and to make it so shiny 05:54.900 --> 05:59.260 and dominate the academic world 05:59.260 --> 06:02.380 and make it to be the standard. 06:02.380 --> 06:05.180 If you look at the area of natural language processing, 06:06.420 --> 06:09.140 it is well known fact that the reason that statistics 06:09.140 --> 06:13.980 in NLP took such a long time to become mainstream 06:13.980 --> 06:16.860 because there were quite a number of personalities 06:16.860 --> 06:18.460 which didn't believe in this idea 06:18.460 --> 06:22.060 and didn't stop research progress in this area. 06:22.060 --> 06:25.900 So I do not think that, you know, 06:25.900 --> 06:28.940 kind of asymptotically maybe personalities matters, 06:28.940 --> 06:33.940 but I think locally it does make quite a bit of impact 06:33.940 --> 06:36.900 and it's generally, you know, 06:36.900 --> 06:41.340 speeds up the rate of adoption of the new ideas. 06:41.340 --> 06:43.500 Yeah, and the other interesting question 06:43.500 --> 06:46.540 is in the early days of particular discipline, 06:46.540 --> 06:50.460 I think you mentioned in that book 06:50.460 --> 06:52.340 is ultimately a book of cancer. 06:52.340 --> 06:55.100 It's called The Emperor of All Melodies. 06:55.100 --> 06:58.580 Yeah, and those melodies included the trying to, 06:58.580 --> 07:00.740 the medicine, was it centered around? 07:00.740 --> 07:04.900 So it was actually centered on, you know, 07:04.900 --> 07:07.180 how people thought of curing cancer. 07:07.180 --> 07:10.660 Like for me, it was really a discovery how people, 07:10.660 --> 07:14.140 what was the science of chemistry behind drug development 07:14.140 --> 07:17.220 that it actually grew up out of dying, 07:17.220 --> 07:19.780 like coloring industry that people 07:19.780 --> 07:23.780 who developed chemistry in 19th century in Germany 07:23.780 --> 07:28.140 and Britain to do, you know, the really new dyes. 07:28.140 --> 07:30.180 They looked at the molecules and identified 07:30.180 --> 07:32.140 that they do certain things to cells. 07:32.140 --> 07:34.500 And from there, the process started. 07:34.500 --> 07:35.740 And, you know, like historically saying, 07:35.740 --> 07:36.900 yeah, this is fascinating 07:36.900 --> 07:38.700 that they managed to make the connection 07:38.700 --> 07:42.300 and look under the microscope and do all this discovery. 07:42.300 --> 07:44.340 But as you continue reading about it 07:44.340 --> 07:48.780 and you read about how chemotherapy drugs 07:48.780 --> 07:50.500 which were developed in Boston, 07:50.500 --> 07:52.500 and some of them were developed. 07:52.500 --> 07:57.500 And Farber, Dr. Farber from Dana Farber, 07:57.500 --> 08:00.460 you know, how the experiments were done 08:00.460 --> 08:03.340 that, you know, there was some miscalculation, 08:03.340 --> 08:04.540 let's put it this way. 08:04.540 --> 08:06.740 And they tried it on the patients and they just, 08:06.740 --> 08:09.980 and those were children with leukemia and they died. 08:09.980 --> 08:11.660 And then they tried another modification. 08:11.660 --> 08:15.020 You look at the process, how imperfect is this process? 08:15.020 --> 08:17.500 And, you know, like, if we're again looking back 08:17.500 --> 08:19.180 like 60 years ago, 70 years ago, 08:19.180 --> 08:20.780 you can kind of understand it. 08:20.780 --> 08:23.020 But some of the stories in this book 08:23.020 --> 08:24.620 which were really shocking to me 08:24.620 --> 08:27.980 were really happening, you know, maybe decades ago. 08:27.980 --> 08:30.660 And we still don't have a vehicle 08:30.660 --> 08:35.100 to do it much more fast and effective and, you know, 08:35.100 --> 08:38.220 scientific the way I'm thinking computer science scientific. 08:38.220 --> 08:40.420 So from the perspective of computer science, 08:40.420 --> 08:43.780 you've gotten a chance to work the application to cancer 08:43.780 --> 08:44.860 and to medicine in general. 08:44.860 --> 08:48.420 From a perspective of an engineer and a computer scientist, 08:48.420 --> 08:51.780 how far along are we from understanding the human body, 08:51.780 --> 08:55.140 biology of being able to manipulate it 08:55.140 --> 08:57.940 in a way we can cure some of the maladies, 08:57.940 --> 08:59.740 some of the diseases? 08:59.740 --> 09:02.220 So this is very interesting question. 09:03.460 --> 09:06.020 And if you're thinking as a computer scientist 09:06.020 --> 09:09.820 about this problem, I think one of the reasons 09:09.820 --> 09:11.900 that we succeeded in the areas 09:11.900 --> 09:13.980 we as a computer scientist succeeded 09:13.980 --> 09:16.260 is because we don't have, 09:16.260 --> 09:18.980 we are not trying to understand in some ways. 09:18.980 --> 09:22.260 Like if you're thinking about like eCommerce, Amazon, 09:22.260 --> 09:24.220 Amazon doesn't really understand you. 09:24.220 --> 09:27.700 And that's why it recommends you certain books 09:27.700 --> 09:29.580 or certain products, correct? 09:30.660 --> 09:34.660 And, you know, traditionally when people 09:34.660 --> 09:36.380 were thinking about marketing, you know, 09:36.380 --> 09:39.780 they divided the population to different kind of subgroups, 09:39.780 --> 09:41.740 identify the features of this subgroup 09:41.740 --> 09:43.140 and come up with a strategy 09:43.140 --> 09:45.580 which is specific to that subgroup. 09:45.580 --> 09:47.340 If you're looking about recommendation system, 09:47.340 --> 09:50.580 they're not claiming that they're understanding somebody, 09:50.580 --> 09:52.700 they're just managing to, 09:52.700 --> 09:54.780 from the patterns of your behavior 09:54.780 --> 09:57.540 to recommend you a product. 09:57.540 --> 09:59.580 Now, if you look at the traditional biology, 09:59.580 --> 10:03.180 and obviously I wouldn't say that I 10:03.180 --> 10:06.180 at any way, you know, educated in this field, 10:06.180 --> 10:09.300 but you know what I see, there's really a lot of emphasis 10:09.300 --> 10:10.660 on mechanistic understanding. 10:10.660 --> 10:12.540 And it was very surprising to me 10:12.540 --> 10:13.820 coming from computer science, 10:13.820 --> 10:17.580 how much emphasis is on this understanding. 10:17.580 --> 10:20.740 And given the complexity of the system, 10:20.740 --> 10:23.220 maybe the deterministic full understanding 10:23.220 --> 10:27.380 of this process is, you know, beyond our capacity. 10:27.380 --> 10:29.460 And the same ways in computer science 10:29.460 --> 10:31.540 when we're doing recognition, when you do recommendation 10:31.540 --> 10:32.780 and many other areas, 10:32.780 --> 10:35.940 it's just probabilistic matching process. 10:35.940 --> 10:40.100 And in some way, maybe in certain cases, 10:40.100 --> 10:42.940 we shouldn't even attempt to understand 10:42.940 --> 10:45.780 or we can attempt to understand, but in parallel, 10:45.780 --> 10:48.060 we can actually do this kind of matchings 10:48.060 --> 10:51.060 that would help us to find key role 10:51.060 --> 10:54.100 to do early diagnostics and so on. 10:54.100 --> 10:55.860 And I know that in these communities, 10:55.860 --> 10:59.060 it's really important to understand, 10:59.060 --> 11:00.700 but I'm sometimes wondering, you know, 11:00.700 --> 11:02.940 what exactly does it mean to understand here? 11:02.940 --> 11:05.500 Well, there's stuff that works and, 11:05.500 --> 11:07.620 but that can be, like you said, 11:07.620 --> 11:10.340 separate from this deep human desire 11:10.340 --> 11:12.700 to uncover the mysteries of the universe, 11:12.700 --> 11:16.140 of science, of the way the body works, 11:16.140 --> 11:17.620 the way the mind works. 11:17.620 --> 11:19.540 It's the dream of symbolic AI, 11:19.540 --> 11:24.540 of being able to reduce human knowledge into logic 11:25.220 --> 11:26.900 and be able to play with that logic 11:26.900 --> 11:28.700 in a way that's very explainable 11:28.700 --> 11:30.300 and understandable for us humans. 11:30.300 --> 11:31.780 I mean, that's a beautiful dream. 11:31.780 --> 11:34.860 So I understand it, but it seems that 11:34.860 --> 11:37.900 what seems to work today and we'll talk about it more 11:37.900 --> 11:40.780 is as much as possible, reduce stuff into data, 11:40.780 --> 11:43.900 reduce whatever problem you're interested in to data 11:43.900 --> 11:47.060 and try to apply statistical methods, 11:47.060 --> 11:49.100 apply machine learning to that. 11:49.100 --> 11:51.140 On a personal note, 11:51.140 --> 11:54.140 you were diagnosed with breast cancer in 2014. 11:55.380 --> 11:58.420 What did facing your mortality make you think about? 11:58.420 --> 12:00.260 How did it change you? 12:00.260 --> 12:01.860 You know, this is a great question 12:01.860 --> 12:03.820 and I think that I was interviewed many times 12:03.820 --> 12:05.740 and nobody actually asked me this question. 12:05.740 --> 12:09.700 I think I was 43 at a time. 12:09.700 --> 12:12.860 And the first time I realized in my life that I may die 12:12.860 --> 12:14.460 and I never thought about it before. 12:14.460 --> 12:17.260 And there was a long time since you're diagnosed 12:17.260 --> 12:18.580 until you actually know what you have 12:18.580 --> 12:20.180 and how severe is your disease. 12:20.180 --> 12:23.500 For me, it was like maybe two and a half months. 12:23.500 --> 12:28.340 And I didn't know where I am during this time 12:28.340 --> 12:30.660 because I was getting different tests 12:30.660 --> 12:33.380 and one would say it's bad and I would say, no, it is not. 12:33.380 --> 12:34.900 So until I knew where I am, 12:34.900 --> 12:36.300 I really was thinking about 12:36.300 --> 12:38.220 all these different possible outcomes. 12:38.220 --> 12:39.700 Were you imagining the worst 12:39.700 --> 12:41.940 or were you trying to be optimistic or? 12:41.940 --> 12:43.540 It would be really, 12:43.540 --> 12:47.340 I don't remember what was my thinking. 12:47.340 --> 12:51.100 It was really a mixture with many components at the time 12:51.100 --> 12:54.100 speaking in our terms. 12:54.100 --> 12:59.100 And one thing that I remember, 12:59.340 --> 13:01.500 and every test comes and then you're saying, 13:01.500 --> 13:03.300 oh, it could be this or it may not be this. 13:03.300 --> 13:04.700 And you're hopeful and then you're desperate. 13:04.700 --> 13:07.660 So it's like, there is a whole slew of emotions 13:07.660 --> 13:08.700 that goes through you. 13:09.820 --> 13:14.820 But what I remember is that when I came back to MIT, 13:15.100 --> 13:17.780 I was kind of going the whole time through the treatment 13:17.780 --> 13:19.780 to MIT, but my brain was not really there. 13:19.780 --> 13:21.820 But when I came back, really finished my treatment 13:21.820 --> 13:23.860 and I was here teaching and everything, 13:24.900 --> 13:27.060 I look back at what my group was doing, 13:27.060 --> 13:28.820 what other groups was doing. 13:28.820 --> 13:30.820 And I saw these trivialities. 13:30.820 --> 13:33.260 It's like people are building their careers 13:33.260 --> 13:36.900 on improving some parts around two or 3% or whatever. 13:36.900 --> 13:38.380 I was, it's like, seriously, 13:38.380 --> 13:40.740 I did a work on how to decipher ugaritic, 13:40.740 --> 13:42.860 like a language that nobody speak and whatever, 13:42.860 --> 13:46.140 like what is significance? 13:46.140 --> 13:49.020 When all of a sudden, I walked out of MIT, 13:49.020 --> 13:51.860 which is when people really do care 13:51.860 --> 13:54.500 what happened to your ICLR paper, 13:54.500 --> 13:57.900 what is your next publication to ACL, 13:57.900 --> 14:01.860 to the world where people, you see a lot of suffering 14:01.860 --> 14:04.900 that I'm kind of totally shielded on it on daily basis. 14:04.900 --> 14:07.460 And it's like the first time I've seen like real life 14:07.460 --> 14:08.660 and real suffering. 14:09.700 --> 14:13.260 And I was thinking, why are we trying to improve the parser 14:13.260 --> 14:18.260 or deal with trivialities when we have capacity 14:18.340 --> 14:20.700 to really make a change? 14:20.700 --> 14:24.620 And it was really challenging to me because on one hand, 14:24.620 --> 14:27.420 I have my graduate students really want to do their papers 14:27.420 --> 14:29.860 and their work, and they want to continue to do 14:29.860 --> 14:31.900 what they were doing, which was great. 14:31.900 --> 14:36.300 And then it was me who really kind of reevaluated 14:36.300 --> 14:37.460 what is the importance. 14:37.460 --> 14:40.260 And also at that point, because I had to take some break, 14:42.500 --> 14:47.500 I look back into like my years in science 14:47.740 --> 14:50.460 and I was thinking, like 10 years ago, 14:50.460 --> 14:52.940 this was the biggest thing, I don't know, topic models. 14:52.940 --> 14:55.340 We have like millions of papers on topic models 14:55.340 --> 14:56.500 and variation of topics models. 14:56.500 --> 14:58.580 Now it's totally like irrelevant. 14:58.580 --> 15:02.460 And you start looking at this, what do you perceive 15:02.460 --> 15:04.500 as important at different point of time 15:04.500 --> 15:08.900 and how it fades over time. 15:08.900 --> 15:12.980 And since we have a limited time, 15:12.980 --> 15:14.900 all of us have limited time on us, 15:14.900 --> 15:18.380 it's really important to prioritize things 15:18.380 --> 15:20.540 that really matter to you, maybe matter to you 15:20.540 --> 15:22.020 at that particular point. 15:22.020 --> 15:24.380 But it's important to take some time 15:24.380 --> 15:26.940 and understand what matters to you, 15:26.940 --> 15:28.860 which may not necessarily be the same 15:28.860 --> 15:31.700 as what matters to the rest of your scientific community 15:31.700 --> 15:34.580 and pursue that vision. 15:34.580 --> 15:38.460 So that moment, did it make you cognizant? 15:38.460 --> 15:42.500 You mentioned suffering of just the general amount 15:42.500 --> 15:44.340 of suffering in the world. 15:44.340 --> 15:45.620 Is that what you're referring to? 15:45.620 --> 15:47.420 So as opposed to topic models 15:47.420 --> 15:50.780 and specific detailed problems in NLP, 15:50.780 --> 15:54.460 did you start to think about other people 15:54.460 --> 15:56.940 who have been diagnosed with cancer? 15:56.940 --> 16:00.020 Is that the way you started to see the world perhaps? 16:00.020 --> 16:00.860 Oh, absolutely. 16:00.860 --> 16:04.980 And it actually creates, because like, for instance, 16:04.980 --> 16:05.820 there is parts of the treatment 16:05.820 --> 16:08.500 where you need to go to the hospital every day 16:08.500 --> 16:11.620 and you see the community of people that you see 16:11.620 --> 16:16.100 and many of them are much worse than I was at a time. 16:16.100 --> 16:20.460 And you all of a sudden see it all. 16:20.460 --> 16:23.940 And people who are happier someday 16:23.940 --> 16:25.300 just because they feel better. 16:25.300 --> 16:28.500 And for people who are in our normal realm, 16:28.500 --> 16:30.820 you take it totally for granted that you feel well, 16:30.820 --> 16:32.940 that if you decide to go running, you can go running 16:32.940 --> 16:35.900 and you're pretty much free 16:35.900 --> 16:37.620 to do whatever you want with your body. 16:37.620 --> 16:40.180 Like I saw like a community, 16:40.180 --> 16:42.820 my community became those people. 16:42.820 --> 16:47.460 And I remember one of my friends, Dina Katabi, 16:47.460 --> 16:50.420 took me to Prudential to buy me a gift for my birthday. 16:50.420 --> 16:52.340 And it was like the first time in months 16:52.340 --> 16:54.980 that I went to kind of to see other people. 16:54.980 --> 16:58.180 And I was like, wow, first of all, these people, 16:58.180 --> 16:59.820 they are happy and they're laughing 16:59.820 --> 17:02.620 and they're very different from these other my people. 17:02.620 --> 17:04.620 And second of thing, I think it's totally crazy. 17:04.620 --> 17:06.620 They're like laughing and wasting their money 17:06.620 --> 17:08.420 on some stupid gifts. 17:08.420 --> 17:12.540 And they may die. 17:12.540 --> 17:15.940 They already may have cancer and they don't understand it. 17:15.940 --> 17:20.060 So you can really see how the mind changes 17:20.060 --> 17:22.340 that you can see that, 17:22.340 --> 17:23.180 before that you can ask, 17:23.180 --> 17:24.380 didn't you know that you're gonna die? 17:24.380 --> 17:28.340 Of course I knew, but it was a kind of a theoretical notion. 17:28.340 --> 17:31.060 It wasn't something which was concrete. 17:31.060 --> 17:33.900 And at that point, when you really see it 17:33.900 --> 17:38.060 and see how little means sometimes the system has 17:38.060 --> 17:41.740 to have them, you really feel that we need to take a lot 17:41.740 --> 17:45.420 of our brilliance that we have here at MIT 17:45.420 --> 17:48.020 and translate it into something useful. 17:48.020 --> 17:50.540 Yeah, and you still couldn't have a lot of definitions, 17:50.540 --> 17:53.620 but of course, alleviating, suffering, alleviating, 17:53.620 --> 17:57.460 trying to cure cancer is a beautiful mission. 17:57.460 --> 18:01.940 So I of course know theoretically the notion of cancer, 18:01.940 --> 18:06.940 but just reading more and more about it's 1.7 million 18:07.100 --> 18:09.860 new cancer cases in the United States every year, 18:09.860 --> 18:13.460 600,000 cancer related deaths every year. 18:13.460 --> 18:18.460 So this has a huge impact, United States globally. 18:19.340 --> 18:24.340 When broadly, before we talk about how machine learning, 18:24.340 --> 18:27.180 how MIT can help, 18:27.180 --> 18:32.100 when do you think we as a civilization will cure cancer? 18:32.100 --> 18:34.980 How hard of a problem is it from everything you've learned 18:34.980 --> 18:35.940 from it recently? 18:37.260 --> 18:39.300 I cannot really assess it. 18:39.300 --> 18:42.100 What I do believe will happen with the advancement 18:42.100 --> 18:45.940 in machine learning is that a lot of types of cancer 18:45.940 --> 18:48.500 we will be able to predict way early 18:48.500 --> 18:53.420 and more effectively utilize existing treatments. 18:53.420 --> 18:57.540 I think, I hope at least that with all the advancements 18:57.540 --> 19:01.180 in AI and drug discovery, we would be able 19:01.180 --> 19:04.700 to much faster find relevant molecules. 19:04.700 --> 19:08.220 What I'm not sure about is how long it will take 19:08.220 --> 19:11.940 the medical establishment and regulatory bodies 19:11.940 --> 19:14.780 to kind of catch up and to implement it. 19:14.780 --> 19:17.420 And I think this is a very big piece of puzzle 19:17.420 --> 19:20.420 that is currently not addressed. 19:20.420 --> 19:21.780 That's the really interesting question. 19:21.780 --> 19:25.460 So first a small detail that I think the answer is yes, 19:25.460 --> 19:30.460 but is cancer one of the diseases that when detected earlier 19:33.700 --> 19:37.820 that's a significantly improves the outcomes? 19:37.820 --> 19:41.020 So like, cause we will talk about there's the cure 19:41.020 --> 19:43.020 and then there is detection. 19:43.020 --> 19:45.180 And I think where machine learning can really help 19:45.180 --> 19:46.660 is earlier detection. 19:46.660 --> 19:48.580 So does detection help? 19:48.580 --> 19:49.660 Detection is crucial. 19:49.660 --> 19:53.940 For instance, the vast majority of pancreatic cancer patients 19:53.940 --> 19:57.300 are detected at the stage that they are incurable. 19:57.300 --> 20:02.300 That's why they have such a terrible survival rate. 20:03.740 --> 20:07.300 It's like just few percent over five years. 20:07.300 --> 20:09.820 It's pretty much today the sentence. 20:09.820 --> 20:13.620 But if you can discover this disease early, 20:14.500 --> 20:16.740 there are mechanisms to treat it. 20:16.740 --> 20:20.740 And in fact, I know a number of people who were diagnosed 20:20.740 --> 20:23.580 and saved just because they had food poisoning. 20:23.580 --> 20:25.020 They had terrible food poisoning. 20:25.020 --> 20:28.540 They went to ER, they got scan. 20:28.540 --> 20:30.660 There were early signs on the scan 20:30.660 --> 20:33.540 and that would save their lives. 20:33.540 --> 20:35.820 But this wasn't really an accidental case. 20:35.820 --> 20:40.820 So as we become better, we would be able to help 20:41.260 --> 20:46.260 to many more people that are likely to develop diseases. 20:46.540 --> 20:51.020 And I just want to say that as I got more into this field, 20:51.020 --> 20:53.620 I realized that cancer is of course terrible disease, 20:53.620 --> 20:56.700 but there are really the whole slew of terrible diseases 20:56.700 --> 21:00.820 out there like neurodegenerative diseases and others. 21:01.660 --> 21:04.580 So we, of course, a lot of us are fixated on cancer 21:04.580 --> 21:06.420 because it's so prevalent in our society. 21:06.420 --> 21:08.540 And you see these people where there are a lot of patients 21:08.540 --> 21:10.340 with neurodegenerative diseases 21:10.340 --> 21:12.540 and the kind of aging diseases 21:12.540 --> 21:17.100 that we still don't have a good solution for. 21:17.100 --> 21:22.100 And I felt as a computer scientist, 21:22.860 --> 21:25.460 we kind of decided that it's other people's job 21:25.460 --> 21:29.340 to treat these diseases because it's like traditionally 21:29.340 --> 21:32.420 people in biology or in chemistry or MDs 21:32.420 --> 21:35.340 are the ones who are thinking about it. 21:35.340 --> 21:37.420 And after kind of start paying attention, 21:37.420 --> 21:40.340 I think that it's really a wrong assumption 21:40.340 --> 21:42.940 and we all need to join the battle. 21:42.940 --> 21:46.460 So how it seems like in cancer specifically 21:46.460 --> 21:49.140 that there's a lot of ways that machine learning can help. 21:49.140 --> 21:51.860 So what's the role of machine learning 21:51.860 --> 21:54.100 in the diagnosis of cancer? 21:55.260 --> 21:58.700 So for many cancers today, we really don't know 21:58.700 --> 22:03.460 what is your likelihood to get cancer. 22:03.460 --> 22:06.300 And for the vast majority of patients, 22:06.300 --> 22:07.940 especially on the younger patients, 22:07.940 --> 22:09.580 it really comes as a surprise. 22:09.580 --> 22:11.140 Like for instance, for breast cancer, 22:11.140 --> 22:13.860 80% of the patients are first in their families, 22:13.860 --> 22:15.380 it's like me. 22:15.380 --> 22:18.460 And I never saw that I had any increased risk 22:18.460 --> 22:20.820 because nobody had it in my family. 22:20.820 --> 22:22.300 And for some reason in my head, 22:22.300 --> 22:24.820 it was kind of inherited disease. 22:26.580 --> 22:28.380 But even if I would pay attention, 22:28.380 --> 22:32.420 the very simplistic statistical models 22:32.420 --> 22:34.540 that are currently used in clinical practice, 22:34.540 --> 22:37.460 they really don't give you an answer, so you don't know. 22:37.460 --> 22:40.380 And the same true for pancreatic cancer, 22:40.380 --> 22:45.380 the same true for non smoking lung cancer and many others. 22:45.380 --> 22:47.340 So what machine learning can do here 22:47.340 --> 22:51.620 is utilize all this data to tell us early 22:51.620 --> 22:53.140 who is likely to be susceptible 22:53.140 --> 22:55.980 and using all the information that is already there, 22:55.980 --> 22:59.980 be it imaging, be it your other tests, 22:59.980 --> 23:04.860 and eventually liquid biopsies and others, 23:04.860 --> 23:08.180 where the signal itself is not sufficiently strong 23:08.180 --> 23:11.300 for human eye to do good discrimination 23:11.300 --> 23:12.940 because the signal may be weak, 23:12.940 --> 23:15.620 but by combining many sources, 23:15.620 --> 23:18.100 machine which is trained on large volumes of data 23:18.100 --> 23:20.700 can really detect it early. 23:20.700 --> 23:22.500 And that's what we've seen with breast cancer 23:22.500 --> 23:25.900 and people are reporting it in other diseases as well. 23:25.900 --> 23:28.260 That really boils down to data, right? 23:28.260 --> 23:30.980 And in the different kinds of sources of data. 23:30.980 --> 23:33.740 And you mentioned regulatory challenges. 23:33.740 --> 23:35.180 So what are the challenges 23:35.180 --> 23:39.260 in gathering large data sets in this space? 23:40.860 --> 23:42.660 Again, another great question. 23:42.660 --> 23:45.500 So it took me after I decided that I want to work on it 23:45.500 --> 23:48.740 two years to get access to data. 23:48.740 --> 23:50.580 Any data, like any significant data set? 23:50.580 --> 23:53.580 Any significant amount, like right now in this country, 23:53.580 --> 23:57.060 there is no publicly available data set 23:57.060 --> 23:58.820 of modern mammograms that you can just go 23:58.820 --> 24:01.860 on your computer, sign a document and get it. 24:01.860 --> 24:03.180 It just doesn't exist. 24:03.180 --> 24:06.860 I mean, obviously every hospital has its own collection 24:06.860 --> 24:07.700 of mammograms. 24:07.700 --> 24:11.300 There are data that came out of clinical trials. 24:11.300 --> 24:13.220 What we're talking about here is a computer scientist 24:13.220 --> 24:17.140 who just wants to run his or her model 24:17.140 --> 24:19.060 and see how it works. 24:19.060 --> 24:22.900 This data, like ImageNet, doesn't exist. 24:22.900 --> 24:27.900 And there is a set which is called like Florida data set 24:28.620 --> 24:30.860 which is a film mammogram from 90s 24:30.860 --> 24:32.420 which is totally not representative 24:32.420 --> 24:33.860 of the current developments. 24:33.860 --> 24:35.780 Whatever you're learning on them doesn't scale up. 24:35.780 --> 24:39.300 This is the only resource that is available. 24:39.300 --> 24:42.780 And today there are many agencies 24:42.780 --> 24:44.460 that govern access to data. 24:44.460 --> 24:46.300 Like the hospital holds your data 24:46.300 --> 24:49.260 and the hospital decides whether they would give it 24:49.260 --> 24:52.340 to the researcher to work with this data or not. 24:52.340 --> 24:54.180 Individual hospital? 24:54.180 --> 24:55.020 Yeah. 24:55.020 --> 24:57.220 I mean, the hospital may, you know, 24:57.220 --> 24:59.220 assuming that you're doing research collaboration, 24:59.220 --> 25:01.980 you can submit, you know, 25:01.980 --> 25:05.060 there is a proper approval process guided by RB 25:05.060 --> 25:07.820 and if you go through all the processes, 25:07.820 --> 25:10.140 you can eventually get access to the data. 25:10.140 --> 25:13.540 But if you yourself know our OEI community, 25:13.540 --> 25:16.100 there are not that many people who actually ever got access 25:16.100 --> 25:20.260 to data because it's very challenging process. 25:20.260 --> 25:22.780 And sorry, just in a quick comment, 25:22.780 --> 25:25.780 MGH or any kind of hospital, 25:25.780 --> 25:28.100 are they scanning the data? 25:28.100 --> 25:29.740 Are they digitally storing it? 25:29.740 --> 25:31.580 Oh, it is already digitally stored. 25:31.580 --> 25:34.180 You don't need to do any extra processing steps. 25:34.180 --> 25:38.340 It's already there in the right format is that right now 25:38.340 --> 25:41.180 there are a lot of issues that govern access to the data 25:41.180 --> 25:46.180 because the hospital is legally responsible for the data. 25:46.180 --> 25:51.020 And, you know, they have a lot to lose 25:51.020 --> 25:53.140 if they give the data to the wrong person, 25:53.140 --> 25:56.460 but they may not have a lot to gain if they give it 25:56.460 --> 26:00.580 as a hospital, as a legal entity has given it to you. 26:00.580 --> 26:02.740 And the way, you know, what I would imagine 26:02.740 --> 26:05.220 happening in the future is the same thing that happens 26:05.220 --> 26:06.780 when you're getting your driving license, 26:06.780 --> 26:09.820 you can decide whether you want to donate your organs. 26:09.820 --> 26:13.100 You can imagine that whenever a person goes to the hospital, 26:13.100 --> 26:17.540 they, it should be easy for them to donate their data 26:17.540 --> 26:19.420 for research and it can be different kind of, 26:19.420 --> 26:22.420 do they only give you a test results or only mammogram 26:22.420 --> 26:25.900 or only imaging data or the whole medical record? 26:27.060 --> 26:28.980 Because at the end, 26:30.540 --> 26:33.860 we all will benefit from all this insights. 26:33.860 --> 26:36.060 And it's not like you say, I want to keep my data private, 26:36.060 --> 26:38.780 but I would really love to get it from other people 26:38.780 --> 26:40.740 because other people are thinking the same way. 26:40.740 --> 26:45.740 So if there is a mechanism to do this donation 26:45.740 --> 26:48.020 and the patient has an ability to say 26:48.020 --> 26:50.820 how they want to use their data for research, 26:50.820 --> 26:54.100 it would be really a game changer. 26:54.100 --> 26:56.460 People, when they think about this problem, 26:56.460 --> 26:58.460 there's a, it depends on the population, 26:58.460 --> 27:00.140 depends on the demographics, 27:00.140 --> 27:03.420 but there's some privacy concerns generally, 27:03.420 --> 27:05.860 not just medical data, just any kind of data. 27:05.860 --> 27:09.620 It's what you said, my data, it should belong kind of to me. 27:09.620 --> 27:11.660 I'm worried how it's going to be misused. 27:12.540 --> 27:15.620 How do we alleviate those concerns? 27:17.100 --> 27:19.460 Because that seems like a problem that needs to be, 27:19.460 --> 27:22.980 that problem of trust, of transparency needs to be solved 27:22.980 --> 27:27.260 before we build large data sets that help detect cancer, 27:27.260 --> 27:30.180 help save those very people in the future. 27:30.180 --> 27:31.940 So I think there are two things that could be done. 27:31.940 --> 27:34.460 There is a technical solutions 27:34.460 --> 27:38.220 and there are societal solutions. 27:38.220 --> 27:40.180 So on the technical end, 27:41.460 --> 27:46.460 we today have ability to improve disambiguation. 27:48.140 --> 27:49.740 Like, for instance, for imaging, 27:49.740 --> 27:54.740 it's, you know, for imaging, you can do it pretty well. 27:55.620 --> 27:56.780 What's disambiguation? 27:56.780 --> 27:58.540 And disambiguation, sorry, disambiguation, 27:58.540 --> 27:59.860 removing the identification, 27:59.860 --> 28:02.220 removing the names of the people. 28:02.220 --> 28:04.820 There are other data, like if it is a raw tax, 28:04.820 --> 28:08.180 you cannot really achieve 99.9%, 28:08.180 --> 28:10.060 but there are all these techniques 28:10.060 --> 28:12.460 that actually some of them are developed at MIT, 28:12.460 --> 28:15.460 how you can do learning on the encoded data 28:15.460 --> 28:17.420 where you locally encode the image, 28:17.420 --> 28:22.420 you train a network which only works on the encoded images 28:22.420 --> 28:24.940 and then you send the outcome back to the hospital 28:24.940 --> 28:26.580 and you can open it up. 28:26.580 --> 28:28.020 So those are the technical solutions. 28:28.020 --> 28:30.660 There are a lot of people who are working in this space 28:30.660 --> 28:33.780 where the learning happens in the encoded form. 28:33.780 --> 28:36.180 We are still early, 28:36.180 --> 28:39.260 but this is an interesting research area 28:39.260 --> 28:41.900 where I think we'll make more progress. 28:43.340 --> 28:45.620 There is a lot of work in natural language processing 28:45.620 --> 28:48.620 community how to do the identification better. 28:50.380 --> 28:54.020 But even today, there are already a lot of data 28:54.020 --> 28:55.900 which can be deidentified perfectly, 28:55.900 --> 28:58.780 like your test data, for instance, correct, 28:58.780 --> 29:00.980 where you can just, you know the name of the patient, 29:00.980 --> 29:04.300 you just want to extract the part with the numbers. 29:04.300 --> 29:07.460 The big problem here is again, 29:08.420 --> 29:10.420 hospitals don't see much incentive 29:10.420 --> 29:12.660 to give this data away on one hand 29:12.660 --> 29:14.220 and then there is general concern. 29:14.220 --> 29:17.700 Now, when I'm talking about societal benefits 29:17.700 --> 29:19.660 and about the education, 29:19.660 --> 29:24.340 the public needs to understand that I think 29:25.700 --> 29:29.420 that there are situation and I still remember myself 29:29.420 --> 29:33.380 when I really needed an answer, I had to make a choice. 29:33.380 --> 29:35.220 There was no information to make a choice, 29:35.220 --> 29:36.660 you're just guessing. 29:36.660 --> 29:41.060 And at that moment you feel that your life is at the stake, 29:41.060 --> 29:44.820 but you just don't have information to make the choice. 29:44.820 --> 29:48.740 And many times when I give talks, 29:48.740 --> 29:51.300 I get emails from women who say, 29:51.300 --> 29:52.820 you know, I'm in this situation, 29:52.820 --> 29:55.940 can you please run statistic and see what are the outcomes? 29:57.100 --> 30:01.300 We get almost every week a mammogram that comes by mail 30:01.300 --> 30:03.460 to my office at MIT, I'm serious. 30:04.380 --> 30:07.860 That people ask to run because they need to make 30:07.860 --> 30:10.020 life changing decisions. 30:10.020 --> 30:12.980 And of course, I'm not planning to open a clinic here, 30:12.980 --> 30:16.660 but we do run and give them the results for their doctors. 30:16.660 --> 30:20.100 But the point that I'm trying to make, 30:20.100 --> 30:23.780 that we all at some point or our loved ones 30:23.780 --> 30:26.620 will be in the situation where you need information 30:26.620 --> 30:28.860 to make the best choice. 30:28.860 --> 30:31.860 And if this information is not available, 30:31.860 --> 30:35.100 you would feel vulnerable and unprotected. 30:35.100 --> 30:37.860 And then the question is, you know, what do I care more? 30:37.860 --> 30:40.380 Because at the end, everything is a trade off, correct? 30:40.380 --> 30:41.700 Yeah, exactly. 30:41.700 --> 30:45.580 Just out of curiosity, it seems like one possible solution, 30:45.580 --> 30:47.420 I'd like to see what you think of it, 30:49.340 --> 30:50.660 based on what you just said, 30:50.660 --> 30:52.500 based on wanting to know answers 30:52.500 --> 30:55.060 for when you're yourself in that situation. 30:55.060 --> 30:58.420 Is it possible for patients to own their data 30:58.420 --> 31:01.020 as opposed to hospitals owning their data? 31:01.020 --> 31:04.100 Of course, theoretically, I guess patients own their data, 31:04.100 --> 31:06.620 but can you walk out there with a USB stick 31:07.580 --> 31:10.620 containing everything or upload it to the cloud? 31:10.620 --> 31:14.500 Where a company, you know, I remember Microsoft 31:14.500 --> 31:17.820 had a service, like I try, I was really excited about 31:17.820 --> 31:19.260 and Google Health was there. 31:19.260 --> 31:21.900 I tried to give, I was excited about it. 31:21.900 --> 31:24.780 Basically companies helping you upload your data 31:24.780 --> 31:27.940 to the cloud so that you can move from hospital to hospital 31:27.940 --> 31:29.260 from doctor to doctor. 31:29.260 --> 31:32.700 Do you see a promise of that kind of possibility? 31:32.700 --> 31:34.660 I absolutely think this is, you know, 31:34.660 --> 31:38.180 the right way to exchange the data. 31:38.180 --> 31:41.700 I don't know now who's the biggest player in this field, 31:41.700 --> 31:45.940 but I can clearly see that even for totally selfish 31:45.940 --> 31:49.300 health reasons, when you are going to a new facility 31:49.300 --> 31:52.620 and many of us are sent to some specialized treatment, 31:52.620 --> 31:55.740 they don't easily have access to your data. 31:55.740 --> 31:59.420 And today, you know, we might want to send this mammogram, 31:59.420 --> 32:01.780 need to go to the hospital, find some small office 32:01.780 --> 32:04.820 which gives them the CD and they ship as a CD. 32:04.820 --> 32:08.340 So you can imagine we're looking at kind of decades old 32:08.340 --> 32:10.100 mechanism of data exchange. 32:11.340 --> 32:15.620 So I definitely think this is an area where hopefully 32:15.620 --> 32:20.380 all the right regulatory and technical forces will align 32:20.380 --> 32:23.220 and we will see it actually implemented. 32:23.220 --> 32:27.500 It's sad because unfortunately, and I need to research 32:27.500 --> 32:30.620 why that happened, but I'm pretty sure Google Health 32:30.620 --> 32:32.940 and Microsoft Health Vault or whatever it's called 32:32.940 --> 32:36.100 both closed down, which means that there was 32:36.100 --> 32:39.100 either regulatory pressure or there's not a business case 32:39.100 --> 32:41.820 or there's challenges from hospitals, 32:41.820 --> 32:43.260 which is very disappointing. 32:43.260 --> 32:46.500 So when you say you don't know what the biggest players are, 32:46.500 --> 32:50.540 the two biggest that I was aware of closed their doors. 32:50.540 --> 32:53.140 So I'm hoping, I'd love to see why 32:53.140 --> 32:54.780 and I'd love to see who else can come up. 32:54.780 --> 32:59.620 It seems like one of those Elon Musk style problems 32:59.620 --> 33:01.300 that are obvious needs to be solved 33:01.300 --> 33:02.980 and somebody needs to step up and actually do 33:02.980 --> 33:07.540 this large scale data collection. 33:07.540 --> 33:09.620 So I know there is an initiative in Massachusetts, 33:09.620 --> 33:11.740 I think, which you led by the governor 33:11.740 --> 33:15.460 to try to create this kind of health exchange system 33:15.460 --> 33:17.860 where at least to help people who kind of when you show up 33:17.860 --> 33:20.220 in emergency room and there is no information 33:20.220 --> 33:23.540 about what are your allergies and other things. 33:23.540 --> 33:26.140 So I don't know how far it will go. 33:26.140 --> 33:28.180 But another thing that you said 33:28.180 --> 33:30.780 and I find it very interesting is actually 33:30.780 --> 33:33.780 who are the successful players in this space 33:33.780 --> 33:37.260 and the whole implementation, how does it go? 33:37.260 --> 33:40.300 To me, it is from the anthropological perspective, 33:40.300 --> 33:44.660 it's more fascinating that AI that today goes in healthcare, 33:44.660 --> 33:50.380 we've seen so many attempts and so very little successes. 33:50.380 --> 33:54.220 And it's interesting to understand that I've by no means 33:54.220 --> 33:56.700 have knowledge to assess it, 33:56.700 --> 33:59.620 why we are in the position where we are. 33:59.620 --> 34:02.940 Yeah, it's interesting because data is really fuel 34:02.940 --> 34:04.980 for a lot of successful applications. 34:04.980 --> 34:08.500 And when that data acquires regulatory approval, 34:08.500 --> 34:12.940 like the FDA or any kind of approval, 34:12.940 --> 34:15.740 it seems that the computer scientists 34:15.740 --> 34:17.460 are not quite there yet in being able 34:17.460 --> 34:18.900 to play the regulatory game, 34:18.900 --> 34:21.220 understanding the fundamentals of it. 34:21.220 --> 34:26.500 I think that in many cases when even people do have data, 34:26.500 --> 34:31.300 we still don't know what exactly do you need to demonstrate 34:31.300 --> 34:33.860 to change the standard of care. 34:35.500 --> 34:37.180 Like let me give you an example 34:37.180 --> 34:41.100 related to my breast cancer research. 34:41.100 --> 34:45.500 So in traditional breast cancer risk assessment, 34:45.500 --> 34:47.140 there is something called density, 34:47.140 --> 34:50.500 which determines the likelihood of a woman to get cancer. 34:50.500 --> 34:51.700 And this pretty much says, 34:51.700 --> 34:54.220 how much white do you see on the mammogram? 34:54.220 --> 34:58.980 The whiter it is, the more likely the tissue is dense. 34:58.980 --> 35:03.660 And the idea behind density, it's not a bad idea. 35:03.660 --> 35:08.100 In 1967, a radiologist called Wolf decided to look back 35:08.100 --> 35:09.780 at women who were diagnosed 35:09.780 --> 35:12.420 and see what is special in their images. 35:12.420 --> 35:14.700 Can we look back and say that they're likely to develop? 35:14.700 --> 35:16.180 So he come up with some patterns. 35:16.180 --> 35:20.660 And it was the best that his human eye can identify. 35:20.660 --> 35:22.060 Then it was kind of formalized 35:22.060 --> 35:24.220 and coded into four categories. 35:24.220 --> 35:26.940 And that's what we are using today. 35:26.940 --> 35:31.020 And today this density assessment 35:31.020 --> 35:34.620 is actually a federal law from 2019, 35:34.620 --> 35:36.180 approved by President Trump 35:36.180 --> 35:40.100 and for the previous FDA commissioner, 35:40.100 --> 35:43.620 where women are supposed to be advised by their providers 35:43.620 --> 35:45.100 if they have high density, 35:45.100 --> 35:47.260 putting them into higher risk category. 35:47.260 --> 35:49.460 And in some states, 35:49.460 --> 35:51.260 you can actually get supplementary screening 35:51.260 --> 35:53.700 paid by your insurance because you're in this category. 35:53.700 --> 35:56.780 Now you can say, how much science do we have behind it? 35:56.780 --> 36:00.820 Whatever, biological science or epidemiological evidence. 36:00.820 --> 36:05.140 So it turns out that between 40 and 50% of women 36:05.140 --> 36:06.660 have dense breasts. 36:06.660 --> 36:11.140 So about 40% of patients are coming out of their screening 36:11.140 --> 36:15.020 and somebody tells them, you are in high risk. 36:15.020 --> 36:16.860 Now, what exactly does it mean 36:16.860 --> 36:19.620 if you as half of the population in high risk? 36:19.620 --> 36:22.060 It's from saying, maybe I'm not, 36:22.060 --> 36:23.700 or what do I really need to do with it? 36:23.700 --> 36:27.220 Because the system doesn't provide me 36:27.220 --> 36:28.340 a lot of the solutions 36:28.340 --> 36:30.140 because there are so many people like me, 36:30.140 --> 36:34.620 we cannot really provide very expensive solutions for them. 36:34.620 --> 36:38.740 And the reason this whole density became this big deal, 36:38.740 --> 36:40.820 it's actually advocated by the patients 36:40.820 --> 36:42.500 who felt very unprotected 36:42.500 --> 36:44.900 because many women went and did the mammograms 36:44.900 --> 36:46.260 which were normal. 36:46.260 --> 36:49.460 And then it turns out that they already had cancer, 36:49.460 --> 36:50.580 quite developed cancer. 36:50.580 --> 36:54.420 So they didn't have a way to know who is really at risk 36:54.420 --> 36:56.300 and what is the likelihood that when the doctor tells you, 36:56.300 --> 36:58.060 you're okay, you are not okay. 36:58.060 --> 37:02.140 So at the time, and it was 15 years ago, 37:02.140 --> 37:06.820 this maybe was the best piece of science that we had. 37:06.820 --> 37:11.820 And it took quite 15, 16 years to make it federal law. 37:12.180 --> 37:15.660 But now this is a standard. 37:15.660 --> 37:17.620 Now with a deep learning model, 37:17.620 --> 37:19.660 we can so much more accurately predict 37:19.660 --> 37:21.580 who is gonna develop breast cancer 37:21.580 --> 37:23.700 just because you're trained on a logical thing. 37:23.700 --> 37:26.060 And instead of describing how much white 37:26.060 --> 37:27.380 and what kind of white machine 37:27.380 --> 37:30.140 can systematically identify the patterns, 37:30.140 --> 37:32.780 which was the original idea behind the thought 37:32.780 --> 37:33.700 of the cardiologist, 37:33.700 --> 37:35.740 machines can do it much more systematically 37:35.740 --> 37:38.260 and predict the risk when you're training the machine 37:38.260 --> 37:42.140 to look at the image and to say the risk in one to five years. 37:42.140 --> 37:45.060 Now you can ask me how long it will take 37:45.060 --> 37:46.460 to substitute this density, 37:46.460 --> 37:48.620 which is broadly used across the country 37:48.620 --> 37:53.620 and really is not helping to bring this new models. 37:54.380 --> 37:56.700 And I would say it's not a matter of the algorithm. 37:56.700 --> 37:58.780 Algorithms use already orders of magnitude better 37:58.780 --> 38:00.460 than what is currently in practice. 38:00.460 --> 38:02.500 I think it's really the question, 38:02.500 --> 38:04.380 who do you need to convince? 38:04.380 --> 38:07.460 How many hospitals do you need to run the experiment? 38:07.460 --> 38:11.500 What, you know, all this mechanism of adoption 38:11.500 --> 38:15.180 and how do you explain to patients 38:15.180 --> 38:17.580 and to women across the country 38:17.580 --> 38:20.460 that this is really a better measure? 38:20.460 --> 38:22.740 And again, I don't think it's an AI question. 38:22.740 --> 38:25.940 We can work more and make the algorithm even better, 38:25.940 --> 38:29.300 but I don't think that this is the current, you know, 38:29.300 --> 38:32.060 the barrier, the barrier is really this other piece 38:32.060 --> 38:35.260 that for some reason is not really explored. 38:35.260 --> 38:36.860 It's like anthropological piece. 38:36.860 --> 38:39.860 And coming back to your question about books, 38:39.860 --> 38:42.980 there is a book that I'm reading. 38:42.980 --> 38:47.980 It's called American Sickness by Elizabeth Rosenthal. 38:48.260 --> 38:51.580 And I got this book from my clinical collaborator, 38:51.580 --> 38:53.100 Dr. Connie Lehman. 38:53.100 --> 38:54.820 And I said, I know everything that I need to know 38:54.820 --> 38:56.020 about American health system, 38:56.020 --> 38:59.220 but you know, every page doesn't fail to surprise me. 38:59.220 --> 39:03.140 And I think there is a lot of interesting 39:03.140 --> 39:06.860 and really deep lessons for people like us 39:06.860 --> 39:09.660 from computer science who are coming into this field 39:09.660 --> 39:13.660 to really understand how complex is the system of incentives 39:13.660 --> 39:17.660 in the system to understand how you really need to play 39:17.660 --> 39:18.780 to drive adoption. 39:19.740 --> 39:21.180 You just said it's complex, 39:21.180 --> 39:23.980 but if we're trying to simplify it, 39:23.980 --> 39:27.380 who do you think most likely would be successful 39:27.380 --> 39:29.540 if we push on this group of people? 39:29.540 --> 39:30.780 Is it the doctors? 39:30.780 --> 39:31.820 Is it the hospitals? 39:31.820 --> 39:34.300 Is it the governments or policymakers? 39:34.300 --> 39:37.380 Is it the individual patients, consumers? 39:38.860 --> 39:43.860 Who needs to be inspired to most likely lead to adoption? 39:45.180 --> 39:47.100 Or is there no simple answer? 39:47.100 --> 39:48.260 There's no simple answer, 39:48.260 --> 39:51.980 but I think there is a lot of good people in medical system 39:51.980 --> 39:55.180 who do want to make a change. 39:56.460 --> 40:01.460 And I think a lot of power will come from us as consumers 40:01.540 --> 40:04.260 because we all are consumers or future consumers 40:04.260 --> 40:06.500 of healthcare services. 40:06.500 --> 40:11.500 And I think we can do so much more 40:12.060 --> 40:15.500 in explaining the potential and not in the hype terms 40:15.500 --> 40:17.900 and not saying that we now killed all Alzheimer 40:17.900 --> 40:20.500 and I'm really sick of reading this kind of articles 40:20.500 --> 40:22.100 which make these claims, 40:22.100 --> 40:24.780 but really to show with some examples 40:24.780 --> 40:29.060 what this implementation does and how it changes the care. 40:29.060 --> 40:30.020 Because I can't imagine, 40:30.020 --> 40:33.220 it doesn't matter what kind of politician it is, 40:33.220 --> 40:35.220 we all are susceptible to these diseases. 40:35.220 --> 40:37.740 There is no one who is free. 40:37.740 --> 40:41.060 And eventually, we all are humans 40:41.060 --> 40:44.860 and we're looking for a way to alleviate the suffering. 40:44.860 --> 40:47.260 And this is one possible way 40:47.260 --> 40:49.300 where we currently are under utilizing, 40:49.300 --> 40:50.940 which I think can help. 40:51.860 --> 40:55.100 So it sounds like the biggest problems are outside of AI 40:55.100 --> 40:57.980 in terms of the biggest impact at this point. 40:57.980 --> 41:00.420 But are there any open problems 41:00.420 --> 41:03.780 in the application of ML to oncology in general? 41:03.780 --> 41:07.540 So improving the detection or any other creative methods, 41:07.540 --> 41:09.620 whether it's on the detection segmentations 41:09.620 --> 41:11.780 or the vision perception side 41:11.780 --> 41:16.260 or some other clever of inference? 41:16.260 --> 41:19.620 Yeah, what in general in your view are the open problems 41:19.620 --> 41:20.460 in this space? 41:20.460 --> 41:22.460 Yeah, I just want to mention that beside detection, 41:22.460 --> 41:24.820 not the area where I am kind of quite active 41:24.820 --> 41:28.580 and I think it's really an increasingly important area 41:28.580 --> 41:30.940 in healthcare is drug design. 41:32.260 --> 41:33.100 Absolutely. 41:33.100 --> 41:36.900 Because it's fine if you detect something early, 41:36.900 --> 41:41.100 but you still need to get drugs 41:41.100 --> 41:43.860 and new drugs for these conditions. 41:43.860 --> 41:46.740 And today, all of the drug design, 41:46.740 --> 41:48.300 ML is non existent there. 41:48.300 --> 41:52.980 We don't have any drug that was developed by the ML model 41:52.980 --> 41:54.900 or even not developed, 41:54.900 --> 41:57.060 but at least even knew that ML model 41:57.060 --> 41:59.260 plays some significant role. 41:59.260 --> 42:03.300 I think this area with all the new ability 42:03.300 --> 42:05.780 to generate molecules with desired properties 42:05.780 --> 42:10.780 to do in silica screening is really a big open area. 42:11.460 --> 42:12.740 To be totally honest with you, 42:12.740 --> 42:14.900 when we are doing diagnostics and imaging, 42:14.900 --> 42:17.260 primarily taking the ideas that were developed 42:17.260 --> 42:20.460 for other areas and you applying them with some adaptation, 42:20.460 --> 42:25.460 the area of drug design is really technically interesting 42:26.820 --> 42:27.980 and exciting area. 42:27.980 --> 42:30.380 You need to work a lot with graphs 42:30.380 --> 42:34.580 and capture various 3D properties. 42:34.580 --> 42:37.420 There are lots and lots of opportunities 42:37.420 --> 42:39.820 to be technically creative. 42:39.820 --> 42:44.820 And I think there are a lot of open questions in this area. 42:46.820 --> 42:48.820 We're already getting a lot of successes 42:48.820 --> 42:52.700 even with kind of the first generation of these models, 42:52.700 --> 42:56.500 but there is much more new creative things that you can do. 42:56.500 --> 42:59.260 And what's very nice to see is that actually 42:59.260 --> 43:04.180 the more powerful, the more interesting models 43:04.180 --> 43:05.460 actually do do better. 43:05.460 --> 43:10.460 So there is a place to innovate in machine learning 43:11.300 --> 43:12.540 in this area. 43:13.900 --> 43:16.820 And some of these techniques are really unique to, 43:16.820 --> 43:19.620 let's say, to graph generation and other things. 43:19.620 --> 43:20.820 So... 43:20.820 --> 43:23.980 What, just to interrupt really quick, I'm sorry, 43:23.980 --> 43:28.980 graph generation or graphs, drug discovery in general, 43:30.620 --> 43:31.940 how do you discover a drug? 43:31.940 --> 43:33.340 Is this chemistry? 43:33.340 --> 43:37.500 Is this trying to predict different chemical reactions? 43:37.500 --> 43:39.660 Or is it some kind of... 43:39.660 --> 43:42.100 What do graphs even represent in this space? 43:42.100 --> 43:43.980 Oh, sorry, sorry. 43:43.980 --> 43:45.340 And what's a drug? 43:45.340 --> 43:47.140 Okay, so let's say you're thinking 43:47.140 --> 43:48.540 there are many different types of drugs, 43:48.540 --> 43:50.580 but let's say you're gonna talk about small molecules 43:50.580 --> 43:52.860 because I think today the majority of drugs 43:52.860 --> 43:53.700 are small molecules. 43:53.700 --> 43:55.020 So small molecule is a graph. 43:55.020 --> 43:59.180 The molecule is just where the node in the graph 43:59.180 --> 44:01.500 is an atom and then you have the bonds. 44:01.500 --> 44:03.220 So it's really a graph representation. 44:03.220 --> 44:05.540 If you look at it in 2D, correct, 44:05.540 --> 44:07.460 you can do it 3D, but let's say, 44:07.460 --> 44:09.540 let's keep it simple and stick in 2D. 44:11.500 --> 44:14.740 So pretty much my understanding today, 44:14.740 --> 44:18.620 how it is done at scale in the companies, 44:18.620 --> 44:20.220 without machine learning, 44:20.220 --> 44:22.100 you have high throughput screening. 44:22.100 --> 44:23.740 So you know that you are interested 44:23.740 --> 44:26.540 to get certain biological activity of the compound. 44:26.540 --> 44:28.860 So you scan a lot of compounds, 44:28.860 --> 44:30.700 like maybe hundreds of thousands, 44:30.700 --> 44:32.980 some really big number of compounds. 44:32.980 --> 44:36.060 You identify some compounds which have the right activity 44:36.060 --> 44:39.220 and then at this point, the chemists come 44:39.220 --> 44:43.220 and they're trying to now to optimize 44:43.220 --> 44:45.340 this original heat to different properties 44:45.340 --> 44:47.180 that you want it to be maybe soluble, 44:47.180 --> 44:49.060 you want it to decrease toxicity, 44:49.060 --> 44:51.620 you want it to decrease the side effects. 44:51.620 --> 44:54.020 Are those, sorry again to interrupt, 44:54.020 --> 44:55.500 can that be done in simulation 44:55.500 --> 44:57.700 or just by looking at the molecules 44:57.700 --> 44:59.820 or do you need to actually run reactions 44:59.820 --> 45:02.460 in real labs with lab coats and stuff? 45:02.460 --> 45:04.020 So when you do high throughput screening, 45:04.020 --> 45:06.100 you really do screening. 45:06.100 --> 45:07.020 It's in the lab. 45:07.020 --> 45:09.140 It's really the lab screening. 45:09.140 --> 45:10.980 You screen the molecules, correct? 45:10.980 --> 45:12.580 I don't know what screening is. 45:12.580 --> 45:15.060 The screening is just check them for certain property. 45:15.060 --> 45:17.260 Like in the physical space, in the physical world, 45:17.260 --> 45:18.740 like actually there's a machine probably 45:18.740 --> 45:21.420 that's actually running the reaction. 45:21.420 --> 45:22.900 Actually running the reactions, yeah. 45:22.900 --> 45:25.420 So there is a process where you can run 45:25.420 --> 45:26.660 and that's why it's called high throughput 45:26.660 --> 45:29.580 that it become cheaper and faster 45:29.580 --> 45:33.820 to do it on very big number of molecules. 45:33.820 --> 45:35.820 You run the screening, 45:35.820 --> 45:40.300 you identify potential good starts 45:40.300 --> 45:42.340 and then when the chemists come in 45:42.340 --> 45:44.060 who have done it many times 45:44.060 --> 45:46.180 and then they can try to look at it and say, 45:46.180 --> 45:48.260 how can you change the molecule 45:48.260 --> 45:51.780 to get the desired profile 45:51.780 --> 45:53.460 in terms of all other properties? 45:53.460 --> 45:56.500 So maybe how do I make it more bioactive and so on? 45:56.500 --> 45:59.460 And there the creativity of the chemists 45:59.460 --> 46:03.980 really is the one that determines the success 46:03.980 --> 46:07.460 of this design because again, 46:07.460 --> 46:09.300 they have a lot of domain knowledge 46:09.300 --> 46:12.900 of what works, how do you decrease the CCD and so on 46:12.900 --> 46:15.020 and that's what they do. 46:15.020 --> 46:17.860 So all the drugs that are currently 46:17.860 --> 46:20.220 in the FDA approved drugs 46:20.220 --> 46:22.140 or even drugs that are in clinical trials, 46:22.140 --> 46:27.100 they are designed using these domain experts 46:27.100 --> 46:30.060 which goes through this combinatorial space 46:30.060 --> 46:31.940 of molecules or graphs or whatever 46:31.940 --> 46:35.140 and find the right one or adjust it to be the right ones. 46:35.140 --> 46:38.060 It sounds like the breast density heuristic 46:38.060 --> 46:40.460 from 67 to the same echoes. 46:40.460 --> 46:41.820 It's not necessarily that. 46:41.820 --> 46:45.380 It's really driven by deep understanding. 46:45.380 --> 46:46.820 It's not like they just observe it. 46:46.820 --> 46:48.540 I mean, they do deeply understand chemistry 46:48.540 --> 46:50.460 and they do understand how different groups 46:50.460 --> 46:53.140 and how does it changes the properties. 46:53.140 --> 46:56.660 So there is a lot of science that gets into it 46:56.660 --> 46:58.740 and a lot of kind of simulation, 46:58.740 --> 47:00.940 how do you want it to behave? 47:01.900 --> 47:03.900 It's very, very complex. 47:03.900 --> 47:06.140 So they're quite effective at this design, obviously. 47:06.140 --> 47:08.420 Now effective, yeah, we have drugs. 47:08.420 --> 47:10.780 Like depending on how do you measure effective, 47:10.780 --> 47:13.940 if you measure it in terms of cost, it's prohibitive. 47:13.940 --> 47:15.820 If you measure it in terms of times, 47:15.820 --> 47:18.420 we have lots of diseases for which we don't have any drugs 47:18.420 --> 47:20.060 and we don't even know how to approach 47:20.060 --> 47:23.460 and don't need to mention few drugs 47:23.460 --> 47:27.140 or neurodegenerative disease drugs that fail. 47:27.140 --> 47:32.140 So there are lots of trials that fail in later stages, 47:32.180 --> 47:35.180 which is really catastrophic from the financial perspective. 47:35.180 --> 47:39.540 So is it the effective, the most effective mechanism? 47:39.540 --> 47:42.700 Absolutely no, but this is the only one that currently works. 47:44.300 --> 47:47.900 And I was closely interacting 47:47.900 --> 47:49.260 with people in pharmaceutical industry. 47:49.260 --> 47:51.340 I was really fascinated on how sharp 47:51.340 --> 47:55.260 and what a deep understanding of the domain do they have. 47:55.260 --> 47:57.020 It's not observation driven. 47:57.020 --> 48:00.220 There is really a lot of science behind what they do. 48:00.220 --> 48:02.300 But if you ask me, can machine learning change it, 48:02.300 --> 48:05.300 I firmly believe yes, 48:05.300 --> 48:07.860 because even the most experienced chemists 48:07.860 --> 48:11.100 cannot hold in their memory and understanding 48:11.100 --> 48:12.500 everything that you can learn 48:12.500 --> 48:15.420 from millions of molecules and reactions. 48:17.220 --> 48:19.900 And the space of graphs is a totally new space. 48:19.900 --> 48:22.060 I mean, it's a really interesting space 48:22.060 --> 48:23.980 for machine learning to explore, graph generation. 48:23.980 --> 48:26.260 Yeah, so there are a lot of things that you can do here. 48:26.260 --> 48:28.740 So we do a lot of work. 48:28.740 --> 48:31.620 So the first tool that we started with 48:31.620 --> 48:36.300 was the tool that can predict properties of the molecules. 48:36.300 --> 48:39.420 So you can just give the molecule and the property. 48:39.420 --> 48:41.340 It can be by activity property, 48:41.340 --> 48:44.300 or it can be some other property. 48:44.300 --> 48:46.460 And you train the molecules 48:46.460 --> 48:50.020 and you can now take a new molecule 48:50.020 --> 48:52.180 and predict this property. 48:52.180 --> 48:54.860 Now, when people started working in this area, 48:54.860 --> 48:55.980 it is something very simple. 48:55.980 --> 48:58.580 They do kind of existing fingerprints, 48:58.580 --> 49:00.740 which is kind of handcrafted features of the molecule. 49:00.740 --> 49:02.980 When you break the graph to substructures 49:02.980 --> 49:05.980 and then you run it in a feed forward neural network. 49:05.980 --> 49:08.500 And what was interesting to see that clearly, 49:08.500 --> 49:11.020 this was not the most effective way to proceed. 49:11.020 --> 49:14.060 And you need to have much more complex models 49:14.060 --> 49:16.300 that can induce a representation, 49:16.300 --> 49:19.220 which can translate this graph into the embeddings 49:19.220 --> 49:21.300 and do these predictions. 49:21.300 --> 49:23.220 So this is one direction. 49:23.220 --> 49:25.260 Then another direction, which is kind of related 49:25.260 --> 49:29.180 is not only to stop by looking at the embedding itself, 49:29.180 --> 49:32.780 but actually modify it to produce better molecules. 49:32.780 --> 49:36.020 So you can think about it as machine translation 49:36.020 --> 49:38.140 that you can start with a molecule 49:38.140 --> 49:40.580 and then there is an improved version of molecule. 49:40.580 --> 49:42.860 And you can again, with encoder translate it 49:42.860 --> 49:45.380 into the hidden space and then learn how to modify it 49:45.380 --> 49:49.340 to improve the in some ways version of the molecules. 49:49.340 --> 49:52.620 So that's, it's kind of really exciting. 49:52.620 --> 49:54.740 We already have seen that the property prediction 49:54.740 --> 49:56.140 works pretty well. 49:56.140 --> 49:59.780 And now we are generating molecules 49:59.780 --> 50:01.820 and there is actually labs 50:01.820 --> 50:04.180 which are manufacturing this molecule. 50:04.180 --> 50:06.340 So we'll see where it will get us. 50:06.340 --> 50:07.780 Okay, that's really exciting. 50:07.780 --> 50:08.860 There's a lot of promise. 50:08.860 --> 50:11.820 Speaking of machine translation and embeddings, 50:11.820 --> 50:15.580 I think you have done a lot of really great research 50:15.580 --> 50:17.540 in NLP, natural language processing. 50:19.260 --> 50:21.540 Can you tell me your journey through NLP? 50:21.540 --> 50:25.100 What ideas, problems, approaches were you working on? 50:25.100 --> 50:28.180 Were you fascinated with, did you explore 50:28.180 --> 50:33.180 before this magic of deep learning reemerged and after? 50:34.020 --> 50:37.180 So when I started my work in NLP, it was in 97. 50:38.180 --> 50:39.460 This was very interesting time. 50:39.460 --> 50:42.620 It was exactly the time that I came to ACL. 50:43.500 --> 50:46.140 And at the time I could barely understand English, 50:46.140 --> 50:48.500 but it was exactly like the transition point 50:48.500 --> 50:53.500 because half of the papers were really rule based approaches 50:53.500 --> 50:56.180 where people took more kind of heavy linguistic approaches 50:56.180 --> 51:00.060 for small domains and try to build up from there. 51:00.060 --> 51:02.220 And then there were the first generation of papers 51:02.220 --> 51:04.500 which were corpus based papers. 51:04.500 --> 51:06.420 And they were very simple in our terms 51:06.420 --> 51:07.900 when you collect some statistics 51:07.900 --> 51:10.020 and do prediction based on them. 51:10.020 --> 51:13.100 And I found it really fascinating that one community 51:13.100 --> 51:18.100 can think so very differently about the problem. 51:19.220 --> 51:22.820 And I remember my first paper that I wrote, 51:22.820 --> 51:24.460 it didn't have a single formula. 51:24.460 --> 51:25.740 It didn't have evaluation. 51:25.740 --> 51:28.340 It just had examples of outputs. 51:28.340 --> 51:32.020 And this was a standard of the field at the time. 51:32.020 --> 51:35.860 In some ways, I mean, people maybe just started emphasizing 51:35.860 --> 51:38.940 the empirical evaluation, but for many applications 51:38.940 --> 51:42.780 like summarization, you just show some examples of outputs. 51:42.780 --> 51:45.460 And then increasingly you can see that how 51:45.460 --> 51:48.300 the statistical approaches dominated the field 51:48.300 --> 51:52.100 and we've seen increased performance 51:52.100 --> 51:56.020 across many basic tasks. 51:56.020 --> 52:00.420 The sad part of the story maybe that if you look again 52:00.420 --> 52:05.100 through this journey, we see that the role of linguistics 52:05.100 --> 52:07.460 in some ways greatly diminishes. 52:07.460 --> 52:11.580 And I think that you really need to look 52:11.580 --> 52:14.540 through the whole proceeding to find one or two papers 52:14.540 --> 52:17.260 which make some interesting linguistic references. 52:17.260 --> 52:18.100 It's really big. 52:18.100 --> 52:18.920 Today, yeah. 52:18.920 --> 52:19.760 Today, today. 52:19.760 --> 52:20.600 This was definitely one of the. 52:20.600 --> 52:23.140 Things like syntactic trees, just even basically 52:23.140 --> 52:26.180 against our conversation about human understanding 52:26.180 --> 52:30.300 of language, which I guess what linguistics would be 52:30.300 --> 52:34.300 structured, hierarchical representing language 52:34.300 --> 52:37.140 in a way that's human explainable, understandable 52:37.140 --> 52:39.500 is missing today. 52:39.500 --> 52:42.380 I don't know if it is, what is explainable 52:42.380 --> 52:43.620 and understandable. 52:43.620 --> 52:47.360 In the end, we perform functions and it's okay 52:47.360 --> 52:50.140 to have machine which performs a function. 52:50.140 --> 52:53.200 Like when you're thinking about your calculator, correct? 52:53.200 --> 52:56.100 Your calculator can do calculation very different 52:56.100 --> 52:57.620 from you would do the calculation, 52:57.620 --> 52:58.860 but it's very effective in it. 52:58.860 --> 53:02.560 And this is fine if we can achieve certain tasks 53:02.560 --> 53:05.760 with high accuracy, doesn't necessarily mean 53:05.760 --> 53:09.300 that it has to understand it the same way as we understand. 53:09.300 --> 53:11.260 In some ways, it's even naive to request 53:11.260 --> 53:14.940 because you have so many other sources of information 53:14.940 --> 53:17.900 that are absent when you are training your system. 53:17.900 --> 53:19.220 So it's okay. 53:19.220 --> 53:20.060 Is it delivered? 53:20.060 --> 53:21.500 And I would tell you one application 53:21.500 --> 53:22.780 that is really fascinating. 53:22.780 --> 53:25.060 In 97, when it came to ACL, there were some papers 53:25.060 --> 53:25.900 on machine translation. 53:25.900 --> 53:27.440 They were like primitive. 53:27.440 --> 53:31.060 Like people were trying really, really simple. 53:31.060 --> 53:34.260 And the feeling, my feeling was that, you know, 53:34.260 --> 53:36.260 to make real machine translation system, 53:36.260 --> 53:39.580 it's like to fly at the moon and build a house there 53:39.580 --> 53:41.580 and the garden and live happily ever after. 53:41.580 --> 53:42.600 I mean, it's like impossible. 53:42.600 --> 53:46.740 I never could imagine that within, you know, 10 years, 53:46.740 --> 53:48.540 we would already see the system working. 53:48.540 --> 53:51.420 And now, you know, nobody is even surprised 53:51.420 --> 53:54.420 to utilize the system on daily basis. 53:54.420 --> 53:56.220 So this was like a huge, huge progress, 53:56.220 --> 53:57.860 saying that people for very long time 53:57.860 --> 54:00.820 tried to solve using other mechanisms. 54:00.820 --> 54:03.220 And they were unable to solve it. 54:03.220 --> 54:06.140 That's why coming back to your question about biology, 54:06.140 --> 54:10.800 that, you know, in linguistics, people try to go this way 54:10.800 --> 54:13.500 and try to write the syntactic trees 54:13.500 --> 54:17.500 and try to abstract it and to find the right representation. 54:17.500 --> 54:22.240 And, you know, they couldn't get very far 54:22.240 --> 54:26.580 with this understanding while these models using, 54:26.580 --> 54:29.640 you know, other sources actually capable 54:29.640 --> 54:31.680 to make a lot of progress. 54:31.680 --> 54:33.960 Now, I'm not naive to think 54:33.960 --> 54:36.780 that we are in this paradise space in NLP. 54:36.780 --> 54:38.580 And sure as you know, 54:38.580 --> 54:40.860 that when we slightly change the domain 54:40.860 --> 54:42.620 and when we decrease the amount of training, 54:42.620 --> 54:44.740 it can do like really bizarre and funny thing. 54:44.740 --> 54:46.500 But I think it's just a matter 54:46.500 --> 54:48.540 of improving generalization capacity, 54:48.540 --> 54:51.500 which is just a technical question. 54:51.500 --> 54:54.340 Wow, so that's the question. 54:54.340 --> 54:57.720 How much of language understanding can be solved 54:57.720 --> 54:59.180 with deep neural networks? 54:59.180 --> 55:03.740 In your intuition, I mean, it's unknown, I suppose. 55:03.740 --> 55:07.660 But as we start to creep towards romantic notions 55:07.660 --> 55:10.620 of the spirit of the Turing test 55:10.620 --> 55:14.220 and conversation and dialogue 55:14.220 --> 55:18.340 and something that maybe to me or to us, 55:18.340 --> 55:21.620 so the humans feels like it needs real understanding. 55:21.620 --> 55:23.500 How much can that be achieved 55:23.500 --> 55:27.180 with these neural networks or statistical methods? 55:27.180 --> 55:32.180 So I guess I am very much driven by the outcomes. 55:33.340 --> 55:35.420 Can we achieve the performance 55:35.420 --> 55:40.420 which would be satisfactory for us for different tasks? 55:40.700 --> 55:43.020 Now, if you again look at machine translation system, 55:43.020 --> 55:46.020 which are trained on large amounts of data, 55:46.020 --> 55:48.780 they really can do a remarkable job 55:48.780 --> 55:51.300 relatively to where they've been a few years ago. 55:51.300 --> 55:54.620 And if you project into the future, 55:54.620 --> 55:59.380 if it will be the same speed of improvement, you know, 55:59.380 --> 56:00.220 this is great. 56:00.220 --> 56:01.060 Now, does it bother me 56:01.060 --> 56:04.860 that it's not doing the same translation as we are doing? 56:04.860 --> 56:06.620 Now, if you go to cognitive science, 56:06.620 --> 56:09.460 we still don't really understand what we are doing. 56:10.460 --> 56:11.860 I mean, there are a lot of theories 56:11.860 --> 56:13.840 and there's obviously a lot of progress and studying, 56:13.840 --> 56:17.540 but our understanding what exactly goes on in our brains 56:17.540 --> 56:21.020 when we process language is still not crystal clear 56:21.020 --> 56:25.460 and precise that we can translate it into machines. 56:25.460 --> 56:29.220 What does bother me is that, you know, 56:29.220 --> 56:31.700 again, that machines can be extremely brittle 56:31.700 --> 56:33.980 when you go out of your comfort zone 56:33.980 --> 56:36.060 of when there is a distributional shift 56:36.060 --> 56:37.300 between training and testing. 56:37.300 --> 56:39.020 And it have been years and years, 56:39.020 --> 56:41.320 every year when I teach an LP class, 56:41.320 --> 56:43.560 now show them some examples of translation 56:43.560 --> 56:47.300 from some newspaper in Hebrew or whatever, it was perfect. 56:47.300 --> 56:51.300 And then I have a recipe that Tomi Yakel's system 56:51.300 --> 56:53.900 sent me a while ago and it was written in Finnish 56:53.900 --> 56:55.720 of Karelian pies. 56:55.720 --> 56:59.280 And it's just a terrible translation. 56:59.280 --> 57:01.460 You cannot understand anything what it does. 57:01.460 --> 57:04.180 It's not like some syntactic mistakes, it's just terrible. 57:04.180 --> 57:07.020 And year after year, I tried and will translate 57:07.020 --> 57:08.980 and year after year, it does this terrible work 57:08.980 --> 57:10.980 because I guess, you know, the recipes 57:10.980 --> 57:14.580 are not a big part of their training repertoire. 57:14.580 --> 57:19.020 So, but in terms of outcomes, that's a really clean, 57:19.020 --> 57:20.240 good way to look at it. 57:21.100 --> 57:23.140 I guess the question I was asking is, 57:24.060 --> 57:27.700 do you think, imagine a future, 57:27.700 --> 57:30.540 do you think the current approaches can pass 57:30.540 --> 57:32.460 the Turing test in the way, 57:34.700 --> 57:37.060 in the best possible formulation of the Turing test? 57:37.060 --> 57:39.460 Which is, would you wanna have a conversation 57:39.460 --> 57:42.340 with a neural network for an hour? 57:42.340 --> 57:45.820 Oh God, no, no, there are not that many people 57:45.820 --> 57:48.380 that I would want to talk for an hour, but. 57:48.380 --> 57:51.500 There are some people in this world, alive or not, 57:51.500 --> 57:53.260 that you would like to talk to for an hour. 57:53.260 --> 57:56.700 Could a neural network achieve that outcome? 57:56.700 --> 57:58.860 So I think it would be really hard to create 57:58.860 --> 58:02.300 a successful training set, which would enable it 58:02.300 --> 58:04.980 to have a conversation, a contextual conversation 58:04.980 --> 58:05.820 for an hour. 58:05.820 --> 58:08.140 Do you think it's a problem of data, perhaps? 58:08.140 --> 58:09.940 I think in some ways it's not a problem of data, 58:09.940 --> 58:13.620 it's a problem both of data and the problem of 58:13.620 --> 58:15.780 the way we're training our systems, 58:15.780 --> 58:18.060 their ability to truly, to generalize, 58:18.060 --> 58:19.300 to be very compositional. 58:19.300 --> 58:23.220 In some ways it's limited in the current capacity, 58:23.220 --> 58:27.980 at least we can translate well, 58:27.980 --> 58:32.540 we can find information well, we can extract information. 58:32.540 --> 58:35.180 So there are many capacities in which it's doing very well. 58:35.180 --> 58:38.000 And you can ask me, would you trust the machine 58:38.000 --> 58:39.820 to translate for you and use it as a source? 58:39.820 --> 58:42.580 I would say absolutely, especially if we're talking about 58:42.580 --> 58:45.660 newspaper data or other data which is in the realm 58:45.660 --> 58:47.900 of its own training set, I would say yes. 58:48.900 --> 58:52.900 But having conversations with the machine, 58:52.900 --> 58:56.460 it's not something that I would choose to do. 58:56.460 --> 58:59.420 But I would tell you something, talking about Turing tests 58:59.420 --> 59:02.940 and about all this kind of ELISA conversations, 59:02.940 --> 59:05.540 I remember visiting Tencent in China 59:05.540 --> 59:07.620 and they have this chat board and they claim 59:07.620 --> 59:10.780 there is really humongous amount of the local population 59:10.780 --> 59:12.940 which for hours talks to the chat board. 59:12.940 --> 59:15.340 To me it was, I cannot believe it, 59:15.340 --> 59:18.000 but apparently it's documented that there are some people 59:18.000 --> 59:20.760 who enjoy this conversation. 59:20.760 --> 59:24.540 And it brought to me another MIT story 59:24.540 --> 59:26.980 about ELISA and Weisenbaum. 59:26.980 --> 59:29.340 I don't know if you're familiar with the story. 59:29.340 --> 59:31.020 So Weisenbaum was a professor at MIT 59:31.020 --> 59:32.580 and when he developed this ELISA, 59:32.580 --> 59:34.620 which was just doing string matching, 59:34.620 --> 59:38.540 very trivial, like restating of what you said 59:38.540 --> 59:41.260 with very few rules, no syntax. 59:41.260 --> 59:43.740 Apparently there were secretaries at MIT 59:43.740 --> 59:48.180 that would sit for hours and converse with this trivial thing 59:48.180 --> 59:50.180 and at the time there was no beautiful interfaces 59:50.180 --> 59:51.820 so you actually need to go through the pain 59:51.820 --> 59:53.540 of communicating. 59:53.540 --> 59:56.940 And Weisenbaum himself was so horrified by this phenomenon 59:56.940 --> 59:59.300 that people can believe enough to the machine 59:59.300 --> 1:00:00.820 that you just need to give them the hint 1:00:00.820 --> 1:00:03.940 that machine understands you and you can complete the rest 1:00:03.940 --> 1:00:05.420 that he kind of stopped this research 1:00:05.420 --> 1:00:08.660 and went into kind of trying to understand 1:00:08.660 --> 1:00:11.480 what this artificial intelligence can do to our brains. 1:00:12.740 --> 1:00:14.380 So my point is, you know, 1:00:14.380 --> 1:00:19.300 how much, it's not how good is the technology, 1:00:19.300 --> 1:00:22.620 it's how ready we are to believe 1:00:22.620 --> 1:00:25.580 that it delivers the goods that we are trying to get. 1:00:25.580 --> 1:00:27.200 That's a really beautiful way to put it. 1:00:27.200 --> 1:00:29.800 I, by the way, I'm not horrified by that possibility, 1:00:29.800 --> 1:00:33.140 but inspired by it because, 1:00:33.140 --> 1:00:35.920 I mean, human connection, 1:00:35.920 --> 1:00:38.220 whether it's through language or through love, 1:00:39.860 --> 1:00:44.860 it seems like it's very amenable to machine learning 1:00:44.900 --> 1:00:49.340 and the rest is just challenges of psychology. 1:00:49.340 --> 1:00:52.460 Like you said, the secretaries who enjoy spending hours. 1:00:52.460 --> 1:00:55.020 I would say I would describe most of our lives 1:00:55.020 --> 1:00:58.020 as enjoying spending hours with those we love 1:00:58.020 --> 1:01:00.820 for very silly reasons. 1:01:00.820 --> 1:01:02.780 All we're doing is keyword matching as well. 1:01:02.780 --> 1:01:05.100 So I'm not sure how much intelligence 1:01:05.100 --> 1:01:08.140 we exhibit to each other with the people we love 1:01:08.140 --> 1:01:09.820 that we're close with. 1:01:09.820 --> 1:01:12.660 So it's a very interesting point 1:01:12.660 --> 1:01:16.020 of what it means to pass the Turing test with language. 1:01:16.020 --> 1:01:16.860 I think you're right. 1:01:16.860 --> 1:01:18.220 In terms of conversation, 1:01:18.220 --> 1:01:20.180 I think machine translation 1:01:21.420 --> 1:01:24.420 has very clear performance and improvement, right? 1:01:24.420 --> 1:01:28.020 What it means to have a fulfilling conversation 1:01:28.020 --> 1:01:32.660 is very person dependent and context dependent 1:01:32.660 --> 1:01:33.580 and so on. 1:01:33.580 --> 1:01:36.340 That's, yeah, it's very well put. 1:01:36.340 --> 1:01:40.740 But in your view, what's a benchmark in natural language, 1:01:40.740 --> 1:01:43.640 a test that's just out of reach right now, 1:01:43.640 --> 1:01:46.020 but we might be able to, that's exciting. 1:01:46.020 --> 1:01:49.100 Is it in perfecting machine translation 1:01:49.100 --> 1:01:51.900 or is there other, is it summarization? 1:01:51.900 --> 1:01:52.740 What's out there just out of reach? 1:01:52.740 --> 1:01:55.820 I think it goes across specific application. 1:01:55.820 --> 1:01:59.500 It's more about the ability to learn from few examples 1:01:59.500 --> 1:02:03.300 for real, what we call few short learning and all these cases 1:02:03.300 --> 1:02:05.940 because the way we publish these papers today, 1:02:05.940 --> 1:02:09.900 we say, if we have like naively, we get 55, 1:02:09.900 --> 1:02:12.500 but now we had a few example and we can move to 65. 1:02:12.500 --> 1:02:13.540 None of these methods 1:02:13.540 --> 1:02:15.980 actually are realistically doing anything useful. 1:02:15.980 --> 1:02:18.540 You cannot use them today. 1:02:18.540 --> 1:02:23.540 And the ability to be able to generalize and to move 1:02:25.460 --> 1:02:28.940 or to be autonomous in finding the data 1:02:28.940 --> 1:02:30.300 that you need to learn, 1:02:31.340 --> 1:02:34.280 to be able to perfect new tasks or new language, 1:02:35.300 --> 1:02:38.100 this is an area where I think we really need 1:02:39.200 --> 1:02:43.020 to move forward to and we are not yet there. 1:02:43.020 --> 1:02:45.060 Are you at all excited, 1:02:45.060 --> 1:02:46.540 curious by the possibility 1:02:46.540 --> 1:02:48.520 of creating human level intelligence? 1:02:49.900 --> 1:02:52.540 Is this, cause you've been very in your discussion. 1:02:52.540 --> 1:02:54.340 So if we look at oncology, 1:02:54.340 --> 1:02:58.100 you're trying to use machine learning to help the world 1:02:58.100 --> 1:02:59.700 in terms of alleviating suffering. 1:02:59.700 --> 1:03:02.340 If you look at natural language processing, 1:03:02.340 --> 1:03:05.300 you're focused on the outcomes of improving practical things 1:03:05.300 --> 1:03:06.820 like machine translation. 1:03:06.820 --> 1:03:09.880 But human level intelligence is a thing 1:03:09.880 --> 1:03:13.800 that our civilization has dreamed about creating, 1:03:13.800 --> 1:03:15.740 super human level intelligence. 1:03:15.740 --> 1:03:16.940 Do you think about this? 1:03:16.940 --> 1:03:19.040 Do you think it's at all within our reach? 1:03:20.380 --> 1:03:22.660 So as you said yourself, Elie, 1:03:22.660 --> 1:03:26.140 talking about how do you perceive 1:03:26.140 --> 1:03:28.980 our communications with each other, 1:03:28.980 --> 1:03:31.940 that we're matching keywords and certain behaviors 1:03:31.940 --> 1:03:33.020 and so on. 1:03:33.020 --> 1:03:36.860 So at the end, whenever one assesses, 1:03:36.860 --> 1:03:38.680 let's say relations with another person, 1:03:38.680 --> 1:03:41.460 you have separate kind of measurements and outcomes 1:03:41.460 --> 1:03:43.620 inside your head that determine 1:03:43.620 --> 1:03:45.860 what is the status of the relation. 1:03:45.860 --> 1:03:48.580 So one way, this is this classical level, 1:03:48.580 --> 1:03:49.600 what is the intelligence? 1:03:49.600 --> 1:03:51.860 Is it the fact that now we are gonna do the same way 1:03:51.860 --> 1:03:52.940 as human is doing, 1:03:52.940 --> 1:03:55.500 when we don't even understand what the human is doing? 1:03:55.500 --> 1:03:59.100 Or we now have an ability to deliver these outcomes, 1:03:59.100 --> 1:04:01.300 but not in one area, not in NLP, 1:04:01.300 --> 1:04:03.940 not just to translate or just to answer questions, 1:04:03.940 --> 1:04:05.380 but across many, many areas 1:04:05.380 --> 1:04:08.100 that we can achieve the functionalities 1:04:08.100 --> 1:04:11.060 that humans can achieve with their ability to learn 1:04:11.060 --> 1:04:12.380 and do other things. 1:04:12.380 --> 1:04:15.500 I think this is, and this we can actually measure 1:04:15.500 --> 1:04:17.560 how far we are. 1:04:17.560 --> 1:04:21.580 And that's what makes me excited that we, 1:04:21.580 --> 1:04:23.780 in my lifetime, at least so far what we've seen, 1:04:23.780 --> 1:04:25.840 it's like tremendous progress 1:04:25.840 --> 1:04:28.700 across these different functionalities. 1:04:28.700 --> 1:04:32.260 And I think it will be really exciting 1:04:32.260 --> 1:04:35.540 to see where we will be. 1:04:35.540 --> 1:04:39.300 And again, one way to think about it, 1:04:39.300 --> 1:04:41.820 there are machines which are improving their functionality. 1:04:41.820 --> 1:04:44.940 Another one is to think about us with our brains, 1:04:44.940 --> 1:04:46.420 which are imperfect, 1:04:46.420 --> 1:04:51.420 how they can be accelerated by this technology 1:04:51.420 --> 1:04:55.900 as it becomes stronger and stronger. 1:04:55.900 --> 1:04:57.260 Coming back to another book 1:04:57.260 --> 1:05:01.060 that I love, Flowers for Algernon. 1:05:01.060 --> 1:05:02.100 Have you read this book? 1:05:02.100 --> 1:05:02.940 Yes. 1:05:02.940 --> 1:05:05.700 So there is this point that the patient gets 1:05:05.700 --> 1:05:07.980 this miracle cure, which changes his brain. 1:05:07.980 --> 1:05:11.020 And all of a sudden they see life in a different way 1:05:11.020 --> 1:05:13.300 and can do certain things better, 1:05:13.300 --> 1:05:14.860 but certain things much worse. 1:05:14.860 --> 1:05:19.860 So you can imagine this kind of computer augmented cognition 1:05:22.400 --> 1:05:24.800 where it can bring you that now in the same way 1:05:24.800 --> 1:05:28.120 as the cars enable us to get to places 1:05:28.120 --> 1:05:30.080 where we've never been before, 1:05:30.080 --> 1:05:31.640 can we think differently? 1:05:31.640 --> 1:05:33.600 Can we think faster? 1:05:33.600 --> 1:05:36.680 And we already see a lot of it happening 1:05:36.680 --> 1:05:38.260 in how it impacts us, 1:05:38.260 --> 1:05:42.200 but I think we have a long way to go there. 1:05:42.200 --> 1:05:45.040 So that's sort of artificial intelligence 1:05:45.040 --> 1:05:47.280 and technology affecting our, 1:05:47.280 --> 1:05:50.440 augmenting our intelligence as humans. 1:05:50.440 --> 1:05:55.440 Yesterday, a company called Neuralink announced, 1:05:55.520 --> 1:05:56.800 they did this whole demonstration. 1:05:56.800 --> 1:05:57.980 I don't know if you saw it. 1:05:57.980 --> 1:06:01.000 It's, they demonstrated brain computer, 1:06:01.000 --> 1:06:02.680 brain machine interface, 1:06:02.680 --> 1:06:06.360 where there's like a sewing machine for the brain. 1:06:06.360 --> 1:06:11.120 Do you, you know, a lot of that is quite out there 1:06:11.120 --> 1:06:14.040 in terms of things that some people would say 1:06:14.040 --> 1:06:16.340 are impossible, but they're dreamers 1:06:16.340 --> 1:06:18.080 and want to engineer systems like that. 1:06:18.080 --> 1:06:20.360 Do you see, based on what you just said, 1:06:20.360 --> 1:06:23.820 a hope for that more direct interaction with the brain? 1:06:25.120 --> 1:06:27.040 I think there are different ways. 1:06:27.040 --> 1:06:29.000 One is a direct interaction with the brain. 1:06:29.000 --> 1:06:30.900 And again, there are lots of companies 1:06:30.900 --> 1:06:32.280 that work in this space 1:06:32.280 --> 1:06:35.080 and I think there will be a lot of developments. 1:06:35.080 --> 1:06:36.600 But I'm just thinking that many times 1:06:36.600 --> 1:06:39.080 we are not aware of our feelings, 1:06:39.080 --> 1:06:41.400 of motivation, what drives us. 1:06:41.400 --> 1:06:44.200 Like, let me give you a trivial example, our attention. 1:06:45.520 --> 1:06:47.260 There are a lot of studies that demonstrate 1:06:47.260 --> 1:06:49.200 that it takes a while to a person to understand 1:06:49.200 --> 1:06:51.080 that they are not attentive anymore. 1:06:51.080 --> 1:06:52.160 And we know that there are people 1:06:52.160 --> 1:06:54.520 who really have strong capacity to hold attention. 1:06:54.520 --> 1:06:57.080 There are other end of the spectrum people with ADD 1:06:57.080 --> 1:06:58.800 and other issues that they have problem 1:06:58.800 --> 1:07:00.760 to regulate their attention. 1:07:00.760 --> 1:07:03.520 Imagine to yourself that you have like a cognitive aid 1:07:03.520 --> 1:07:06.280 that just alerts you based on your gaze, 1:07:06.280 --> 1:07:09.280 that your attention is now not on what you are doing. 1:07:09.280 --> 1:07:10.560 And instead of writing a paper, 1:07:10.560 --> 1:07:12.760 you're now dreaming of what you're gonna do in the evening. 1:07:12.760 --> 1:07:16.360 So even this kind of simple measurement things, 1:07:16.360 --> 1:07:17.840 how they can change us. 1:07:17.840 --> 1:07:22.400 And I see it even in simple ways with myself. 1:07:22.400 --> 1:07:26.480 I have my zone app that I got in MIT gym. 1:07:26.480 --> 1:07:28.800 It kind of records, you know, how much did you run 1:07:28.800 --> 1:07:29.800 and you have some points 1:07:29.800 --> 1:07:32.880 and you can get some status, whatever. 1:07:32.880 --> 1:07:35.840 Like, I said, what is this ridiculous thing? 1:07:35.840 --> 1:07:38.800 Who would ever care about some status in some app? 1:07:38.800 --> 1:07:39.640 Guess what? 1:07:39.640 --> 1:07:41.560 So to maintain the status, 1:07:41.560 --> 1:07:44.640 you have to do set a number of points every month. 1:07:44.640 --> 1:07:48.040 And not only is that I do it every single month 1:07:48.040 --> 1:07:50.560 for the last 18 months, 1:07:50.560 --> 1:07:54.160 it went to the point that I was injured. 1:07:54.160 --> 1:07:56.160 And when I could run again, 1:07:58.120 --> 1:08:02.560 in two days, I did like some humongous amount of running 1:08:02.560 --> 1:08:04.080 just to complete the points. 1:08:04.080 --> 1:08:05.920 It was like really not safe. 1:08:05.920 --> 1:08:08.440 It was like, I'm not gonna lose my status 1:08:08.440 --> 1:08:10.240 because I want to get there. 1:08:10.240 --> 1:08:13.320 So you can already see that this direct measurement 1:08:13.320 --> 1:08:15.160 and the feedback is, you know, 1:08:15.160 --> 1:08:16.320 we're looking at video games 1:08:16.320 --> 1:08:18.720 and see why, you know, the addiction aspect of it, 1:08:18.720 --> 1:08:21.200 but you can imagine that the same idea can be expanded 1:08:21.200 --> 1:08:23.640 to many other areas of our life. 1:08:23.640 --> 1:08:25.960 When we really can get feedback 1:08:25.960 --> 1:08:28.480 and imagine in your case in relations, 1:08:29.880 --> 1:08:31.240 when we are doing keyword matching, 1:08:31.240 --> 1:08:36.120 imagine that the person who is generating the keywords, 1:08:36.120 --> 1:08:37.720 that person gets direct feedback 1:08:37.720 --> 1:08:39.560 before the whole thing explodes. 1:08:39.560 --> 1:08:42.000 Is it maybe at this happy point, 1:08:42.000 --> 1:08:44.000 we are going in the wrong direction. 1:08:44.000 --> 1:08:48.040 Maybe it will be really a behavior modifying moment. 1:08:48.040 --> 1:08:51.360 So yeah, it's a relationship management too. 1:08:51.360 --> 1:08:54.200 So yeah, that's a fascinating whole area 1:08:54.200 --> 1:08:56.120 of psychology actually as well, 1:08:56.120 --> 1:08:58.240 of seeing how our behavior has changed 1:08:58.240 --> 1:09:01.840 with basically all human relations now have 1:09:01.840 --> 1:09:06.200 other nonhuman entities helping us out. 1:09:06.200 --> 1:09:09.440 So you teach a large, 1:09:09.440 --> 1:09:12.600 a huge machine learning course here at MIT. 1:09:14.000 --> 1:09:15.360 I can ask you a million questions, 1:09:15.360 --> 1:09:17.560 but you've seen a lot of students. 1:09:17.560 --> 1:09:20.920 What ideas do students struggle with the most 1:09:20.920 --> 1:09:23.920 as they first enter this world of machine learning? 1:09:23.920 --> 1:09:26.520 Actually, this year was the first time 1:09:26.520 --> 1:09:28.480 I started teaching a small machine learning class. 1:09:28.480 --> 1:09:31.160 And it came as a result of what I saw 1:09:31.160 --> 1:09:34.640 in my big machine learning class that Tomi Yakel and I built 1:09:34.640 --> 1:09:36.640 maybe six years ago. 1:09:38.040 --> 1:09:40.360 What we've seen that as this area become more 1:09:40.360 --> 1:09:43.440 and more popular, more and more people at MIT 1:09:43.440 --> 1:09:45.360 want to take this class. 1:09:45.360 --> 1:09:48.320 And while we designed it for computer science majors, 1:09:48.320 --> 1:09:50.760 there were a lot of people who really are interested 1:09:50.760 --> 1:09:52.600 to learn it, but unfortunately, 1:09:52.600 --> 1:09:55.720 their background was not enabling them 1:09:55.720 --> 1:09:57.200 to do well in the class. 1:09:57.200 --> 1:09:59.360 And many of them associated machine learning 1:09:59.360 --> 1:10:01.360 with the word struggle and failure, 1:10:02.480 --> 1:10:04.640 primarily for non majors. 1:10:04.640 --> 1:10:06.840 And that's why we actually started a new class 1:10:06.840 --> 1:10:10.800 which we call machine learning from algorithms to modeling, 1:10:10.800 --> 1:10:15.000 which emphasizes more the modeling aspects of it 1:10:15.000 --> 1:10:20.000 and focuses on, it has majors and non majors. 1:10:20.000 --> 1:10:23.480 So we kind of try to extract the relevant parts 1:10:23.480 --> 1:10:25.560 and make it more accessible, 1:10:25.560 --> 1:10:27.800 because the fact that we're teaching 20 classifiers 1:10:27.800 --> 1:10:29.240 in standard machine learning class, 1:10:29.240 --> 1:10:32.200 it's really a big question to really need it. 1:10:32.200 --> 1:10:34.520 But it was interesting to see this 1:10:34.520 --> 1:10:36.480 from first generation of students, 1:10:36.480 --> 1:10:39.080 when they came back from their internships 1:10:39.080 --> 1:10:42.320 and from their jobs, 1:10:42.320 --> 1:10:45.560 what different and exciting things they can do. 1:10:45.560 --> 1:10:47.600 I would never think that you can even apply 1:10:47.600 --> 1:10:50.800 machine learning to, some of them are like matching, 1:10:50.800 --> 1:10:53.480 the relations and other things like variety. 1:10:53.480 --> 1:10:56.080 Everything is amenable as the machine learning. 1:10:56.080 --> 1:10:58.320 That actually brings up an interesting point 1:10:58.320 --> 1:11:00.680 of computer science in general. 1:11:00.680 --> 1:11:03.520 It almost seems, maybe I'm crazy, 1:11:03.520 --> 1:11:06.520 but it almost seems like everybody needs to learn 1:11:06.520 --> 1:11:08.160 how to program these days. 1:11:08.160 --> 1:11:11.400 If you're 20 years old, or if you're starting school, 1:11:11.400 --> 1:11:14.200 even if you're an English major, 1:11:14.200 --> 1:11:19.200 it seems like programming unlocks so much possibility 1:11:20.480 --> 1:11:21.880 in this world. 1:11:21.880 --> 1:11:25.000 So when you interacted with those non majors, 1:11:25.000 --> 1:11:30.000 is there skills that they were simply lacking at the time 1:11:30.280 --> 1:11:33.000 that you wish they had and that they learned 1:11:33.000 --> 1:11:34.680 in high school and so on? 1:11:34.680 --> 1:11:37.520 Like how should education change 1:11:37.520 --> 1:11:41.320 in this computerized world that we live in? 1:11:41.320 --> 1:11:44.320 I think because I knew that there is a Python component 1:11:44.320 --> 1:11:47.000 in the class, their Python skills were okay 1:11:47.000 --> 1:11:49.160 and the class isn't really heavy on programming. 1:11:49.160 --> 1:11:52.400 They primarily kind of add parts to the programs. 1:11:52.400 --> 1:11:55.440 I think it was more of the mathematical barriers 1:11:55.440 --> 1:11:58.200 and the class, again, with the design on the majors 1:11:58.200 --> 1:12:01.200 was using the notation, like big O for complexity 1:12:01.200 --> 1:12:04.520 and others, people who come from different backgrounds 1:12:04.520 --> 1:12:05.800 just don't have it in the lexical, 1:12:05.800 --> 1:12:09.120 so necessarily very challenging notion, 1:12:09.120 --> 1:12:11.480 but they were just not aware. 1:12:12.360 --> 1:12:16.240 So I think that kind of linear algebra and probability, 1:12:16.240 --> 1:12:19.120 the basics, the calculus, multivariate calculus, 1:12:19.120 --> 1:12:20.840 things that can help. 1:12:20.840 --> 1:12:23.520 What advice would you give to students 1:12:23.520 --> 1:12:25.280 interested in machine learning, 1:12:25.280 --> 1:12:29.240 interested, you've talked about detecting, 1:12:29.240 --> 1:12:31.360 curing cancer, drug design, 1:12:31.360 --> 1:12:34.520 if they want to get into that field, what should they do? 1:12:36.320 --> 1:12:39.040 Get into it and succeed as researchers 1:12:39.040 --> 1:12:42.080 and entrepreneurs. 1:12:43.320 --> 1:12:45.240 The first good piece of news is that right now 1:12:45.240 --> 1:12:47.400 there are lots of resources 1:12:47.400 --> 1:12:50.160 that are created at different levels 1:12:50.160 --> 1:12:54.800 and you can find online in your school classes 1:12:54.800 --> 1:12:57.560 which are more mathematical, more applied and so on. 1:12:57.560 --> 1:13:01.320 So you can find a kind of a preacher 1:13:01.320 --> 1:13:02.760 which preaches in your own language 1:13:02.760 --> 1:13:04.520 where you can enter the field 1:13:04.520 --> 1:13:06.720 and you can make many different types of contribution 1:13:06.720 --> 1:13:09.640 depending of what is your strengths. 1:13:10.760 --> 1:13:13.720 And the second point, I think it's really important 1:13:13.720 --> 1:13:18.160 to find some area which you really care about 1:13:18.160 --> 1:13:20.240 and it can motivate your learning 1:13:20.240 --> 1:13:22.640 and it can be for somebody curing cancer 1:13:22.640 --> 1:13:25.360 or doing self driving cars or whatever, 1:13:25.360 --> 1:13:29.680 but to find an area where there is data 1:13:29.680 --> 1:13:31.320 where you believe there are strong patterns 1:13:31.320 --> 1:13:33.600 and we should be doing it and we're still not doing it 1:13:33.600 --> 1:13:35.280 or you can do it better 1:13:35.280 --> 1:13:39.680 and just start there and see where it can bring you. 1:13:40.800 --> 1:13:45.600 So you've been very successful in many directions in life, 1:13:46.480 --> 1:13:48.840 but you also mentioned Flowers of Argonon. 1:13:51.200 --> 1:13:53.840 And I think I've read or listened to you mention somewhere 1:13:53.840 --> 1:13:55.360 that researchers often get lost 1:13:55.360 --> 1:13:56.720 in the details of their work. 1:13:56.720 --> 1:14:00.240 This is per our original discussion with cancer and so on 1:14:00.240 --> 1:14:02.200 and don't look at the bigger picture, 1:14:02.200 --> 1:14:05.320 bigger questions of meaning and so on. 1:14:05.320 --> 1:14:07.440 So let me ask you the impossible question 1:14:08.640 --> 1:14:11.560 of what's the meaning of this thing, 1:14:11.560 --> 1:14:16.560 of life, of your life, of research. 1:14:16.720 --> 1:14:21.440 Why do you think we descendant of great apes 1:14:21.440 --> 1:14:24.480 are here on this spinning ball? 1:14:26.800 --> 1:14:30.320 You know, I don't think that I have really a global answer. 1:14:30.320 --> 1:14:32.800 You know, maybe that's why I didn't go to humanities 1:14:33.760 --> 1:14:36.480 and I didn't take humanities classes in my undergrad. 1:14:39.480 --> 1:14:43.560 But the way I'm thinking about it, 1:14:43.560 --> 1:14:48.200 each one of us inside of them have their own set of, 1:14:48.200 --> 1:14:51.120 you know, things that we believe are important. 1:14:51.120 --> 1:14:53.360 And it just happens that we are busy 1:14:53.360 --> 1:14:56.240 with achieving various goals, busy listening to others 1:14:56.240 --> 1:15:00.960 and to kind of try to conform and to be part of the crowd, 1:15:00.960 --> 1:15:03.680 that we don't listen to that part. 1:15:04.600 --> 1:15:09.600 And, you know, we all should find some time to understand 1:15:09.600 --> 1:15:11.840 what is our own individual missions. 1:15:11.840 --> 1:15:14.080 And we may have very different missions 1:15:14.080 --> 1:15:18.200 and to make sure that while we are running 10,000 things, 1:15:18.200 --> 1:15:21.920 we are not, you know, missing out 1:15:21.920 --> 1:15:26.800 and we're putting all the resources to satisfy 1:15:26.800 --> 1:15:28.440 our own mission. 1:15:28.440 --> 1:15:32.400 And if I look over my time, when I was younger, 1:15:32.400 --> 1:15:35.000 most of these missions, you know, 1:15:35.000 --> 1:15:38.600 I was primarily driven by the external stimulus, 1:15:38.600 --> 1:15:41.520 you know, to achieve this or to be that. 1:15:41.520 --> 1:15:46.520 And now a lot of what I do is driven by really thinking 1:15:47.640 --> 1:15:51.360 what is important for me to achieve independently 1:15:51.360 --> 1:15:55.160 of the external recognition. 1:15:55.160 --> 1:16:00.080 And, you know, I don't mind to be viewed in certain ways. 1:16:01.400 --> 1:16:05.760 The most important thing for me is to be true to myself, 1:16:05.760 --> 1:16:07.520 to what I think is right. 1:16:07.520 --> 1:16:08.680 How long did it take? 1:16:08.680 --> 1:16:13.240 How hard was it to find the you that you have to be true to? 1:16:14.160 --> 1:16:15.520 So it takes time. 1:16:15.520 --> 1:16:17.760 And even now, sometimes, you know, 1:16:17.760 --> 1:16:20.880 the vanity and the triviality can take, you know. 1:16:20.880 --> 1:16:22.560 At MIT. 1:16:22.560 --> 1:16:25.080 Yeah, it can everywhere, you know, 1:16:25.080 --> 1:16:26.960 it's just the vanity at MIT is different, 1:16:26.960 --> 1:16:28.160 the vanity in different places, 1:16:28.160 --> 1:16:30.920 but we all have our piece of vanity. 1:16:30.920 --> 1:16:35.920 But I think actually for me, many times the place 1:16:38.720 --> 1:16:43.720 to get back to it is, you know, when I'm alone 1:16:43.800 --> 1:16:45.800 and also when I read. 1:16:45.800 --> 1:16:47.760 And I think by selecting the right books, 1:16:47.760 --> 1:16:52.760 you can get the right questions and learn from what you read. 1:16:54.880 --> 1:16:58.080 So, but again, it's not perfect. 1:16:58.080 --> 1:17:02.040 Like vanity sometimes dominates. 1:17:02.040 --> 1:17:04.800 Well, that's a beautiful way to end. 1:17:04.800 --> 1:17:06.400 Thank you so much for talking today. 1:17:06.400 --> 1:17:07.240 Thank you. 1:17:07.240 --> 1:17:08.080 That was fun. 1:17:08.080 --> 1:17:28.080 That was fun.