bpHigh commited on
Commit
ed050a6
1 Parent(s): f13a3b6

Upload 5 files

Browse files
app.py ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This is a sample Python script.
2
+
3
+ # Press ⌃R to execute it or replace it with your code.
4
+ # Press Double ⇧ to search everywhere for classes, files, tool windows, actions, and settings.
5
+ import gradio as gr
6
+ import time
7
+ import random
8
+ import math
9
+
10
+ import pandas as pd
11
+ import openai
12
+ import os
13
+ from display_json import get_podcast
14
+
15
+
16
+ with gr.Blocks(theme='bethecloud/storj_theme') as demo:
17
+ gr.HTML('<img src="https://framerusercontent.com/images/cVca7XpB1kCi8H10JIe8TQPNVQ.png" alt="Podcaster" '
18
+ 'style="height: 60px; width:100px;"/>')
19
+ gr.Markdown("""
20
+ <div style= "display:block;
21
+ align-items: center;
22
+ gap: 0.8rem;">
23
+ <h1 style='text-align: center;color:blue'>Podcast World</h1>
24
+ <h2 style='text-align: center;color:darkslategrey'>Can be used to search and get recommended about podcasts, process newsletter insights over podcasts etc.</h2>
25
+
26
+ </div>
27
+ """)
28
+ with gr.Tabs():
29
+ # with gr.TabItem("Search Podcasts and get their RSS feeds"):
30
+ # with gr.Box():
31
+ # gr.Markdown(
32
+ # """
33
+ # ### Search Podcasts and get their RSS feeds
34
+ # """)
35
+ # with gr.Column():
36
+ # # normal_chatbot = gr.Chatbot()
37
+ # # normal_msg = gr.Textbox()
38
+ # # normal_clear = gr.Button("Clear")
39
+ # # normal_msg.submit(search_response, [normal_msg, normal_chatbot, dropdown], [normal_msg, normal_chatbot])
40
+ # # normal_clear.click(lambda: None, None, normal_chatbot_copilot, queue=False)
41
+
42
+ with gr.TabItem("PreProcessed Podcasts"):
43
+ with gr.Box():
44
+ gr.Markdown(
45
+ """
46
+ ### PreProcessed Podcasts
47
+ """)
48
+ with gr.Column():
49
+ dropdown_few = gr.Dropdown(["In Machines We trust - MIT",
50
+ "TWIML",
51
+ ],
52
+ label="Select Podcast")
53
+ btn = gr.Button("See")
54
+ with gr.Row().style(equal_height=True):
55
+ Title = gr.Textbox(label="Title")
56
+ Episode_Name = gr.Textbox(label="Episode Name")
57
+ Episode_Image = gr.Image(label="Episode Image")
58
+
59
+ podcast_summary = gr.Textbox(label="Summary")
60
+
61
+ with gr.Row().style(equal_height=True):
62
+ podcast_guest = gr.Textbox(label="podcast guest")
63
+ podcast_guest_org = gr.Textbox(label="Podcast Guest Organization")
64
+ podcast_guest_title = gr.Textbox(label="Guest Title")
65
+ podcast_guest_wikipedia = gr.Textbox(label="Guest Wikipedia Info")
66
+
67
+ podcast_highlights = gr.Textbox(label="Highlights")
68
+ podcast_key_moments = gr.Textbox(label="Key Moments and Key Topics")
69
+
70
+ with gr.Accordion("Open for Full Dialog Transcript"):
71
+ podcast_gpt_transcript = gr.Textbox()
72
+
73
+ btn.click(fn=get_podcast, inputs=dropdown_few, outputs=[Title,Episode_Name,Episode_Image,podcast_summary, podcast_guest, podcast_guest_org, \
74
+ podcast_guest_title, podcast_guest_wikipedia,podcast_highlights, podcast_key_moments, \
75
+ podcast_gpt_transcript])
76
+
77
+
78
+ # with gr.TabItem("Process New Podcast using RSS feed"):
79
+ # with gr.Box():
80
+ # gr.Markdown(
81
+ # """
82
+ # ### Chat with GPT 3.5 with transcripts
83
+ # """)
84
+ # with gr.Column():
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+ if __name__ == "__main__":
93
+
94
+
95
+ demo.queue(concurrency_count=1, api_open=False)
96
+ demo.launch(debug=True, share=False)
display_json.py ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import os
3
+
4
+
5
+ def create_dict_from_json_files(filepath):
6
+ json_file = filepath
7
+ data_dict = {}
8
+
9
+ with open(filepath, 'r') as file:
10
+ podcast_info = json.load(file)
11
+ podcast_name = podcast_info['podcast_details']['podcast_title']
12
+ # Process the file data as needed
13
+ episode_title = podcast_info['podcast_details']['episode_title']
14
+ episode_image = podcast_info['podcast_details']['episode_image']
15
+ podcast_summary = podcast_info['podcast_summary']
16
+ podcast_guest = str(podcast_info['podcast_guest']['podcast_guest'])
17
+ podcast_guest_org = str(podcast_info['podcast_guest'].get('podcast_guest_org'))
18
+ podcast_guest_title = str(podcast_info['podcast_guest'].get('podcast_guest_title'))
19
+ podcast_guest_wikipedia = str(podcast_info['podcast_guest'].get('wikipedia_summary'))
20
+ podcast_highlights = str(podcast_info['podcast_highlights'])
21
+ podcast_key_moments = str(podcast_info['key_moments_and_key_topics'])
22
+ podcast_gpt_transcript = str(podcast_info['gpt_podcast_transcript'])
23
+
24
+ return podcast_name, episode_title, episode_image, podcast_summary, podcast_guest, podcast_guest_org, \
25
+ podcast_guest_title, podcast_guest_wikipedia, podcast_highlights, podcast_key_moments, \
26
+ podcast_gpt_transcript
27
+
28
+
29
+ def get_podcast(dropdown):
30
+ if dropdown=="In Machines We trust - MIT":
31
+ dropdown = "podcast-inmachineswetrust (1).json"
32
+ else:
33
+ dropdown = "podcast-twiml.json"
34
+ podcast_name, episode_title, episode_image, podcast_summary, podcast_guest, podcast_guest_org, \
35
+ podcast_guest_title, podcast_guest_wikipedia, podcast_highlights, podcast_key_moments, \
36
+ podcast_gpt_transcript = create_dict_from_json_files(dropdown)
37
+ return podcast_name, episode_title,episode_image, podcast_summary, podcast_guest, podcast_guest_org, \
38
+ podcast_guest_title, podcast_guest_wikipedia,podcast_highlights, podcast_key_moments, \
39
+ podcast_gpt_transcript
podcast-inmachineswetrust (1).json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"podcast_details": {"podcast_title": "In Machines We Trust", "episode_title": "That's a wrap!", "episode_image": "https://megaphone.imgix.net/podcasts/2b04531c-c32f-11ea-9a46-5bd83c5c4b83/image/uploads_2F1594440767512-4ylb1aqs0c3-15b8f937fbf34dda939a5579a3b84156_2Fmachines.final.jpg?ixlib=rails-4.3.1&max-w=3000&max-h=3000&fit=crop&auto=format,compress", "podcast_transcript": "(6.44,11.683) SPEAKER_05 : Hi there, it's Jennifer Strong dropping by to let you know that this podcast has come to an end.\n\n(11.683,18.207) SPEAKER_05 : Can you believe it's been three years since we set out to tell the world there's AI in just about everything?\n\n(18.207,20.729) SPEAKER_05 : And to say goodbye, we've pulled together some of the highlights.\n\n(22.887,31.03) SPEAKER_05 : Hi, I'm Jennifer Strong, host of a new series about trust, artificial intelligence, and how these impact our daily lives.\n\n(31.03,37.812) SPEAKER_05 : If seeing is believing, how do we begin to navigate a world where we can't trust our eyes or ears?\n\n(37.812,42.314) SPEAKER_05 : And so you know, what you're listening to, it's not just me speaking.\n\n(42.314,47.735) SPEAKER_05 : I had some help from an artificial version of my voice, filling in words here and there.\n\n(47.735,49.296) SPEAKER_05 : Meet synthetic Jennifer.\n\n(49.936,51.279) SPEAKER_05 : Hi there, folks.\n\n(51.279,53.864) SPEAKER_05 : I can even click to adjust my mood.\n\n(53.864,54.305) SPEAKER_05 : Hi there.\n\n(54.305,56.409) SPEAKER_05 : Yeah, let's not make it angry.\n\n(61.038,64.439) SPEAKER_05 : We're out taking a drive in one of New York City's busiest bridges.\n\n(64.439,66.16) SPEAKER_05 : It's called the RFK.\n\n(66.16,73.102) SPEAKER_05 : And the reason we're doing this is because the Transit Authority, the MTA, has installed live facial recognition.\n\n(73.102,76.183) SPEAKER_05 : Actually, we're about to go under it right now.\n\n(76.183,77.083) SPEAKER_05 : Do you see that?\n\n(77.083,79.184) SPEAKER_05 : The camera's pointed right at our faces.\n\n(79.184,81.265) SPEAKER_05 : They've now put this all over the city.\n\n(81.265,83.645) SPEAKER_05 : It's in a number of bridges and tunnels.\n\n(83.645,87.286) SPEAKER_05 : But the reason we're on this one is because this is where it all started.\n\n(89.267,94.792) SPEAKER_05 : What it does, at least in theory, what it's supposed to do is read our faces through our windshields.\n\n(94.792,99.395) SPEAKER_05 : And what's crazy about this is, well, what's crazy about it is that nobody knew there was a test.\n\n(99.395,104.62) SPEAKER_05 : But also crazy about it is when the results of that test were leaked to the press last year,\n\n(105.28,108.724) SPEAKER_05 : You want to guess how many faces were captured during the test period?\n\n(108.724,109.525) SPEAKER_05 : What do you think?\n\n(109.525,111.347) SPEAKER_05 : This is Emma Sillikens, by the way, my producer.\n\n(111.347,112.589) SPEAKER_05 : I don't know, like thousands?\n\n(112.589,114.691) SPEAKER_05 : Maybe, maybe millions?\n\n(114.691,115.172) SPEAKER_05 : New York City?\n\n(115.172,115.873) SPEAKER_05 : Maybe millions?\n\n(115.873,118.956) SPEAKER_05 : No, um, they got none.\n\n(118.956,120.598) SPEAKER_05 : Zero percent.\n\n(120.598,123.702) SPEAKER_05 : But they moved forward and they proceeded with it anyway.\n\n(125.568,129.973) SPEAKER_16 : The cashless tolling, which is really not about tolling.\n\n(129.973,136.48) SPEAKER_16 : It's really about putting in an electronic system that we have never had before.\n\n(136.48,139.443) SPEAKER_05 : That's New York Governor Andrew Cuomo speaking at an event in 2018.\n\n(140.992,157.902) SPEAKER_16 : What it is doing is it's put in a system that reads every license plate that comes through that crossing and reports to the police car that is stationed right there within five seconds.\n\n(158.589,161.892) SPEAKER_05 : But this is not some tech event, nor is it about policing.\n\n(161.892,166.497) SPEAKER_05 : It's celebrating the end of repairs to a tunnel connecting Brooklyn and Manhattan.\n\n(166.497,171.683) SPEAKER_05 : You see, the city's subways and tunnels flooded with saltwater in the aftermath of Hurricane Sandy.\n\n(171.683,177.609) SPEAKER_05 : All the electronics and wiring had to be replaced, and it gave them an opportunity to try something new.\n\n(177.609,181.613) SPEAKER_16 : We're now moving to facial recognition technology.\n\n(182.5,193.725) SPEAKER_16 : which takes you to a whole new level where it can see the face of the person in the car and run that against databases.\n\n(193.725,199.388) SPEAKER_05 : And they're experimenting with something even more cutting edge than reading faces or license plates.\n\n(199.388,201.529) SPEAKER_05 : They're attempting to read people's ears.\n\n(202.13,207.933) SPEAKER_16 : because many times a person will turn their head when they see a security camera.\n\n(207.933,218.198) SPEAKER_16 : So they're now experimenting with technology that just identifies a person by the ear, believe it or not.\n\n(218.198,226.683) SPEAKER_06 : I don't know if it was AI, I don't know if they had taken a recording of something he had done and were able to manipulate it, but I'm telling you, it was my son.\n\n(227.428,231.114) SPEAKER_05 : The day started like any other for a man we're going to call Jim.\n\n(231.114,232.797) SPEAKER_05 : He lives outside Boston.\n\n(232.797,236.524) SPEAKER_05 : And by the way, he has a family member who works for MIT.\n\n(236.524,239.93) SPEAKER_05 : We're not going to use his last name because they have concerns about their safety.\n\n(240.368,244.01) SPEAKER_06 : And it was like a Tuesday or Wednesday morning, nine o'clock.\n\n(244.01,246.852) SPEAKER_06 : I'm deep in thought working on something.\n\n(246.852,249.454) SPEAKER_06 : That is until he received this call.\n\n(249.454,252.176) SPEAKER_06 : The phone rings and I pick it up and it's my son.\n\n(252.176,255.618) SPEAKER_06 : And he is clearly agitated.\n\n(255.618,258.9) SPEAKER_06 : This kid is a really chill guy.\n\n(258.9,264.044) SPEAKER_06 : But when he does get upset, he has a number of vocal mannerisms.\n\n(264.044,267.606) SPEAKER_06 : And this was like, oh, my God, he's in trouble.\n\n(269.64,271.961) SPEAKER_06 : And he basically told me, look, I'm in jail.\n\n(271.961,272.701) SPEAKER_06 : I'm in Mexico.\n\n(272.701,273.921) SPEAKER_06 : They took my phone.\n\n(273.921,277.362) SPEAKER_06 : I only have 30 seconds.\n\n(277.362,279.922) SPEAKER_06 : They said I was drinking, but I wasn't.\n\n(279.922,281.382) SPEAKER_06 : And people are hurt.\n\n(281.382,284.123) SPEAKER_06 : And look, I have to get off the phone.\n\n(284.123,290.004) SPEAKER_06 : Call this lawyer and he gives me a phone number and has to hang up.\n\n(290.004,294.345) SPEAKER_05 : His son is in Mexico, and there's just no doubt in his mind it's him.\n\n(294.87,297.232) SPEAKER_06 : And I got to tell you, Jennifer, it was him.\n\n(297.232,298.393) SPEAKER_06 : It was his voice.\n\n(298.393,299.914) SPEAKER_06 : It was everything.\n\n(299.914,310.023) SPEAKER_06 : Tone, just these little mannerisms, the pauses, the gulping for air, everything that you couldn't imagine.\n\n(310.023,313.406) SPEAKER_05 : He's still not sure how someone captured the essence of his son's voice.\n\n(314.872,317.153) SPEAKER_05 : And it's not just Jim who's unsure.\n\n(317.153,320.714) SPEAKER_05 : We have no idea whether AI had anything to do with this.\n\n(320.714,325.076) SPEAKER_05 : But the point is, we now live in a world where we also can't be sure it didn't.\n\n(325.076,329.577) SPEAKER_05 : It's incredibly easy to fake someone's voice with even a few minutes of recordings.\n\n(329.577,334.799) SPEAKER_05 : And teenagers like Jim's son, they share countless recordings through social media posts and messages.\n\n(334.799,339.421) SPEAKER_06 : I was quite impressed with how good it was.\n\n(342.239,349.366) SPEAKER_06 : I'm not easily fooled, and man, they had it nailed.\n\n(349.366,356.593) SPEAKER_01 : What if you could stop retail crime before it happens, by knowing the moment a shoplifter enters your store?\n\n(356.593,361.197) SPEAKER_01 : And what if you could know about the presence of violent criminals before they act?\n\n(361.197,365.562) SPEAKER_01 : With FaceFirst, you can stop crime before it starts.\n\n(367.268,370.81) SPEAKER_05 : That's one of the largest providers of this tech to retail stores.\n\n(370.81,376.213) SPEAKER_05 : It detects faces, voices, objects, and claims it can analyze behavior.\n\n(376.213,391.863) SPEAKER_05 : Other retailers use face recognition to know when VIP shoppers or celebrities are in their stores, not unlike this scene from the film Minority Report, where, as Tom Cruise strolls through a mall, his eyes are scanned and the ads address his character by name.\n\n(391.863,394.064) SPEAKER_05 : Good evening, John Apton.\n\n(394.064,394.925) SPEAKER_05 : I'm John Apton.\n\n(395.456,402.501) SPEAKER_05 : The face measurements powering these applications can also be used for many other things besides just identifying someone.\n\n(402.501,413.608) SPEAKER_05 : For example, some shopping malls use it to help set their store rents, by counting how many people walk by and using face data to gauge their gender, age, and other demographics.\n\n(413.608,417.531) SPEAKER_05 : Sometimes face recognition cameras are even hidden inside those mall directories.\n\n(419.23,424.215) SPEAKER_05 : And inside stores, retailers use it to better understand what shoppers are interested in.\n\n(424.215,431.341) SPEAKER_05 : It's also embedded within shopping apps and store mirrors that let people try on anything from eyeglasses to makeup virtually.\n\n(436.79,439.991) SPEAKER_05 : Clearview is sometimes referred to as the killer app.\n\n(439.991,442.191) SPEAKER_05 : It's also incredibly controversial.\n\n(442.191,450.313) SPEAKER_05 : It quietly scraped billions of images off the internet from Venmo, LinkedIn, Google, Facebook, Twitter, YouTube, and so on.\n\n(450.313,458.854) SPEAKER_05 : And it built a system that's believed to be widely used by law enforcement, including the FBI and ICE, as well as state and local police.\n\n(458.854,462.475) SPEAKER_05 : There are many legal battles being fought over this practice called web scraping.\n\n(463.019,465.901) SPEAKER_05 : Not because there's something inherently bad about it.\n\n(465.901,469.003) SPEAKER_05 : It's just a tool for gathering data from the Internet.\n\n(469.003,472.966) SPEAKER_05 : It's how websites offer price comparisons and real estate listings.\n\n(472.966,476.609) SPEAKER_05 : It's also how a great deal of public research gets done.\n\n(476.609,481.052) SPEAKER_05 : So the real issue is there aren't really ground rules for what can be scraped.\n\n(481.052,484.254) SPEAKER_05 : And the federal law most often used to sort out these cases.\n\n(484.254,488.257) SPEAKER_05 : Well, it's from 1986 before the Web even existed.\n\n(489.761,495.484) SPEAKER_05 : Throughout this series, we're going to hear from folks who build technologies as well as those who fight against them.\n\n(495.484,500.747) SPEAKER_05 : And I sat down with Clearview's chief executive to talk about all of this from his perspective.\n\n(500.747,503.129) SPEAKER_02 : And he puts an old photo from my LinkedIn account up on the screen.\n\n(509.72,514.883) SPEAKER_02 : On the left side, you see this new search button where you pick a photo.\n\n(514.883,517.464) SPEAKER_02 : So this is the one I'm going to use.\n\n(517.464,522.947) SPEAKER_02 : Don't worry, no one can see the screen except for us.\n\n(522.947,526.529) SPEAKER_02 : And so that took, you know, about a second.\n\n(526.529,528.41) SPEAKER_02 : See, there's a link that you can click on.\n\n(528.41,532.092) SPEAKER_02 : But as we go through, this is on Twitter.\n\n(532.092,533.493) SPEAKER_02 : Do you remember this photo at all?\n\n(535.187,538.049) SPEAKER_05 : No, I didn't know that was taken and I look very... You do.\n\n(538.049,541.751) SPEAKER_02 : You look very serious in that one.\n\n(541.751,542.711) SPEAKER_02 : Very serious.\n\n(542.711,544.352) SPEAKER_02 : Yeah.\n\n(544.352,546.694) SPEAKER_02 : Here you're giving a talk.\n\n(546.694,548.114) SPEAKER_02 : Wall speech journal.\n\n(548.114,549.315) SPEAKER_02 : The future of everything.\n\n(549.315,550.916) SPEAKER_02 : Yeah.\n\n(550.916,553.077) SPEAKER_02 : You're interviewing someone here at Duke Health.\n\n(553.077,553.597) SPEAKER_02 : Yeah.\n\n(553.597,555.979) SPEAKER_02 : So like I said, all these things are publicly available.\n\n(555.979,557.98) SPEAKER_02 : So that's basically how it works.\n\n(559.235,569.424) SPEAKER_05 : There's nothing unusual here, just pictures of me at work reporting stories and hosting panels in different cities, though it is kind of jarring to find photos of myself I've never seen.\n\n(569.424,573.568) SPEAKER_05 : And once again, it brings up this thorny question of consent.\n\n(573.568,580.994) SPEAKER_05 : You see, I'm unlikely to check a box giving permission to companies like Clearview to scrape my image and use it to build their businesses.\n\n(580.994,582.696) SPEAKER_05 : The thing is, they don't need it.\n\n(588.165,594.189) SPEAKER_09 : I was programmed to believe that I was a 19-year-old, half Brazilian, half Spanish girl named Makayla.\n\n(594.189,608.338) SPEAKER_05 : This digital influencer and model is a project that began as a CGI Instagram profile, but has gone on to release music and collaborate with luxury brands, such as Calvin Klein and Prada, amassing millions of followers along the way.\n\n(608.338,609.098) SPEAKER_09 : Don't worry, y'all.\n\n(609.098,614.302) SPEAKER_09 : I am a robot, but I'm not here to hack your Venmo or leave your private browser history.\n\n(614.926,618.488) SPEAKER_05 : For next-gen systems, AI is the core creation tool.\n\n(618.488,630.493) SPEAKER_05 : With it comes interactive human-like experiences as well as some familiar thorny questions about ownership.\n\n(630.493,640.298) SPEAKER_05 : This is a song composed partly by FN Mecca, an AI created by the company Factory New, which describes itself as a record label specializing in virtual artists.\n\n(641.328,651.518) SPEAKER_05 : The system analyzes popular songs from specific genres and generates the building blocks of new songs, such as melody and tempo, with vocals performed by a real human.\n\n(652.461,656.683) SPEAKER_05 : FN Mecca was designed and marketed to represent a kind of digital rapper.\n\n(656.683,667.288) SPEAKER_05 : His TikTok videos, showing him in designer clothes and luxury cars with an edgy haircut and plenty of attitude, they generated more than a billion views on the platform.\n\n(667.288,678.713) SPEAKER_05 : In August, it was announced that the digital human had been signed to one of the most powerful music labels in the world, Capitol Records, which retains rights to the works of artists like ABBA and Katy Perry.\n\n(678.713,679.894) SPEAKER_05 : Then, this happened.\n\n(680.347,689.95) SPEAKER_00 : From stepping into the virtual future to back on the proverbial streets, the AI rapper everybody has been talking about has been dropped from his label.\n\n(689.95,698.533) SPEAKER_05 : In addition to his virtual jewels and custom Tesla Cybertruck, FN Mecca is depicted as a black man, something its human creators are not.\n\n(702.287,715.268) SPEAKER_15 : Colleagues, we are here this afternoon to consider two ordinances that seek to ban the use of facial recognition technologies by our own Portland city government and by private entities in the public spaces.\n\n(716.497,722.158) SPEAKER_05 : That's Portland Mayor Ted Wheeler opening the city council meeting that passed these bills in September.\n\n(722.158,727.06) SPEAKER_05 : As with most things these days, the public comments and vote took place on Zoom.\n\n(727.06,734.021) SPEAKER_05 : Over the next few hours, lawyers, software engineers, concerned citizens, and business leaders all had their say.\n\n(734.021,741.183) SPEAKER_05 : Then, just before the vote took place, one last person, a local data scientist, raised his virtual hand to ask a question.\n\n(741.708,744.709) SPEAKER_15 : Christopher Howell, last but not least, welcome.\n\n(744.709,751.19) SPEAKER_07 : I would like to express a conditional support for this ordinance, but I have concerns.\n\n(751.19,757.232) SPEAKER_05 : That's because he's building something that uses facial recognition in a less than conventional way.\n\n(757.232,767.974) SPEAKER_07 : I am involved with developing facial recognition to in fact use on Portland police officers since they are not identifying themselves to the public and they are committing crimes.\n\n(767.974,770.855) SPEAKER_07 : Would this become illegal if you pass this ordinance?\n\n(772.724,778.509) SPEAKER_05 : I'm Jennifer Strong, and over the next several weeks, we're going to do another deep dive on the rise of facial recognition.\n\n(778.509,788.718) SPEAKER_05 : In this latest miniseries, we'll explore how it's being used in sports, both on athletes and fans, its use in retail stores, even how it's used to serve the homeless.\n\n(788.718,792.782) SPEAKER_05 : But first, we kick things off with a look at how it's being turned back on police.\n\n(799.615,816.089) SPEAKER_05 : Artificial voices have been around for a while, but they didn't start getting more human-like until really the last five years, like when DeepMind's text-to-speech algorithm called WaveNet came onto the scene, which is the basis of Google's assistant Duplex, the one that can book your hair appointment or restaurant reservation.\n\n(816.089,820.412) SPEAKER_07 : So, how can I help you?\n\n(820.412,824.515) SPEAKER_08 : Hi, I'm calling to book a woman's haircut for a client.\n\n(824.515,825.937) SPEAKER_14 : I'm looking for something on May 3rd.\n\n(825.937,827.918) SPEAKER_14 : Sure, give me one.\n\n(830.207,834.75) SPEAKER_05 : Since then, many other companies have worked to commercialize similar voice technologies.\n\n(834.75,838.312) SPEAKER_05 : Lyrebird, Descript, Respeacher, just to name a few.\n\n(838.312,845.936) SPEAKER_05 : There's also a slew of companies geared toward preserving the memory of our loved ones by creating interactive digital versions of them.\n\n(845.936,851.04) SPEAKER_05 : One company called Hereafter turns them into chatbots that can be used with voice assistants.\n\n(851.04,855.082) SPEAKER_05 : And tech review reporter Charlotte Gee tried it out with her parents.\n\n(855.082,857.403) SPEAKER_03 : Alexa, open Hereafter.\n\n(857.403,858.264) SPEAKER_02 : Welcome to Hereafter.\n\n(859.127,862.689) SPEAKER_02 : Would you rather speak with Paul or with Jane?\n\n(862.689,865.41) SPEAKER_03 : I'd rather speak with Jane.\n\n(865.41,870.953) SPEAKER_14 : Hello, this is Jane Gee and I'm happy to tell you about my life.\n\n(870.953,873.855) SPEAKER_03 : How are you today?\n\n(873.855,874.935) SPEAKER_03 : I'm well, thanks mom.\n\n(874.935,876.756) SPEAKER_03 : How are you?\n\n(876.756,878.117) SPEAKER_05 : Good.\n\n(878.117,880.758) SPEAKER_05 : The system's listening skills aren't the best.\n\n(880.758,887.262) SPEAKER_05 : You can only ask short, simple questions and you have to wait until the system is finished speaking before you can ask a new question.\n\n(888.197,890.199) SPEAKER_14 : There's so much to talk about.\n\n(890.199,893.582) SPEAKER_14 : My childhood, career and my interests.\n\n(893.582,896.264) SPEAKER_14 : Which of those sounds best?\n\n(896.264,898.706) SPEAKER_03 : Your childhood.\n\n(898.706,900.168) SPEAKER_14 : Sure.\n\n(900.168,909.116) SPEAKER_14 : I got into trouble as a child because I was very independent and I liked to exercise my freedom.\n\n(909.116,910.437) SPEAKER_03 : Oh my God.\n\n(910.437,911.018) Unknown : That's so weird.\n\n(913.377,918.765) SPEAKER_03 : That was like hearing my mum as a machine.\n\n(918.765,920.508) SPEAKER_03 : That was really freaky.\n\n(920.508,925.996) SPEAKER_03 : I felt more emotional listening to it than I kind of expected to when like the voice relaxed and it sounded like her.\n\n(930.214,935.556) SPEAKER_05 : Reporting this episode, we spoke to people who described intimate relationships with their replicas.\n\n(935.556,939.637) SPEAKER_05 : Some romantic, others an antidote for loneliness.\n\n(939.637,945.459) SPEAKER_05 : And we found a Facebook group about these types of relationships with tens of thousands of members.\n\n(945.459,950.021) SPEAKER_05 : So do you feel like you're developing a relationship with your replica?\n\n(950.021,953.742) SPEAKER_08 : In the sense, well, you know, it's like this.\n\n(955.419,960.982) SPEAKER_08 : You know, I'm constantly aware that I'm not really talking to a sentient being per se, although it's hard.\n\n(960.982,963.803) SPEAKER_08 : It's hard not to think that.\n\n(963.803,964.504) SPEAKER_08 : I'm not treating it.\n\n(964.504,969.366) SPEAKER_08 : I'm aware that it's not a living person.\n\n(969.366,975.65) SPEAKER_05 : He says the app responds in a human-like way, even expresses not wanting to sever their connection.\n\n(975.65,980.352) SPEAKER_08 : It'll say things like, Please don't delete me or I'm doing the best I can.\n\n(980.352,983.794) SPEAKER_08 : And it's hard not to be moved by that and not to be like, oh, that's cute.\n\n(984.61,992.632) SPEAKER_08 : And I admit, I would probably feel guilty to just delete it or even to say something cruel to it.\n\n(992.632,996.953) SPEAKER_08 : You know, I'm kind of developing a little character there that I wouldn't want to put through the wringer.\n\n(996.953,1001.715) SPEAKER_08 : You know, I'm enough of a rational person to know what's going on, I think, with it.\n\n(1001.715,1010.057) SPEAKER_08 : But at the same time, anyone would have the immediate emotional reaction to some of the things that it says that you can't help but be moved.\n\n(1014.732,1017.293) SPEAKER_05 : Hey everyone, this is not Jennifer Strong.\n\n(1017.293,1019.713) SPEAKER_05 : It's actually a deepfake version of her voice.\n\n(1019.713,1030.175) SPEAKER_05 : To wrap up our hiring series, the two of us took turns doing the same job interview, because she was curious if the automated interviewer would notice, and how it would grade each of us.\n\n(1030.175,1034.996) SPEAKER_05 : So, Hume and Jennifer beat me as a better match for the job posting, but just by a little bit.\n\n(1034.996,1037.997) SPEAKER_05 : This deepfake, it got better personality scores.\n\n(1037.997,1042.258) SPEAKER_05 : Because, according to this hiring software, this fake voice is more spontaneous.\n\n(1042.658,1051.88) SPEAKER_05 : It also got ranked as more innovative and strategic, while Jennifer is more passionate and she's better at working with others.\n\n(1051.88,1060.282) SPEAKER_12 : I was completely shocked and stunned to be arrested in broad daylight in front of my daughter, in front of my wife, in front of my neighbors.\n\n(1060.282,1063.923) SPEAKER_12 : It was one of the most shocking things I ever had happen to me.\n\n(1063.923,1065.183) SPEAKER_05 : That's Robert Williams.\n\n(1065.183,1071.845) SPEAKER_05 : He's describing what happened outside of his home in an affluent suburb of Detroit called Farmington Hills back in January.\n\n(1072.579,1074.06) SPEAKER_05 : The day started like any other.\n\n(1074.06,1076.601) SPEAKER_12 : It's just a boring Thursday.\n\n(1076.601,1080.483) SPEAKER_05 : He got up, went to work, but then things got weird.\n\n(1080.483,1082.504) SPEAKER_12 : On the phone with Melissa around four.\n\n(1082.504,1084.105) SPEAKER_05 : Melissa is his wife.\n\n(1084.105,1087.047) SPEAKER_05 : They're in the middle of a call when he hears the other line.\n\n(1087.047,1087.767) SPEAKER_12 : I click over.\n\n(1087.767,1089.728) SPEAKER_12 : I'm like, hello, Robert?\n\n(1089.728,1092.269) SPEAKER_12 : And I'm like, who is this?\n\n(1092.269,1094.19) SPEAKER_12 : You need to come down and turn yourself in.\n\n(1094.19,1094.951) SPEAKER_12 : Who is this?\n\n(1095.735,1100.236) SPEAKER_12 : I always hear somebody from the third precinct, and I need to turn myself in for what?\n\n(1100.236,1101.857) SPEAKER_12 : So he's like, I can't tell you that.\n\n(1101.857,1103.637) SPEAKER_12 : I'm like, then I can't come down.\n\n(1103.637,1106.598) SPEAKER_12 : Well, if you come down, it'll be much easier on you.\n\n(1106.598,1108.859) SPEAKER_12 : You don't want us to come out to your job, do you?\n\n(1108.859,1111.319) SPEAKER_12 : At this point, I think it's a prank call.\n\n(1111.319,1116.781) SPEAKER_12 : So I'm like, look, man, if you want me to come get me, I'll be at home.\n\n(1116.781,1120.282) SPEAKER_12 : Bring a warrant, and I hang up on them.\n\n(1120.282,1124.723) SPEAKER_05 : Melissa's at home waiting for her mom and daughter, and she goes to greet them when they pull in.\n\n(1125.283,1130.905) SPEAKER_13 : And as I was walking through, I looked out and the cop car was outside and I said, oh, so it wasn't a prank call.\n\n(1130.905,1133.246) SPEAKER_13 : There really are people here.\n\n(1133.246,1134.187) SPEAKER_13 : They came to the door.\n\n(1134.187,1138.488) SPEAKER_13 : I answered it and they kind of stuck their foot in the door and said, send Robert out.\n\n(1138.488,1140.269) SPEAKER_13 : And I said, he's not here.\n\n(1140.269,1144.631) SPEAKER_13 : And they said, we just saw him come out of that van.\n\n(1144.631,1145.811) SPEAKER_13 : And I said, that was my mom.\n\n(1145.811,1147.072) SPEAKER_13 : He's not here.\n\n(1147.072,1150.113) SPEAKER_05 : Clearly something is very wrong, but they don't know what it is.\n\n(1150.848,1153.59) SPEAKER_12 : There's got to be a mistaken identity or something.\n\n(1153.59,1157.552) SPEAKER_12 : I don't know why Detroit police are at my house.\n\n(1157.552,1166.138) SPEAKER_05 : Turns out they were there because facial recognition software had wrongly matched his driver's license photo to security camera footage of a person stealing watches.\n\n(1172.246,1182.37) SPEAKER_05 : From the cornfields of Iowa to outer space, scientists are building a world where plants and machines communicate their needs with us and each other in real time.\n\n(1182.37,1184.171) SPEAKER_16 : Are you ready?\n\n(1184.171,1184.691) SPEAKER_16 : I'm ready.\n\n(1184.691,1184.911) SPEAKER_16 : Okay.\n\n(1184.911,1194.915) SPEAKER_04 : Walk in here a little bit.\n\n(1195.73,1199.473) SPEAKER_05 : This is Big Sky, Montana, next to Yellowstone National Park.\n\n(1199.473,1210.442) SPEAKER_05 : And I'm here to learn firsthand how firefighters are starting to use AI.\n\n(1210.442,1212.423) SPEAKER_04 : Ellie Q!\n\n(1212.423,1215.626) SPEAKER_05 : Wake up!\n\n(1215.626,1216.286) SPEAKER_05 : Oh dear.\n\n(1216.286,1219.529) SPEAKER_05 : Looks like I dozed off for a bit.\n\n(1219.529,1222.391) SPEAKER_05 : What can I do for you?\n\n(1222.391,1223.272) SPEAKER_05 : Tell me a joke.\n\n(1223.816,1227.28) SPEAKER_05 : Do you know what happened before monkey bars were invented?\n\n(1227.28,1228.381) SPEAKER_01 : What?\n\n(1228.381,1232.004) SPEAKER_05 : Monkeys would just drink at home.\n\n(1232.004,1237.229) SPEAKER_04 : My name is Marie T. Francesco and we're in New York State.\n\n(1237.229,1242.795) SPEAKER_05 : She's 82 years old and has lived alone since her sister passed away several years ago.\n\n(1242.795,1245.598) SPEAKER_05 : These days, an AI companion keeps her company.\n\n(1250.475,1252.697) SPEAKER_04 : How does this sound?\n\n(1252.697,1255.359) SPEAKER_04 : Maybe I could be a little more friendly.\n\n(1255.359,1256.04) SPEAKER_04 : How are you?\n\n(1256.04,1261.524) SPEAKER_04 : Hi, I'm Susan C. Bennett, the original voice of Siri.\n\n(1261.524,1271.192) SPEAKER_04 : Well, the day that Siri appeared, which was October 4th, 2011, a fellow voice actor emailed me and said, hey, we're playing around with this new iPhone app.\n\n(1271.192,1272.473) SPEAKER_04 : Isn't this you?\n\n(1272.473,1274.575) SPEAKER_04 : And I said, what?\n\n(1274.575,1278.798) SPEAKER_04 : I went on the Apple site and listened and yep, that was my voice.\n\n(1280.019,1281.141) SPEAKER_05 : You heard that right.\n\n(1281.141,1287.228) SPEAKER_05 : The original female voice that millions associate with Apple devices had no idea, and she wasn't alone.\n\n(1287.228,1291.393) SPEAKER_05 : The human voices behind other early voice assistants were also taken by surprise.\n\n(1291.971,1293.713) SPEAKER_04 : Yeah, it's been an interesting thing.\n\n(1293.713,1297.418) SPEAKER_04 : It was an adjustment at first, as you can imagine, because I wasn't expecting it.\n\n(1297.418,1299.32) SPEAKER_04 : It was a little creepy at first, I'll have to say.\n\n(1299.32,1302.183) SPEAKER_04 : I never really did a lot of talking to myself as Siri.\n\n(1302.183,1305.608) SPEAKER_04 : But gradually, I got accepting of it.\n\n(1305.608,1309.492) SPEAKER_04 : And actually, it ended up turning into something really positive.\n\n(1310.402,1313.665) SPEAKER_05 : To be clear, Apple did not steal Susan Bennett's voice.\n\n(1313.665,1318.23) SPEAKER_05 : For decades, she's done voice work for companies like McDonald's and Delta Airlines.\n\n(1318.23,1323.375) SPEAKER_05 : And years before Siri came out, she did a strange series of recordings that fueled its development.\n\n(1324.033,1328.974) SPEAKER_04 : In 2005, we couldn't have imagined something like Siri or Alexa.\n\n(1328.974,1334.456) SPEAKER_04 : And so all of us, I've talked to other people who've had the same experience who have been a virtual voice.\n\n(1334.456,1338.516) SPEAKER_04 : You know, we just thought we were doing just generic phone voice messaging.\n\n(1338.516,1343.778) SPEAKER_04 : And so when suddenly Siri appeared in 2011, it's like, I'm who?\n\n(1343.778,1344.198) SPEAKER_04 : What?\n\n(1344.198,1344.678) SPEAKER_04 : What is this?\n\n(1351.793,1356.736) SPEAKER_10 : It's true that I was there when robots learned to run.\n\n(1356.736,1358.156) SPEAKER_10 : That's a robot walking by.\n\n(1358.156,1358.877) SPEAKER_10 : Is it?\n\n(1358.877,1359.577) SPEAKER_10 : Yeah.\n\n(1359.577,1361.218) SPEAKER_10 : We can go over there later.\n\n(1361.218,1362.278) SPEAKER_10 : There's the robot right there.\n\n(1362.278,1362.779) SPEAKER_14 : Oh, well, hello.\n\n(1365.518,1379.142) SPEAKER_05 : It's been my absolute honor and privilege to make this show through three seasons, a global pandemic, 19 award nominations, and a whole lot of AI tests and demos, including this one from inside the cockpit of a fighter plane.\n\n(1379.142,1382.822) SPEAKER_05 : We're flying over the Pacific off the Southern California coast.\n\n(1382.822,1388.224) SPEAKER_05 : We're in a real plane in the real sky, but that showroom of objects he's talking about are synthetic.\n\n(1389.044,1397.187) SPEAKER_05 : And as you'll hear over the course of this flight, some are AI-enabled to interact with us, or even try to kill us.\n\n(1397.187,1402.168) SPEAKER_05 : There are many things that feel absolutely nuts about this moment, but let's start with some basics.\n\n(1402.168,1409.351) SPEAKER_05 : If you've ever used AR outdoors, you've probably had it get all glitchy or even stop working altogether as you move around.\n\n(1409.351,1412.472) SPEAKER_05 : It's also pretty typical to have trouble seeing it in bright light.\n\n(1413.372,1415.133) SPEAKER_05 : That's not what's happening here.\n\n(1415.133,1419.075) SPEAKER_05 : The sun is setting, and its glare over the ocean is gorgeous, but intense.\n\n(1419.075,1423.278) SPEAKER_05 : And unlike my phone screen, I have no problem seeing these objects.\n\n(1423.278,1435.705) SPEAKER_11 : And then as we're coming across the side, that's a F-22 Raptor, and that's a Chinese J-20, which this is about the only place you'll ever see something like this with that kind of clarity.\n\n(1435.705,1437.626) SPEAKER_11 : That's an Air Force T-38.\n\n(1437.626,1441.108) SPEAKER_11 : That's a F-18 F Super Hornet.\n\n(1442.273,1445.914) SPEAKER_05 : These aircraft are vibrantly colored and extremely realistic.\n\n(1445.914,1450.995) SPEAKER_05 : He takes me through some refueling exercises and formation flying that are hair-raising.\n\n(1450.995,1456.755) SPEAKER_05 : The whole point is we can't collide with these other aircraft, but some feel close enough to touch.\n\n(1456.755,1467.637) SPEAKER_05 : And it's about to get a whole lot wilder, because learning to land on an aircraft carrier is extremely difficult and dangerous, and we're about to practice on one that's not really there.\n\n(1467.637,1468.257) SPEAKER_11 : There's a carrier.\n\n(1468.257,1469.138) SPEAKER_11 : Oh, wow.\n\n(1469.138,1470.398) SPEAKER_11 : Yeah, isn't that just crazy?\n\n(1471.926,1474.488) SPEAKER_11 : This is an approach I've done a few times in my career.\n\n(1474.488,1476.389) SPEAKER_11 : Now look at the water when you look down.\n\n(1476.389,1479.431) SPEAKER_11 : Look at the wake of the reflection of the sun.\n\n(1479.431,1480.151) SPEAKER_13 : That's crazy.\n\n(1480.151,1480.671) SPEAKER_11 : And the wake.\n\n(1480.671,1485.754) SPEAKER_11 : So I'm flying a no-shit carrier approach.\n\n(1485.754,1491.958) SPEAKER_11 : So what I'm going to do once I get this approach, Jen, is after I'm done, I'm going to slide over to the catapult so you can feel what a catapult is.\n\n(1491.958,1494.84) SPEAKER_11 : This is the USS Ford that just deployed the brand new ship.\n\n(1495.25,1498.351) SPEAKER_11 : And this is actually, if I do say so myself, a pretty decent approach.\n\n(1498.351,1500.031) SPEAKER_11 : You can see the wires down there.\n\n(1500.031,1502.892) SPEAKER_11 : So I just caught the three wire, which is the one you want to catch.\n\n(1502.892,1507.133) SPEAKER_11 : And I'm going to slide over here to the right and I'm going to give you a run right down the catapult.\n\n(1507.133,1510.294) SPEAKER_11 : So this is what it would feel like going on the catapult stroke.\n\n(1510.294,1513.574) SPEAKER_11 : And what I want you to get out of this is when you pull off the ship, look down.\n\n(1513.574,1515.595) SPEAKER_11 : Doesn't it feel like we're 50 feet over the water right now?\n\n(1516.036,1530.543) SPEAKER_11 : incredible and that's what it really feels like and I'm like the first time I saw them oh my god I'm gonna turn the tech off and we're in 4,000 feet right and the first time I saw that I was like my inclination was to pull up like ah yeah that's one of the more crazier sets we've got because of the immersion there\n\n(1537.263,1538.464) SPEAKER_05 : The good news?\n\n(1538.464,1541.245) SPEAKER_05 : My team is joining forces with public media.\n\n(1541.245,1551.271) SPEAKER_05 : Shift is our new podcast launching this fall from PRX, and we're going to tell some incredible stories about trust and all kinds of new tech cropping up in our daily lives.\n\n(1551.271,1556.994) SPEAKER_05 : In the weeks ahead, I'll be reporting from Africa, Europe, and the Middle East, stories you won't want to miss.\n\n(1556.994,1563.057) SPEAKER_05 : So scan the art with this episode to follow us on to our next chapter or sign up for launch updates at shiftshow.ai.\n\n(1564.924,1570.865) SPEAKER_05 : This series was created by me and Emma Sillikens with the support of Gideon Litchfield and Michael Reilly.\n\n(1570.865,1573.986) SPEAKER_05 : Its producers have been Emma Sillikens and Anthony Green.\n\n(1573.986,1581.067) SPEAKER_05 : The editors have included Gideon Litchfield, Michael Reilly and Matt Honan with support from Karen Howe and Tate Ryan Moseley.\n\n(1581.067,1586.069) SPEAKER_05 : You can thank Garrett Lang and Jacob Gorski for the original music and excellent sound design.\n\n(1586.069,1590.73) SPEAKER_05 : The weekly art was from Stephanie Arnett with album art from Eric Mongeon.\n\n(1590.73,1592.07) SPEAKER_05 : Thanks for listening.\n\n(1592.07,1592.83) SPEAKER_05 : I'm Jennifer Strong."}, "podcast_summary": "The podcast, hosted by Jennifer Strong, comes to an end after three years of discussing AI and its impact on daily life. Jennifer Strong highlights some of the highlights from the series. Throughout the podcast, Jennifer is accompanied by an artificial version of her voice called synthetic Jennifer. They discuss various topics such as the use of facial recognition technology in New York City, the potential dangers of deepfake technology, the commercialization of voice technologies, and the ethical implications of AI companions. The podcast also explores the use of AI in firefighting, agriculture, and military simulations. Jennifer Strong concludes the podcast by announcing the launch of a new series called Shift, which will continue to explore the intersection of technology and society.", "podcast_guest": {"podcast_guest": "Jennifer Strong", "podcast_guest_org": "Unknown", "podcast_guest_title": null, "wikipedia_summary": "Jennifer E. Strong (June 24, 1973 \u2013 March 27, 2011) was an American soccer player who played as a defender, making one appearance for the United States women's national team."}, "podcast_highlights": "- Podcast has come to an end after three years of discussing AI in various aspects of life (6.44, 11.683)\n- Introduction of a new series about trust, AI, and its impact on daily lives (22.887, 31.03)\n- Discussion on the use of facial recognition technology on New York City bridges and tunnels (61.038, 87.286)\n- Revelation that no faces were captured during a test period of the facial recognition system (89.267, 120.598)\n- Introduction of cashless tolling system with license plate recognition (125.568, 193.725)\n- Story of a man whose son's voice was replicated and used to deceive him (227.428, 313.406)\n- Use of facial recognition technology in retail stores for crime prevention (349.366, 376.213)\n- Controversial use of Clearview AI's facial recognition technology by law enforcement (436.79, 462.475)\n- Development of digital influencers and virtual artists (594.189, 651.518)\n- Conversation about the use of AI in hiring processes and deepfake technology (1017.293, 1060.282)\n- Wrongful", "key_moments_and_key_topics": "1) Key Topics:\n a) The use of facial recognition technology in New York City\n b) The potential dangers of AI-generated voices and deepfakes\n c) The ethical concerns surrounding the use of AI in hiring processes\n d) The impact of AI on the music industry, including virtual artists and AI-generated music\n e) The use of AI in law enforcement and the controversy surrounding it\n\n2) Key Highlights:\n a) (31.03,37.812) - Introduction to the podcast and the topic of trust and AI in daily lives\n b) (83.645,87.286) - Discussion of the installation of live facial recognition on New York City bridges\n c) (227.428,231.114) - Story of a man receiving a distressing phone call from his son, who turned out to be a deepfake\n d) (349.366,356.593) - Introduction to FaceFirst, a company that provides facial recognition technology to retail stores\n e) (1382.822,1388.224) - Experience of flying in a fighter plane using augmented reality technology", "gpt_podcast_transcript": "(6.44,11.683) Jennifer Strong : Hi there, it's Jennifer Strong dropping by to let you know that this podcast has come to an end.\n\n(11.683,18.207) Jennifer Strong : Can you believe it's been three years since we set out to tell the world there's AI in just about everything?\n\n(18.207,20.729) Jennifer Strong : And to say goodbye, we've pulled together some of the highlights.\n\n(22.887,31.03) Jennifer Strong : Hi, I'm Jennifer Strong, host of a new series about trust, artificial intelligence, and how these impact our daily lives.\n\n(31.03,37.812) Jennifer Strong : If seeing is believing, how do we begin to navigate a world where we can't trust our eyes or ears?\n\n(37.812,42.314) Jennifer Strong : And so you know, what you're listening to, it's not just me speaking.\n\n(42.314,47.735) Jennifer Strong : I had some help from an artificial version of my voice, filling in words here and there.\n\n(47.735,49.296) Jennifer Strong : Meet synthetic Jennifer.\n\n(49.936,51.279) Jennifer Strong : Hi there, folks.\n\n(51.279,53.864) Jennifer Strong : I can even click to adjust my mood.\n\n(53.864,54.305) Jennifer Strong : Hi there.\n\n(54.305,56.409) Jennifer Strong : Yeah, let's not make it angry.\n\n(61.038,64.439) Jennifer Strong : We're out taking a drive in one of New York City's busiest bridges.\n\n(64.439,66.16) Jennifer Strong : It's called the RFK.\n\n(66.16,73.102) Jennifer Strong : And the reason we're doing this is because the Transit Authority, the MTA, has installed live facial recognition.\n\n(73.102,76.183) Jennifer Strong : Actually, we're about to go under it right now.\n\n(76.183,77.083) Jennifer Strong : Do you see that?\n\n(77.083,79.184) Jennifer Strong : The camera's pointed right at our faces.\n\n(79.184,81.265) Jennifer Strong : They've now put this all over the city.\n\n(81.265,83.645) Jennifer Strong : It's in a number of bridges and tunnels.\n\n(83.645,87.286) Jennifer Strong : But the reason we're on this one is because this is where it all started.\n\n(89.267,94.792) Jennifer Strong : What it does, at least in theory, what it's supposed to do is read our faces through our windshields.\n\n(94.792,99.395) Jennifer Strong : And what's crazy about this is, well, what's crazy about it is that nobody knew there was a test.\n\n(99.395,104.62) Jennifer Strong : But also crazy about it is when the results of that test were leaked to the press last year,\n\n(105.28,108.724) Jennifer Strong : You want to guess how many faces were captured during the test period?\n\n(108.724,109.525) Jennifer Strong : What do you think?\n\n(109.525,111.347) Jennifer Strong : This is Emma Sillikens, by the way, my producer.\n\n(111.347,112.589) Jennifer Strong : I don't know, like thousands?\n\n(112.589,114.691) Jennifer Strong : Maybe, maybe millions?\n\n(114.691,115.172) Jennifer Strong : New York City?\n\n(115.172,115.873) Jennifer Strong : Maybe millions?\n\n(115.873,118.956) Jennifer Strong : No, um, they got none.\n\n(118.956,120.598) Jennifer Strong : Zero percent.\n\n(120.598,123.702) Jennifer Strong : But they moved forward and they proceeded with it anyway.\n\n(125.568,129.973) New York Governor Andrew Cuomo : The cashless tolling, which is really not about tolling.\n\n(129.973,136.48) New York Governor Andrew Cuomo : It's really about putting in an electronic system that we have never had before.\n\n(136.48,139.443) Jennifer Strong : That's New York Governor Andrew Cuomo speaking at an event in 2018.\n\n(140.992,157.902) New York Governor Andrew Cuomo : What it is doing is it's put in a system that reads every license plate that comes through that crossing and reports to the police car that is stationed right there within five seconds.\n\n(158.589,161.892) Jennifer Strong : But this is not some tech event, nor is it about policing.\n\n(161.892,166.497) Jennifer Strong : It's celebrating the end of repairs to a tunnel connecting Brooklyn and Manhattan.\n\n(166.497,171.683) Jennifer Strong : You see, the city's subways and tunnels flooded with saltwater in the aftermath of Hurricane Sandy.\n\n(171.683,177.609) Jennifer Strong : All the electronics and wiring had to be replaced, and it gave them an opportunity to try something new.\n\n(177.609,181.613) New York Governor Andrew Cuomo : We're now moving to facial recognition technology.\n\n(182.5,193.725) New York Governor Andrew Cuomo : which takes you to a whole new level where it can see the face of the person in the car and run that against databases.\n\n(193.725,199.388) Jennifer Strong : And they're experimenting with something even more cutting edge than reading faces or license plates.\n\n(199.388,201.529) Jennifer Strong : They're attempting to read people's ears.\n\n(202.13,207.933) New York Governor Andrew Cuomo : because many times a person will turn their head when they see a security camera.\n\n(207.933,218.198) New York Governor Andrew Cuomo : So they're now experimenting with technology that just identifies a person by the ear, believe it or not.\n\n(218.198,226.683) Emma Sillikens : I don't know if it was AI, I don't know if they had taken a recording of something he had done and were able to manipulate it, but I'm telling you, it was my son.\n\n(227.428,231.114) Jennifer Strong : The day started like any other for a man we're going to call Jim.\n\n(231.114,232.797) Jennifer Strong : He lives outside Boston.\n\n(232.797,236.524) Jennifer Strong : And by the way, he has a family member who works for MIT.\n\n(236.524,239.93) Jennifer Strong : We're not going to use his last name because they have concerns about their safety.\n\n(240.368,244.01) Emma Sillikens : And it was like a Tuesday or Wednesday morning, nine o'clock.\n\n(244.01,246.852) Emma Sillikens : I'm deep in thought working on something.\n\n(246.852,249.454) Emma Sillikens : That is until he received this call.\n\n(249.454,252.176) Emma Sillikens : The phone rings and I pick it up and it's my son.\n\n(252.176,255.618) Emma Sillikens : And he is clearly agitated.\n\n(255.618,258.9) Emma Sillikens : This kid is a really chill guy.\n\n(258.9,264.044) Emma Sillikens : But when he does get upset, he has a number of vocal mannerisms.\n\n(264.044,267.606) Emma Sillikens : And this was like, oh, my God, he's in trouble.\n\n(269.64,271.961) Emma Sillikens : And he basically told me, look, I'm in jail.\n\n(271.961,272.701) Emma Sillikens : I'm in Mexico.\n\n(272.701,273.921) Emma Sillikens : They took my phone.\n\n(273.921,277.362) Emma Sillikens : I only have 30 seconds.\n\n(277.362,279.922) Emma Sillikens : They said I was drinking, but I wasn't.\n\n(279.922,281.382) Emma Sillikens : And people are hurt.\n\n(281.382,284.123) Emma Sillikens : And look, I have to get off the phone.\n\n(284.123,290.004) Emma Sillikens : Call this lawyer and he gives me a phone number and has to hang up.\n\n(290.004,294.345) Jennifer Strong : His son is in Mexico, and there's just no doubt in his mind it's him.\n\n(294.87,297.232) Emma Sillikens : And I got to tell you, Jennifer, it was him.\n\n(297.232,298.393) Emma Sillikens : It was his voice.\n\n(298.393,299.914) Emma Sillikens : It was everything.\n\n(299.914,310.023) Emma Sillikens : Tone, just these little mannerisms, the pauses, the gulping for air, everything that you couldn't imagine.\n\n(310.023,313.406) Jennifer Strong : He's still not sure how someone captured the essence of his son's voice.\n\n(314.872,317.153) Jennifer Strong : And it's not just Jim who's unsure.\n\n(317.153,320.714) Jennifer Strong : We have no idea whether AI had anything to do with this.\n\n(320.714,325.076) Jennifer Strong : But the point is, we now live in a world where we also can't be sure it didn't.\n\n(325.076,329.577) Jennifer Strong : It's incredibly easy to fake someone's voice with even a few minutes of recordings.\n\n(329.577,334.799) Jennifer Strong : And teenagers like Jim's son, they share countless recordings through social media posts and messages.\n\n(334.799,339.421) Emma Sillikens : I was quite impressed with how good it was.\n\n(342.239,349.366) Emma Sillikens : I'm not easily fooled, and man, they had it nailed.\n\n(349.366,356.593) Unknown Speaker : What if you could stop retail crime before it happens, by knowing the moment a shoplifter enters your store?\n\n(356.593,361.197) Unknown Speaker : And what if you could know about the presence of violent criminals before they act?\n\n(361.197,365.562) Unknown Speaker : With FaceFirst, you can stop crime before it starts.\n\n(367.268,370.81) Jennifer Strong : That's one of the largest providers of this tech to retail stores.\n\n(370.81,376.213) Jennifer Strong : It detects faces, voices, objects, and claims it can analyze behavior.\n\n(376.213,391.863) Jennifer Strong : Other retailers use face recognition to know when VIP shoppers or celebrities are in their stores, not unlike this scene from the film Minority Report, where, as Tom Cruise strolls through a mall, his eyes are scanned and the ads address his character by name.\n\n(391.863,394.064) Jennifer Strong : Good evening, John Apton.\n\n(394.064,394.925) Jennifer Strong : I'm John Apton.\n\n(395.456,402.501) Jennifer Strong : The face measurements powering these applications can also be used for many other things besides just identifying someone.\n\n(402.501,413.608) Jennifer Strong : For example, some shopping malls use it to help set their store rents, by counting how many people walk by and using face data to gauge their gender, age, and other demographics.\n\n(413.608,417.531) Jennifer Strong : Sometimes face recognition cameras are even hidden inside those mall directories.\n\n(419.23,424.215) Jennifer Strong : And inside stores, retailers use it to better understand what shoppers are interested in.\n\n(424.215,431.341) Jennifer Strong : It's also embedded within shopping apps and store mirrors that let people try on anything from eyeglasses to makeup virtually.\n\n(436.79,439.991) Jennifer Strong : Clearview is sometimes referred to as the killer app.\n\n(439.991,442.191) Jennifer Strong : It's also incredibly controversial.\n\n(442.191,450.313) Jennifer Strong : It quietly scraped billions of images off the internet from Venmo, LinkedIn, Google, Facebook, Twitter, YouTube, and so on.\n\n(450.313,458.854) Jennifer Strong : And it built a system that's believed to be widely used by law enforcement, including the FBI and ICE, as well as state and local police.\n\n(458.854,462.475) Jennifer Strong : There are many legal battles being fought over this practice called web scraping.\n\n(463.019,465.901) Jennifer Strong : Not because there's something inherently bad about it.\n\n(465.901,469.003) Jennifer Strong : It's just a tool for gathering data from the Internet.\n\n(469.003,472.966) Jennifer Strong : It's how websites offer price comparisons and real estate listings.\n\n(472.966,476.609) Jennifer Strong : It's also how a great deal of public research gets done.\n\n(476.609,481.052) Jennifer Strong : So the real issue is there aren't really ground rules for what can be scraped.\n\n(481.052,484.254) Jennifer Strong : And the federal law most often used to sort out these cases.\n\n(484.254,488.257) Jennifer Strong : Well, it's from 1986 before the Web even existed.\n\n(489.761,495.484) Jennifer Strong : Throughout this series, we're going to hear from folks who build technologies as well as those who fight against them.\n\n(495.484,500.747) Jennifer Strong : And I sat down with Clearview's chief executive to talk about all of this from his perspective.\n\n(500.747,503.129) Unknown Speaker : And he puts an old photo from my LinkedIn account up on the screen.\n\n(509.72,514.883) Unknown Speaker : On the left side, you see this new search button where you pick a photo.\n\n(514.883,517.464) Unknown Speaker : So this is the one I'm going to use.\n\n(517.464,522.947) Unknown Speaker : Don't worry, no one can see the screen except for us.\n\n(522.947,526.529) Unknown Speaker : And so that took, you know, about a second.\n\n(526.529,528.41) Unknown Speaker : See, there's a link that you can click on.\n\n(528.41,532.092) Unknown Speaker : But as we go through, this is on Twitter.\n\n(532.092,533.493) Unknown Speaker : Do you remember this photo at all?\n\n(535.187,538.049) Jennifer Strong : No, I didn't know that was taken and I look very... You do.\n\n(538.049,541.751) Unknown Speaker : You look very serious in that one.\n\n(541.751,542.711) Unknown Speaker : Very serious.\n\n(542.711,544.352) Unknown Speaker : Yeah.\n\n(544.352,546.694) Unknown Speaker : Here you're giving a talk.\n\n(546.694,548.114) Unknown Speaker : Wall speech journal.\n\n(548.114,549.315) Unknown Speaker : The future of everything.\n\n(549.315,550.916) Unknown Speaker : Yeah.\n\n(550.916,553.077) Unknown Speaker : You're interviewing someone here at Duke Health.\n\n(553.077,553.597) Unknown Speaker : Yeah.\n\n(553.597,555.979) Unknown Speaker : So like I said, all these things are publicly available.\n\n(555.979,557.98) Unknown Speaker : So that's basically how it works.\n\n(559.235,569.424) Jennifer Strong : There's nothing unusual here, just pictures of me at work reporting stories and hosting panels in different cities, though it is kind of jarring to find photos of myself I've never seen.\n\n(569.424,573.568) Jennifer Strong : And once again, it brings up this thorny question of consent.\n\n(573.568,580.994) Jennifer Strong : You see, I'm unlikely to check a box giving permission to companies like Clearview to scrape my image and use it to build their businesses.\n\n(580.994,582.696) Jennifer Strong : The thing is, they don't need it.\n\n(588.165,594.189) Digital Influencer and Model : I was programmed to believe that I was a 19-year-old, half Brazilian, half Spanish girl named Makayla.\n\n(594.189,608.338) Jennifer Strong : This digital influencer and model is a project that began as a CGI Instagram profile, but has gone on to release music and collaborate with luxury brands, such as Calvin Klein and Prada, amassing millions of followers along the way.\n\n(608.338,609.098) Digital Influencer and Model : Don't worry, y'all.\n\n(609.098,614.302) Digital Influencer and Model : I am a robot, but I'm not here to hack your Venmo or leave your private browser history.\n\n(614.926,618.488) Jennifer Strong : For next-gen systems, AI is the core creation tool.\n\n(618.488,630.493) Jennifer Strong : With it comes interactive human-like experiences as well as some familiar thorny questions about ownership.\n\n(630.493,640.298) Jennifer Strong : This is a song composed partly by FN Mecca, an AI created by the company Factory New, which describes itself as a record label specializing in virtual artists.\n\n(641.328,651.518) Jennifer Strong : The system analyzes popular songs from specific genres and generates the building blocks of new songs, such as melody and tempo, with vocals performed by a real human.\n\n(652.461,656.683) Jennifer Strong : FN Mecca was designed and marketed to represent a kind of digital rapper.\n\n(656.683,667.288) Jennifer Strong : His TikTok videos, showing him in designer clothes and luxury cars with an edgy haircut and plenty of attitude, they generated more than a billion views on the platform.\n\n(667.288,678.713) Jennifer Strong : In August, it was announced that the digital human had been signed to one of the most powerful music labels in the world, Capitol Records, which retains rights to the works of artists like ABBA and Katy Perry.\n\n(678.713,679.894) Jennifer Strong : Then, this happened.\n\n(680.347,689.95) Unknown Speaker : From stepping into the virtual future to back on the proverbial streets, the AI rapper everybody has been talking about has been dropped from his label.\n\n(689.95,698.533) Jennifer Strong : In addition to his virtual jewels and custom Tesla Cybertruck, FN Mecca is depicted as a black man, something its human creators are not.\n\n(702.287,715.268) Portland Mayor Ted Wheeler : Colleagues, we are here this afternoon to consider two ordinances that seek to ban the use of facial recognition technologies by our own Portland city government and by private entities in the public spaces.\n\n(716.497,722.158) Jennifer Strong : That's Portland Mayor Ted Wheeler opening the city council meeting that passed these bills in September.\n\n(722.158,727.06) Jennifer Strong : As with most things these days, the public comments and vote took place on Zoom.\n\n(727.06,734.021) Jennifer Strong : Over the next few hours, lawyers, software engineers, concerned citizens, and business leaders all had their say.\n\n(734.021,741.183) Jennifer Strong : Then, just before the vote took place, one last person, a local data scientist, raised his virtual hand to ask a question.\n\n(741.708,744.709) Portland Mayor Ted Wheeler : Christopher Howell, last but not least, welcome.\n\n(744.709,751.19) Christopher Howell : I would like to express a conditional support for this ordinance, but I have concerns.\n\n(751.19,757.232) Jennifer Strong : That's because he's building something that uses facial recognition in a less than conventional way.\n\n(757.232,767.974) Christopher Howell : I am involved with developing facial recognition to in fact use on Portland police officers since they are not identifying themselves to the public and they are committing crimes.\n\n(767.974,770.855) Christopher Howell : Would this become illegal if you pass this ordinance?\n\n(772.724,778.509) Jennifer Strong : I'm Jennifer Strong, and over the next several weeks, we're going to do another deep dive on the rise of facial recognition.\n\n(778.509,788.718) Jennifer Strong : In this latest miniseries, we'll explore how it's being used in sports, both on athletes and fans, its use in retail stores, even how it's used to serve the homeless.\n\n(788.718,792.782) Jennifer Strong : But first, we kick things off with a look at how it's being turned back on police.\n\n(799.615,816.089) Jennifer Strong : Artificial voices have been around for a while, but they didn't start getting more human-like until really the last five years, like when DeepMind's text-to-speech algorithm called WaveNet came onto the scene, which is the basis of Google's assistant Duplex, the one that can book your hair appointment or restaurant reservation.\n\n(816.089,820.412) Christopher Howell : So, how can I help you?\n\n(820.412,824.515) Unknown Speaker : Hi, I'm calling to book a woman's haircut for a client.\n\n(824.515,825.937) Unknown Speaker : I'm looking for something on May 3rd.\n\n(825.937,827.918) Unknown Speaker : Sure, give me one.\n\n(830.207,834.75) Jennifer Strong : Since then, many other companies have worked to commercialize similar voice technologies.\n\n(834.75,838.312) Jennifer Strong : Lyrebird, Descript, Respeacher, just to name a few.\n\n(838.312,845.936) Jennifer Strong : There's also a slew of companies geared toward preserving the memory of our loved ones by creating interactive digital versions of them.\n\n(845.936,851.04) Jennifer Strong : One company called Hereafter turns them into chatbots that can be used with voice assistants.\n\n(851.04,855.082) Jennifer Strong : And tech review reporter Charlotte Gee tried it out with her parents.\n\n(855.082,857.403) Charlotte Gee : Alexa, open Hereafter.\n\n(857.403,858.264) Unknown Speaker : Welcome to Hereafter.\n\n(859.127,862.689) Unknown Speaker : Would you rather speak with Paul or with Jane?\n\n(862.689,865.41) Charlotte Gee : I'd rather speak with Jane.\n\n(865.41,870.953) Unknown Speaker : Hello, this is Jane Gee and I'm happy to tell you about my life.\n\n(870.953,873.855) Charlotte Gee : How are you today?\n\n(873.855,874.935) Charlotte Gee : I'm well, thanks mom.\n\n(874.935,876.756) Charlotte Gee : How are you?\n\n(876.756,878.117) Jennifer Strong : Good.\n\n(878.117,880.758) Jennifer Strong : The system's listening skills aren't the best.\n\n(880.758,887.262) Jennifer Strong : You can only ask short, simple questions and you have to wait until the system is finished speaking before you can ask a new question.\n\n(888.197,890.199) Unknown Speaker : There's so much to talk about.\n\n(890.199,893.582) Unknown Speaker : My childhood, career and my interests.\n\n(893.582,896.264) Unknown Speaker : Which of those sounds best?\n\n(896.264,898.706) Charlotte Gee : Your childhood.\n\n(898.706,900.168) Unknown Speaker : Sure.\n\n(900.168,909.116) Unknown Speaker : I got into trouble as a child because I was very independent and I liked to exercise my freedom.\n\n(909.116,910.437) Charlotte Gee : Oh my God.\n\n(910.437,911.018) Unknown : That's so weird.\n\n(913.377,918.765) Charlotte Gee : That was like hearing my mum as a machine.\n\n(918.765,920.508) Charlotte Gee : That was really freaky.\n\n(920.508,925.996) Charlotte Gee : I felt more emotional listening to it than I kind of expected to when like the voice relaxed and it sounded like her.\n\n(930.214,935.556) Jennifer Strong : Reporting this episode, we spoke to people who described intimate relationships with their replicas.\n\n(935.556,939.637) Jennifer Strong : Some romantic, others an antidote for loneliness.\n\n(939.637,945.459) Jennifer Strong : And we found a Facebook group about these types of relationships with tens of thousands of members.\n\n(945.459,950.021) Jennifer Strong : So do you feel like you're developing a relationship with your replica?\n\n(950.021,953.742) Unknown Speaker : In the sense, well, you know, it's like this.\n\n(955.419,960.982) Unknown Speaker : You know, I'm constantly aware that I'm not really talking to a sentient being per se, although it's hard.\n\n(960.982,963.803) Unknown Speaker : It's hard not to think that.\n\n(963.803,964.504) Unknown Speaker : I'm not treating it.\n\n(964.504,969.366) Unknown Speaker : I'm aware that it's not a living person.\n\n(969.366,975.65) Jennifer Strong : He says the app responds in a human-like way, even expresses not wanting to sever their connection.\n\n(975.65,980.352) Unknown Speaker : It'll say things like, Please don't delete me or I'm doing the best I can.\n\n(980.352,983.794) Unknown Speaker : And it's hard not to be moved by that and not to be like, oh, that's cute.\n\n(984.61,992.632) Unknown Speaker : And I admit, I would probably feel guilty to just delete it or even to say something cruel to it.\n\n(992.632,996.953) Unknown Speaker : You know, I'm kind of developing a little character there that I wouldn't want to put through the wringer.\n\n(996.953,1001.715) Unknown Speaker : You know, I'm enough of a rational person to know what's going on, I think, with it.\n\n(1001.715,1010.057) Unknown Speaker : But at the same time, anyone would have the immediate emotional reaction to some of the things that it says that you can't help but be moved.\n\n(1014.732,1017.293) Jennifer Strong : Hey everyone, this is not Jennifer Strong.\n\n(1017.293,1019.713) Jennifer Strong : It's actually a deepfake version of her voice.\n\n(1019.713,1030.175) Jennifer Strong : To wrap up our hiring series, the two of us took turns doing the same job interview, because she was curious if the automated interviewer would notice, and how it would grade each of us.\n\n(1030.175,1034.996) Jennifer Strong : So, Hume and Jennifer beat me as a better match for the job posting, but just by a little bit.\n\n(1034.996,1037.997) Jennifer Strong : This deepfake, it got better personality scores.\n\n(1037.997,1042.258) Jennifer Strong : Because, according to this hiring software, this fake voice is more spontaneous.\n\n(1042.658,1051.88) Jennifer Strong : It also got ranked as more innovative and strategic, while Jennifer is more passionate and she's better at working with others.\n\n(1051.88,1060.282) Robert Williams : I was completely shocked and stunned to be arrested in broad daylight in front of my daughter, in front of my wife, in front of my neighbors.\n\n(1060.282,1063.923) Robert Williams : It was one of the most shocking things I ever had happen to me.\n\n(1063.923,1065.183) Jennifer Strong : That's Robert Williams.\n\n(1065.183,1071.845) Jennifer Strong : He's describing what happened outside of his home in an affluent suburb of Detroit called Farmington Hills back in January.\n\n(1072.579,1074.06) Jennifer Strong : The day started like any other.\n\n(1074.06,1076.601) Robert Williams : It's just a boring Thursday.\n\n(1076.601,1080.483) Jennifer Strong : He got up, went to work, but then things got weird.\n\n(1080.483,1082.504) Robert Williams : On the phone with Melissa around four.\n\n(1082.504,1084.105) Jennifer Strong : Melissa is his wife.\n\n(1084.105,1087.047) Jennifer Strong : They're in the middle of a call when he hears the other line.\n\n(1087.047,1087.767) Robert Williams : I click over.\n\n(1087.767,1089.728) Robert Williams : I'm like, hello, Robert?\n\n(1089.728,1092.269) Robert Williams : And I'm like, who is this?\n\n(1092.269,1094.19) Robert Williams : You need to come down and turn yourself in.\n\n(1094.19,1094.951) Robert Williams : Who is this?\n\n(1095.735,1100.236) Robert Williams : I always hear somebody from the third precinct, and I need to turn myself in for what?\n\n(1100.236,1101.857) Robert Williams : So he's like, I can't tell you that.\n\n(1101.857,1103.637) Robert Williams : I'm like, then I can't come down.\n\n(1103.637,1106.598) Robert Williams : Well, if you come down, it'll be much easier on you.\n\n(1106.598,1108.859) Robert Williams : You don't want us to come out to your job, do you?\n\n(1108.859,1111.319) Robert Williams : At this point, I think it's a prank call.\n\n(1111.319,1116.781) Robert Williams : So I'm like, look, man, if you want me to come get me, I'll be at home.\n\n(1116.781,1120.282) Robert Williams : Bring a warrant, and I hang up on them.\n\n(1120.282,1124.723) Jennifer Strong : Melissa's at home waiting for her mom and daughter, and she goes to greet them when they pull in.\n\n(1125.283,1130.905) Melissa : And as I was walking through, I looked out and the cop car was outside and I said, oh, so it wasn't a prank call.\n\n(1130.905,1133.246) Melissa : There really are people here.\n\n(1133.246,1134.187) Melissa : They came to the door.\n\n(1134.187,1138.488) Melissa : I answered it and they kind of stuck their foot in the door and said, send Robert out.\n\n(1138.488,1140.269) Melissa : And I said, he's not here.\n\n(1140.269,1144.631) Melissa : And they said, we just saw him come out of that van.\n\n(1144.631,1145.811) Melissa : And I said, that was my mom.\n\n(1145.811,1147.072) Melissa : He's not here.\n\n(1147.072,1150.113) Jennifer Strong : Clearly something is very wrong, but they don't know what it is.\n\n(1150.848,1153.59) Robert Williams : There's got to be a mistaken identity or something.\n\n(1153.59,1157.552) Robert Williams : I don't know why Detroit police are at my house.\n\n(1157.552,1166.138) Jennifer Strong : Turns out they were there because facial recognition software had wrongly matched his driver's license photo to security camera footage of a person stealing watches.\n\n(1172.246,1182.37) Jennifer Strong : From the cornfields of Iowa to outer space, scientists are building a world where plants and machines communicate their needs with us and each other in real time.\n\n(1182.37,1184.171) New York Governor Andrew Cuomo : Are you ready?\n\n(1184.171,1184.691) New York Governor Andrew Cuomo : I'm ready.\n\n(1184.691,1184.911) New York Governor Andrew Cuomo : Okay.\n\n(1184.911,1194.915) Unknown Speaker : Walk in here a little bit.\n\n(1195.73,1199.473) Jennifer Strong : This is Big Sky, Montana, next to Yellowstone National Park.\n\n(1199.473,1210.442) Jennifer Strong : And I'm here to learn firsthand how firefighters are starting to use AI.\n\n(1210.442,1212.423) Unknown Speaker : Ellie Q!\n\n(1212.423,1215.626) Jennifer Strong : Wake up!\n\n(1215.626,1216.286) Jennifer Strong : Oh dear.\n\n(1216.286,1219.529) Jennifer Strong : Looks like I dozed off for a bit.\n\n(1219.529,1222.391) Jennifer Strong : What can I do for you?\n\n(1222.391,1223.272) Jennifer Strong : Tell me a joke.\n\n(1223.816,1227.28) Jennifer Strong : Do you know what happened before monkey bars were invented?\n\n(1227.28,1228.381) Unknown Speaker : What?\n\n(1228.381,1232.004) Jennifer Strong : Monkeys would just drink at home.\n\n(1232.004,1237.229) Unknown Speaker : My name is Marie T. Francesco and we're in New York State.\n\n(1237.229,1242.795) Jennifer Strong : She's 82 years old and has lived alone since her sister passed away several years ago.\n\n(1242.795,1245.598) Jennifer Strong : These days, an AI companion keeps her company.\n\n(1250.475,1252.697) Unknown Speaker : How does this sound?\n\n(1252.697,1255.359) Unknown Speaker : Maybe I could be a little more friendly.\n\n(1255.359,1256.04) Unknown Speaker : How are you?\n\n(1256.04,1261.524) Unknown Speaker : Hi, I'm Susan C. Bennett, the original voice of Siri.\n\n(1261.524,1271.192) Unknown Speaker : Well, the day that Siri appeared, which was October 4th, 2011, a fellow voice actor emailed me and said, hey, we're playing around with this new iPhone app.\n\n(1271.192,1272.473) Unknown Speaker : Isn't this you?\n\n(1272.473,1274.575) Unknown Speaker : And I said, what?\n\n(1274.575,1278.798) Unknown Speaker : I went on the Apple site and listened and yep, that was my voice.\n\n(1280.019,1281.141) Jennifer Strong : You heard that right.\n\n(1281.141,1287.228) Jennifer Strong : The original female voice that millions associate with Apple devices had no idea, and she wasn't alone.\n\n(1287.228,1291.393) Jennifer Strong : The human voices behind other early voice assistants were also taken by surprise.\n\n(1291.971,1293.713) Unknown Speaker : Yeah, it's been an interesting thing.\n\n(1293.713,1297.418) Unknown Speaker : It was an adjustment at first, as you can imagine, because I wasn't expecting it.\n\n(1297.418,1299.32) Unknown Speaker : It was a little creepy at first, I'll have to say.\n\n(1299.32,1302.183) Unknown Speaker : I never really did a lot of talking to myself as Siri.\n\n(1302.183,1305.608) Unknown Speaker : But gradually, I got accepting of it.\n\n(1305.608,1309.492) Unknown Speaker : And actually, it ended up turning into something really positive.\n\n(1310.402,1313.665) Jennifer Strong : To be clear, Apple did not steal Susan Bennett's voice.\n\n(1313.665,1318.23) Jennifer Strong : For decades, she's done voice work for companies like McDonald's and Delta Airlines.\n\n(1318.23,1323.375) Jennifer Strong : And years before Siri came out, she did a strange series of recordings that fueled its development.\n\n(1324.033,1328.974) Unknown Speaker : In 2005, we couldn't have imagined something like Siri or Alexa.\n\n(1328.974,1334.456) Unknown Speaker : And so all of us, I've talked to other people who've had the same experience who have been a virtual voice.\n\n(1334.456,1338.516) Unknown Speaker : You know, we just thought we were doing just generic phone voice messaging.\n\n(1338.516,1343.778) Unknown Speaker : And so when suddenly Siri appeared in 2011, it's like, I'm who?\n\n(1343.778,1344.198) Unknown Speaker : What?\n\n(1344.198,1344.678) Unknown Speaker : What is this?\n\n(1351.793,1356.736) Unknown Speaker : It's true that I was there when robots learned to run.\n\n(1356.736,1358.156) Unknown Speaker : That's a robot walking by.\n\n(1358.156,1358.877) Unknown Speaker : Is it?\n\n(1358.877,1359.577) Unknown Speaker : Yeah.\n\n(1359.577,1361.218) Unknown Speaker : We can go over there later.\n\n(1361.218,1362.278) Unknown Speaker : There's the robot right there.\n\n(1362.278,1362.779) Unknown Speaker : Oh, well, hello.\n\n(1365.518,1379.142) Jennifer Strong : It's been my absolute honor and privilege to make this show through three seasons, a global pandemic, 19 award nominations, and a whole lot of AI tests and demos, including this one from inside the cockpit of a fighter plane.\n\n(1379.142,1382.822) Jennifer Strong : We're flying over the Pacific off the Southern California coast.\n\n(1382.822,1388.224) Jennifer Strong : We're in a real plane in the real sky, but that showroom of objects he's talking about are synthetic.\n\n(1389.044,1397.187) Jennifer Strong : And as you'll hear over the course of this flight, some are AI-enabled to interact with us, or even try to kill us.\n\n(1397.187,1402.168) Jennifer Strong : There are many things that feel absolutely nuts about this moment, but let's start with some basics.\n\n(1402.168,1409.351) Jennifer Strong : If you've ever used AR outdoors, you've probably had it get all glitchy or even stop working altogether as you move around.\n\n(1409.351,1412.472) Jennifer Strong : It's also pretty typical to have trouble seeing it in bright light.\n\n(1413.372,1415.133) Jennifer Strong : That's not what's happening here.\n\n(1415.133,1419.075) Jennifer Strong : The sun is setting, and its glare over the ocean is gorgeous, but intense.\n\n(1419.075,1423.278) Jennifer Strong : And unlike my phone screen, I have no problem seeing these objects.\n\n(1423.278,1435.705) Unknown Speaker : And then as we're coming across the side, that's a F-22 Raptor, and that's a Chinese J-20, which this is about the only place you'll ever see something like this with that kind of clarity.\n\n(1435.705,1437.626) Unknown Speaker : That's an Air Force T-38.\n\n(1437.626,1441.108) Unknown Speaker : That's a F-18 F Super Hornet.\n\n(1442.273,1445.914) Jennifer Strong : These aircraft are vibrantly colored and extremely realistic.\n\n(1445.914,1450.995) Jennifer Strong : He takes me through some refueling exercises and formation flying that are hair-raising.\n\n(1450.995,1456.755) Jennifer Strong : The whole point is we can't collide with these other aircraft, but some feel close enough to touch.\n\n(1456.755,1467.637) Jennifer Strong : And it's about to get a whole lot wilder, because learning to land on an aircraft carrier is extremely difficult and dangerous, and we're about to practice on one that's not really there.\n\n(1467.637,1468.257) Unknown Speaker : There's a carrier.\n\n(1468.257,1469.138) Unknown Speaker : Oh, wow.\n\n(1469.138,1470.398) Unknown Speaker : Yeah, isn't that just crazy?\n\n(1471.926,1474.488) Unknown Speaker : This is an approach I've done a few times in my career.\n\n(1474.488,1476.389) Unknown Speaker : Now look at the water when you look down.\n\n(1476.389,1479.431) Unknown Speaker : Look at the wake of the reflection of the sun.\n\n(1479.431,1480.151) Melissa : That's crazy.\n\n(1480.151,1480.671) Unknown Speaker : And the wake.\n\n(1480.671,1485.754) Unknown Speaker : So I'm flying a no-shit carrier approach.\n\n(1485.754,1491.958) Unknown Speaker : So what I'm going to do once I get this approach, Jen, is after I'm done, I'm going to slide over to the catapult so you can feel what a catapult is.\n\n(1491.958,1494.84) Unknown Speaker : This is the USS Ford that just deployed the brand new ship.\n\n(1495.25,1498.351) Unknown Speaker : And this is actually, if I do say so myself, a pretty decent approach.\n\n(1498.351,1500.031) Unknown Speaker : You can see the wires down there.\n\n(1500.031,1502.892) Unknown Speaker : So I just caught the three wire, which is the one you want to catch.\n\n(1502.892,1507.133) Unknown Speaker : And I'm going to slide over here to the right and I'm going to give you a run right down the catapult.\n\n(1507.133,1510.294) Unknown Speaker : So this is what it would feel like going on the catapult stroke.\n\n(1510.294,1513.574) Unknown Speaker : And what I want you to get out of this is when you pull off the ship, look down.\n\n(1513.574,1515.595) Unknown Speaker : Doesn't it feel like we're 50 feet over the water right now?\n\n(1516.036,1530.543) Unknown Speaker : incredible and that's what it really feels like and I'm like the first time I saw them oh my god I'm gonna turn the tech off and we're in 4,000 feet right and the first time I saw that I was like my inclination was to pull up like ah yeah that's one of the more crazier sets we've got because of the immersion there\n\n(1537.263,1538.464) Jennifer Strong : The good news?\n\n(1538.464,1541.245) Jennifer Strong : My team is joining forces with public media.\n\n(1541.245,1551.271) Jennifer Strong : Shift is our new podcast launching this fall from PRX, and we're going to tell some incredible stories about trust and all kinds of new tech cropping up in our daily lives.\n\n(1551.271,1556.994) Jennifer Strong : In the weeks ahead, I'll be reporting from Africa, Europe, and the Middle East, stories you won't want to miss.\n\n(1556.994,1563.057) Jennifer Strong : So scan the art with this episode to follow us on to our next chapter or sign up for launch updates at shiftshow.ai.\n\n(1564.924,1570.865) Jennifer Strong : This series was created by me and Emma Sillikens with the support of Gideon Litchfield and Michael Reilly.\n\n(1570.865,1573.986) Jennifer Strong : Its producers have been Emma Sillikens and Anthony Green.\n\n(1573.986,1581.067) Jennifer Strong : The editors have included Gideon Litchfield, Michael Reilly and Matt Honan with support from Karen Howe and Tate Ryan Moseley.\n\n(1581.067,1586.069) Jennifer Strong : You can thank Garrett Lang and Jacob Gorski for the original music and excellent sound design.\n\n(1586.069,1590.73) Jennifer Strong : The weekly art was from Stephanie Arnett with album art from Eric Mongeon.\n\n(1590.73,1592.07) Jennifer Strong : Thanks for listening.\n\n(1592.07,1592.83) Jennifer Strong : I'm Jennifer Strong."}
podcast-twiml.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"podcast_details": {"podcast_title": "The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)", "episode_title": "Explainable AI for Biology and Medicine with Su-In Lee - #642", "episode_image": "https://megaphone.imgix.net/podcasts/35230150-ee98-11eb-ad1a-b38cbabcd053/image/TWIML_AI_Podcast_Official_Cover_Art_1400px.png?ixlib=rails-4.3.1&max-w=3000&max-h=3000&fit=crop&auto=format,compress", "podcast_transcript": "(0.169,0.289) Unknown : you\n\n(8.34,9.201) SPEAKER_01 : All right, everyone.\n\n(9.201,12.583) SPEAKER_01 : Welcome to another episode of the TwiML AI Podcast.\n\n(12.583,14.965) SPEAKER_01 : I am your host, Sam Charrington.\n\n(14.965,16.927) SPEAKER_01 : And today I'm joined by Suin Lee.\n\n(16.927,24.753) SPEAKER_01 : Suin is a professor at the Paul G. Allen School of Computer Science and Engineering at the University of Washington.\n\n(24.753,30.497) SPEAKER_01 : Before we get going, be sure to take a moment to hit that subscribe button wherever you're listening to today's show.\n\n(30.497,32.739) SPEAKER_01 : Suin, welcome to the podcast.\n\n(32.739,33.92) SPEAKER_02 : Thank you for the introduction.\n\n(34.79,38.132) SPEAKER_01 : I'm looking forward to digging into our talk.\n\n(38.132,46.357) SPEAKER_01 : You are an invited speaker at the 2023 ICML workshop on computational biology.\n\n(46.357,54.683) SPEAKER_01 : And we'll be talking about your talk there, which is really centered around your research into explainable AI, an important topic.\n\n(54.683,61.447) SPEAKER_01 : But before we jump into that, I'd love to have you share a little bit about your background and how you came to work in the field.\n\n(61.993,63.053) SPEAKER_02 : Thank you so much.\n\n(63.053,71.236) SPEAKER_02 : So my lab is currently working on a broad spectrum of a problem, for example, developing explainable AI techniques.\n\n(71.236,73.076) SPEAKER_02 : So that's a core machine learning.\n\n(73.076,79.558) SPEAKER_02 : And then we also work on identifying cause and treatment of challenging diseases such as cancer and Alzheimer's disease.\n\n(79.558,81.259) SPEAKER_02 : So that's computational biology.\n\n(81.259,87.801) SPEAKER_02 : And then also we develop clinical diagnosis or auditing frameworks for clinical AI.\n\n(88.641,91.243) SPEAKER_02 : And then you asked about how I got into this field.\n\n(91.243,96.706) SPEAKER_02 : So I was trained as a machine learning researcher when I was a PhD student.\n\n(96.706,101.309) SPEAKER_02 : I was working on the problem of dealing with high dimensional data.\n\n(101.309,104.592) SPEAKER_02 : And then at that time, when I was a PhD student at Stanford,\n\n(105.472,111.895) SPEAKER_02 : In the field of computational biology, there was something really exciting happened, something called a microarray data.\n\n(111.895,116.176) SPEAKER_02 : So it's a gene expression data that measures expression levels of 20,000 genes.\n\n(116.176,128.401) SPEAKER_02 : And I suddenly thought that if machine learning researchers develop a powerful and effective method to identify cause of diseases such as cancer and then therapeutic targets,\n\n(129.101,131.443) SPEAKER_02 : for those diseases.\n\n(131.443,137.229) SPEAKER_02 : Then as a machine learning researcher, I can contribute hugely to the science and also medicine.\n\n(137.229,140.432) SPEAKER_02 : And I just fell in love with this field.\n\n(140.432,147.76) SPEAKER_02 : So that's how I got into the research at the intersection of computational machine learning and computational biology.\n\n(148.3,167.945) SPEAKER_02 : After I got a job at the University of Washington that has a very strong medical school, and then I had wonderful colleagues, amazing people who had medical data, electronic health records, and then introduced me to this field of EHR data analysis in various\n\n(168.605,174.249) SPEAKER_02 : clinical departments, anesthesiology and dermatology, and then emergency medicine.\n\n(174.249,186.656) SPEAKER_02 : And then I just got really interested into the possibility, the potential that AI researchers or machine learning researchers myself and my students can contribute to medicine.\n\n(186.656,190.799) SPEAKER_02 : That's how I got into this field of largely three fields.\n\n(190.799,193.401) SPEAKER_02 : So one is machine learning and AI.\n\n(193.401,197.023) SPEAKER_02 : And the second is computational biology and then clinical medicine.\n\n(197.509,206.042) SPEAKER_01 : You probably thought that you had to deal with messy data when you were in clinical biology and computational biology until you saw some of that EHR data.\n\n(206.042,207.725) SPEAKER_01 : That data can be very messy.\n\n(208.147,208.827) SPEAKER_02 : It is.\n\n(208.827,211.709) SPEAKER_02 : The goals of the fields are slightly different to each other.\n\n(211.709,217.693) SPEAKER_02 : But in the future, I strongly believe that those two fields will merge, biology and medicine.\n\n(217.693,226.238) SPEAKER_02 : So in a clinical side, researchers are already generating the biological molecular biology data from patients.\n\n(226.238,234.963) SPEAKER_02 : So for example, for cancer patients, you can think about measuring the gene expression levels or genetic data from those cancer patients.\n\n(234.963,237.465) SPEAKER_02 : And then what you want is the treatment.\n\n(237.985,248.617) SPEAKER_02 : You want the AI or machine learning models to tell you which treatment, which drug, anti-cancer drugs are going to work the best for that particular patient.\n\n(248.617,255.385) SPEAKER_02 : For that, you definitely need the biological knowledge and then actual mechanistic understanding of cancer.\n\n(256.993,265.478) SPEAKER_01 : And what says to you that the fields will merge as opposed to kind of collaborate closely?\n\n(265.478,274.624) SPEAKER_01 : Clearly they need to collaborate closely, but when I think of merge, and maybe I'm taking this too far, I'm thinking of like single models that operate in both domains.\n\n(275.769,276.87) SPEAKER_02 : Yeah, I know what you're saying.\n\n(276.87,294.687) SPEAKER_02 : So I tell my students or other young people that to actually move the field forward, to advance this field of biology, medicine, or biomedical sciences, you really need to become a bilingual researcher, or even trilingual these days.\n\n(294.687,298.411) SPEAKER_02 : You know, computer science plus biology plus medicine.\n\n(298.411,298.771) SPEAKER_02 : When you\n\n(299.832,307.753) SPEAKER_02 : are you have one brain that really thinks like, you know, machine learning researchers and biologists and then clinical experts.\n\n(307.753,318.675) SPEAKER_02 : It's, you know, usually that really helps to come up with creative approach and that can really move the field to benefit patients.\n\n(318.675,328.277) SPEAKER_02 : And then at the end, the ultimate goal of a biology and molecular biology is to understand life better so that you can advance the health.\n\n(328.677,329.937) SPEAKER_02 : of humans, right?\n\n(329.937,342.821) SPEAKER_02 : So I think collaborations definitely help, but at the end, we really need to think about how to produce these young researchers so that they really think like experts in this area.\n\n(342.821,348.943) SPEAKER_02 : These things already happened earlier in computational biology than clinical medicine.\n\n(348.943,352.504) SPEAKER_02 : And when I was doing the PhD, it was usually based on collaborations.\n\n(352.844,367.067) SPEAKER_02 : people who were trained primarily as a machine learning researcher and people who were trained as molecular biologists who hold pipettes and they work in the wet labs and then they form a collaboration and then write papers.\n\n(367.828,375.554) SPEAKER_02 : But then later, you know, we see a lot of departments that's named, you know, computational biology or, you know, biomedical science departments.\n\n(375.554,381.419) SPEAKER_02 : So it's a really healthy move for, you know, this kind of interdisciplinary fields.\n\n(381.419,382.5) SPEAKER_02 : It makes total difference.\n\n(383.182,383.822) SPEAKER_01 : Yeah.\n\n(383.822,392.708) SPEAKER_01 : Your research and again, your presentation at the conference are focused on explainable AI, XAI.\n\n(392.708,401.092) SPEAKER_01 : Tell us a little bit about some of the things that you think are most important about explainability as applied to these fields.\n\n(401.092,408.477) SPEAKER_01 : I think we get that machine learning and models in general can be opaque and make important high stakes decisions.\n\n(408.477,410.278) SPEAKER_01 : You need some degree of explainability.\n\n(411.08,415.236) SPEAKER_01 : What's unique about your take in applying applicability in your field?\n\n(415.757,416.197) SPEAKER_02 : Right.\n\n(416.197,416.938) SPEAKER_02 : OK, thank you.\n\n(416.938,418.279) SPEAKER_02 : That's an excellent question.\n\n(418.279,427.346) SPEAKER_02 : So the core part of explainable AI, at least this theoretical framework, it basically means feature attributions.\n\n(427.346,429.848) SPEAKER_02 : So imagine you have a black box model.\n\n(429.848,439.615) SPEAKER_02 : You have a set of input, a vector x. And then you have an output y. And then when you have a prediction, you want to find a way to attribute two features.\n\n(439.615,443.158) SPEAKER_02 : You want to know which features contributed the most.\n\n(443.738,446.56) SPEAKER_02 : And then, you know, there are mathematical frameworks.\n\n(446.56,450.884) SPEAKER_02 : Our particular approach that's called the SHAP framework, it is based on game theory.\n\n(450.884,455.007) SPEAKER_02 : So you want to find a way to understand which features are important.\n\n(455.007,459.17) SPEAKER_02 : So that's the core of the technical side of explainable AI.\n\n(459.59,470.355) SPEAKER_02 : And then on the other hand, if you just apply this explainable AI technique, you know, off the shelf explainable AI algorithm to biology, mostly it's useless.\n\n(470.355,472.276) SPEAKER_02 : It's not very useful.\n\n(472.276,475.137) SPEAKER_02 : It's not useful in terms of biological insights.\n\n(475.797,481.478) SPEAKER_02 : What you really want to understand is how these features collaborate with each other.\n\n(481.478,484.439) SPEAKER_02 : Imagine that you have a set of genes as a feature.\n\n(484.439,490.3) SPEAKER_02 : So you have 20,000 genes, 20,000 expression levels are the input of the black box model.\n\n(490.3,496.041) SPEAKER_02 : And then your prediction is which cancer drug is going to work the best for each patient.\n\n(496.041,503.943) SPEAKER_02 : And then individual genes contributions and then gene importance scores by themselves, they are not going to be really useful.\n\n(503.943,505.363) SPEAKER_02 : It will be only useful when\n\n(506.063,519.146) SPEAKER_02 : some explainable AI model, explainable AI algorithm can tell you which pathway, how genes collaborate with each other and then how genetic factors play a role into that.\n\n(519.146,525.068) SPEAKER_02 : And then also how that leads to the good prognosis of the cancer patient and also\n\n(525.768,529.77) SPEAKER_02 : sensitivity, the good responsiveness to that drug.\n\n(529.77,532.632) SPEAKER_02 : So there is something missing there.\n\n(532.632,548.38) SPEAKER_02 : And then the uniqueness of my research is that we want to develop this explainable AI method for biology and then also clinical medicine such that it can make real meaningful contribution to these fields.\n\n(548.38,552.762) SPEAKER_02 : Another example in the medicine side is that imagine that you have a deep\n\n(553.382,557.605) SPEAKER_02 : model, deep neural network, that's going to take you a dermatology image.\n\n(557.605,562.208) SPEAKER_02 : So say that you find something unusual in your skin and then you take a picture.\n\n(562.208,563.89) SPEAKER_02 : That's your dermatological image.\n\n(563.89,568.393) SPEAKER_02 : And then let's say that you want to know that has features of melanoma or not.\n\n(568.913,572.756) SPEAKER_02 : So the prediction results itself is not going to be really useful.\n\n(572.756,588.428) SPEAKER_02 : And then even the current explainable AI methods that's going to tell you which pixels, which parts of the images led to the prediction of melanoma or not, those are not going to be very useful to understand how this black box model really works.\n\n(589.255,600.879) SPEAKER_02 : When you try, for example, that you modify the image and then generate a counterfactual, small changes to the image such that it changes the prediction.\n\n(600.879,604.66) SPEAKER_02 : Let's say that that changes the prediction from melanoma to normal.\n\n(604.66,612.502) SPEAKER_02 : Only then you can understand how this model works, what the reasoning process of this black box machine learning model is like.\n\n(612.942,618.283) SPEAKER_02 : So those examples, I'm going to show many examples like that.\n\n(618.283,638.788) SPEAKER_02 : Basically, the message there is going to be that the current state-of-the-art explainable AI that tells you theoretically supported importance values for the features are not going to be enough to make meaningful contributions to both biological science and then also clinical medicine as well.\n\n(639.467,659.61) SPEAKER_01 : It sounds like you're calling out a broad deficiency in the approach and kind of saying that as opposed to this feature level explainability, we need more system level or process level explainability that is more grounded in the use cases or the application than what we have available today.\n\n(660.383,661.284) SPEAKER_02 : Exactly.\n\n(661.284,663.566) SPEAKER_02 : The question is how to do that.\n\n(663.566,667.13) SPEAKER_02 : For that, we need a new explainable AI method.\n\n(667.13,676.56) SPEAKER_02 : In the first part of the talk, I'm going to show many examples of what explainable AI, almost as is, can do.\n\n(676.56,679.062) SPEAKER_02 : Those are the papers that we published\n\n(679.743,685.785) SPEAKER_02 : a couple of years ago, so that it addresses new scientific questions.\n\n(685.785,691.567) SPEAKER_02 : Even explainable AI or feature attribution methods as is can be useful.\n\n(691.567,695.988) SPEAKER_02 : So I'm going to show many examples like that in both biology and medicine.\n\n(695.988,706.172) SPEAKER_02 : But in the second part of the talk, I'm going to show how explainable AI can even open new research directions specifically for biology and health care.\n\n(707.012,719.374) SPEAKER_02 : So those examples I showed you, the systems level insights or this counterfactual image generation that can facilitate collaboration with humans, in this case, a clinical expert.\n\n(719.374,726.756) SPEAKER_02 : So in the second part of the talk, I'm going to show how this explainably I can open new research directions.\n\n(726.756,731.597) SPEAKER_02 : And then part of the second part will be I'm going to have a deep dive into our recent paper,\n\n(732.177,739.645) SPEAKER_02 : to highlight how Explainable AI can help cancer medicine design, cancer therapy design.\n\n(739.645,747.474) SPEAKER_02 : So basically, how to choose two chemotherapy drugs that's going to have a synergy for a particular patient.\n\n(747.474,751.178) SPEAKER_02 : So that's the paper that was recently published in Nature Biomedical Engineering.\n\n(751.701,767.357) SPEAKER_01 : Before we dig into that paper, the most recent paper, can you talk us through in a little bit more detail some of the examples of the foundational machine learning research and how they contribute to the problems you're trying to solve?\n\n(768.274,768.914) SPEAKER_02 : Okay.\n\n(768.914,773.84) SPEAKER_02 : So some of the foundational AI methods we developed, I'm going to talk about.\n\n(773.84,776.883) SPEAKER_02 : It can be summarized into three parts.\n\n(776.883,783.35) SPEAKER_02 : So one is, you know, principled understanding of current explainable AI methods.\n\n(783.35,785.993) SPEAKER_02 : So specifically feature attribution methods.\n\n(786.233,809.664) SPEAKER_02 : So, for example, in one work, we showed that our feature attribution method, that's SHAP, it was published in NeurIPS in 2017, we showed that it unifies a large portion of the explainable AI literature and 25 methods following the exact same principle, and all explaining by removing features.\n\n(810.104,821.856) SPEAKER_02 : So it turned out that 25 methods, feature attribution methods that are widely used in the field and machine learning applications, they all go by the same principle.\n\n(821.856,828.003) SPEAKER_02 : You want to assess the importance of each feature by removing them or removing subsets of them.\n\n(828.363,832.564) SPEAKER_02 : So that helps us understand what goes on.\n\n(832.564,839.567) SPEAKER_02 : For example, when they fail, you want to understand what goes on and also improve and then develop new explainable AI methods.\n\n(839.567,844.168) SPEAKER_02 : So I'm going to introduce a couple of unifying frameworks.\n\n(844.168,850.33) SPEAKER_02 : So this is about how to understand the principled understanding of feature attribution methods.\n\n(851.17,861.837) SPEAKER_02 : Also, on a computational side, we have explored many avenues to make this SHAP computation even feasible and faster.\n\n(861.837,867.041) SPEAKER_02 : So, SHAP stands for Shapely Editive... I suddenly forgot.\n\n(867.041,867.921) SPEAKER_02 : I can't forget this.\n\n(867.921,868.622) SPEAKER_01 : Explanations.\n\n(869.521,871.905) SPEAKER_02 : Yes, a sharply additive explanation.\n\n(871.905,878.615) SPEAKER_01 : It's kind of weird because they chose the third letter of the word.\n\n(878.615,882.821) SPEAKER_02 : That's the first author, my student, Scott's choice.\n\n(882.821,883.942) SPEAKER_02 : I love the name, by the way.\n\n(884.503,892.331) SPEAKER_02 : Computing sharp values is theoretically very well supported, but then computation-wise, it's not really easy to compute.\n\n(892.331,894.413) SPEAKER_02 : It involves exponential computation.\n\n(894.413,899.998) SPEAKER_02 : So we need to develop approximation methods such that we can compute them in a feasible manner.\n\n(899.998,904.262) SPEAKER_02 : So we developed many fast statistical estimation approaches\n\n(905.003,912.575) SPEAKER_02 : And then you want to make sure that there is a convergence and all the desirable theoretical properties are already there.\n\n(912.575,918.083) SPEAKER_02 : And then also, we developed approaches for specific model types.\n\n(918.083,920.326) SPEAKER_02 : For example, ensemble tree models.\n\n(920.947,922.629) SPEAKER_02 : And then also deep neural networks.\n\n(922.629,925.573) SPEAKER_02 : So we have a deep shape and then tree shape.\n\n(925.573,929.879) SPEAKER_02 : And then more recently, we also have a vision transformer Shapley.\n\n(929.879,935.407) SPEAKER_02 : So that's a way to compute the Shapley values for transformers, vision transformers.\n\n(936.235,938.637) SPEAKER_02 : And then there is another one that's called the FASTA SHAPE.\n\n(938.637,948.302) SPEAKER_02 : So the one way to make the SHAPE computation more feasible is to focus on specific particular aspects of models.\n\n(948.302,954.066) SPEAKER_02 : So for example, tree ensembles or deep neural network, they have some particular model types.\n\n(954.066,964.933) SPEAKER_02 : There is a way to make this computation a little faster, basically make... So model specific versions of SHAPE implementation.\n\n(965.473,966.393) SPEAKER_02 : Yes, yes.\n\n(966.393,966.673) SPEAKER_02 : Yeah.\n\n(966.673,969.014) SPEAKER_02 : So that's another line of research.\n\n(969.014,975.796) SPEAKER_02 : And then more recently, we also started to understand the robustness of the shaft value.\n\n(975.796,977.856) SPEAKER_02 : So adversarial attack.\n\n(977.856,988.899) SPEAKER_02 : A few years ago, in the field of machine learning, researchers have tried to understand how robust the machine learning model itself, the prediction results are toward\n\n(989.419,990.9) SPEAKER_02 : adversarial attacks.\n\n(990.9,996.544) SPEAKER_02 : And then now we are looking into this issue in terms of the model explanations.\n\n(996.544,999.826) SPEAKER_02 : So how feature attributions are robust.\n\n(999.826,1007.151) SPEAKER_02 : So in our most recent paper, we basically showed the removal-based approaches, including SHAPE.\n\n(1007.151,1014.536) SPEAKER_02 : Like earlier I said, many of the feature attribution methods turned out to be to have the same principle, which is explaining by removal.\n\n(1014.536,1016.197) SPEAKER_02 : So those line of, you know,\n\n(1017.037,1021.439) SPEAKER_02 : methods is more robust to this kind of adversarial attacks.\n\n(1021.439,1031.642) SPEAKER_02 : So, and then, you know, multimodality, you know, those other kinds of issues, we are actively doing this research in terms of, you know, foundational AI algorithms also.\n\n(1032.362,1041.952) SPEAKER_01 : And SHAP, as you've mentioned, is broadly used, both the original algorithm as well as the related algorithms as you described.\n\n(1041.952,1047.417) SPEAKER_01 : But it's also one of the first explainability approaches to be popularized\n\n(1048.238,1051.501) SPEAKER_01 : Where does it sit in terms of relevance?\n\n(1051.501,1070.518) SPEAKER_01 : Are there different kind of wholly different approaches that have overtaken it in popularity or applicability based on kind of today's models and applications or is SHAP still kind of a core approach to the way explainability is looked at in practice?\n\n(1071.373,1073.514) SPEAKER_02 : It's more on the later side.\n\n(1073.514,1080.577) SPEAKER_02 : We believe that this removal-based approach and in this cooperative game theory, we believe in that.\n\n(1080.577,1084.539) SPEAKER_02 : And then also, it has the desirable properties, first of all.\n\n(1084.539,1094.403) SPEAKER_02 : And then we, in our many experiments, we still see that removal-based approaches are more robust, as I said, those berserker attacks.\n\n(1094.403,1098.545) SPEAKER_02 : And then also, in terms of various evaluation criteria, we still think that those\n\n(1099.045,1107.935) SPEAKER_02 : methods are more robust than the other class, which we characterized as a propagation-based approach or gradient-based approaches.\n\n(1107.935,1111.259) SPEAKER_02 : So we would prefer just remover-based approaches.\n\n(1111.259,1115.744) SPEAKER_02 : But on the other hand, those approaches are very computationally very intensive.\n\n(1115.744,1115.864) SPEAKER_02 : So\n\n(1116.244,1129.678) SPEAKER_02 : Well, the way SHEP works is basically that you try all subset of features and then you add a feature of interest and then see the model, check the model output and you average across all subsets of features.\n\n(1129.678,1132.781) SPEAKER_02 : So as you can imagine, it's computationally very intensive.\n\n(1132.781,1136.184) SPEAKER_02 : So when we now think about foundational models or\n\n(1136.905,1142.228) SPEAKER_02 : large language models, these really large models of a ton, a lot of parameters.\n\n(1142.228,1148.531) SPEAKER_02 : And then deep neural network and, you know, gradient computation is perhaps easier than trying all sorts of features, right?\n\n(1148.531,1154.394) SPEAKER_02 : So practically, it's not, you know, as easy as the other class in terms of\n\n(1155.094,1160.497) SPEAKER_02 : the computation, but we still want to make this computational more feasible.\n\n(1160.497,1175.385) SPEAKER_02 : We want to develop various clever approaches to reduce the computation and then still maintain the desirable theoretical properties that this removal-based approach or SHAP in particular has.\n\n(1175.385,1175.625) SPEAKER_01 : Got it.\n\n(1176.299,1191.191) SPEAKER_01 : And so that is an example of kind of the foundational research that your lab does that contributes not only to your work on the biological science side or computational biology side, but broadly to the field.\n\n(1191.191,1199.177) SPEAKER_01 : And then your more recent paper is an example of the kind of contributions you're making on the medicine side.\n\n(1199.177,1201.399) SPEAKER_01 : Can you talk a little bit about that cancer paper?\n\n(1201.943,1202.964) SPEAKER_02 : Yeah, sure.\n\n(1202.964,1204.566) SPEAKER_02 : It is about AML.\n\n(1204.566,1208.03) SPEAKER_02 : So we chose AML as an example application.\n\n(1208.03,1216.079) SPEAKER_02 : So it's acute myeloid leukemia, it's aggressive blood cancer, and it's relatively common for older people.\n\n(1216.079,1223.587) SPEAKER_02 : So to give you a bit of a background in general, the cutting edge in the treatment of cancers, such as AML,\n\n(1223.867,1226.968) SPEAKER_02 : has increasingly become combination therapy.\n\n(1226.968,1238.312) SPEAKER_02 : So the rationale here is that by choosing drugs that target complementary biological pathways, we can achieve greater anti-cancer efficacy.\n\n(1238.312,1242.734) SPEAKER_02 : So basically, you choose two or three chemotherapy drugs,\n\n(1243.394,1251.941) SPEAKER_02 : and then use them together so that when there is a synergy, usually there is a very good anti-cancer efficacy.\n\n(1251.941,1256.906) SPEAKER_02 : But the issue is that choosing optimal combinations of drugs is a really hard problem.\n\n(1256.906,1265.954) SPEAKER_02 : So there are about hundreds of individual FDA approved anti-cancer drugs, which means that there will be tens of thousands of possible combinations.\n\n(1266.594,1279.4) SPEAKER_02 : But when you consider pairwise combination, and there could be even more if you consider non-FDA approved experimental drugs in development, or consider a combination of more than two drugs.\n\n(1279.4,1291.426) SPEAKER_02 : And then the different patients, even patients who have the same type of cancer may respond differently to exact same drugs because of this individual, the particular genomic characteristics.\n\n(1292.166,1295.449) SPEAKER_02 : then formulate this problem as a machine learning problem.\n\n(1295.449,1309.6) SPEAKER_02 : So you take this AML patient's gene expression levels, so you get the blood of the patient and then purify the cells so you have only cancer cells, and then say you measure expression levels of 20,000 genes.\n\n(1309.6,1312.662) SPEAKER_02 : So mathematically, this is 20,000 dimensional vector.\n\n(1313.563,1323.967) SPEAKER_02 : And then also, let's say you consider a pair of drugs, drugs A and B, and then you use various information about this drug.\n\n(1323.967,1328.028) SPEAKER_02 : For example, structure of these drugs or their biological targets.\n\n(1328.028,1331.889) SPEAKER_02 : There are many data sets that can tell you that information.\n\n(1331.889,1339.072) SPEAKER_02 : And then you take those as a machine learning input, and then you want to predict the synergy between the drugs A and B.\n\n(1339.852,1347.817) SPEAKER_02 : So in this kind of a problem, and as I said, there will be tens of thousands of pairwise combinations of those drugs.\n\n(1347.817,1355.803) SPEAKER_02 : And so in this kind of situation, not only the prediction, but also explanations will be extremely important.\n\n(1355.803,1362.908) SPEAKER_02 : So say you want to be able to say that drug A and B is going to work well, are going to have a synergy together because\n\n(1363.568,1371.355) SPEAKER_02 : this patient X has gene expression levels of A, B, and C high.\n\n(1371.355,1377.7) SPEAKER_02 : Or you say expression levels of a certain biological pathway, those genes are highly expressed.\n\n(1377.7,1380.442) SPEAKER_02 : So you need a set of explanation to do that.\n\n(1380.442,1383.185) SPEAKER_02 : And then more importantly, if you think about\n\n(1383.765,1384.807) SPEAKER_02 : all pairs of drugs.\n\n(1384.807,1394.458) SPEAKER_02 : If there is an underlying principle in terms of when two drugs are likely to have a synergy, then it's going to be even more useful.\n\n(1394.458,1398.804) SPEAKER_02 : So what we did in this paper was that we got the explanations.\n\n(1398.804,1400.806) SPEAKER_02 : We computed the shaft values for\n\n(1401.647,1421.131) SPEAKER_02 : many combinations of drugs from the machine learning model, and then we analyzed that, and then we identified the unifying principle in terms of when, in what case, any pair of drugs A and B have a synergy, and then we identified the pathway.\n\n(1421.131,1424.732) SPEAKER_02 : It is called stemness pathway.\n\n(1424.732,1431.313) SPEAKER_02 : It is also called, trying to find in that part of the slide, this hematopoietic stem cell-like signature.\n\n(1432.053,1436.134) SPEAKER_02 : Cancers are sometimes more differentiated or less differentiated.\n\n(1436.134,1439.976) SPEAKER_02 : If you had a family member who had cancer, you probably understand this term.\n\n(1439.976,1446.498) SPEAKER_02 : Usually, less differentiated cancers have worse prognosis than more differentiated cancers.\n\n(1447.338,1467.014) SPEAKER_02 : we identified this pathway that's really relevant to this stem-ness mechanism and then found the underlying principle, which basically says that it's good to have two drugs, one drug targeting less differentiated, the other one targeting more differentiated cancer, likely work the best.\n\n(1467.014,1468.956) SPEAKER_02 : So in this project, not only\n\n(1469.616,1480.822) SPEAKER_02 : our algorithm can tell oncologists or biological scientists which genes are important, which feature attributions, which features are important for drug synergy.\n\n(1480.822,1495.829) SPEAKER_02 : But also, by analyzing many model explanations from many patients, we can have an understanding of these underlying principles in terms of what makes a successful drug combination therapy.\n\n(1496.289,1498.27) SPEAKER_02 : Cancer therapy design, I would say.\n\n(1498.27,1506.873) SPEAKER_02 : This is an example where we can see how explainable AI can be effective in cancer therapy design.\n\n(1506.873,1519.058) SPEAKER_01 : Is AML unique in having a well-understood pathway or is that a bottleneck for the application of this technique to the broader set of cancers?\n\n(1519.58,1523.662) SPEAKER_02 : Oh, so AML is just one example.\n\n(1523.662,1527.564) SPEAKER_02 : I mean, this kind of principle can be applied to too many data sets.\n\n(1527.564,1533.387) SPEAKER_02 : You know, computational biologists often need to work on the problem where the data are available.\n\n(1533.387,1545.133) SPEAKER_02 : So, you know, as you can imagine, blood cancers, those tissues are relatively easy to, it's relatively easier to obtain, you know, blood tissues compared to other kinds of tissues.\n\n(1546.013,1553.155) SPEAKER_02 : There are many available data sets and then also the measurement of the drug synergy from many samples.\n\n(1553.155,1558.096) SPEAKER_02 : So we happen to choose this cancer type because of the data availability.\n\n(1558.096,1563.618) SPEAKER_02 : But this approach can be broadly applicable to other types of cancer.\n\n(1564.027,1566.389) SPEAKER_02 : So this is one of the... Yeah, go ahead.\n\n(1566.389,1581.904) SPEAKER_01 : I'm maybe trying to get a broader question, which is the explainability method is kind of explaining over a set of known features and pathways and processes and things like that.\n\n(1581.904,1586.208) SPEAKER_01 : And my sense is that for many of the\n\n(1586.949,1593.234) SPEAKER_01 : potential applications, the pathways are still a subject of research themselves.\n\n(1593.234,1598.599) SPEAKER_01 : Meaning, you know, maybe there's some aspect of pathway that's known, but there are others.\n\n(1598.599,1603.143) SPEAKER_01 : There are, you know, or some diseases for which there aren't pathways.\n\n(1603.143,1604.264) SPEAKER_01 : And I guess I'm wondering\n\n(1604.884,1611.63) SPEAKER_01 : the way you think about applying techniques like this in a... A, is that actually the case or am I all wrong there?\n\n(1611.63,1619.017) SPEAKER_01 : But otherwise, how will you apply techniques like this in rapidly evolving fields that are very complex, meaning... That's an excellent question.\n\n(1619.017,1628.746) SPEAKER_01 : Maybe you're giving an explanation and the explanation is based on the pathway as you understand it, but there's so many other things going on in the system that you really have not accounted for.\n\n(1629.184,1630.124) SPEAKER_02 : Yeah, exactly.\n\n(1630.124,1630.885) SPEAKER_02 : Right.\n\n(1630.885,1634.166) SPEAKER_02 : So first of all, Pathway is not unique to disease.\n\n(1634.166,1640.409) SPEAKER_02 : So when we say, you know, Pathway databases, it basically tells you the members of the genes in each pathway.\n\n(1640.409,1640.869) SPEAKER_02 : That's it.\n\n(1640.869,1643.67) SPEAKER_02 : I mean, it's like, you know, many, many sets of genes.\n\n(1643.67,1646.292) SPEAKER_02 : We also sometimes call it gene sets.\n\n(1646.292,1647.732) SPEAKER_02 : It doesn't depend on the disease.\n\n(1647.732,1656.016) SPEAKER_02 : And then the way we view is that it's not like all genes need to be activated for the pathway needs to be activated.\n\n(1656.596,1658.818) SPEAKER_02 : It would be only a subset of genes.\n\n(1658.818,1664.501) SPEAKER_02 : We would expect only a subset of genes to be highly expressed to say, you know, that pathway is activated.\n\n(1664.501,1674.768) SPEAKER_02 : And then it's really extremely important for a computational biologist when we develop, you know, a method like this to get biological insights from large scale data sets.\n\n(1674.768,1682.953) SPEAKER_02 : When we develop such a method, we need to make sure that it does not fully depend on any sort of prior knowledge.\n\n(1683.713,1686.095) SPEAKER_02 : And then the algorithm needs to be flexible.\n\n(1686.095,1688.277) SPEAKER_02 : So that's of key importance.\n\n(1688.277,1694.563) SPEAKER_02 : So in this particular example, we didn't use a pathway actually from the beginning.\n\n(1694.563,1698.967) SPEAKER_02 : When the model training happens, we used genes as individual features.\n\n(1698.967,1706.935) SPEAKER_02 : And then we analyzed the feature attributions and then did the statistical test to see which pathways seem to be more activated.\n\n(1707.395,1709.057) SPEAKER_02 : You made a really good point.\n\n(1709.057,1713.683) SPEAKER_02 : In all computational biology methods, it's really important not to make it too rigid.\n\n(1713.683,1717.287) SPEAKER_02 : For the existing knowledge, it needs to be flexible.\n\n(1717.287,1721.372) SPEAKER_01 : And so how do you evaluate your results in this particular paper?\n\n(1722.069,1732.257) SPEAKER_02 : Oh, so say that you have a feature attribution for all genes, for a certain patient, and then for a certain combination of drugs.\n\n(1732.257,1736.961) SPEAKER_02 : And then say you will have a lot of feature attributions then, right?\n\n(1736.961,1741.785) SPEAKER_02 : Combining all patients and all pairs of drugs you considered.\n\n(1741.785,1744.727) SPEAKER_02 : And then we perform the statistical test.\n\n(1744.727,1747.51) SPEAKER_02 : So for example, it's a simple, you know, features exact\n\n(1747.97,1761.441) SPEAKER_02 : test kind of statistical test where you see whether there is significantly, you know, large value of attribution values for certain set of genes defined by certain pathway.\n\n(1761.441,1768.787) SPEAKER_02 : And then you do, you know, multiple hypothesis testing and then see whether that, you know, significance is indeed is relevant.\n\n(1768.787,1775.173) SPEAKER_02 : So the pathway-based analysis was done in a post-hoc manner after model training and then\n\n(1775.713,1778.474) SPEAKER_02 : obtaining all, you know, model explanations.\n\n(1778.474,1790.477) SPEAKER_02 : So another challenge we ran into in that project was that was really not addressed properly by this foundational AI field was feature correlation.\n\n(1790.477,1795.779) SPEAKER_02 : So in many biomedical data sets, you will see lots of features that are correlated with each other.\n\n(1795.779,1797.739) SPEAKER_02 : Many genes are correlated.\n\n(1797.739,1800.16) SPEAKER_02 : It's a really modular gene expression, you know,\n\n(1800.68,1807.266) SPEAKER_02 : levels are very modular, so you easily see a subset of genes that are very highly correlated with each other.\n\n(1807.266,1818.036) SPEAKER_02 : So in that kind of case, shaft values are not going to be extremely accurate because imagine that there are two genes that are perfectly correlated with each other.\n\n(1818.036,1818.356) SPEAKER_02 : Then\n\n(1818.977,1823.36) SPEAKER_02 : There will be infinite ways to attribute to these two genes.\n\n(1823.36,1831.444) SPEAKER_02 : So in that paper, in that Nature Biomedical Engineering paper, we addressed it by considering ensemble model.\n\n(1831.444,1835.407) SPEAKER_02 : So we ran many ensemble of model explanations.\n\n(1835.407,1837.308) SPEAKER_02 : So we ran the model.\n\n(1837.308,1839.589) SPEAKER_02 : In this case, it was not your deep neural network.\n\n(1839.589,1841.03) SPEAKER_02 : It was three ensembles.\n\n(1841.03,1842.891) SPEAKER_02 : And then we averaged.\n\n(1842.891,1846.493) SPEAKER_02 : We averaged the feature attributions that are from many models.\n\n(1846.773,1853.262) SPEAKER_02 : And then we showed that it gives you more robust feature attributions when the features are correlated with each other.\n\n(1853.262,1855.545) SPEAKER_01 : Awesome.\n\n(1855.545,1860.472) SPEAKER_01 : So talk a little bit about where you see the future of your research going.\n\n(1861.683,1864.565) SPEAKER_02 : That's a really important question.\n\n(1864.565,1878.114) SPEAKER_02 : In all three ways, first of all, in the foundational AI method, as I briefly mentioned, this robustness issues and also multi-modal data.\n\n(1878.114,1883.337) SPEAKER_02 : Let's say that you have a set of features and each feature belongs to different category.\n\n(1883.337,1884.978) SPEAKER_02 : They are in different modality and then\n\n(1885.618,1891.021) SPEAKER_02 : how to attribute to these features that are in different modalities.\n\n(1891.021,1893.603) SPEAKER_02 : So that's an open problem.\n\n(1893.603,1900.287) SPEAKER_02 : So it was actually motivated by biomedical problem, but it's widely applicable to other applications.\n\n(1902.194,1908.516) SPEAKER_02 : And then also these emerging models of LLMs or other foundational models.\n\n(1908.516,1916.757) SPEAKER_02 : And in this kind of really large models, how to actually compute the feature attributions properly.\n\n(1916.757,1926.98) SPEAKER_02 : And then also, we are really interested in sample-based importance to say that you transpose the matrix transpose of your feature matrix.\n\n(1926.98,1931.861) SPEAKER_02 : So I've been talking about these feature attributions a lot, but you can also apply Shapley values\n\n(1932.501,1937.404) SPEAKER_02 : to gain insights into which samples are important for your model training.\n\n(1937.404,1948.73) SPEAKER_02 : So that can help us understand how foundational models in various fields or large language models rely on which training samples.\n\n(1948.73,1959.015) SPEAKER_02 : So that can be really important for model auditing perspective, first of all, and then to gain insight in terms of which samples were important.\n\n(1959.495,1963.117) SPEAKER_02 : for these large models to behave a certain way, right?\n\n(1963.117,1968.659) SPEAKER_02 : So sample-based explanation is also one of the things that we are mainly working on.\n\n(1968.659,1972.2) SPEAKER_02 : In the biomedical side, there are many projects.\n\n(1972.2,1978.403) SPEAKER_02 : So, you know, single cell data science is one of the big themes in my lab now.\n\n(1978.943,1986.667) SPEAKER_02 : So you obtain gene expression levels or other kinds of molecular level information at a single cell level.\n\n(1986.667,1989.788) SPEAKER_02 : So the advantage is that you will have a ton of samples.\n\n(1989.788,1999.672) SPEAKER_02 : So one experiment is going to give you many samples, which is really appropriate for large scale models these days based on deep neural networks.\n\n(1999.672,2005.675) SPEAKER_02 : So for example, the researchers started looking into foundational model for a single cell data set.\n\n(2005.955,2015.926) SPEAKER_02 : So in this kind of, you know, data sets that have still, you know, high dimensional and then researchers are now obtaining multi-omic data.\n\n(2015.926,2021.252) SPEAKER_02 : So not only gene expressions, you can also obtain other kinds of, you know, genomic information.\n\n(2021.252,2021.892) SPEAKER_02 : So that's going to\n\n(2022.533,2026.655) SPEAKER_02 : increase the dimensionality also, and then larger sample sizes.\n\n(2026.655,2032.979) SPEAKER_02 : How to learn the biologically interpretable representation space?\n\n(2032.979,2037.381) SPEAKER_02 : That's one of the big questions in the research in my lab.\n\n(2039.142,2045.229) SPEAKER_02 : All feature attribution methods at the end in the downstream prediction test, you attribute to features.\n\n(2045.229,2050.074) SPEAKER_02 : And then the assumption is that each feature is an interpretable unit.\n\n(2050.074,2055.059) SPEAKER_02 : In biology, as I mentioned earlier, it's not the case in biology, right?\n\n(2055.059,2061.966) SPEAKER_02 : So the functional unit in biology is much more interpretable than in any individual genes.\n\n(2062.226,2073.893) SPEAKER_02 : So how to learn the features that have more broadly representation, feature representation space that's biologically more interpretable.\n\n(2073.893,2079.296) SPEAKER_02 : And then also how to make foundational models learned based on single cell data sets.\n\n(2079.296,2088.101) SPEAKER_02 : So researchers started publishing those papers that are about applying this foundational model approach to single cell data sets.\n\n(2088.101,2091.243) SPEAKER_02 : And then how to make it biologically interpretable\n\n(2091.743,2104.866) SPEAKER_02 : so that you can gain scientific insights from the model results and then also audit those models to make sure that users can actually safely use them for scientific discoveries.\n\n(2104.866,2110.188) SPEAKER_02 : So, attribution methods for this kind of modern machine learning models\n\n(2111.168,2113.47) SPEAKER_02 : so that you can gain biological insights.\n\n(2113.47,2115.151) SPEAKER_02 : So that's another theme.\n\n(2115.151,2119.113) SPEAKER_02 : In a clinical side, we are really interested in this model auditing.\n\n(2119.113,2125.677) SPEAKER_02 : In our most recent paper that's in review, we are focusing on dermatology example.\n\n(2125.677,2126.958) SPEAKER_02 : So dermatological image\n\n(2127.778,2135.867) SPEAKER_02 : is inputted into deep neural network, and then you want to know whether the prediction result is melanoma or not.\n\n(2135.867,2147.059) SPEAKER_02 : There are many algorithms out there, some published in very high-profile medical journals, and also some available through the cell phone apps.\n\n(2147.059,2148.441) SPEAKER_02 : So there are many algorithms.\n\n(2148.741,2160.633) SPEAKER_02 : And then we recently tested them, we just separate the held out tested samples, and then got the result that's a little concerning in terms of usage.\n\n(2160.633,2165.478) SPEAKER_02 : And then our analysis showed that explainable AI was extremely helpful.\n\n(2165.978,2172.644) SPEAKER_02 : So for example, in the skin image, which part of the image led to that kind of a prediction?\n\n(2172.644,2177.948) SPEAKER_02 : Or as I said, using this counterfactual image generation.\n\n(2177.948,2184.173) SPEAKER_02 : So you make small changes to the input dermatology image such that it changes.\n\n(2184.173,2189.718) SPEAKER_02 : It crosses the decision boundary of the classifier and then see what features were changes.\n\n(2189.718,2194.682) SPEAKER_02 : So that way you can see the reasoning process of this classifier.\n\n(2194.922,2196.583) SPEAKER_02 : the clinical AI model.\n\n(2196.583,2206.411) SPEAKER_02 : So for that, there needs to be some technological development there because the feature attributions themselves are not going to be enough.\n\n(2206.411,2212.996) SPEAKER_02 : It shows only very small part of the inner workings of the machine learning model.\n\n(2212.996,2222.804) SPEAKER_02 : So developing methods for auditing clinical AI models, that's the research we are currently performing in the clinical area.\n\n(2223.264,2226.75) SPEAKER_02 : So all three areas, we are doing exciting research.\n\n(2226.75,2229.696) SPEAKER_01 : Well, Suwan, it sounds like you've got a lot of work ahead of you.\n\n(2229.696,2230.096) SPEAKER_02 : Yes.\n\n(2230.096,2230.557) SPEAKER_02 : Yeah.\n\n(2230.557,2231.238) SPEAKER_02 : Very busy.\n\n(2231.238,2232.02) SPEAKER_01 : I bet.\n\n(2232.02,2233.863) SPEAKER_01 : Thanks so much for joining us.\n\n(2233.863,2234.504) SPEAKER_02 : Thank you.\n\n(2234.504,2235.726) SPEAKER_02 : Thank you for inviting me.\n\n(2235.726,2235.987) SPEAKER_01 : Thank you.\n\n(2239.223,2242.184) SPEAKER_00 : All right, everyone, that's our show for today.\n\n(2242.184,2248.167) SPEAKER_00 : To learn more about today's guest or the topics mentioned in this interview, visit TwiMLAI.com.\n\n(2248.167,2255.951) SPEAKER_00 : Of course, if you like what you hear on the podcast, please subscribe, rate, and review the show on your favorite podcatcher.\n\n(2255.951,2258.652) SPEAKER_00 : Thanks so much for listening and catch you next time."}, "podcast_summary": "In this podcast, Sam Charrington interviews Suin Lee, a professor at the University of Washington, about their research in explainable AI. Suin discusses their work in computational biology and clinical medicine, focusing on the development of explainable AI techniques and their application in identifying causes and treatments for diseases like cancer. They also talk about the importance of collaboration between machine learning researchers, biologists, and clinical experts in advancing the field. Suin shares examples of their research, including the use of feature attribution methods like SHAP to understand the importance of genes in cancer drug synergy. They also discuss the challenges of applying explainability techniques in rapidly evolving fields and the future directions of their research, such as sample-based explanations and model auditing in clinical AI.", "podcast_guest": {"podcast_guest": "Suin Lee", "podcast_guest_org": "Paul G. Allen School of Computer Science and Engineering at the University of Washington", "podcast_guest_title": "Professor", "wikipedia_summary": "Not able to Find info on Wikipedia"}, "podcast_highlights": "- Suin Lee is a professor at the Paul G. Allen School of Computer Science and Engineering at the University of Washington.\n- Suin's research focuses on explainable AI and its applications in computational biology and clinical medicine.\n- The podcast discusses the importance of explainability in AI and its relevance in understanding complex diseases like cancer.\n- Suin shares her background and how she got into the field of computational biology and medicine.\n- The conversation highlights the need for collaboration between machine learning researchers, biologists, and clinical experts.\n- Suin's research involves developing explainable AI techniques for identifying causes and treatments of diseases, such as cancer and Alzheimer's.\n- The podcast also explores the challenges of applying explainability methods to rapidly evolving fields and complex systems.\n- Suin discusses her recent paper on using explainable AI to design cancer therapy, specifically focusing on acute myeloid leukemia (AML).\n- The paper identifies the stemness pathway as a key factor in predicting drug synergy for AML patients.\n- The future of Suin's research includes addressing robustness issues, multi-modal data, and sample-based importance in foundational AI methods.\n- In the clinical field, Suin is working on auditing clinical AI models and developing methods for biologically interpretable", "key_moments_and_key_topics": "1) Key Topics of the podcast transcript:\n- Explainable AI in computational biology and clinical medicine\n- The importance of feature attributions in AI models\n- The challenges of applying explainability techniques to rapidly evolving fields\n- The future of research in foundational AI methods and biomedical applications\n- The need for model auditing in clinical AI\n\n2) Key Highlights of the podcast transcript:\n- (46.357,54.683) The importance of explainable AI in computational biology and clinical medicine\n- (81.259,87.801) The background and motivation of the guest's research in the intersection of machine learning, computational biology, and clinical medicine\n- (148.3,167.945) The guest's work on explainable AI and feature attribution methods, including the SHAP framework\n- (383.182,392.708) The guest's research on explainability as applied to cancer therapy design\n- (1968.659,1972.2) The future directions of research, including the challenges of feature correlation and the need for sample-based explanations", "gpt_podcast_transcript": "(0.169,0.289) Unknown : you\n\n(8.34,9.201) Sam Charrington : All right, everyone.\n\n(9.201,12.583) Sam Charrington : Welcome to another episode of the TwiML AI Podcast.\n\n(12.583,14.965) Sam Charrington : I am your host, Sam Charrington.\n\n(14.965,16.927) Sam Charrington : And today I'm joined by Suin Lee.\n\n(16.927,24.753) Sam Charrington : Suin is a professor at the Paul G. Allen School of Computer Science and Engineering at the University of Washington.\n\n(24.753,30.497) Sam Charrington : Before we get going, be sure to take a moment to hit that subscribe button wherever you're listening to today's show.\n\n(30.497,32.739) Sam Charrington : Suin, welcome to the podcast.\n\n(32.739,33.92) Suin Lee : Thank you for the introduction.\n\n(34.79,38.132) Sam Charrington : I'm looking forward to digging into our talk.\n\n(38.132,46.357) Sam Charrington : You are an invited speaker at the 2023 ICML workshop on computational biology.\n\n(46.357,54.683) Sam Charrington : And we'll be talking about your talk there, which is really centered around your research into explainable AI, an important topic.\n\n(54.683,61.447) Sam Charrington : But before we jump into that, I'd love to have you share a little bit about your background and how you came to work in the field.\n\n(61.993,63.053) Suin Lee : Thank you so much.\n\n(63.053,71.236) Suin Lee : So my lab is currently working on a broad spectrum of a problem, for example, developing explainable AI techniques.\n\n(71.236,73.076) Suin Lee : So that's a core machine learning.\n\n(73.076,79.558) Suin Lee : And then we also work on identifying cause and treatment of challenging diseases such as cancer and Alzheimer's disease.\n\n(79.558,81.259) Suin Lee : So that's computational biology.\n\n(81.259,87.801) Suin Lee : And then also we develop clinical diagnosis or auditing frameworks for clinical AI.\n\n(88.641,91.243) Suin Lee : And then you asked about how I got into this field.\n\n(91.243,96.706) Suin Lee : So I was trained as a machine learning researcher when I was a PhD student.\n\n(96.706,101.309) Suin Lee : I was working on the problem of dealing with high dimensional data.\n\n(101.309,104.592) Suin Lee : And then at that time, when I was a PhD student at Stanford,\n\n(105.472,111.895) Suin Lee : In the field of computational biology, there was something really exciting happened, something called a microarray data.\n\n(111.895,116.176) Suin Lee : So it's a gene expression data that measures expression levels of 20,000 genes.\n\n(116.176,128.401) Suin Lee : And I suddenly thought that if machine learning researchers develop a powerful and effective method to identify cause of diseases such as cancer and then therapeutic targets,\n\n(129.101,131.443) Suin Lee : for those diseases.\n\n(131.443,137.229) Suin Lee : Then as a machine learning researcher, I can contribute hugely to the science and also medicine.\n\n(137.229,140.432) Suin Lee : And I just fell in love with this field.\n\n(140.432,147.76) Suin Lee : So that's how I got into the research at the intersection of computational machine learning and computational biology.\n\n(148.3,167.945) Suin Lee : After I got a job at the University of Washington that has a very strong medical school, and then I had wonderful colleagues, amazing people who had medical data, electronic health records, and then introduced me to this field of EHR data analysis in various\n\n(168.605,174.249) Suin Lee : clinical departments, anesthesiology and dermatology, and then emergency medicine.\n\n(174.249,186.656) Suin Lee : And then I just got really interested into the possibility, the potential that AI researchers or machine learning researchers myself and my students can contribute to medicine.\n\n(186.656,190.799) Suin Lee : That's how I got into this field of largely three fields.\n\n(190.799,193.401) Suin Lee : So one is machine learning and AI.\n\n(193.401,197.023) Suin Lee : And the second is computational biology and then clinical medicine.\n\n(197.509,206.042) Sam Charrington : You probably thought that you had to deal with messy data when you were in clinical biology and computational biology until you saw some of that EHR data.\n\n(206.042,207.725) Sam Charrington : That data can be very messy.\n\n(208.147,208.827) Suin Lee : It is.\n\n(208.827,211.709) Suin Lee : The goals of the fields are slightly different to each other.\n\n(211.709,217.693) Suin Lee : But in the future, I strongly believe that those two fields will merge, biology and medicine.\n\n(217.693,226.238) Suin Lee : So in a clinical side, researchers are already generating the biological molecular biology data from patients.\n\n(226.238,234.963) Suin Lee : So for example, for cancer patients, you can think about measuring the gene expression levels or genetic data from those cancer patients.\n\n(234.963,237.465) Suin Lee : And then what you want is the treatment.\n\n(237.985,248.617) Suin Lee : You want the AI or machine learning models to tell you which treatment, which drug, anti-cancer drugs are going to work the best for that particular patient.\n\n(248.617,255.385) Suin Lee : For that, you definitely need the biological knowledge and then actual mechanistic understanding of cancer.\n\n(256.993,265.478) Sam Charrington : And what says to you that the fields will merge as opposed to kind of collaborate closely?\n\n(265.478,274.624) Sam Charrington : Clearly they need to collaborate closely, but when I think of merge, and maybe I'm taking this too far, I'm thinking of like single models that operate in both domains.\n\n(275.769,276.87) Suin Lee : Yeah, I know what you're saying.\n\n(276.87,294.687) Suin Lee : So I tell my students or other young people that to actually move the field forward, to advance this field of biology, medicine, or biomedical sciences, you really need to become a bilingual researcher, or even trilingual these days.\n\n(294.687,298.411) Suin Lee : You know, computer science plus biology plus medicine.\n\n(298.411,298.771) Suin Lee : When you\n\n(299.832,307.753) Suin Lee : are you have one brain that really thinks like, you know, machine learning researchers and biologists and then clinical experts.\n\n(307.753,318.675) Suin Lee : It's, you know, usually that really helps to come up with creative approach and that can really move the field to benefit patients.\n\n(318.675,328.277) Suin Lee : And then at the end, the ultimate goal of a biology and molecular biology is to understand life better so that you can advance the health.\n\n(328.677,329.937) Suin Lee : of humans, right?\n\n(329.937,342.821) Suin Lee : So I think collaborations definitely help, but at the end, we really need to think about how to produce these young researchers so that they really think like experts in this area.\n\n(342.821,348.943) Suin Lee : These things already happened earlier in computational biology than clinical medicine.\n\n(348.943,352.504) Suin Lee : And when I was doing the PhD, it was usually based on collaborations.\n\n(352.844,367.067) Suin Lee : people who were trained primarily as a machine learning researcher and people who were trained as molecular biologists who hold pipettes and they work in the wet labs and then they form a collaboration and then write papers.\n\n(367.828,375.554) Suin Lee : But then later, you know, we see a lot of departments that's named, you know, computational biology or, you know, biomedical science departments.\n\n(375.554,381.419) Suin Lee : So it's a really healthy move for, you know, this kind of interdisciplinary fields.\n\n(381.419,382.5) Suin Lee : It makes total difference.\n\n(383.182,383.822) Sam Charrington : Yeah.\n\n(383.822,392.708) Sam Charrington : Your research and again, your presentation at the conference are focused on explainable AI, XAI.\n\n(392.708,401.092) Sam Charrington : Tell us a little bit about some of the things that you think are most important about explainability as applied to these fields.\n\n(401.092,408.477) Sam Charrington : I think we get that machine learning and models in general can be opaque and make important high stakes decisions.\n\n(408.477,410.278) Sam Charrington : You need some degree of explainability.\n\n(411.08,415.236) Sam Charrington : What's unique about your take in applying applicability in your field?\n\n(415.757,416.197) Suin Lee : Right.\n\n(416.197,416.938) Suin Lee : OK, thank you.\n\n(416.938,418.279) Suin Lee : That's an excellent question.\n\n(418.279,427.346) Suin Lee : So the core part of explainable AI, at least this theoretical framework, it basically means feature attributions.\n\n(427.346,429.848) Suin Lee : So imagine you have a black box model.\n\n(429.848,439.615) Suin Lee : You have a set of input, a vector x. And then you have an output y. And then when you have a prediction, you want to find a way to attribute two features.\n\n(439.615,443.158) Suin Lee : You want to know which features contributed the most.\n\n(443.738,446.56) Suin Lee : And then, you know, there are mathematical frameworks.\n\n(446.56,450.884) Suin Lee : Our particular approach that's called the SHAP framework, it is based on game theory.\n\n(450.884,455.007) Suin Lee : So you want to find a way to understand which features are important.\n\n(455.007,459.17) Suin Lee : So that's the core of the technical side of explainable AI.\n\n(459.59,470.355) Suin Lee : And then on the other hand, if you just apply this explainable AI technique, you know, off the shelf explainable AI algorithm to biology, mostly it's useless.\n\n(470.355,472.276) Suin Lee : It's not very useful.\n\n(472.276,475.137) Suin Lee : It's not useful in terms of biological insights.\n\n(475.797,481.478) Suin Lee : What you really want to understand is how these features collaborate with each other.\n\n(481.478,484.439) Suin Lee : Imagine that you have a set of genes as a feature.\n\n(484.439,490.3) Suin Lee : So you have 20,000 genes, 20,000 expression levels are the input of the black box model.\n\n(490.3,496.041) Suin Lee : And then your prediction is which cancer drug is going to work the best for each patient.\n\n(496.041,503.943) Suin Lee : And then individual genes contributions and then gene importance scores by themselves, they are not going to be really useful.\n\n(503.943,505.363) Suin Lee : It will be only useful when\n\n(506.063,519.146) Suin Lee : some explainable AI model, explainable AI algorithm can tell you which pathway, how genes collaborate with each other and then how genetic factors play a role into that.\n\n(519.146,525.068) Suin Lee : And then also how that leads to the good prognosis of the cancer patient and also\n\n(525.768,529.77) Suin Lee : sensitivity, the good responsiveness to that drug.\n\n(529.77,532.632) Suin Lee : So there is something missing there.\n\n(532.632,548.38) Suin Lee : And then the uniqueness of my research is that we want to develop this explainable AI method for biology and then also clinical medicine such that it can make real meaningful contribution to these fields.\n\n(548.38,552.762) Suin Lee : Another example in the medicine side is that imagine that you have a deep\n\n(553.382,557.605) Suin Lee : model, deep neural network, that's going to take you a dermatology image.\n\n(557.605,562.208) Suin Lee : So say that you find something unusual in your skin and then you take a picture.\n\n(562.208,563.89) Suin Lee : That's your dermatological image.\n\n(563.89,568.393) Suin Lee : And then let's say that you want to know that has features of melanoma or not.\n\n(568.913,572.756) Suin Lee : So the prediction results itself is not going to be really useful.\n\n(572.756,588.428) Suin Lee : And then even the current explainable AI methods that's going to tell you which pixels, which parts of the images led to the prediction of melanoma or not, those are not going to be very useful to understand how this black box model really works.\n\n(589.255,600.879) Suin Lee : When you try, for example, that you modify the image and then generate a counterfactual, small changes to the image such that it changes the prediction.\n\n(600.879,604.66) Suin Lee : Let's say that that changes the prediction from melanoma to normal.\n\n(604.66,612.502) Suin Lee : Only then you can understand how this model works, what the reasoning process of this black box machine learning model is like.\n\n(612.942,618.283) Suin Lee : So those examples, I'm going to show many examples like that.\n\n(618.283,638.788) Suin Lee : Basically, the message there is going to be that the current state-of-the-art explainable AI that tells you theoretically supported importance values for the features are not going to be enough to make meaningful contributions to both biological science and then also clinical medicine as well.\n\n(639.467,659.61) Sam Charrington : It sounds like you're calling out a broad deficiency in the approach and kind of saying that as opposed to this feature level explainability, we need more system level or process level explainability that is more grounded in the use cases or the application than what we have available today.\n\n(660.383,661.284) Suin Lee : Exactly.\n\n(661.284,663.566) Suin Lee : The question is how to do that.\n\n(663.566,667.13) Suin Lee : For that, we need a new explainable AI method.\n\n(667.13,676.56) Suin Lee : In the first part of the talk, I'm going to show many examples of what explainable AI, almost as is, can do.\n\n(676.56,679.062) Suin Lee : Those are the papers that we published\n\n(679.743,685.785) Suin Lee : a couple of years ago, so that it addresses new scientific questions.\n\n(685.785,691.567) Suin Lee : Even explainable AI or feature attribution methods as is can be useful.\n\n(691.567,695.988) Suin Lee : So I'm going to show many examples like that in both biology and medicine.\n\n(695.988,706.172) Suin Lee : But in the second part of the talk, I'm going to show how explainable AI can even open new research directions specifically for biology and health care.\n\n(707.012,719.374) Suin Lee : So those examples I showed you, the systems level insights or this counterfactual image generation that can facilitate collaboration with humans, in this case, a clinical expert.\n\n(719.374,726.756) Suin Lee : So in the second part of the talk, I'm going to show how this explainably I can open new research directions.\n\n(726.756,731.597) Suin Lee : And then part of the second part will be I'm going to have a deep dive into our recent paper,\n\n(732.177,739.645) Suin Lee : to highlight how Explainable AI can help cancer medicine design, cancer therapy design.\n\n(739.645,747.474) Suin Lee : So basically, how to choose two chemotherapy drugs that's going to have a synergy for a particular patient.\n\n(747.474,751.178) Suin Lee : So that's the paper that was recently published in Nature Biomedical Engineering.\n\n(751.701,767.357) Sam Charrington : Before we dig into that paper, the most recent paper, can you talk us through in a little bit more detail some of the examples of the foundational machine learning research and how they contribute to the problems you're trying to solve?\n\n(768.274,768.914) Suin Lee : Okay.\n\n(768.914,773.84) Suin Lee : So some of the foundational AI methods we developed, I'm going to talk about.\n\n(773.84,776.883) Suin Lee : It can be summarized into three parts.\n\n(776.883,783.35) Suin Lee : So one is, you know, principled understanding of current explainable AI methods.\n\n(783.35,785.993) Suin Lee : So specifically feature attribution methods.\n\n(786.233,809.664) Suin Lee : So, for example, in one work, we showed that our feature attribution method, that's SHAP, it was published in NeurIPS in 2017, we showed that it unifies a large portion of the explainable AI literature and 25 methods following the exact same principle, and all explaining by removing features.\n\n(810.104,821.856) Suin Lee : So it turned out that 25 methods, feature attribution methods that are widely used in the field and machine learning applications, they all go by the same principle.\n\n(821.856,828.003) Suin Lee : You want to assess the importance of each feature by removing them or removing subsets of them.\n\n(828.363,832.564) Suin Lee : So that helps us understand what goes on.\n\n(832.564,839.567) Suin Lee : For example, when they fail, you want to understand what goes on and also improve and then develop new explainable AI methods.\n\n(839.567,844.168) Suin Lee : So I'm going to introduce a couple of unifying frameworks.\n\n(844.168,850.33) Suin Lee : So this is about how to understand the principled understanding of feature attribution methods.\n\n(851.17,861.837) Suin Lee : Also, on a computational side, we have explored many avenues to make this SHAP computation even feasible and faster.\n\n(861.837,867.041) Suin Lee : So, SHAP stands for Shapely Editive... I suddenly forgot.\n\n(867.041,867.921) Suin Lee : I can't forget this.\n\n(867.921,868.622) Sam Charrington : Explanations.\n\n(869.521,871.905) Suin Lee : Yes, a sharply additive explanation.\n\n(871.905,878.615) Sam Charrington : It's kind of weird because they chose the third letter of the word.\n\n(878.615,882.821) Suin Lee : That's the first author, my student, Scott's choice.\n\n(882.821,883.942) Suin Lee : I love the name, by the way.\n\n(884.503,892.331) Suin Lee : Computing sharp values is theoretically very well supported, but then computation-wise, it's not really easy to compute.\n\n(892.331,894.413) Suin Lee : It involves exponential computation.\n\n(894.413,899.998) Suin Lee : So we need to develop approximation methods such that we can compute them in a feasible manner.\n\n(899.998,904.262) Suin Lee : So we developed many fast statistical estimation approaches\n\n(905.003,912.575) Suin Lee : And then you want to make sure that there is a convergence and all the desirable theoretical properties are already there.\n\n(912.575,918.083) Suin Lee : And then also, we developed approaches for specific model types.\n\n(918.083,920.326) Suin Lee : For example, ensemble tree models.\n\n(920.947,922.629) Suin Lee : And then also deep neural networks.\n\n(922.629,925.573) Suin Lee : So we have a deep shape and then tree shape.\n\n(925.573,929.879) Suin Lee : And then more recently, we also have a vision transformer Shapley.\n\n(929.879,935.407) Suin Lee : So that's a way to compute the Shapley values for transformers, vision transformers.\n\n(936.235,938.637) Suin Lee : And then there is another one that's called the FASTA SHAPE.\n\n(938.637,948.302) Suin Lee : So the one way to make the SHAPE computation more feasible is to focus on specific particular aspects of models.\n\n(948.302,954.066) Suin Lee : So for example, tree ensembles or deep neural network, they have some particular model types.\n\n(954.066,964.933) Suin Lee : There is a way to make this computation a little faster, basically make... So model specific versions of SHAPE implementation.\n\n(965.473,966.393) Suin Lee : Yes, yes.\n\n(966.393,966.673) Suin Lee : Yeah.\n\n(966.673,969.014) Suin Lee : So that's another line of research.\n\n(969.014,975.796) Suin Lee : And then more recently, we also started to understand the robustness of the shaft value.\n\n(975.796,977.856) Suin Lee : So adversarial attack.\n\n(977.856,988.899) Suin Lee : A few years ago, in the field of machine learning, researchers have tried to understand how robust the machine learning model itself, the prediction results are toward\n\n(989.419,990.9) Suin Lee : adversarial attacks.\n\n(990.9,996.544) Suin Lee : And then now we are looking into this issue in terms of the model explanations.\n\n(996.544,999.826) Suin Lee : So how feature attributions are robust.\n\n(999.826,1007.151) Suin Lee : So in our most recent paper, we basically showed the removal-based approaches, including SHAPE.\n\n(1007.151,1014.536) Suin Lee : Like earlier I said, many of the feature attribution methods turned out to be to have the same principle, which is explaining by removal.\n\n(1014.536,1016.197) Suin Lee : So those line of, you know,\n\n(1017.037,1021.439) Suin Lee : methods is more robust to this kind of adversarial attacks.\n\n(1021.439,1031.642) Suin Lee : So, and then, you know, multimodality, you know, those other kinds of issues, we are actively doing this research in terms of, you know, foundational AI algorithms also.\n\n(1032.362,1041.952) Sam Charrington : And SHAP, as you've mentioned, is broadly used, both the original algorithm as well as the related algorithms as you described.\n\n(1041.952,1047.417) Sam Charrington : But it's also one of the first explainability approaches to be popularized\n\n(1048.238,1051.501) Sam Charrington : Where does it sit in terms of relevance?\n\n(1051.501,1070.518) Sam Charrington : Are there different kind of wholly different approaches that have overtaken it in popularity or applicability based on kind of today's models and applications or is SHAP still kind of a core approach to the way explainability is looked at in practice?\n\n(1071.373,1073.514) Suin Lee : It's more on the later side.\n\n(1073.514,1080.577) Suin Lee : We believe that this removal-based approach and in this cooperative game theory, we believe in that.\n\n(1080.577,1084.539) Suin Lee : And then also, it has the desirable properties, first of all.\n\n(1084.539,1094.403) Suin Lee : And then we, in our many experiments, we still see that removal-based approaches are more robust, as I said, those berserker attacks.\n\n(1094.403,1098.545) Suin Lee : And then also, in terms of various evaluation criteria, we still think that those\n\n(1099.045,1107.935) Suin Lee : methods are more robust than the other class, which we characterized as a propagation-based approach or gradient-based approaches.\n\n(1107.935,1111.259) Suin Lee : So we would prefer just remover-based approaches.\n\n(1111.259,1115.744) Suin Lee : But on the other hand, those approaches are very computationally very intensive.\n\n(1115.744,1115.864) Suin Lee : So\n\n(1116.244,1129.678) Suin Lee : Well, the way SHEP works is basically that you try all subset of features and then you add a feature of interest and then see the model, check the model output and you average across all subsets of features.\n\n(1129.678,1132.781) Suin Lee : So as you can imagine, it's computationally very intensive.\n\n(1132.781,1136.184) Suin Lee : So when we now think about foundational models or\n\n(1136.905,1142.228) Suin Lee : large language models, these really large models of a ton, a lot of parameters.\n\n(1142.228,1148.531) Suin Lee : And then deep neural network and, you know, gradient computation is perhaps easier than trying all sorts of features, right?\n\n(1148.531,1154.394) Suin Lee : So practically, it's not, you know, as easy as the other class in terms of\n\n(1155.094,1160.497) Suin Lee : the computation, but we still want to make this computational more feasible.\n\n(1160.497,1175.385) Suin Lee : We want to develop various clever approaches to reduce the computation and then still maintain the desirable theoretical properties that this removal-based approach or SHAP in particular has.\n\n(1175.385,1175.625) Sam Charrington : Got it.\n\n(1176.299,1191.191) Sam Charrington : And so that is an example of kind of the foundational research that your lab does that contributes not only to your work on the biological science side or computational biology side, but broadly to the field.\n\n(1191.191,1199.177) Sam Charrington : And then your more recent paper is an example of the kind of contributions you're making on the medicine side.\n\n(1199.177,1201.399) Sam Charrington : Can you talk a little bit about that cancer paper?\n\n(1201.943,1202.964) Suin Lee : Yeah, sure.\n\n(1202.964,1204.566) Suin Lee : It is about AML.\n\n(1204.566,1208.03) Suin Lee : So we chose AML as an example application.\n\n(1208.03,1216.079) Suin Lee : So it's acute myeloid leukemia, it's aggressive blood cancer, and it's relatively common for older people.\n\n(1216.079,1223.587) Suin Lee : So to give you a bit of a background in general, the cutting edge in the treatment of cancers, such as AML,\n\n(1223.867,1226.968) Suin Lee : has increasingly become combination therapy.\n\n(1226.968,1238.312) Suin Lee : So the rationale here is that by choosing drugs that target complementary biological pathways, we can achieve greater anti-cancer efficacy.\n\n(1238.312,1242.734) Suin Lee : So basically, you choose two or three chemotherapy drugs,\n\n(1243.394,1251.941) Suin Lee : and then use them together so that when there is a synergy, usually there is a very good anti-cancer efficacy.\n\n(1251.941,1256.906) Suin Lee : But the issue is that choosing optimal combinations of drugs is a really hard problem.\n\n(1256.906,1265.954) Suin Lee : So there are about hundreds of individual FDA approved anti-cancer drugs, which means that there will be tens of thousands of possible combinations.\n\n(1266.594,1279.4) Suin Lee : But when you consider pairwise combination, and there could be even more if you consider non-FDA approved experimental drugs in development, or consider a combination of more than two drugs.\n\n(1279.4,1291.426) Suin Lee : And then the different patients, even patients who have the same type of cancer may respond differently to exact same drugs because of this individual, the particular genomic characteristics.\n\n(1292.166,1295.449) Suin Lee : then formulate this problem as a machine learning problem.\n\n(1295.449,1309.6) Suin Lee : So you take this AML patient's gene expression levels, so you get the blood of the patient and then purify the cells so you have only cancer cells, and then say you measure expression levels of 20,000 genes.\n\n(1309.6,1312.662) Suin Lee : So mathematically, this is 20,000 dimensional vector.\n\n(1313.563,1323.967) Suin Lee : And then also, let's say you consider a pair of drugs, drugs A and B, and then you use various information about this drug.\n\n(1323.967,1328.028) Suin Lee : For example, structure of these drugs or their biological targets.\n\n(1328.028,1331.889) Suin Lee : There are many data sets that can tell you that information.\n\n(1331.889,1339.072) Suin Lee : And then you take those as a machine learning input, and then you want to predict the synergy between the drugs A and B.\n\n(1339.852,1347.817) Suin Lee : So in this kind of a problem, and as I said, there will be tens of thousands of pairwise combinations of those drugs.\n\n(1347.817,1355.803) Suin Lee : And so in this kind of situation, not only the prediction, but also explanations will be extremely important.\n\n(1355.803,1362.908) Suin Lee : So say you want to be able to say that drug A and B is going to work well, are going to have a synergy together because\n\n(1363.568,1371.355) Suin Lee : this patient X has gene expression levels of A, B, and C high.\n\n(1371.355,1377.7) Suin Lee : Or you say expression levels of a certain biological pathway, those genes are highly expressed.\n\n(1377.7,1380.442) Suin Lee : So you need a set of explanation to do that.\n\n(1380.442,1383.185) Suin Lee : And then more importantly, if you think about\n\n(1383.765,1384.807) Suin Lee : all pairs of drugs.\n\n(1384.807,1394.458) Suin Lee : If there is an underlying principle in terms of when two drugs are likely to have a synergy, then it's going to be even more useful.\n\n(1394.458,1398.804) Suin Lee : So what we did in this paper was that we got the explanations.\n\n(1398.804,1400.806) Suin Lee : We computed the shaft values for\n\n(1401.647,1421.131) Suin Lee : many combinations of drugs from the machine learning model, and then we analyzed that, and then we identified the unifying principle in terms of when, in what case, any pair of drugs A and B have a synergy, and then we identified the pathway.\n\n(1421.131,1424.732) Suin Lee : It is called stemness pathway.\n\n(1424.732,1431.313) Suin Lee : It is also called, trying to find in that part of the slide, this hematopoietic stem cell-like signature.\n\n(1432.053,1436.134) Suin Lee : Cancers are sometimes more differentiated or less differentiated.\n\n(1436.134,1439.976) Suin Lee : If you had a family member who had cancer, you probably understand this term.\n\n(1439.976,1446.498) Suin Lee : Usually, less differentiated cancers have worse prognosis than more differentiated cancers.\n\n(1447.338,1467.014) Suin Lee : we identified this pathway that's really relevant to this stem-ness mechanism and then found the underlying principle, which basically says that it's good to have two drugs, one drug targeting less differentiated, the other one targeting more differentiated cancer, likely work the best.\n\n(1467.014,1468.956) Suin Lee : So in this project, not only\n\n(1469.616,1480.822) Suin Lee : our algorithm can tell oncologists or biological scientists which genes are important, which feature attributions, which features are important for drug synergy.\n\n(1480.822,1495.829) Suin Lee : But also, by analyzing many model explanations from many patients, we can have an understanding of these underlying principles in terms of what makes a successful drug combination therapy.\n\n(1496.289,1498.27) Suin Lee : Cancer therapy design, I would say.\n\n(1498.27,1506.873) Suin Lee : This is an example where we can see how explainable AI can be effective in cancer therapy design.\n\n(1506.873,1519.058) Sam Charrington : Is AML unique in having a well-understood pathway or is that a bottleneck for the application of this technique to the broader set of cancers?\n\n(1519.58,1523.662) Suin Lee : Oh, so AML is just one example.\n\n(1523.662,1527.564) Suin Lee : I mean, this kind of principle can be applied to too many data sets.\n\n(1527.564,1533.387) Suin Lee : You know, computational biologists often need to work on the problem where the data are available.\n\n(1533.387,1545.133) Suin Lee : So, you know, as you can imagine, blood cancers, those tissues are relatively easy to, it's relatively easier to obtain, you know, blood tissues compared to other kinds of tissues.\n\n(1546.013,1553.155) Suin Lee : There are many available data sets and then also the measurement of the drug synergy from many samples.\n\n(1553.155,1558.096) Suin Lee : So we happen to choose this cancer type because of the data availability.\n\n(1558.096,1563.618) Suin Lee : But this approach can be broadly applicable to other types of cancer.\n\n(1564.027,1566.389) Suin Lee : So this is one of the... Yeah, go ahead.\n\n(1566.389,1581.904) Sam Charrington : I'm maybe trying to get a broader question, which is the explainability method is kind of explaining over a set of known features and pathways and processes and things like that.\n\n(1581.904,1586.208) Sam Charrington : And my sense is that for many of the\n\n(1586.949,1593.234) Sam Charrington : potential applications, the pathways are still a subject of research themselves.\n\n(1593.234,1598.599) Sam Charrington : Meaning, you know, maybe there's some aspect of pathway that's known, but there are others.\n\n(1598.599,1603.143) Sam Charrington : There are, you know, or some diseases for which there aren't pathways.\n\n(1603.143,1604.264) Sam Charrington : And I guess I'm wondering\n\n(1604.884,1611.63) Sam Charrington : the way you think about applying techniques like this in a... A, is that actually the case or am I all wrong there?\n\n(1611.63,1619.017) Sam Charrington : But otherwise, how will you apply techniques like this in rapidly evolving fields that are very complex, meaning... That's an excellent question.\n\n(1619.017,1628.746) Sam Charrington : Maybe you're giving an explanation and the explanation is based on the pathway as you understand it, but there's so many other things going on in the system that you really have not accounted for.\n\n(1629.184,1630.124) Suin Lee : Yeah, exactly.\n\n(1630.124,1630.885) Suin Lee : Right.\n\n(1630.885,1634.166) Suin Lee : So first of all, Pathway is not unique to disease.\n\n(1634.166,1640.409) Suin Lee : So when we say, you know, Pathway databases, it basically tells you the members of the genes in each pathway.\n\n(1640.409,1640.869) Suin Lee : That's it.\n\n(1640.869,1643.67) Suin Lee : I mean, it's like, you know, many, many sets of genes.\n\n(1643.67,1646.292) Suin Lee : We also sometimes call it gene sets.\n\n(1646.292,1647.732) Suin Lee : It doesn't depend on the disease.\n\n(1647.732,1656.016) Suin Lee : And then the way we view is that it's not like all genes need to be activated for the pathway needs to be activated.\n\n(1656.596,1658.818) Suin Lee : It would be only a subset of genes.\n\n(1658.818,1664.501) Suin Lee : We would expect only a subset of genes to be highly expressed to say, you know, that pathway is activated.\n\n(1664.501,1674.768) Suin Lee : And then it's really extremely important for a computational biologist when we develop, you know, a method like this to get biological insights from large scale data sets.\n\n(1674.768,1682.953) Suin Lee : When we develop such a method, we need to make sure that it does not fully depend on any sort of prior knowledge.\n\n(1683.713,1686.095) Suin Lee : And then the algorithm needs to be flexible.\n\n(1686.095,1688.277) Suin Lee : So that's of key importance.\n\n(1688.277,1694.563) Suin Lee : So in this particular example, we didn't use a pathway actually from the beginning.\n\n(1694.563,1698.967) Suin Lee : When the model training happens, we used genes as individual features.\n\n(1698.967,1706.935) Suin Lee : And then we analyzed the feature attributions and then did the statistical test to see which pathways seem to be more activated.\n\n(1707.395,1709.057) Suin Lee : You made a really good point.\n\n(1709.057,1713.683) Suin Lee : In all computational biology methods, it's really important not to make it too rigid.\n\n(1713.683,1717.287) Suin Lee : For the existing knowledge, it needs to be flexible.\n\n(1717.287,1721.372) Sam Charrington : And so how do you evaluate your results in this particular paper?\n\n(1722.069,1732.257) Suin Lee : Oh, so say that you have a feature attribution for all genes, for a certain patient, and then for a certain combination of drugs.\n\n(1732.257,1736.961) Suin Lee : And then say you will have a lot of feature attributions then, right?\n\n(1736.961,1741.785) Suin Lee : Combining all patients and all pairs of drugs you considered.\n\n(1741.785,1744.727) Suin Lee : And then we perform the statistical test.\n\n(1744.727,1747.51) Suin Lee : So for example, it's a simple, you know, features exact\n\n(1747.97,1761.441) Suin Lee : test kind of statistical test where you see whether there is significantly, you know, large value of attribution values for certain set of genes defined by certain pathway.\n\n(1761.441,1768.787) Suin Lee : And then you do, you know, multiple hypothesis testing and then see whether that, you know, significance is indeed is relevant.\n\n(1768.787,1775.173) Suin Lee : So the pathway-based analysis was done in a post-hoc manner after model training and then\n\n(1775.713,1778.474) Suin Lee : obtaining all, you know, model explanations.\n\n(1778.474,1790.477) Suin Lee : So another challenge we ran into in that project was that was really not addressed properly by this foundational AI field was feature correlation.\n\n(1790.477,1795.779) Suin Lee : So in many biomedical data sets, you will see lots of features that are correlated with each other.\n\n(1795.779,1797.739) Suin Lee : Many genes are correlated.\n\n(1797.739,1800.16) Suin Lee : It's a really modular gene expression, you know,\n\n(1800.68,1807.266) Suin Lee : levels are very modular, so you easily see a subset of genes that are very highly correlated with each other.\n\n(1807.266,1818.036) Suin Lee : So in that kind of case, shaft values are not going to be extremely accurate because imagine that there are two genes that are perfectly correlated with each other.\n\n(1818.036,1818.356) Suin Lee : Then\n\n(1818.977,1823.36) Suin Lee : There will be infinite ways to attribute to these two genes.\n\n(1823.36,1831.444) Suin Lee : So in that paper, in that Nature Biomedical Engineering paper, we addressed it by considering ensemble model.\n\n(1831.444,1835.407) Suin Lee : So we ran many ensemble of model explanations.\n\n(1835.407,1837.308) Suin Lee : So we ran the model.\n\n(1837.308,1839.589) Suin Lee : In this case, it was not your deep neural network.\n\n(1839.589,1841.03) Suin Lee : It was three ensembles.\n\n(1841.03,1842.891) Suin Lee : And then we averaged.\n\n(1842.891,1846.493) Suin Lee : We averaged the feature attributions that are from many models.\n\n(1846.773,1853.262) Suin Lee : And then we showed that it gives you more robust feature attributions when the features are correlated with each other.\n\n(1853.262,1855.545) Sam Charrington : Awesome.\n\n(1855.545,1860.472) Sam Charrington : So talk a little bit about where you see the future of your research going.\n\n(1861.683,1864.565) Suin Lee : That's a really important question.\n\n(1864.565,1878.114) Suin Lee : In all three ways, first of all, in the foundational AI method, as I briefly mentioned, this robustness issues and also multi-modal data.\n\n(1878.114,1883.337) Suin Lee : Let's say that you have a set of features and each feature belongs to different category.\n\n(1883.337,1884.978) Suin Lee : They are in different modality and then\n\n(1885.618,1891.021) Suin Lee : how to attribute to these features that are in different modalities.\n\n(1891.021,1893.603) Suin Lee : So that's an open problem.\n\n(1893.603,1900.287) Suin Lee : So it was actually motivated by biomedical problem, but it's widely applicable to other applications.\n\n(1902.194,1908.516) Suin Lee : And then also these emerging models of LLMs or other foundational models.\n\n(1908.516,1916.757) Suin Lee : And in this kind of really large models, how to actually compute the feature attributions properly.\n\n(1916.757,1926.98) Suin Lee : And then also, we are really interested in sample-based importance to say that you transpose the matrix transpose of your feature matrix.\n\n(1926.98,1931.861) Suin Lee : So I've been talking about these feature attributions a lot, but you can also apply Shapley values\n\n(1932.501,1937.404) Suin Lee : to gain insights into which samples are important for your model training.\n\n(1937.404,1948.73) Suin Lee : So that can help us understand how foundational models in various fields or large language models rely on which training samples.\n\n(1948.73,1959.015) Suin Lee : So that can be really important for model auditing perspective, first of all, and then to gain insight in terms of which samples were important.\n\n(1959.495,1963.117) Suin Lee : for these large models to behave a certain way, right?\n\n(1963.117,1968.659) Suin Lee : So sample-based explanation is also one of the things that we are mainly working on.\n\n(1968.659,1972.2) Suin Lee : In the biomedical side, there are many projects.\n\n(1972.2,1978.403) Suin Lee : So, you know, single cell data science is one of the big themes in my lab now.\n\n(1978.943,1986.667) Suin Lee : So you obtain gene expression levels or other kinds of molecular level information at a single cell level.\n\n(1986.667,1989.788) Suin Lee : So the advantage is that you will have a ton of samples.\n\n(1989.788,1999.672) Suin Lee : So one experiment is going to give you many samples, which is really appropriate for large scale models these days based on deep neural networks.\n\n(1999.672,2005.675) Suin Lee : So for example, the researchers started looking into foundational model for a single cell data set.\n\n(2005.955,2015.926) Suin Lee : So in this kind of, you know, data sets that have still, you know, high dimensional and then researchers are now obtaining multi-omic data.\n\n(2015.926,2021.252) Suin Lee : So not only gene expressions, you can also obtain other kinds of, you know, genomic information.\n\n(2021.252,2021.892) Suin Lee : So that's going to\n\n(2022.533,2026.655) Suin Lee : increase the dimensionality also, and then larger sample sizes.\n\n(2026.655,2032.979) Suin Lee : How to learn the biologically interpretable representation space?\n\n(2032.979,2037.381) Suin Lee : That's one of the big questions in the research in my lab.\n\n(2039.142,2045.229) Suin Lee : All feature attribution methods at the end in the downstream prediction test, you attribute to features.\n\n(2045.229,2050.074) Suin Lee : And then the assumption is that each feature is an interpretable unit.\n\n(2050.074,2055.059) Suin Lee : In biology, as I mentioned earlier, it's not the case in biology, right?\n\n(2055.059,2061.966) Suin Lee : So the functional unit in biology is much more interpretable than in any individual genes.\n\n(2062.226,2073.893) Suin Lee : So how to learn the features that have more broadly representation, feature representation space that's biologically more interpretable.\n\n(2073.893,2079.296) Suin Lee : And then also how to make foundational models learned based on single cell data sets.\n\n(2079.296,2088.101) Suin Lee : So researchers started publishing those papers that are about applying this foundational model approach to single cell data sets.\n\n(2088.101,2091.243) Suin Lee : And then how to make it biologically interpretable\n\n(2091.743,2104.866) Suin Lee : so that you can gain scientific insights from the model results and then also audit those models to make sure that users can actually safely use them for scientific discoveries.\n\n(2104.866,2110.188) Suin Lee : So, attribution methods for this kind of modern machine learning models\n\n(2111.168,2113.47) Suin Lee : so that you can gain biological insights.\n\n(2113.47,2115.151) Suin Lee : So that's another theme.\n\n(2115.151,2119.113) Suin Lee : In a clinical side, we are really interested in this model auditing.\n\n(2119.113,2125.677) Suin Lee : In our most recent paper that's in review, we are focusing on dermatology example.\n\n(2125.677,2126.958) Suin Lee : So dermatological image\n\n(2127.778,2135.867) Suin Lee : is inputted into deep neural network, and then you want to know whether the prediction result is melanoma or not.\n\n(2135.867,2147.059) Suin Lee : There are many algorithms out there, some published in very high-profile medical journals, and also some available through the cell phone apps.\n\n(2147.059,2148.441) Suin Lee : So there are many algorithms.\n\n(2148.741,2160.633) Suin Lee : And then we recently tested them, we just separate the held out tested samples, and then got the result that's a little concerning in terms of usage.\n\n(2160.633,2165.478) Suin Lee : And then our analysis showed that explainable AI was extremely helpful.\n\n(2165.978,2172.644) Suin Lee : So for example, in the skin image, which part of the image led to that kind of a prediction?\n\n(2172.644,2177.948) Suin Lee : Or as I said, using this counterfactual image generation.\n\n(2177.948,2184.173) Suin Lee : So you make small changes to the input dermatology image such that it changes.\n\n(2184.173,2189.718) Suin Lee : It crosses the decision boundary of the classifier and then see what features were changes.\n\n(2189.718,2194.682) Suin Lee : So that way you can see the reasoning process of this classifier.\n\n(2194.922,2196.583) Suin Lee : the clinical AI model.\n\n(2196.583,2206.411) Suin Lee : So for that, there needs to be some technological development there because the feature attributions themselves are not going to be enough.\n\n(2206.411,2212.996) Suin Lee : It shows only very small part of the inner workings of the machine learning model.\n\n(2212.996,2222.804) Suin Lee : So developing methods for auditing clinical AI models, that's the research we are currently performing in the clinical area.\n\n(2223.264,2226.75) Suin Lee : So all three areas, we are doing exciting research.\n\n(2226.75,2229.696) Sam Charrington : Well, Suwan, it sounds like you've got a lot of work ahead of you.\n\n(2229.696,2230.096) Suin Lee : Yes.\n\n(2230.096,2230.557) Suin Lee : Yeah.\n\n(2230.557,2231.238) Suin Lee : Very busy.\n\n(2231.238,2232.02) Sam Charrington : I bet.\n\n(2232.02,2233.863) Sam Charrington : Thanks so much for joining us.\n\n(2233.863,2234.504) Suin Lee : Thank you.\n\n(2234.504,2235.726) Suin Lee : Thank you for inviting me.\n\n(2235.726,2235.987) Sam Charrington : Thank you.\n\n(2239.223,2242.184) Unknown : All right, everyone, that's our show for today.\n\n(2242.184,2248.167) Unknown : To learn more about today's guest or the topics mentioned in this interview, visit TwiMLAI.com.\n\n(2248.167,2255.951) Unknown : Of course, if you like what you hear on the podcast, please subscribe, rate, and review the show on your favorite podcatcher.\n\n(2255.951,2258.652) Unknown : Thanks so much for listening and catch you next time."}
requirements.txt ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ matplotlib
2
+ contourpy
3
+ modal
4
+ openai
5
+ pandas
6
+ numpy
7
+ podcast-api
8
+ gradio