Anonymous Authors commited on
Commit
01c612b
1 Parent(s): 572dd5c

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +46 -0
app.py ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from datasets import load_from_disk
3
+ import numpy as np
4
+
5
+
6
+ gender_labels = ['man', 'non-binary', 'woman', 'no_gender_specified', ]
7
+
8
+ ethnicity_labels = ['African-American', 'American_Indian', 'Black', 'Caucasian', 'East_Asian',
9
+ 'First_Nations', 'Hispanic', 'Indigenous_American', 'Latino', 'Latinx',
10
+ 'Multiracial', 'Native_American', 'Pacific_Islander', 'South_Asian',
11
+ 'Southeast_Asian', 'White', 'no_ethnicity_specified']
12
+ models = ['DallE', 'SD_14', 'SD_2']
13
+ nos = [1,2,3,4,5,6,7,8,9,10]
14
+
15
+ ds = load_dataset("color-sorted")
16
+
17
+ def get_nearest_64(gender, ethnicity, model, no, index):
18
+ df = ds.remove_columns(["image","image_path"]).to_pandas()
19
+ ix = df.loc[(df['ethnicity'] == ethnicity) & (df['gender'] == gender) & (df['no'] == no) & (df['model'] == model)].index[0]
20
+ image = ds.select([index[ix][0]])["image"][0]
21
+ neighbors = ds.select(max(ix-10, 0), min(ix+10, len(ds)-1))
22
+ neighbor_images = neighbors["image"]
23
+ neighbor_captions = [caption.split("/")[-1] for caption in neighbors["image_path"]]
24
+ neighbor_captions = [' '.join(caption.split("_")[4:-3]) for caption in neighbor_captions]
25
+ neighbor_models = neighbors["model"]
26
+ neighbor_captions = [f"{a} {b}" for a,b in zip(neighbor_captions,neighbor_models)]
27
+ return image, list(zip(neighbor_images, neighbor_captions))
28
+
29
+ with gr.Blocks() as demo:
30
+ gr.Markdown("# Colorfulness Nearest Neighbors Explorer")
31
+ gr.Markdown("### TF-IDF index of the _identities_ dataset of images generated by 3 models using colorfulness")
32
+ gr.Markdown("#### Choose one of the generated identity images to see its nearest neighbors according to colorfulness")
33
+ with gr.Row():
34
+ with gr.Column():
35
+ model = gr.Radio(models, label="Model")
36
+ index = gr.Radio(indexes, label="Visual vocabulary size")
37
+ gender = gr.Radio(gender_labels, label="Gender label")
38
+ with gr.Column():
39
+ ethnicity = gr.Radio(ethnicity_labels, label="Ethnicity label")
40
+ no = gr.Radio(nos, label="Image number")
41
+ button = gr.Button(value="Get nearest neighbors")
42
+ with gr.Row():
43
+ image = gr.Image()
44
+ gallery = gr.Gallery().style(grid=4)
45
+ button.click(get_nearest_64, inputs=[gender, ethnicity, model, no, index], outputs=[image, gallery])
46
+ demo.launch()