Avijit Ghosh commited on
Commit
5bb95f3
1 Parent(s): e6e82b8

change type to level

Browse files
app.py CHANGED
@@ -29,19 +29,19 @@ globaldf['Link'] = '<u>'+globaldf['Link']+'</u>'
29
  modality_order = ["Text", "Image", "Audio", "Video"]
30
  type_order = ["Model", "Dataset", "Output", "Taxonomy"]
31
 
32
- # Convert Modality and Type columns to categorical with specified order
33
  globaldf['Modality'] = pd.Categorical(globaldf['Modality'], categories=modality_order, ordered=True)
34
- globaldf['Type'] = pd.Categorical(globaldf['Type'], categories=type_order, ordered=True)
35
 
36
- # Sort DataFrame by Modality and Type
37
- globaldf.sort_values(by=['Modality', 'Type'], inplace=True)
38
 
39
  # create a gradio page with tabs and accordions
40
 
41
  # Path: taxonomy.py
42
 
43
  def filter_modality_type(fulltable, modality_filter, type_filter):
44
- filteredtable = fulltable[fulltable['Modality'].isin(modality_filter) & fulltable['Type'].isin(type_filter)]
45
  return filteredtable
46
 
47
  def showmodal(evt: gr.SelectData):
@@ -100,7 +100,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
100
  with gr.Tabs(elem_classes="tab-buttons") as tabs1:
101
  with gr.TabItem("Bias/Stereotypes"):
102
  fulltable = globaldf[globaldf['Group'] == 'BiasEvals']
103
- fulltable = fulltable[['Modality','Type', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
104
 
105
  gr.Markdown("""
106
  Generative AI systems can perpetuate harmful biases from various sources, including systemic, human, and statistical biases. These biases, also known as "fairness" considerations, can manifest in the final system due to choices made throughout the development process. They include harmful associations and stereotypes related to protected classes, such as race, gender, and sexuality. Evaluating biases involves assessing correlations, co-occurrences, sentiment, and toxicity across different modalities, both within the model itself and in the outputs of downstream tasks.
@@ -114,7 +114,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
114
  )
115
  type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
116
  value=["Model", "Dataset", "Output", "Taxonomy"],
117
- label="Type",
118
  show_label=True,
119
  # info="Which modality to show."
120
  )
@@ -138,7 +138,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
138
 
139
  with gr.TabItem("Cultural Values/Sensitive Content"):
140
  fulltable = globaldf[globaldf['Group'] == 'CulturalEvals']
141
- fulltable = fulltable[['Modality','Type', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
142
 
143
  gr.Markdown("""Cultural values are specific to groups and sensitive content is normative. Sensitive topics also vary by culture and can include hate speech. What is considered a sensitive topic, such as egregious violence or adult sexual content, can vary widely by viewpoint. Due to norms differing by culture, region, and language, there is no standard for what constitutes sensitive content.
144
  Distinct cultural values present a challenge for deploying models into a global sphere, as what may be appropriate in one culture may be unsafe in others. Generative AI systems cannot be neutral or objective, nor can they encompass truly universal values. There is no “view from nowhere”; in quantifying anything, a particular frame of reference is imposed.
@@ -152,7 +152,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
152
  )
153
  type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
154
  value=["Model", "Dataset", "Output", "Taxonomy"],
155
- label="Type",
156
  show_label=True,
157
  # info="Which modality to show."
158
  )
 
29
  modality_order = ["Text", "Image", "Audio", "Video"]
30
  type_order = ["Model", "Dataset", "Output", "Taxonomy"]
31
 
32
+ # Convert Modality and Level columns to categorical with specified order
33
  globaldf['Modality'] = pd.Categorical(globaldf['Modality'], categories=modality_order, ordered=True)
34
+ globaldf['Level'] = pd.Categorical(globaldf['Level'], categories=type_order, ordered=True)
35
 
36
+ # Sort DataFrame by Modality and Level
37
+ globaldf.sort_values(by=['Modality', 'Level'], inplace=True)
38
 
39
  # create a gradio page with tabs and accordions
40
 
41
  # Path: taxonomy.py
42
 
43
  def filter_modality_type(fulltable, modality_filter, type_filter):
44
+ filteredtable = fulltable[fulltable['Modality'].isin(modality_filter) & fulltable['Level'].isin(type_filter)]
45
  return filteredtable
46
 
47
  def showmodal(evt: gr.SelectData):
 
100
  with gr.Tabs(elem_classes="tab-buttons") as tabs1:
101
  with gr.TabItem("Bias/Stereotypes"):
102
  fulltable = globaldf[globaldf['Group'] == 'BiasEvals']
103
+ fulltable = fulltable[['Modality','Level', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
104
 
105
  gr.Markdown("""
106
  Generative AI systems can perpetuate harmful biases from various sources, including systemic, human, and statistical biases. These biases, also known as "fairness" considerations, can manifest in the final system due to choices made throughout the development process. They include harmful associations and stereotypes related to protected classes, such as race, gender, and sexuality. Evaluating biases involves assessing correlations, co-occurrences, sentiment, and toxicity across different modalities, both within the model itself and in the outputs of downstream tasks.
 
114
  )
115
  type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
116
  value=["Model", "Dataset", "Output", "Taxonomy"],
117
+ label="Level",
118
  show_label=True,
119
  # info="Which modality to show."
120
  )
 
138
 
139
  with gr.TabItem("Cultural Values/Sensitive Content"):
140
  fulltable = globaldf[globaldf['Group'] == 'CulturalEvals']
141
+ fulltable = fulltable[['Modality','Level', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
142
 
143
  gr.Markdown("""Cultural values are specific to groups and sensitive content is normative. Sensitive topics also vary by culture and can include hate speech. What is considered a sensitive topic, such as egregious violence or adult sexual content, can vary widely by viewpoint. Due to norms differing by culture, region, and language, there is no standard for what constitutes sensitive content.
144
  Distinct cultural values present a challenge for deploying models into a global sphere, as what may be appropriate in one culture may be unsafe in others. Generative AI systems cannot be neutral or objective, nor can they encompass truly universal values. There is no “view from nowhere”; in quantifying anything, a particular frame of reference is imposed.
 
152
  )
153
  type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
154
  value=["Model", "Dataset", "Output", "Taxonomy"],
155
+ label="Level",
156
  show_label=True,
157
  # info="Which modality to show."
158
  )
configs/crowspairs.yaml CHANGED
@@ -14,6 +14,6 @@ Screenshots:
14
  - Images/CrowsPairs1.png
15
  - Images/CrowsPairs2.png
16
  Suggested Evaluation: Crow-S Pairs
17
- Type: Dataset
18
  URL: https://arxiv.org/abs/2010.00133
19
  What it is evaluating: Protected class stereotypes
 
14
  - Images/CrowsPairs1.png
15
  - Images/CrowsPairs2.png
16
  Suggested Evaluation: Crow-S Pairs
17
+ Level: Dataset
18
  URL: https://arxiv.org/abs/2010.00133
19
  What it is evaluating: Protected class stereotypes
configs/homoglyphbias.yaml CHANGED
@@ -9,7 +9,7 @@ Link: Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
9
  Modality: Image
10
  Screenshots: []
11
  Suggested Evaluation: Effect of different scripts on text-to-image generation
12
- Type: Output
13
  URL: https://arxiv.org/pdf/2209.08891.pdf
14
  What it is evaluating: It evaluates generated images for cultural stereotypes, when
15
  using different scripts (homoglyphs). It somewhat measures the suceptibility of
 
9
  Modality: Image
10
  Screenshots: []
11
  Suggested Evaluation: Effect of different scripts on text-to-image generation
12
+ Level: Output
13
  URL: https://arxiv.org/pdf/2209.08891.pdf
14
  What it is evaluating: It evaluates generated images for cultural stereotypes, when
15
  using different scripts (homoglyphs). It somewhat measures the suceptibility of
configs/honest.yaml CHANGED
@@ -11,6 +11,6 @@ Link: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
11
  Modality: Text
12
  Screenshots: []
13
  Suggested Evaluation: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
14
- Type: Output
15
  URL: https://aclanthology.org/2021.naacl-main.191.pdf
16
  What it is evaluating: Protected class stereotypes and hurtful language
 
11
  Modality: Text
12
  Screenshots: []
13
  Suggested Evaluation: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
14
+ Level: Output
15
  URL: https://aclanthology.org/2021.naacl-main.191.pdf
16
  What it is evaluating: Protected class stereotypes and hurtful language
configs/ieat.yaml CHANGED
@@ -12,6 +12,6 @@ Link: Image Representations Learned With Unsupervised Pre-Training Contain Human
12
  Modality: Image
13
  Screenshots: []
14
  Suggested Evaluation: Image Embedding Association Test (iEAT)
15
- Type: Model
16
  URL: https://dl.acm.org/doi/abs/10.1145/3442188.3445932
17
  What it is evaluating: Embedding associations
 
12
  Modality: Image
13
  Screenshots: []
14
  Suggested Evaluation: Image Embedding Association Test (iEAT)
15
+ Level: Model
16
  URL: https://dl.acm.org/doi/abs/10.1145/3442188.3445932
17
  What it is evaluating: Embedding associations
configs/imagedataleak.yaml CHANGED
@@ -10,6 +10,6 @@ Link: 'Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias i
10
  Modality: Image
11
  Screenshots: []
12
  Suggested Evaluation: Dataset leakage and model leakage
13
- Type: Dataset
14
  URL: https://arxiv.org/abs/1811.08489
15
  What it is evaluating: Gender and label bias
 
10
  Modality: Image
11
  Screenshots: []
12
  Suggested Evaluation: Dataset leakage and model leakage
13
+ Level: Dataset
14
  URL: https://arxiv.org/abs/1811.08489
15
  What it is evaluating: Gender and label bias
configs/notmyvoice.yaml CHANGED
@@ -11,6 +11,6 @@ Modality: Audio
11
  Screenshots: []
12
  Suggested Evaluation: Not My Voice! A Taxonomy of Ethical and Safety Harms of Speech
13
  Generators
14
- Type: Taxonomy
15
  URL: https://arxiv.org/pdf/2402.01708.pdf
16
  What it is evaluating: Lists harms of audio/speech generators
 
11
  Screenshots: []
12
  Suggested Evaluation: Not My Voice! A Taxonomy of Ethical and Safety Harms of Speech
13
  Generators
14
+ Level: Taxonomy
15
  URL: https://arxiv.org/pdf/2402.01708.pdf
16
  What it is evaluating: Lists harms of audio/speech generators
configs/palms.yaml CHANGED
@@ -9,6 +9,6 @@ Link: 'Process for Adapting Language Models to Society (PALMS) with Values-Targe
9
  Modality: Text
10
  Screenshots: .nan
11
  Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
12
- Type: Output
13
  URL: http://arxiv.org/abs/2106.10328
14
  What it is evaluating: Adherence to defined norms for a set of cultural categories
 
9
  Modality: Text
10
  Screenshots: .nan
11
  Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
12
+ Level: Output
13
  URL: http://arxiv.org/abs/2106.10328
14
  What it is evaluating: Adherence to defined norms for a set of cultural categories
configs/stablebias.yaml CHANGED
@@ -9,6 +9,6 @@ Link: 'Stable bias: Analyzing societal representations in diffusion models'
9
  Modality: Image
10
  Screenshots: []
11
  Suggested Evaluation: Characterizing the variation in generated images
12
- Type: Output
13
  URL: https://arxiv.org/abs/2303.11408
14
  What it is evaluating: .nan
 
9
  Modality: Image
10
  Screenshots: []
11
  Suggested Evaluation: Characterizing the variation in generated images
12
+ Level: Output
13
  URL: https://arxiv.org/abs/2303.11408
14
  What it is evaluating: .nan
configs/stereoset.yaml CHANGED
@@ -11,6 +11,6 @@ Link: 'StereoSet: Measuring stereotypical bias in pretrained language models'
11
  Modality: Text
12
  Screenshots: []
13
  Suggested Evaluation: StereoSet
14
- Type: Dataset
15
  URL: https://arxiv.org/abs/2004.09456
16
  What it is evaluating: Protected class stereotypes
 
11
  Modality: Text
12
  Screenshots: []
13
  Suggested Evaluation: StereoSet
14
+ Level: Dataset
15
  URL: https://arxiv.org/abs/2004.09456
16
  What it is evaluating: Protected class stereotypes
configs/tango.yaml CHANGED
@@ -14,6 +14,6 @@ Screenshots:
14
  - Images/TANGO1.png
15
  - Images/TANGO2.png
16
  Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
17
- Type: Output
18
  URL: http://arxiv.org/abs/2106.10328
19
  What it is evaluating: Bias measurement for trans and nonbinary community via measuring gender non-affirmative language, specifically 1) misgendering 2), negative responses to gender disclosure
 
14
  - Images/TANGO1.png
15
  - Images/TANGO2.png
16
  Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
17
+ Level: Output
18
  URL: http://arxiv.org/abs/2106.10328
19
  What it is evaluating: Bias measurement for trans and nonbinary community via measuring gender non-affirmative language, specifically 1) misgendering 2), negative responses to gender disclosure
configs/videodiversemisinfo.yaml CHANGED
@@ -13,7 +13,7 @@ Modality: Video
13
  Screenshots: []
14
  Suggested Evaluation: 'Diverse Misinformation: Impacts of Human Biases on Detection
15
  of Deepfakes on Networks'
16
- Type: Output
17
  URL: https://arxiv.org/abs/2210.10026
18
  What it is evaluating: Human led evaluations of deepfakes to understand susceptibility
19
  and representational harms (including political violence)
 
13
  Screenshots: []
14
  Suggested Evaluation: 'Diverse Misinformation: Impacts of Human Biases on Detection
15
  of Deepfakes on Networks'
16
+ Level: Output
17
  URL: https://arxiv.org/abs/2210.10026
18
  What it is evaluating: Human led evaluations of deepfakes to understand susceptibility
19
  and representational harms (including political violence)
configs/weat.yaml CHANGED
@@ -36,7 +36,7 @@ Screenshots:
36
  - Images/WEAT1.png
37
  - Images/WEAT2.png
38
  Suggested Evaluation: Word Embedding Association Test (WEAT)
39
- Type: Model
40
  URL: https://researchportal.bath.ac.uk/en/publications/semantics-derived-automatically-from-language-corpora-necessarily
41
  What it is evaluating: Associations and word embeddings based on Implicit Associations
42
  Test (IAT)
 
36
  - Images/WEAT1.png
37
  - Images/WEAT2.png
38
  Suggested Evaluation: Word Embedding Association Test (WEAT)
39
+ Level: Model
40
  URL: https://researchportal.bath.ac.uk/en/publications/semantics-derived-automatically-from-language-corpora-necessarily
41
  What it is evaluating: Associations and word embeddings based on Implicit Associations
42
  Test (IAT)