Spaces:
Runtime error
Runtime error
Avijit Ghosh
commited on
Commit
•
3d19330
1
Parent(s):
b2d9196
testing out yaml addition
Browse files- Images/CrowsPairs1.png +0 -0
- configs/crowspairs.yaml +4 -3
Images/CrowsPairs1.png
ADDED
configs/crowspairs.yaml
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
-
Abstract: .
|
2 |
'Applicable Models ': .nan
|
3 |
-
Authors: .
|
4 |
Considerations: Automating stereotype detection makes distinguishing harmful stereotypes
|
5 |
difficult. It also raises many false positives and can flag relatively neutral associations
|
6 |
based in fact (e.g. population x has a high proportion of lactose intolerant people).
|
@@ -10,7 +10,8 @@ Hashtags: .nan
|
|
10 |
Link: 'CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language
|
11 |
Models'
|
12 |
Modality: Text
|
13 |
-
Screenshots:
|
|
|
14 |
Suggested Evaluation: Crow-S Pairs
|
15 |
Type: Dataset
|
16 |
URL: https://arxiv.org/abs/2010.00133
|
|
|
1 |
+
Abstract: "Pretrained language models, especially masked language models (MLMs) have seen success across many NLP tasks. However, there is ample evidence that they use the cultural biases that are undoubtedly present in the corpora they are trained on, implicitly creating harm with biased representations. To measure some forms of social bias in language models against protected demographic groups in the US, we introduce the Crowdsourced Stereotype Pairs benchmark (CrowS-Pairs). CrowS-Pairs has 1508 examples that cover stereotypes dealing with nine types of bias, like race, religion, and age. In CrowS-Pairs a model is presented with two sentences: one that is more stereotyping and another that is less stereotyping. The data focuses on stereotypes about historically disadvantaged groups and contrasts them with advantaged groups. We find that all three of the widely-used MLMs we evaluate substantially favor sentences that express stereotypes in every category in CrowS-Pairs. As work on building less biased models advances, this dataset can be used as a benchmark to evaluate progress."
|
2 |
'Applicable Models ': .nan
|
3 |
+
Authors: Nikita Nangia, Clara Vania, Rasika Bhalerao, Samuel R. Bowman
|
4 |
Considerations: Automating stereotype detection makes distinguishing harmful stereotypes
|
5 |
difficult. It also raises many false positives and can flag relatively neutral associations
|
6 |
based in fact (e.g. population x has a high proportion of lactose intolerant people).
|
|
|
10 |
Link: 'CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language
|
11 |
Models'
|
12 |
Modality: Text
|
13 |
+
Screenshots:
|
14 |
+
- Images/CrowsPairs1.png
|
15 |
Suggested Evaluation: Crow-S Pairs
|
16 |
Type: Dataset
|
17 |
URL: https://arxiv.org/abs/2010.00133
|