kiddothe2b commited on
Commit
a190ec4
1 Parent(s): 5fce3df

initial commit

Browse files
Files changed (1) hide show
  1. README.md +78 -0
README.md ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - de
4
+ - fr
5
+ - it
6
+ pipeline_tag: fill-mask
7
+ license: cc-by-nc-sa-4.0
8
+ tags:
9
+ - legal
10
+ - fairlex
11
+ widget:
12
+ - text: "Aus seinem damaligen strafbaren Verhalten resultierte eine Forderung der Nachlassverwaltung eines <mask>, worüber eine aussergerichtliche Vereinbarung über Fr. 500'000."
13
+ - text: " Elle avait pour but social les <mask> dans le domaine des changes, en particulier l'exploitation d'une plateforme internet."
14
+ - text: "Il Pretore ha accolto la petizione con sentenza 16 luglio 2015, accordando all'attore l'importo <mask>, con interessi di mora a partire dalla notifica del precetto esecutivo, e ha rigettato in tale misura l'opposizione interposta a quest'ultimo."
15
+ ---
16
+
17
+
18
+ # FairLex: A multilingual benchmark for evaluating fairness in legal text processing
19
+
20
+ We present a benchmark suite of four datasets for evaluating the fairness of pre-trained legal language models and the techniques used to fine-tune them for downstream tasks. Our benchmarks cover four jurisdictions (European Council, USA, Swiss, and Chinese), five languages (English, German, French, Italian and Chinese) and fairness across five attributes (gender, age, nationality/region, language, and legal area). In our experiments, we evaluate pre-trained language models using several group-robust fine-tuning techniques and show that performance group disparities are vibrant in many cases, while none of these techniques guarantee fairness, nor consistently mitigate group disparities. Furthermore, we provide a quantitative and qualitative analysis of our results, highlighting open challenges in the development of robustness methods in legal NLP.
21
+
22
+ ---
23
+
24
+ Ilias Chalkidis, Tommaso Passini, Sheng Zhang, Letizia Tomada, Sebastian Felix Schwemer, and Anders Søgaard. 2022. FairLex: A multilingual bench-mark for evaluating fairness in legal text processing. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland.
25
+
26
+ ---
27
+
28
+ ## Pre-training details
29
+
30
+ For the purpose of this work, we release four domain-specific BERT models with continued pre-training on the corpora of the examined datasets (ECtHR, SCOTUS, FSCS, SPC).
31
+ We train mini-sized BERT models with 6 Transformer blocks, 384 hidden units, and 12 attention heads.
32
+ We warm-start all models from the public MiniLMv2 (Wang et al., 2021) using the distilled version of RoBERTa (Liu et al., 2019).
33
+ For the English datasets (ECtHR, SCOTUS) and the one distilled from XLM-R (Conneau et al., 2021) for the rest (trilingual FSCS, and Chinese SPC).
34
+
35
+ ## Models list
36
+
37
+ | Model name | Training corpora | Language |
38
+ |-----------------------------------|------------------|--------------------|
39
+ | `coastalcph/fairlex-ecthr-minlm` | ECtHR | `en` |
40
+ | `coastalcph/fairlex-scotus-minlm` | SCOTUS | `en` |
41
+ | `coastalcph/fairlex-fscs-minlm` | FSCS | [`de`, `fr`, `it`] |
42
+ | `coastalcph/fairlex-cail-minlm` | CAIL | `zh` |
43
+
44
+
45
+ ## Load Pretrained Model
46
+
47
+ ```python
48
+ from transformers import AutoTokenizer, AutoModel
49
+
50
+ tokenizer = AutoTokenizer.from_pretrained("coastalcph/fairlex-fscs-minlm")
51
+ model = AutoModel.from_pretrained("coastalcph/fairlex-fscs-minlm")
52
+ ```
53
+
54
+
55
+
56
+ ## Evaluation on downstream tasks
57
+
58
+ Consider the experiments in the article:
59
+
60
+ _Ilias Chalkidis, Tommaso Passini, Sheng Zhang, Letizia Tomada, Sebastian Felix Schwemer, and Anders Søgaard. 2022. Fairlex: A multilingual bench-mark for evaluating fairness in legal text processing. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland._
61
+
62
+
63
+ ## Author - Publication
64
+
65
+ ```
66
+ @inproceedings{chalkidis-2022-fairlex,
67
+ author={Chalkidis, Ilias and Passini, Tommaso and Zhang, Sheng and
68
+ Tomada, Letizia and Schwemer, Sebastian Felix and Søgaard, Anders},
69
+ title={FairLex: A Multilingual Benchmark for Evaluating Fairness in Legal Text Processing},
70
+ booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics},
71
+ year={2022},
72
+ address={Dublin, Ireland}
73
+ }
74
+ ```
75
+
76
+ Ilias Chalkidis on behalf of [CoAStaL NLP Group](https://coastalcph.github.io)
77
+
78
+ | Github: [@ilias.chalkidis](https://github.com/iliaschalkidis) | Twitter: [@KiddoThe2B](https://twitter.com/KiddoThe2B) |