jarodrigues commited on
Commit
7054784
1 Parent(s): a968e64

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -11
README.md CHANGED
@@ -31,22 +31,22 @@ widget:
31
 
32
  ---
33
 
34
- # Albertina PT-PT base
35
 
36
 
37
- **Albertina PT-PT base** is a foundation, large language model for European **Portuguese** from **Portugal**.
38
 
39
  It is an **encoder** of the BERT family, based on the neural architecture Transformer and
40
  developed over the DeBERTa model, with most competitive performance for this language.
41
  It is distributed free of charge and under a most permissible license.
42
 
43
- You may be also interested in [**Albertina 900m PT-PT**](https://huggingface.co/PORTULAN/albertina-ptpt).
44
  This is a larger version,
45
  and to the best of our knowledge, this is an encoder specifically for this language and variant
46
  that, at the time of its initial distribution, sets a new state of the art for it, and is made publicly available
47
  and distributed for reuse.
48
 
49
- **Albertina PT-PT base** is developed by a joint team from the University of Lisbon and the University of Porto, Portugal.
50
  For further details, check the respective [publication](https://arxiv.org/abs/?):
51
 
52
  ``` latex
@@ -68,7 +68,7 @@ Please use the above cannonical reference when using or citing this model.
68
 
69
  # Model Description
70
 
71
- **This model card is for Albertina-PT-PT base**, with 100M parameters, 12 layers and a hidden size of 768.
72
 
73
  Albertina-PT-PT base is distributed under an [MIT license](https://huggingface.co/PORTULAN/albertina-ptpt/blob/main/LICENSE).
74
 
@@ -79,7 +79,7 @@ DeBERTa is distributed under an [MIT license](https://github.com/microsoft/DeBER
79
 
80
  # Training Data
81
 
82
- [**Albertina-PT-PT base**](https://huggingface.co/PORTULAN/albertina-ptpt-base) was trained over a 2.2 billion token data set that resulted from gathering some openly available corpora of European Portuguese from the following sources:
83
 
84
  - [OSCAR](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301): the OSCAR data set includes documents in more than one hundred languages, including Portuguese, and it is widely used in the literature. It is the result of a selection performed over the [Common Crawl](https://commoncrawl.org/) data set, crawled from the Web, that retains only pages whose metadata indicates permission to be crawled, that performs deduplication, and that removes some boilerplate, among other filters. Given that it does not discriminate between the Portuguese variants, we performed extra filtering by retaining only documents whose meta-data indicate the Internet country code top-level domain of Portugal. We used the January 2023 version of OSCAR, which is based on the November/December 2022 version of Common Crawl.
85
  - [DCEP](https://joint-research-centre.ec.europa.eu/language-technology-resources/dcep-digital-corpus-european-parliament_en): the Digital Corpus of the European Parliament is a multilingual corpus including documents in all official EU languages published on the European Parliament's official website. We retained its European Portuguese portion.
@@ -89,7 +89,7 @@ DeBERTa is distributed under an [MIT license](https://github.com/microsoft/DeBER
89
 
90
  ## Preprocessing
91
 
92
- We filtered the PT-PT corpora using the [BLOOM pre-processing](https://github.com/bigscience-workshop/data-preparation) pipeline.
93
  We skipped the default filtering of stopwords since it would disrupt the syntactic structure, and also the filtering for language identification given the corpus was pre-selected as Portuguese.
94
 
95
 
@@ -97,7 +97,7 @@ We skipped the default filtering of stopwords since it would disrupt the syntact
97
 
98
  As codebase, we resorted to the [DeBERTa V1 base](https://huggingface.co/microsoft/deberta-base), for English.
99
 
100
- To train [**Albertina-PT-PT base**](https://huggingface.co/PORTULAN/albertina-ptpt-base), the data set was tokenized with the original DeBERTa tokenizer with a 128 token sequence truncation and dynamic padding.
101
  The model was trained using the maximum available memory capacity resulting in a batch size of 3072 samples (192 samples per GPU).
102
  We opted for a learning rate of 1e-5 with linear decay and 10k warm-up steps.
103
  A total of 200 training epochs were performed resulting in approximately 180k steps.
@@ -114,13 +114,13 @@ The base model version was evaluated on downstream tasks, namely the translation
114
  ## GLUE tasks translated
115
 
116
 
117
- We resorted to [GLUE-PT](https://huggingface.co/datasets/PORTULAN/glue-ptpt), a **PT-PT version of the GLUE** benchmark.
118
  We automatically translated the same four tasks from GLUE using [DeepL Translate](https://www.deepl.com/), which specifically provides translation from English to PT-PT as an option.
119
 
120
  | Model | RTE (Accuracy) | WNLI (Accuracy)| MRPC (F1) | STS-B (Pearson) |
121
  |--------------------------|----------------|----------------|-----------|-----------------|
122
- | **Albertina 900m PT-PT** | **0.8339** | 0.4225 | **0.9171**| **0.8801** |
123
- | **Albertina-PT-PT base** | 0.6787 | **0.4507** | 0.8829 | 0.8581 |
124
 
125
  <br>
126
 
 
31
 
32
  ---
33
 
34
+ # Albertina 100M PTPT
35
 
36
 
37
+ **Albertina 100M PTPT** is a foundation, large language model for European **Portuguese** from **Portugal**.
38
 
39
  It is an **encoder** of the BERT family, based on the neural architecture Transformer and
40
  developed over the DeBERTa model, with most competitive performance for this language.
41
  It is distributed free of charge and under a most permissible license.
42
 
43
+ You may be also interested in [**Albertina 900m PTPT**](https://huggingface.co/PORTULAN/albertina-ptpt).
44
  This is a larger version,
45
  and to the best of our knowledge, this is an encoder specifically for this language and variant
46
  that, at the time of its initial distribution, sets a new state of the art for it, and is made publicly available
47
  and distributed for reuse.
48
 
49
+ **Albertina 100M PTPT** is developed by a joint team from the University of Lisbon and the University of Porto, Portugal.
50
  For further details, check the respective [publication](https://arxiv.org/abs/?):
51
 
52
  ``` latex
 
68
 
69
  # Model Description
70
 
71
+ **This model card is for Albertina 100M PTPT base**, with 100M parameters, 12 layers and a hidden size of 768.
72
 
73
  Albertina-PT-PT base is distributed under an [MIT license](https://huggingface.co/PORTULAN/albertina-ptpt/blob/main/LICENSE).
74
 
 
79
 
80
  # Training Data
81
 
82
+ [**Albertina 100M PTPT**](https://huggingface.co/PORTULAN/albertina-ptpt-base) was trained over a 2.2 billion token data set that resulted from gathering some openly available corpora of European Portuguese from the following sources:
83
 
84
  - [OSCAR](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301): the OSCAR data set includes documents in more than one hundred languages, including Portuguese, and it is widely used in the literature. It is the result of a selection performed over the [Common Crawl](https://commoncrawl.org/) data set, crawled from the Web, that retains only pages whose metadata indicates permission to be crawled, that performs deduplication, and that removes some boilerplate, among other filters. Given that it does not discriminate between the Portuguese variants, we performed extra filtering by retaining only documents whose meta-data indicate the Internet country code top-level domain of Portugal. We used the January 2023 version of OSCAR, which is based on the November/December 2022 version of Common Crawl.
85
  - [DCEP](https://joint-research-centre.ec.europa.eu/language-technology-resources/dcep-digital-corpus-european-parliament_en): the Digital Corpus of the European Parliament is a multilingual corpus including documents in all official EU languages published on the European Parliament&#39;s official website. We retained its European Portuguese portion.
 
89
 
90
  ## Preprocessing
91
 
92
+ We filtered the PTPT corpora using the [BLOOM pre-processing](https://github.com/bigscience-workshop/data-preparation) pipeline.
93
  We skipped the default filtering of stopwords since it would disrupt the syntactic structure, and also the filtering for language identification given the corpus was pre-selected as Portuguese.
94
 
95
 
 
97
 
98
  As codebase, we resorted to the [DeBERTa V1 base](https://huggingface.co/microsoft/deberta-base), for English.
99
 
100
+ To train [**Albertina 100M PTPT**](https://huggingface.co/PORTULAN/albertina-ptpt-base), the data set was tokenized with the original DeBERTa tokenizer with a 128 token sequence truncation and dynamic padding.
101
  The model was trained using the maximum available memory capacity resulting in a batch size of 3072 samples (192 samples per GPU).
102
  We opted for a learning rate of 1e-5 with linear decay and 10k warm-up steps.
103
  A total of 200 training epochs were performed resulting in approximately 180k steps.
 
114
  ## GLUE tasks translated
115
 
116
 
117
+ We resorted to [GLUE-PT](https://huggingface.co/datasets/PORTULAN/glue-ptpt), a **PTPT version of the GLUE** benchmark.
118
  We automatically translated the same four tasks from GLUE using [DeepL Translate](https://www.deepl.com/), which specifically provides translation from English to PT-PT as an option.
119
 
120
  | Model | RTE (Accuracy) | WNLI (Accuracy)| MRPC (F1) | STS-B (Pearson) |
121
  |--------------------------|----------------|----------------|-----------|-----------------|
122
+ | **Albertina 900m PTPT** | **0.8339** | 0.4225 | **0.9171**| **0.8801** |
123
+ | **Albertina 100m PTPT** | 0.6787 | **0.4507** | 0.8829 | 0.8581 |
124
 
125
  <br>
126