nicholasKluge commited on
Commit
d902840
1 Parent(s): 5a66b9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -141,26 +141,26 @@ trainer.train()
141
 
142
  ## Fine-Tuning Comparisons
143
 
144
- | Models | [FaQuAD-NLI](https://huggingface.co/datasets/ruanchaves/faquad-nli) |
145
- |--------------------------------------------------------------------------------------------|---------------------------------------------------------------------|
146
- | [Bert-base-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased) | 93.07 |
147
- | [Bert-large-portuguese-cased](https://huggingface.co/neuralmind/bert-base-portuguese-cased)| 92.26 |
148
- | [Teeny Tiny Llama 460m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m) | 91.18 |
149
- | [Teeny Tiny Llama 160m](https://huggingface.co/nicholasKluge/TeenyTinyLlama-160m) | 90.00 |
150
- | [Gpt2-small-portuguese](https://huggingface.co/pierreguillou/gpt2-small-portuguese) | 86.46 |
 
 
 
151
 
152
  ## Cite as 🤗
153
 
154
  ```latex
155
 
156
- @misc{nicholas22llama,
157
- doi = {10.5281/zenodo.6989727},
158
- url = {https://huggingface.co/nicholasKluge/TeenyTinyLlama-160m},
159
- author = {Nicholas Kluge Corrêa},
160
- title = {TeenyTinyLlama},
161
- year = {2023},
162
- publisher = {HuggingFace},
163
- journal = {HuggingFace repository},
164
  }
165
 
166
  ```
 
141
 
142
  ## Fine-Tuning Comparisons
143
 
144
+ To further evaluate the downstream capabilities of our models, we decided to employ a basic fine-tuning procedure for our TTL pair on a subset of tasks from the Poeta benchmark. We apply the same procedure for comparison purposes on both [BERTimbau](https://huggingface.co/neuralmind/bert-base-portuguese-cased) models, given that they are also LLM trained from scratch in Brazilian Portuguese and have a similar size range to our models. We used these comparisons to assess if our pre-training runs produced LLM capable of producing good results ("good" here means "close to BERTimbau") when utilized for downstream applications.
145
+
146
+ | Models | IMDB | FaQuAD-NLI | HateBr | Assin2 | AgNews | Average |
147
+ |-----------------|-----------|------------|-----------|-----------|-----------|---------|
148
+ | BERTimbau-large | **93.58** | 92.26 | 91.57 | **88.97** | 94.11 | 92.10 |
149
+ | BERTimbau-small | 92.22 | **93.07** | 91.28 | 87.45 | 94.19 | 91.64 |
150
+ | **TTL-460m** | 91.64 | 91.18 | **92.28** | 86.43 | **94.42** | 91.19 |
151
+ | **TTL-160m** | 91.14 | 90.00 | 90.71 | 85.78 | 94.05 | 90.34 |
152
+
153
+ All the shown results are the higher accuracy scores achieved on the respective task test sets after fine-tuning the models on the training sets. All fine-tuning runs used the same hyperparameters, and the code implementation can be found in the [model cards](https://huggingface.co/nicholasKluge/TeenyTinyLlama-460m-HateBR) of our fine-tuned models.
154
 
155
  ## Cite as 🤗
156
 
157
  ```latex
158
 
159
+ @misc{correa24ttllama,
160
+ title = {TeenyTinyLlama: a pair of open-source tiny language models trained in Brazilian Portuguese},
161
+ author = {Corr{\^e}a, Nicholas Kluge and Falk, Sophia and Fatimah, Shiza and Sen, Aniket and De Oliveira, Nythamar},
162
+ journal={arXiv},
163
+ year = {2024},
 
 
 
164
  }
165
 
166
  ```