Edit model card

The Cabrita model is a collection of continued pre-trained and tokenizer-adapted models for the Portuguese language. This artifact is the 3 billion size variant.

The weights were initially obtained from the open-llama project (https://github.com/openlm-research/open_llama) in the open_llama_3b option.

@misc{larcher2023cabrita,
      title={Cabrita: closing the gap for foreign languages}, 
      author={Celio Larcher and Marcos Piau and Paulo Finardi and Pedro Gengo and Piero Esposito and Vinicius Caridรก},
      year={2023},
      eprint={2308.11878},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 35.54
AI2 Reasoning Challenge (25-Shot) 33.79
HellaSwag (10-Shot) 55.35
MMLU (5-Shot) 25.16
TruthfulQA (0-shot) 38.50
Winogrande (5-shot) 59.43
GSM8k (5-shot) 0.99
Downloads last month
5,021

Spaces using 22h/open-cabrita3b 3

Evaluation results