--- license: apache-2.0 language: - pt - en --- The Cabrita model is a collection of continued pre-trained and tokenizer-adapted models for the Portuguese language. This artifact is the 3 billion size variant. The weights were initially obtained from the open-llama project (https://github.com/openlm-research/open_llama) in the open_llama_3b option. ``` @misc{larcher2023cabrita, title={Cabrita: closing the gap for foreign languages}, author={Celio Larcher and Marcos Piau and Paulo Finardi and Pedro Gengo and Piero Esposito and Vinicius Caridá}, year={2023}, eprint={2308.11878}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```