open-cabrita3b / README.md
celiolarcher's picture
Update README.md
fc2a2de
|
raw
history blame
675 Bytes
metadata
license: apache-2.0
language:
  - pt
  - en

The Cabrita model is a collection of continued pre-trained and tokenizer-adapted models for the Portuguese language. This artifact is the 3 billion size variant.

The weights were initially obtained from the open-llama project (https://github.com/openlm-research/open_llama) in the open_llama_3b option.

@misc{larcher2023cabrita,
      title={Cabrita: closing the gap for foreign languages}, 
      author={Celio Larcher and Marcos Piau and Paulo Finardi and Pedro Gengo and Piero Esposito and Vinicius Caridá},
      year={2023},
      eprint={2308.11878},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}