--- license: apache-2.0 language: - en library_name: transformers --- # OpenLLaMA-3B-Chat: Chat Model on top of Open Reproduction of LLaMA ## Training Traces [wandb](https://wandb.ai/autoai-org/fmengine/runs/3ddwtzyl/overview?workspace=user-xzyaoi) ## Prompt Format ``` : Who is Alan Turing?<|endoftext|>: ``` ## Reference If you found OpenLLaMA useful in your research or applications, please cite using the following BibTeX: ``` @software{Yao_FMEngine_Library_for_2023, author = {Yao, Xiaozhe}, doi = {10.5281/zenodo.8314779}, month = sep, title = {{FMEngine: Library for Training Foundation Models}}, url = {https://github.com/eth-easl/fmengine}, version = {0.0.1}, year = {2023} } @software{openlm2023openllama, author = {Geng, Xinyang and Liu, Hao}, title = {OpenLLaMA: An Open Reproduction of LLaMA}, month = May, year = 2023, url = {https://github.com/openlm-research/open_llama} } @software{together2023redpajama, author = {Together Computer}, title = {RedPajama-Data: An Open Source Recipe to Reproduce LLaMA training dataset}, month = April, year = 2023, url = {https://github.com/togethercomputer/RedPajama-Data} } @article{touvron2023llama, title={Llama: Open and efficient foundation language models}, author={Touvron, Hugo and Lavril, Thibaut and Izacard, Gautier and Martinet, Xavier and Lachaux, Marie-Anne and Lacroix, Timoth{\'e}e and Rozi{\`e}re, Baptiste and Goyal, Naman and Hambro, Eric and Azhar, Faisal and others}, journal={arXiv preprint arXiv:2302.13971}, year={2023} } ``` ## Limitations and Bias As with all language models, `openllama-3b-chat` may generate incorrect or biased content. It's important to keep this in mind when using the model.