File size: 1,635 Bytes
2551873 57469e1 2551873 a2caade fd409bf a2caade 57469e1 a2caade 57469e1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 |
---
license: apache-2.0
---
# OLMo 7B-Instruct-hf
> For more details on OLMO-7B-Instruct, refer to [Allen AI's OLMo-7B-Instruct model card](https://huggingface.co/allenai/OLMo-7B-Instruct).
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset.
The Instruct version is trained on the [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned).
OLMo 7B Instruct is trained for better question answering. They show the performance gain that OLMo base models can achieve with existing fine-tuning techniques.
**This version is for direct use with HuggingFace Transformers** from v4.40 on.
Run instructions are forthcoming.
For faster inference with llama.cpp or similar software supporting the GGUF format,
you can find this model as GGUF at [ssec-uw/OLMo-7B-Instruct-GGUF](https://huggingface.co/ssec-uw/OLMo-7B-Instruct-GGUF).
## Contact
For errors in this model card, contact Don or Anant, {landungs, anmittal} at uw dot edu.
## Acknowledgement
We would like to thank the hardworking folks at [Allen AI](https://huggingface.co/allenai) for providing the original model.
Additionally, the work to convert the model to the new `hf` version was done by the
[University of Washington Scientific Software Engineering Center (SSEC)](https://escience.washington.edu/software-engineering/ssec/),
as part of the [Schmidt Futures Virtual Institute for Scientific Software (VISS)](https://www.schmidtsciences.org/viss/). |