--- license: apache-2.0 --- # OLMo 7B-Instruct-hf > For more details on OLMO-7B-Instruct, refer to [Allen AI's OLMo-7B-Instruct model card](https://huggingface.co/allenai/OLMo-7B-Instruct). OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models. The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset. The Instruct version is trained on the [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned). OLMo 7B Instruct is trained for better question answering. They show the performance gain that OLMo base models can achieve with existing fine-tuning techniques. **This version is for direct use with HuggingFace Transformers** from v4.40 on. Run instructions are forthcoming. For faster inference with llama.cpp or similar software supporting the GGUF format, you can find this model as GGUF at [ssec-uw/OLMo-7B-Instruct-GGUF](https://huggingface.co/ssec-uw/OLMo-7B-Instruct-GGUF). ## Contact For errors in this model card, contact Don or Anant, {landungs, anmittal} at uw dot edu. ## Acknowledgement We would like to thank the hardworking folks at [Allen AI](https://huggingface.co/allenai) for providing the original model. Additionally, the work to convert the model to the new `hf` version was done by the [University of Washington Scientific Software Engineering Center (SSEC)](https://escience.washington.edu/software-engineering/ssec/), as part of the [Schmidt Futures Virtual Institute for Scientific Software (VISS)](https://www.schmidtsciences.org/viss/).