File size: 1,635 Bytes
2551873
 
 
 
57469e1
2551873
 
 
 
 
 
 
 
 
a2caade
 
fd409bf
 
 
 
a2caade
57469e1
a2caade
57469e1
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
license: apache-2.0
---

# OLMo 7B-Instruct-hf

> For more details on OLMO-7B-Instruct, refer to [Allen AI's OLMo-7B-Instruct model card](https://huggingface.co/allenai/OLMo-7B-Instruct).

OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
The OLMo base models are trained on the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset.
The Instruct version is trained on the [cleaned version of the UltraFeedback dataset](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned).

OLMo 7B Instruct is trained for better question answering. They show the performance gain that OLMo base models can achieve with existing fine-tuning techniques.

**This version is for direct use with HuggingFace Transformers** from v4.40 on.

Run instructions are forthcoming.

For faster inference with llama.cpp or similar software supporting the GGUF format,
you can find this model as GGUF at [ssec-uw/OLMo-7B-Instruct-GGUF](https://huggingface.co/ssec-uw/OLMo-7B-Instruct-GGUF).

## Contact

For errors in this model card, contact Don or Anant, {landungs, anmittal} at uw dot edu.

## Acknowledgement

We would like to thank the hardworking folks at [Allen AI](https://huggingface.co/allenai) for providing the original model.

Additionally, the work to convert the model to the new `hf` version was done by the
[University of Washington Scientific Software Engineering Center (SSEC)](https://escience.washington.edu/software-engineering/ssec/),
as part of the [Schmidt Futures Virtual Institute for Scientific Software (VISS)](https://www.schmidtsciences.org/viss/).