|
--- |
|
language: |
|
- en |
|
library_name: transformers |
|
tags: |
|
- orpo |
|
- llama 3 |
|
- sft |
|
datasets: |
|
- Open-Orca/OpenOrca |
|
--- |
|
|
|
# Model description |
|
|
|
Meta-Llama-3-8B-OpenOrca is a fine-tuned version of the [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on 1.5k subsamples of the |
|
[OpenOrca](https://huggingface.co/datasets/Open-Orca/OpenOrca) dataset. |
|
|
|
This model follows the ChatML chat template! |
|
|
|
## How to use |
|
|
|
```` |
|
import torch |
|
from transformers import AutoTokenizer, pipeline |
|
|
|
model = "MuntasirHossain/Meta-Llama-3-8B-OpenOrca" |
|
tokenizer = AutoTokenizer.from_pretrained(model) |
|
|
|
llm = pipeline( |
|
task = "text-generation", |
|
model=model, |
|
eos_token_id=tokenizer.eos_token_id, |
|
torch_dtype=torch.float16, |
|
max_new_tokens=256, |
|
do_sample=True, |
|
device_map="auto", |
|
) |
|
|
|
def generate(input_text): |
|
system_prompt = "You are a helpful AI assistant." |
|
messages = [ |
|
{"role": "system", "content": system_prompt}, |
|
{"role": "user", "content": input_text}, |
|
] |
|
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
|
outputs = llm(prompt) |
|
return outputs[0]["generated_text"][len(prompt):] |
|
|
|
generate("What is a large language model?") |
|
```` |
|
|
|
|