Edit model card

license: llama2 language: - ko - en library_name: transformers base_model: mncai/llama2-13b-dpo-v7 pipeline_tag: text-generation

mnsim-dpo-peftmerged-2-eos

Our Team

Research & Engineering Product Management
David Sohn David Sohn

Model Details

Base Model

mncai/llama2-13b-dpo-v7

Trained On

  • OS: Ubuntu 22.04
  • GPU: A100 40GB 1ea
  • transformers: v4.35.2

Instruction format

It follows Custom format.

E.g.

text = """\
<s>
<|user|>
๊ฑด๊ฐ•ํ•œ ์‹์Šต๊ด€์„ ๋งŒ๋“ค๊ธฐ ์œ„ํ•ด์„œ๋Š” ์–ด๋–ป๊ฒŒ ํ•˜๋Š”๊ฒƒ์ด ์ข‹์„๊นŒ์š”?
<|assistant|>
"""

Implementation Code

This model contains the chat_template instruction format.
You can use the code below.

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="msy127/mnsim-dpo-peftmerged-2-eos")

# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("msy127/mnsim-dpo-peftmerged-2-eos")
model = AutoModelForCausalLM.from_pretrained("msy127/mnsim-dpo-peftmerged-2-eos")
Downloads last month
3,923
Safetensors
Model size
13.2B params
Tensor type
BF16
ยท