Edit model card

license: llama2 language: - ko library_name: transformers base_model: beomi/llama-2-ko-7b pipeline_tag: text-generation

msy127/ft_240201_01

Our Team

Research & Engineering Product Management
David Sohn David Sohn

Model Details

Base Model

beomi/llama-2-ko-7b

Trained On

  • OS: Ubuntu 22.04
  • GPU: A100 40GB 1ea
  • transformers: v4.37

Instruction format

It follows Custom format.

E.g.

text = """\
<|user|>
๊ฑด๊ฐ•ํ•œ ์‹์Šต๊ด€์„ ๋งŒ๋“ค๊ธฐ ์œ„ํ•ด์„œ๋Š” ์–ด๋–ป๊ฒŒ ํ•˜๋Š”๊ฒƒ์ด ์ข‹์„๊นŒ์š”?
<|assistant|>
"""

Implementation Code

This model contains the chat_template instruction format.
You can use the code below.

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="msy127/ft_240201_01")

# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("msy127/ft_240201_01")
model = AutoModelForCausalLM.from_pretrained("msy127/ft_240201_01")
Downloads last month
1,247
Safetensors
Model size
6.86B params
Tensor type
BF16
ยท