Edit model card

"We must sleep, but AI Never Sleeps!"

 

Prompt Template

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: {prompt}
Assistant:

Simple-Usage

from transformers import AutoTokenizer
from transformers import AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("yanolja/EEVE-Korean-Instruct-2.8B-v1.0", trust_remote_code=True)

prompt_template = "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.\nHuman: {prompt}\nAssistant:\n"
text = '๋‹ค์ด์–ดํŠธ์‹ ๋ฉ”๋‰ด๋ฅผ ์ถ”์ฒœํ•ด์ฃผ์„ธ์š”.\n\n(A) ์ƒ๋Ÿฌ๋“œ\n(B) ์น˜ํ‚จ\n(C) ํ”ผ์ž\n(D) ํŒŒ์Šคํƒ€'
model_inputs = tokenizer(prompt_template.format(prompt=text), return_tensors='pt')

outputs = model.generate(**model_inputs, max_new_tokens=256)
output_text = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
print(output_text)

Example Output

A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
Human: ๋‹ค์ด์–ดํŠธ์‹ ๋ฉ”๋‰ด๋ฅผ ์ถ”์ฒœํ•ด์ฃผ์„ธ์š”.

(A) ์ƒ๋Ÿฌ๋“œ
(B) ์น˜ํ‚จ
(C) ํ”ผ์ž
(D) ํŒŒ์Šคํƒ€
Assistant:
(A) ์ƒ๋Ÿฌ๋“œ๋ฅผ ์ถ”์ฒœ๋“œ๋ฆฝ๋‹ˆ๋‹ค. ์ƒ๋Ÿฌ๋“œ๋Š” ์ €์นผ๋กœ๋ฆฌ์ด๋ฉด์„œ๋„ ์˜์–‘์†Œ๊ฐ€ ํ’๋ถ€ํ•ด ๋‹ค์ด์–ดํŠธ์‹์œผ๋กœ ์ ํ•ฉํ•ฉ๋‹ˆ๋‹ค. ๋‹ค์–‘ํ•œ ์ฑ„์†Œ์™€ ๋‹จ๋ฐฑ์งˆ์„ ์ถ”๊ฐ€ํ•˜์—ฌ ๊ท ํ˜• ์žกํžŒ ์‹์‚ฌ๋ฅผ ๋งŒ๋“œ์‹ค ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

About the Model

First of all, Overwhelming gratitude to 'yanolja/EEVE' Model & Team! This model is a fine-tuned version of crimsonjoo/Neversleep-3B-v0.1, which is a Korean vocabulary-extended version of microsoft/phi-2. Specifically, we utilized Direct Preference Optimization (DPO) through the use of Axolotl.

For more details, please refer to our technical report: Efficient and Effective Vocabulary Expansion Towards Multilingual Large Language Models.

Training Data

Downloads last month
9
Safetensors
Model size
2.82B params
Tensor type
BF16
ยท
Inference API
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for crimsonjoo/Neversleep-3B-Instruct-v0.1

Base model

microsoft/phi-2
Finetuned
this model