Edit model card
ENERGY-DRINK-LOVE/DataVortexS_dpov3
Our Team
- Youjin Chung
- Jingyeom Kim
Model
Base Model
Hardware and Software
- Hardware: A100 * 8 for training our model
- Deepspeed library & Huggingface TRL Trainer
Dataset
- DPO_dataset
- ์์ฒด ์ ์ dpo dataset(AI-hub dataset ํ์ฉ)
- OpenOrca DPO ๋ฑ ์์ด ๋ฐ์ดํฐ์
๋ฒ์ญ(ENERGY-DRINK-LOVE/translate_share_gpt_dedup_llama_SFT_1024, ์์ฒด๋ชจ๋ธ ํ์ฉ)
Training Method
Benchmark
Ko LM Eval Harness
Ko-LLM-Leaderboard
- (240316๊ธฐ์ค 7๋ฑ)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6551c0e37bbfce18781a8748/S4cpra6iTlzCdN7PP6A3o.png)
Average |
Ko-ARC |
Ko-HellaSwag |
Ko-MMLU |
Ko-TruthfulQA |
Ko-CommonGen V2 |
60.18 |
56.23 |
69.15 |
52.76 |
67.87 |
54.9 |
- Downloads last month
- 1,169
Finetuned from