Cartinoe5930
commited on
Commit
โข
3293663
1
Parent(s):
f6d72bd
Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,6 @@ language:
|
|
6 |
- ko
|
7 |
library_name: transformers
|
8 |
---
|
9 |
-
|
10 |
## KoRAE
|
11 |
|
12 |
<p align="center"><img src="https://cdn-uploads.huggingface.co/production/uploads/63e087b6a98d931aa90c1b9c/XQ-pNzRDRccd7UFgYDOrx.png", width='300', height='300'></p>
|
@@ -18,12 +17,13 @@ We utilized special data filtering methods which introduced in [AlpaGasus](https
|
|
18 |
We finetuned [Korean Llama-2](https://huggingface.co/beomi/llama-2-koen-13b) that introduced by [@beomi](https://huggingface.co/beomi) on the filtered dataset.
|
19 |
The Flash-Attention2 and LoRA were utilized for efficient finetuning.
|
20 |
|
21 |
-
The
|
22 |
-
In addition, the DPO trained version of KoRAE will be uploaded soon too!
|
23 |
-
Stay tuned for the update of KoRAE!
|
24 |
|
25 |
-
|
|
|
|
|
26 |
|
|
|
27 |
- **Developed by:** [Cartinoe5930](https://huggingface.co/Cartinoe5930)
|
28 |
- **Base model:** [beomi/llama-2-koen-13b](https://huggingface.co/beomi/llama-2-koen-13b)
|
29 |
- **Repository:** [gauss5930/KoRAE](https://github.com/gauss5930/KoRAE)
|
@@ -34,6 +34,7 @@ For more details, please check the GitHub Repository!
|
|
34 |
|
35 |
- **Hardward:** We utilized A100 80G for finetuning
|
36 |
- **Training factors:** The [Transformers Trainer](https://huggingface.co/docs/transformers/main_classes/trainer) and [Huggingface PEFT](https://huggingface.co/docs/peft/index) were utilized for finetuning.
|
|
|
37 |
|
38 |
For more details, please check the GitHub Repository!
|
39 |
|
@@ -45,6 +46,10 @@ For more information, please refer to the [dataset card](https://huggingface.co/
|
|
45 |
|
46 |
## Open Ko-LLM Leaderboard
|
47 |
|
|
|
|
|
|
|
|
|
48 |
## Prompt Template
|
49 |
|
50 |
```
|
@@ -73,7 +78,6 @@ messages = [
|
|
73 |
},
|
74 |
{"role": "user", "content": "์คํธ๋ ์ค๋ฅผ ํด์ํ๋ 5๊ฐ์ง ๋ฐฉ๋ฒ์ ๋ํด์ ์ค๋ช
ํด์ค."}
|
75 |
]
|
76 |
-
|
77 |
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
78 |
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
79 |
print(outputs[0]["generated_text"])
|
|
|
6 |
- ko
|
7 |
library_name: transformers
|
8 |
---
|
|
|
9 |
## KoRAE
|
10 |
|
11 |
<p align="center"><img src="https://cdn-uploads.huggingface.co/production/uploads/63e087b6a98d931aa90c1b9c/XQ-pNzRDRccd7UFgYDOrx.png", width='300', height='300'></p>
|
|
|
17 |
We finetuned [Korean Llama-2](https://huggingface.co/beomi/llama-2-koen-13b) that introduced by [@beomi](https://huggingface.co/beomi) on the filtered dataset.
|
18 |
The Flash-Attention2 and LoRA were utilized for efficient finetuning.
|
19 |
|
20 |
+
The finding of KoRAE is as follows:
|
|
|
|
|
21 |
|
22 |
+
1. The finetuning in some epochs showed that high-quality filtered data has positive effects on model's performance. However, finetuning in a few epochs, the quantity of data is more matter than quality. It seems to be due to the lack of performance of the Korean base model. Therefore, the research to improve the Korean base model must continue.
|
23 |
+
2. The model trained with DPO showed best performance among KoRAE variants. This shows that DPO is clearly effective in the Korean LLM.
|
24 |
+
3. The model finetuned with filtered high-quality KoRAE showed better performance than without. Therefore, for better LLM, we should try to finetune the LLM with high-quality data.
|
25 |
|
26 |
+
## Model Details
|
27 |
- **Developed by:** [Cartinoe5930](https://huggingface.co/Cartinoe5930)
|
28 |
- **Base model:** [beomi/llama-2-koen-13b](https://huggingface.co/beomi/llama-2-koen-13b)
|
29 |
- **Repository:** [gauss5930/KoRAE](https://github.com/gauss5930/KoRAE)
|
|
|
34 |
|
35 |
- **Hardward:** We utilized A100 80G for finetuning
|
36 |
- **Training factors:** The [Transformers Trainer](https://huggingface.co/docs/transformers/main_classes/trainer) and [Huggingface PEFT](https://huggingface.co/docs/peft/index) were utilized for finetuning.
|
37 |
+
- **Training Details:** Supervised finetuning 1 epoch on [filtered KoRAE](https://huggingface.co/datasets/Cartinoe5930/KoRAE_filtered_12k) dataset
|
38 |
|
39 |
For more details, please check the GitHub Repository!
|
40 |
|
|
|
46 |
|
47 |
## Open Ko-LLM Leaderboard
|
48 |
|
49 |
+
|Model|Average|Ko-ARC|Ko-HellaSwag|Ko-MMLU|Ko-TruthfulQA|Ko-CommonGen V2|
|
50 |
+
|---|---|---|---|---|---|---|
|
51 |
+
|weak-KoRAE-13b|48.1|45.22|56.79|42|40.4|56.08|
|
52 |
+
|
53 |
## Prompt Template
|
54 |
|
55 |
```
|
|
|
78 |
},
|
79 |
{"role": "user", "content": "์คํธ๋ ์ค๋ฅผ ํด์ํ๋ 5๊ฐ์ง ๋ฐฉ๋ฒ์ ๋ํด์ ์ค๋ช
ํด์ค."}
|
80 |
]
|
|
|
81 |
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
82 |
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
83 |
print(outputs[0]["generated_text"])
|