File size: 3,273 Bytes
19dd71b
 
3aac946
 
 
 
 
 
 
19dd71b
 
 
 
 
3aac946
19dd71b
3aac946
19dd71b
3aac946
19dd71b
23f93ad
 
19dd71b
3aac946
19dd71b
 
 
 
 
 
 
23f93ad
 
 
3aac946
23f93ad
 
3aac946
 
23f93ad
3aac946
 
19dd71b
 
 
 
 
23f93ad
 
 
19dd71b
 
23f93ad
 
 
 
 
 
 
 
 
 
19dd71b
23f93ad
19dd71b
 
 
 
 
 
 
23f93ad
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19dd71b
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
library_name: transformers
license: apache-2.0
datasets:
- squarelike/OpenOrca-gugugo-ko
language:
- ko
- en
pipeline_tag: text-generation
---

# Model Card for Model ID

<!-- Provide a quick summary of what the model is/does. -->
LLaMA2 7b 모델의 한국어 CP(Continual Pre-trained)모델인 https://huggingface.co/beomi/llama-2-ko-7b 모델의 instruction tuning 모델

transformers와 trl을 이용하여 QLoRA로 훈련 진행

Dataset은 https://huggingface.co/datasets/squarelike/OpenOrca-gugugo-ko

QLoRA 훈련 테스트를 위한 훈련 결과물


## Model Details

### Model Sources [optional]


## Uses

<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->

### Prompt template

```
"""
### instruction:
### intput:
### output:
"""
```


## Training Details

#### Training Hyperparameters

- **Training regime:** [bf16 mixed precision] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->

- <PAD>토큰 추가 후 right 패딩사이드 지정하여 진행


- LoRA config
    ```
peft_config = LoraConfig(
    lora_alpha=16,
    lora_dropout=0.1,
    r=64,
    bias="none",
    task_type="CAUSAL_LM"
)
    ```

    

## Evaluation

<!-- This section describes the evaluation protocols and provides the results. -->

### Testing Data, Factors & Metrics

link: https://github.com/Beomi/ko-lm-evaluation-harness

results/all/aeolian83/llama_ko_sft_gugugo_experi_01
|                                  |        0 |        5 |
|:---------------------------------|---------:|---------:|
| kobest_boolq (macro_f1)          | 0.588382 | 0.384051 |
| kobest_copa (macro_f1)           | 0.749558 | 0.778787 |
| kobest_hellaswag (macro_f1)      | 0.439247 | 0.439444 |
| kobest_sentineg (macro_f1)       | 0.448283 | 0.934415 |
| kohatespeech (macro_f1)          | 0.244828 | 0.371245 |
| kohatespeech_apeach (macro_f1)   | 0.337434 | 0.394607 |
| kohatespeech_gen_bias (macro_f1) | 0.135272 | 0.461714 |
| korunsmile (f1)                  | 0.254562 | 0.315907 |
| nsmc (acc)                       | 0.61248  | 0.84256  |
| pawsx_ko (acc)                   | 0.5615   | 0.5365   |


results/all/beomi/llama-2-ko-7b
|                                  |        0 |        5 |       10 |       50 |
|:---------------------------------|---------:|---------:|---------:|---------:|
| kobest_boolq (macro_f1)          | 0.612147 | 0.682832 | 0.713392 | 0.71622  |
| kobest_copa (macro_f1)           | 0.759784 | 0.799843 | 0.807907 | 0.829976 |
| kobest_hellaswag (macro_f1)      | 0.447951 | 0.460632 | 0.464623 | 0.458628 |
| kobest_sentineg (macro_f1)       | 0.3517   | 0.969773 | 0.977329 | 0.97481  |
| kohatespeech (macro_f1)          | 0.314636 | 0.383336 | 0.357491 | 0.366585 |
| kohatespeech_apeach (macro_f1)   | 0.346127 | 0.567627 | 0.583391 | 0.629269 |
| kohatespeech_gen_bias (macro_f1) | 0.204651 | 0.509189 | 0.471078 | 0.451119 |
| korunsmile (f1)                  | 0.290663 | 0.306208 | 0.304279 | 0.343946 |
| nsmc (acc)                       | 0.57942  | 0.84242  | 0.87368  | 0.8939   |
| pawsx_ko (acc)                   | 0.538    | 0.52     | 0.5275   | 0.5195   |