taeminlee commited on
Commit
ffe1905
1 Parent(s): faf6aaa

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - nlpai-lab/kullm-v2
5
+ language:
6
+ - ko
7
+ ---
8
+
9
+ # KULLM-Polyglot-5.8B-v2
10
+
11
+ This model is a parameter-efficient fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-5.8b) on a KULLM v2
12
+
13
+ Detail Codes are available at [KULLM Github Repository](https://github.com/nlpai-lab/KULLM)
14
+
15
+
16
+ ## Training procedure
17
+
18
+ ### Training hyperparameters
19
+
20
+ The following hyperparameters were used during training:
21
+
22
+ - learning_rate: 3e-4
23
+ - train_batch_size: 128
24
+ - seed: 42
25
+ - distributed_type: multi-GPU (A100 80G)
26
+ - num_devices: 4
27
+ - gradient_accumulation_steps: 8
28
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
29
+ - lr_scheduler_type: linear
30
+ - num_epochs: 8.0
31
+
32
+ ### Framework versions
33
+
34
+ - Transformers 4.28.1
35
+ - Pytorch 2.0.0+cu117
36
+ - Datasets 2.11.0
37
+ - Tokenizers 0.13.3