mssongit commited on
Commit
4fe2628
1 Parent(s): 9a4fd78

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -6,7 +6,7 @@ tags:
6
  - gpt-neox
7
  - KoAlpaca
8
  model-index:
9
- - name: KoAlpaca-Polyglot-12.8B
10
  results: []
11
  language:
12
  - ko
@@ -16,11 +16,9 @@ pipeline_tag: text-generation
16
  ---
17
 
18
 
19
- # KoAlpaca-Polyglot-12.8B (v1.1b)
20
 
21
- This model is a fine-tuned version of [EleutherAI/polyglot-ko-12.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b) on a KoAlpaca Dataset v1.1b
22
-
23
- Detail Codes are available at [KoAlpaca Github Repository](https://github.com/Beomi/KoAlpaca)
24
 
25
 
26
  ## Training procedure
@@ -31,8 +29,8 @@ The following hyperparameters were used during training:
31
  - learning_rate: 5e-05
32
  - train_batch_size: 1
33
  - seed: 42
34
- - distributed_type: multi-GPU (A100 80G)
35
- - num_devices: 4
36
  - gradient_accumulation_steps: 64
37
  - total_train_batch_size: 256
38
  - total_eval_batch_size: 32
 
6
  - gpt-neox
7
  - KoAlpaca
8
  model-index:
9
+ - name: KoAlpaca-Polyglot-5.8B
10
  results: []
11
  language:
12
  - ko
 
16
  ---
17
 
18
 
19
+ # KoAlpaca-Polyglot-5.8B (v1.1b)
20
 
21
+ This model is a fine-tuned version of [EleutherAI/polyglot-ko-5.8b](https://huggingface.co/EleutherAI/polyglot-ko-12.8b) on a KoAlpaca Dataset v1.1b
 
 
22
 
23
 
24
  ## Training procedure
 
29
  - learning_rate: 5e-05
30
  - train_batch_size: 1
31
  - seed: 42
32
+ - distributed_type: multi-GPU (A40 40G)
33
+ - num_devices: 8
34
  - gradient_accumulation_steps: 64
35
  - total_train_batch_size: 256
36
  - total_eval_batch_size: 32