chlee10 commited on
Commit
155c575
โ€ข
1 Parent(s): b6f894e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
 
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ pipeline_tag: text-generation
3
  license: apache-2.0
4
+ language:
5
+ - en
6
+ tags:
7
+ - Open-platypus-Commercial
8
+ base_model: bardsai/jaskier-7b-dpo-v6.1
9
+ datasets:
10
+ - kyujinpy/Open-platypus-Commercial
11
+ model-index:
12
+ - name: T3Q-Platypus-Mistral7B
13
+ results: []
14
  ---
15
+ Update @ 2024.03.07
16
+
17
+ ## T3Q-Platypus-Mistral7B
18
+
19
+ This model is a fine-tuned version of bardsai/jaskier-7b-dpo-v6.1
20
+
21
+ **Model Developers** Chihoon Lee(chlee10), T3Q
22
+
23
+ ## Training hyperparameters
24
+
25
+ The following hyperparameters were used during training:
26
+
27
+ ```python
28
+ # ๋ฐ์ดํ„ฐ์…‹๊ณผ ํ›ˆ๋ จ ํšŸ์ˆ˜์™€ ๊ด€๋ จ๋œ ํ•˜์ดํผ ํŒŒ๋ผ๋ฏธํ„ฐ
29
+ batch_size = 16
30
+ num_epochs = 1
31
+ micro_batch = 1
32
+ gradient_accumulation_steps = batch_size // micro_batch
33
+
34
+ # ํ›ˆ๋ จ ๋ฐฉ๋ฒ•์— ๋Œ€ํ•œ ํ•˜์ดํผ ํŒŒ๋ผ๋ฏธํ„ฐ
35
+ cutoff_len = 4096
36
+ lr_scheduler = 'cosine'
37
+ warmup_ratio = 0.06 # warmup_steps = 100
38
+ learning_rate = 4e-4
39
+ optimizer = 'adamw_torch'
40
+ weight_decay = 0.01
41
+ max_grad_norm = 1.0
42
+
43
+ # LoRA config
44
+ lora_r = 16
45
+ lora_alpha = 16
46
+ lora_dropout = 0.05
47
+ lora_target_modules = ["gate_proj", "down_proj", "up_proj"]
48
+
49
+ # Tokenizer์—์„œ ๋‚˜์˜ค๋Š” input๊ฐ’ ์„ค์ • ์˜ต์…˜
50
+ train_on_inputs = False
51
+ add_eos_token = False
52
+
53
+ # NEFTune params
54
+ noise_alpha: int = 5
55
+ ```
56
+
57
+