KrafterDen commited on
Commit
63e5ac8
·
verified ·
1 Parent(s): 1166ee0

Training in progress, step 200, checkpoint

Browse files
checkpoint-200/README.md CHANGED
@@ -201,5 +201,5 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
201
 
202
  ### Framework versions
203
 
204
- - PEFT 0.8.2
205
- - PEFT 0.7.1
 
201
 
202
  ### Framework versions
203
 
204
+ - PEFT 0.10.0
205
+ - PEFT 0.8.2
checkpoint-200/adapter_config.json CHANGED
@@ -6,6 +6,7 @@
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
@@ -22,5 +23,6 @@
22
  "c_attn"
23
  ],
24
  "task_type": "CAUSAL_LM",
 
25
  "use_rslora": false
26
  }
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
+ "layer_replication": null,
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
 
23
  "c_attn"
24
  ],
25
  "task_type": "CAUSAL_LM",
26
+ "use_dora": false,
27
  "use_rslora": false
28
  }
checkpoint-200/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9589f3acca05c13aa4c73b4a2cf488f113c6aa7efb90ad3c443f0eb3d7148b85
3
  size 9443384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed0f8a25c99bf0c4ece72eaba3a91a8ce363d040554758830c308008a85b8220
3
  size 9443384
checkpoint-200/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:393ca752c28abe23d40f30f7207bc6ede7e063b8e8165452969f9159cf788557
3
  size 18914450
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41a6fc26fd7b5026ca3631c40fc12ae6e5d9e13b4c209b07c2b04a2930045f4a
3
  size 18914450
checkpoint-200/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9725a066742779d96c921f5f7073a1a23056dd3248a6947bd5720b0306ff606c
3
  size 14168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1eea912470e6b4cd46b7542517282753ef83722547642edfbc4744d8b77eaf0
3
  size 14168
checkpoint-200/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eeaff60190187c5a099d5ee7c23e25fa6f5dd8d654469a0c8ba0019b00e60fc
3
  size 1056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a4c70499dd3107fca70fe56610df257c36dff9d7702a8aa9ddd4d6235c99938
3
  size 1056
checkpoint-200/trainer_state.json CHANGED
@@ -9,132 +9,152 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.02,
 
13
  "learning_rate": 2.9999999999999997e-05,
14
- "loss": 3.623,
15
  "step": 10
16
  },
17
  {
18
- "epoch": 0.05,
 
19
  "learning_rate": 5.9999999999999995e-05,
20
- "loss": 3.605,
21
  "step": 20
22
  },
23
  {
24
- "epoch": 0.07,
 
25
  "learning_rate": 8.999999999999999e-05,
26
- "loss": 3.5168,
27
  "step": 30
28
  },
29
  {
30
- "epoch": 0.1,
 
31
  "learning_rate": 0.00011999999999999999,
32
- "loss": 3.3511,
33
  "step": 40
34
  },
35
  {
36
- "epoch": 0.12,
 
37
  "learning_rate": 0.00015,
38
- "loss": 3.1952,
39
  "step": 50
40
  },
41
  {
42
- "epoch": 0.15,
 
43
  "learning_rate": 0.00017999999999999998,
44
- "loss": 3.0575,
45
  "step": 60
46
  },
47
  {
48
- "epoch": 0.17,
 
49
  "learning_rate": 0.00020999999999999998,
50
- "loss": 2.8575,
51
  "step": 70
52
  },
53
  {
54
- "epoch": 0.2,
 
55
  "learning_rate": 0.00023999999999999998,
56
- "loss": 2.7029,
57
  "step": 80
58
  },
59
  {
60
- "epoch": 0.22,
 
61
  "learning_rate": 0.00027,
62
- "loss": 2.6392,
63
  "step": 90
64
  },
65
  {
66
- "epoch": 0.25,
 
67
  "learning_rate": 0.0003,
68
- "loss": 2.5679,
69
  "step": 100
70
  },
71
  {
72
  "epoch": 0.27,
73
- "learning_rate": 0.00027,
74
- "loss": 2.5392,
 
75
  "step": 110
76
  },
77
  {
78
  "epoch": 0.3,
79
- "learning_rate": 0.00023999999999999998,
80
- "loss": 2.5153,
 
81
  "step": 120
82
  },
83
  {
84
  "epoch": 0.32,
85
- "learning_rate": 0.00020999999999999998,
86
- "loss": 2.4822,
 
87
  "step": 130
88
  },
89
  {
90
  "epoch": 0.34,
91
- "learning_rate": 0.00017999999999999998,
92
- "loss": 2.4678,
 
93
  "step": 140
94
  },
95
  {
96
  "epoch": 0.37,
97
- "learning_rate": 0.00015,
98
- "loss": 2.4438,
 
99
  "step": 150
100
  },
101
  {
102
  "epoch": 0.39,
103
- "learning_rate": 0.00011999999999999999,
104
- "loss": 2.4351,
 
105
  "step": 160
106
  },
107
  {
108
  "epoch": 0.42,
109
- "learning_rate": 8.999999999999999e-05,
110
- "loss": 2.4147,
 
111
  "step": 170
112
  },
113
  {
114
  "epoch": 0.44,
115
- "learning_rate": 5.9999999999999995e-05,
116
- "loss": 2.3935,
 
117
  "step": 180
118
  },
119
  {
120
  "epoch": 0.47,
121
- "learning_rate": 2.9999999999999997e-05,
122
- "loss": 2.4104,
 
123
  "step": 190
124
  },
125
  {
126
  "epoch": 0.49,
127
- "learning_rate": 0.0,
128
- "loss": 2.3838,
 
129
  "step": 200
130
  }
131
  ],
132
  "logging_steps": 10,
133
- "max_steps": 200,
134
  "num_input_tokens_seen": 0,
135
  "num_train_epochs": 1,
136
  "save_steps": 100,
137
- "total_flos": 2.723845771535155e+16,
138
  "train_batch_size": 4,
139
  "trial_name": null,
140
  "trial_params": null
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.42,
13
+ "grad_norm": 0.1815568506717682,
14
  "learning_rate": 2.9999999999999997e-05,
15
+ "loss": 3.8657,
16
  "step": 10
17
  },
18
  {
19
+ "epoch": 0.84,
20
+ "grad_norm": 0.23455914855003357,
21
  "learning_rate": 5.9999999999999995e-05,
22
+ "loss": 3.8223,
23
  "step": 20
24
  },
25
  {
26
+ "epoch": 1.25,
27
+ "grad_norm": 0.32260793447494507,
28
  "learning_rate": 8.999999999999999e-05,
29
+ "loss": 3.7255,
30
  "step": 30
31
  },
32
  {
33
+ "epoch": 1.67,
34
+ "grad_norm": 0.38705918192863464,
35
  "learning_rate": 0.00011999999999999999,
36
+ "loss": 3.4952,
37
  "step": 40
38
  },
39
  {
40
+ "epoch": 2.09,
41
+ "grad_norm": 0.5273059606552124,
42
  "learning_rate": 0.00015,
43
+ "loss": 3.098,
44
  "step": 50
45
  },
46
  {
47
+ "epoch": 2.51,
48
+ "grad_norm": 0.6030514240264893,
49
  "learning_rate": 0.00017999999999999998,
50
+ "loss": 2.5299,
51
  "step": 60
52
  },
53
  {
54
+ "epoch": 2.92,
55
+ "grad_norm": 0.459722101688385,
56
  "learning_rate": 0.00020999999999999998,
57
+ "loss": 1.899,
58
  "step": 70
59
  },
60
  {
61
+ "epoch": 3.34,
62
+ "grad_norm": 0.1655016839504242,
63
  "learning_rate": 0.00023999999999999998,
64
+ "loss": 1.6018,
65
  "step": 80
66
  },
67
  {
68
+ "epoch": 3.76,
69
+ "grad_norm": 0.10938003659248352,
70
  "learning_rate": 0.00027,
71
+ "loss": 1.4726,
72
  "step": 90
73
  },
74
  {
75
+ "epoch": 4.18,
76
+ "grad_norm": 0.09813433140516281,
77
  "learning_rate": 0.0003,
78
+ "loss": 1.4336,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 0.27,
83
+ "grad_norm": 0.2995990812778473,
84
+ "learning_rate": 0.000285,
85
+ "loss": 3.3878,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 0.3,
90
+ "grad_norm": 0.2459421455860138,
91
+ "learning_rate": 0.00027,
92
+ "loss": 3.0843,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 0.32,
97
+ "grad_norm": 0.2377060502767563,
98
+ "learning_rate": 0.00025499999999999996,
99
+ "loss": 2.8413,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 0.34,
104
+ "grad_norm": 0.1750001609325409,
105
+ "learning_rate": 0.00023999999999999998,
106
+ "loss": 2.7303,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 0.37,
111
+ "grad_norm": 0.1821776032447815,
112
+ "learning_rate": 0.000225,
113
+ "loss": 2.6535,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 0.39,
118
+ "grad_norm": 0.16587179899215698,
119
+ "learning_rate": 0.00020999999999999998,
120
+ "loss": 2.6147,
121
  "step": 160
122
  },
123
  {
124
  "epoch": 0.42,
125
+ "grad_norm": 0.15111136436462402,
126
+ "learning_rate": 0.000195,
127
+ "loss": 2.5929,
128
  "step": 170
129
  },
130
  {
131
  "epoch": 0.44,
132
+ "grad_norm": 0.13922317326068878,
133
+ "learning_rate": 0.00017999999999999998,
134
+ "loss": 2.554,
135
  "step": 180
136
  },
137
  {
138
  "epoch": 0.47,
139
+ "grad_norm": 0.14242495596408844,
140
+ "learning_rate": 0.000165,
141
+ "loss": 2.5381,
142
  "step": 190
143
  },
144
  {
145
  "epoch": 0.49,
146
+ "grad_norm": 0.1816890388727188,
147
+ "learning_rate": 0.00015,
148
+ "loss": 2.5145,
149
  "step": 200
150
  }
151
  ],
152
  "logging_steps": 10,
153
+ "max_steps": 300,
154
  "num_input_tokens_seen": 0,
155
  "num_train_epochs": 1,
156
  "save_steps": 100,
157
+ "total_flos": 3.2322500059336704e+16,
158
  "train_batch_size": 4,
159
  "trial_name": null,
160
  "trial_params": null
checkpoint-200/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d7fe1522993e9d0465d19bb981aba89a2d40513d15ab46cef263ce1ebaa4eb5
3
- size 4768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7189ea50d3101204ed0a14552e3e368c7afb6c90ecc8b189f33ef27c2f23742e
3
+ size 4960