nielsbantilan commited on
Commit
5c94a7f
1 Parent(s): 753bb7c

Upload folder using huggingface_hub

Browse files
flyte_training_config.json CHANGED
@@ -1 +1 @@
1
- {"base_model": "togethercomputer/RedPajama-INCITE-Base-3B-v1", "data_path": "wikipedia", "data_name": "20220301.simple", "num_epochs": 1, "max_steps": 500, "learning_rate": 2e-05, "weight_decay": 0.02, "warmup_ratio": 0.03, "lr_scheduler_type": "cosine", "batch_size": 16, "micro_batch_size": 1, "val_set_size": 0, "group_by_length": false, "instruction_key": "instruction", "input_key": "input", "output_key": "output", "device_map": "auto", "cache_dir": null, "optim": "adamw_torch", "model_max_length": 512, "debug_mode": false, "debug_train_data_size": 1024, "wandb_project": ""}
 
1
+ {"base_model": "togethercomputer/RedPajama-INCITE-Base-3B-v1", "data_path": "wikipedia", "data_name": "20220301.simple", "num_epochs": 1, "max_steps": 100, "learning_rate": 2e-05, "weight_decay": 0.02, "warmup_ratio": 0.03, "lr_scheduler_type": "cosine", "batch_size": 16, "micro_batch_size": 1, "val_set_size": 0, "group_by_length": false, "instruction_key": "instruction", "input_key": "input", "output_key": "output", "device_map": "auto", "cache_dir": null, "optim": "adamw_torch", "model_max_length": 512, "debug_mode": false, "debug_train_data_size": 1024, "wandb_project": ""}
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f6e921b4581ba057c3b200ae0688f99915374b4184b5dc7876c4450b8eee8cf
3
  size 5686106713
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e63cb70257a1b71bfddd7ac332cf76d74e1d295fe8c1bd2752e1b18938257b47
3
  size 5686106713
trainer_state.json CHANGED
@@ -1,325 +1,85 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 333.3333333333333,
5
- "global_step": 500,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 6.67,
12
- "learning_rate": 1.5357481488588927e-05,
13
- "loss": 1.9835,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 13.33,
18
  "learning_rate": 2e-05,
19
- "loss": 1.0763,
20
  "step": 20
21
  },
22
  {
23
  "epoch": 20.0,
24
  "learning_rate": 2e-05,
25
- "loss": 0.2284,
26
  "step": 30
27
  },
28
  {
29
  "epoch": 26.67,
30
  "learning_rate": 2e-05,
31
- "loss": 0.0691,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 33.33,
36
  "learning_rate": 2e-05,
37
- "loss": 0.0442,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 40.0,
42
  "learning_rate": 2e-05,
43
- "loss": 0.0319,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 46.67,
48
  "learning_rate": 2e-05,
49
- "loss": 0.0243,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 53.33,
54
  "learning_rate": 2e-05,
55
- "loss": 0.0186,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 60.0,
60
  "learning_rate": 2e-05,
61
- "loss": 0.0149,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 66.67,
66
  "learning_rate": 2e-05,
67
- "loss": 0.0119,
68
  "step": 100
69
  },
70
  {
71
- "epoch": 73.33,
72
- "learning_rate": 2e-05,
73
- "loss": 0.0103,
74
- "step": 110
75
- },
76
- {
77
- "epoch": 80.0,
78
- "learning_rate": 2e-05,
79
- "loss": 0.0091,
80
- "step": 120
81
- },
82
- {
83
- "epoch": 86.67,
84
- "learning_rate": 2e-05,
85
- "loss": 0.0081,
86
- "step": 130
87
- },
88
- {
89
- "epoch": 93.33,
90
- "learning_rate": 2e-05,
91
- "loss": 0.0073,
92
- "step": 140
93
- },
94
- {
95
- "epoch": 100.0,
96
- "learning_rate": 2e-05,
97
- "loss": 0.0064,
98
- "step": 150
99
- },
100
- {
101
- "epoch": 106.67,
102
- "learning_rate": 2e-05,
103
- "loss": 0.0064,
104
- "step": 160
105
- },
106
- {
107
- "epoch": 113.33,
108
- "learning_rate": 2e-05,
109
- "loss": 0.0059,
110
- "step": 170
111
- },
112
- {
113
- "epoch": 120.0,
114
- "learning_rate": 2e-05,
115
- "loss": 0.0057,
116
- "step": 180
117
- },
118
- {
119
- "epoch": 126.67,
120
- "learning_rate": 2e-05,
121
- "loss": 0.0054,
122
- "step": 190
123
- },
124
- {
125
- "epoch": 133.33,
126
- "learning_rate": 2e-05,
127
- "loss": 0.0052,
128
- "step": 200
129
- },
130
- {
131
- "epoch": 140.0,
132
- "learning_rate": 2e-05,
133
- "loss": 0.0049,
134
- "step": 210
135
- },
136
- {
137
- "epoch": 146.67,
138
- "learning_rate": 2e-05,
139
- "loss": 0.0046,
140
- "step": 220
141
- },
142
- {
143
- "epoch": 153.33,
144
- "learning_rate": 2e-05,
145
- "loss": 0.0045,
146
- "step": 230
147
- },
148
- {
149
- "epoch": 160.0,
150
- "learning_rate": 2e-05,
151
- "loss": 0.0044,
152
- "step": 240
153
- },
154
- {
155
- "epoch": 166.67,
156
- "learning_rate": 2e-05,
157
- "loss": 0.0043,
158
- "step": 250
159
- },
160
- {
161
- "epoch": 173.33,
162
- "learning_rate": 2e-05,
163
- "loss": 0.0044,
164
- "step": 260
165
- },
166
- {
167
- "epoch": 180.0,
168
- "learning_rate": 2e-05,
169
- "loss": 0.0043,
170
- "step": 270
171
- },
172
- {
173
- "epoch": 186.67,
174
- "learning_rate": 2e-05,
175
- "loss": 0.0044,
176
- "step": 280
177
- },
178
- {
179
- "epoch": 193.33,
180
- "learning_rate": 2e-05,
181
- "loss": 0.0041,
182
- "step": 290
183
- },
184
- {
185
- "epoch": 200.0,
186
- "learning_rate": 2e-05,
187
- "loss": 0.0042,
188
- "step": 300
189
- },
190
- {
191
- "epoch": 206.67,
192
- "learning_rate": 2e-05,
193
- "loss": 0.0043,
194
- "step": 310
195
- },
196
- {
197
- "epoch": 213.33,
198
- "learning_rate": 2e-05,
199
- "loss": 0.004,
200
- "step": 320
201
- },
202
- {
203
- "epoch": 220.0,
204
- "learning_rate": 2e-05,
205
- "loss": 0.004,
206
- "step": 330
207
- },
208
- {
209
- "epoch": 226.67,
210
- "learning_rate": 2e-05,
211
- "loss": 0.0041,
212
- "step": 340
213
- },
214
- {
215
- "epoch": 233.33,
216
- "learning_rate": 2e-05,
217
- "loss": 0.0039,
218
- "step": 350
219
- },
220
- {
221
- "epoch": 240.0,
222
- "learning_rate": 2e-05,
223
- "loss": 0.0039,
224
- "step": 360
225
- },
226
- {
227
- "epoch": 246.67,
228
- "learning_rate": 2e-05,
229
- "loss": 0.0038,
230
- "step": 370
231
- },
232
- {
233
- "epoch": 253.33,
234
- "learning_rate": 2e-05,
235
- "loss": 0.0039,
236
- "step": 380
237
- },
238
- {
239
- "epoch": 260.0,
240
- "learning_rate": 2e-05,
241
- "loss": 0.0039,
242
- "step": 390
243
- },
244
- {
245
- "epoch": 266.67,
246
- "learning_rate": 2e-05,
247
- "loss": 0.0038,
248
- "step": 400
249
- },
250
- {
251
- "epoch": 273.33,
252
- "learning_rate": 2e-05,
253
- "loss": 0.0038,
254
- "step": 410
255
- },
256
- {
257
- "epoch": 280.0,
258
- "learning_rate": 2e-05,
259
- "loss": 0.0038,
260
- "step": 420
261
- },
262
- {
263
- "epoch": 286.67,
264
- "learning_rate": 2e-05,
265
- "loss": 0.0037,
266
- "step": 430
267
- },
268
- {
269
- "epoch": 293.33,
270
- "learning_rate": 2e-05,
271
- "loss": 0.0038,
272
- "step": 440
273
- },
274
- {
275
- "epoch": 300.0,
276
- "learning_rate": 2e-05,
277
- "loss": 0.0039,
278
- "step": 450
279
- },
280
- {
281
- "epoch": 306.67,
282
- "learning_rate": 2e-05,
283
- "loss": 0.0037,
284
- "step": 460
285
- },
286
- {
287
- "epoch": 313.33,
288
- "learning_rate": 2e-05,
289
- "loss": 0.0038,
290
- "step": 470
291
- },
292
- {
293
- "epoch": 320.0,
294
- "learning_rate": 2e-05,
295
- "loss": 0.0036,
296
- "step": 480
297
- },
298
- {
299
- "epoch": 326.67,
300
- "learning_rate": 2e-05,
301
- "loss": 0.0037,
302
- "step": 490
303
- },
304
- {
305
- "epoch": 333.33,
306
- "learning_rate": 2e-05,
307
- "loss": 0.0037,
308
- "step": 500
309
- },
310
- {
311
- "epoch": 333.33,
312
- "step": 500,
313
- "total_flos": 210359990353920.0,
314
- "train_loss": 0.07388268795609475,
315
- "train_runtime": 21319.5402,
316
- "train_samples_per_second": 3.002,
317
  "train_steps_per_second": 0.023
318
  }
319
  ],
320
- "max_steps": 500,
321
- "num_train_epochs": 500,
322
- "total_flos": 210359990353920.0,
323
  "trial_name": null,
324
  "trial_params": null
325
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 66.66666666666667,
5
+ "global_step": 100,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
  {
11
  "epoch": 6.67,
12
+ "learning_rate": 2e-05,
13
+ "loss": 1.9669,
14
  "step": 10
15
  },
16
  {
17
  "epoch": 13.33,
18
  "learning_rate": 2e-05,
19
+ "loss": 0.9447,
20
  "step": 20
21
  },
22
  {
23
  "epoch": 20.0,
24
  "learning_rate": 2e-05,
25
+ "loss": 0.1908,
26
  "step": 30
27
  },
28
  {
29
  "epoch": 26.67,
30
  "learning_rate": 2e-05,
31
+ "loss": 0.0666,
32
  "step": 40
33
  },
34
  {
35
  "epoch": 33.33,
36
  "learning_rate": 2e-05,
37
+ "loss": 0.0441,
38
  "step": 50
39
  },
40
  {
41
  "epoch": 40.0,
42
  "learning_rate": 2e-05,
43
+ "loss": 0.0329,
44
  "step": 60
45
  },
46
  {
47
  "epoch": 46.67,
48
  "learning_rate": 2e-05,
49
+ "loss": 0.0251,
50
  "step": 70
51
  },
52
  {
53
  "epoch": 53.33,
54
  "learning_rate": 2e-05,
55
+ "loss": 0.019,
56
  "step": 80
57
  },
58
  {
59
  "epoch": 60.0,
60
  "learning_rate": 2e-05,
61
+ "loss": 0.0166,
62
  "step": 90
63
  },
64
  {
65
  "epoch": 66.67,
66
  "learning_rate": 2e-05,
67
+ "loss": 0.0133,
68
  "step": 100
69
  },
70
  {
71
+ "epoch": 66.67,
72
+ "step": 100,
73
+ "total_flos": 42050959441920.0,
74
+ "train_loss": 0.3319866207242012,
75
+ "train_runtime": 4269.4137,
76
+ "train_samples_per_second": 2.998,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  "train_steps_per_second": 0.023
78
  }
79
  ],
80
+ "max_steps": 100,
81
+ "num_train_epochs": 100,
82
+ "total_flos": 42050959441920.0,
83
  "trial_name": null,
84
  "trial_params": null
85
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2188a20782c485df4937a49e7b018f87feb047b94549b3e8907cdc9eff5c9801
3
- size 5563
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e0c2d603edfa479a939173d26bb532d940f3ac3258651395bfbf4755154a1a
3
+ size 5499