ChiefTheLord commited on
Commit
f83c04e
·
verified ·
1 Parent(s): 0db224b

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -83,3 +83,4 @@ checkpoints-d2.0-t1/checkpoint-17408/eval_state.json filter=lfs diff=lfs merge=l
83
  checkpoints-d1.3/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs -text
84
  checkpoints-d1.4/checkpoint-20480/eval_state.json filter=lfs diff=lfs merge=lfs -text
85
  checkpoints-d3.0/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs -text
 
 
83
  checkpoints-d1.3/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs -text
84
  checkpoints-d1.4/checkpoint-20480/eval_state.json filter=lfs diff=lfs merge=lfs -text
85
  checkpoints-d3.0/checkpoint-21504/eval_state.json filter=lfs diff=lfs merge=lfs -text
86
+ checkpoints-d3.1/checkpoint-8192/eval_state.json filter=lfs diff=lfs merge=lfs -text
checkpoints-d3.1/checkpoint-8192/eval_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39f908e8d74cbb91f5a64c3c6374940831bd52c60bc495cdc3f8f18634f8f260
3
+ size 49040731
checkpoints-d3.1/checkpoint-8192/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92e555e4249466e23cf780a02565a4d9858f9752130d0da1cc375cc8999a71e5
3
+ size 32318104
checkpoints-d3.1/checkpoint-8192/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2d823e33982673a2c6ff40a3445039e3a0892c1cc75f2660853e3289530179
3
+ size 10010635
checkpoints-d3.1/checkpoint-8192/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2614a8364df2314827788266204194df25ce233cbf5d026bdc717b2dab088e6
3
+ size 14645
checkpoints-d3.1/checkpoint-8192/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63797d41b8e8000087d2cb16938b6d93a4a808721fe3acbfc105c9bb8b262526
3
+ size 1383
checkpoints-d3.1/checkpoint-8192/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:633f4a51e3130420dcf105f63b87d79cc31db8e4efd560f5f6c0e0209526c2e9
3
+ size 1465
checkpoints-d3.1/checkpoint-8192/trainer_state.json ADDED
@@ -0,0 +1,394 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.37836589533970716,
6
+ "eval_steps": 1024,
7
+ "global_step": 8192,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.011823934229365849,
14
+ "grad_norm": 0.0007770129013806581,
15
+ "learning_rate": 0.000498046875,
16
+ "loss": 0.638597846031189,
17
+ "step": 256
18
+ },
19
+ {
20
+ "epoch": 0.023647868458731697,
21
+ "grad_norm": 0.0036502168513834476,
22
+ "learning_rate": 0.000998046875,
23
+ "loss": 0.6255820393562317,
24
+ "step": 512
25
+ },
26
+ {
27
+ "epoch": 0.03547180268809755,
28
+ "grad_norm": 0.00819552130997181,
29
+ "learning_rate": 0.000999640996023194,
30
+ "loss": 0.5880061388015747,
31
+ "step": 768
32
+ },
33
+ {
34
+ "epoch": 0.047295736917463395,
35
+ "grad_norm": 0.014988946728408337,
36
+ "learning_rate": 0.0009985588674043958,
37
+ "loss": 0.5405965447425842,
38
+ "step": 1024
39
+ },
40
+ {
41
+ "epoch": 0.047295736917463395,
42
+ "eval_bleu": 0.020278462916643203,
43
+ "eval_ce_loss": 0.4954925418988755,
44
+ "eval_loss": 0.4954925418988755,
45
+ "step": 1024
46
+ },
47
+ {
48
+ "epoch": 0.047295736917463395,
49
+ "eval_bleu": 0.020278462916643203,
50
+ "eval_ce_loss": 0.4954925418988755,
51
+ "eval_loss": 0.4954925418988755,
52
+ "eval_runtime": 131.1195,
53
+ "eval_samples_per_second": 213.492,
54
+ "eval_steps_per_second": 3.34,
55
+ "step": 1024
56
+ },
57
+ {
58
+ "epoch": 0.05911967114682925,
59
+ "grad_norm": 0.018301010131835938,
60
+ "learning_rate": 0.0009967551747861387,
61
+ "loss": 0.4193720519542694,
62
+ "step": 1280
63
+ },
64
+ {
65
+ "epoch": 0.0709436053761951,
66
+ "grad_norm": 0.02350299246609211,
67
+ "learning_rate": 0.000994232528651847,
68
+ "loss": 0.32118111848831177,
69
+ "step": 1536
70
+ },
71
+ {
72
+ "epoch": 0.08276753960556095,
73
+ "grad_norm": 0.024243628606200218,
74
+ "learning_rate": 0.0009909945800260092,
75
+ "loss": 0.27625754475593567,
76
+ "step": 1792
77
+ },
78
+ {
79
+ "epoch": 0.09459147383492679,
80
+ "grad_norm": 0.022370202466845512,
81
+ "learning_rate": 0.0009870460151900522,
82
+ "loss": 0.25225532054901123,
83
+ "step": 2048
84
+ },
85
+ {
86
+ "epoch": 0.09459147383492679,
87
+ "eval_bleu": 0.16173067052738044,
88
+ "eval_ce_loss": 0.24119392792655997,
89
+ "eval_loss": 0.24119392792655997,
90
+ "step": 2048
91
+ },
92
+ {
93
+ "epoch": 0.09459147383492679,
94
+ "eval_bleu": 0.16173067052738044,
95
+ "eval_ce_loss": 0.24119392792655997,
96
+ "eval_loss": 0.24119392792655997,
97
+ "eval_runtime": 124.4827,
98
+ "eval_samples_per_second": 224.875,
99
+ "eval_steps_per_second": 3.519,
100
+ "step": 2048
101
+ },
102
+ {
103
+ "epoch": 0.10641540806429264,
104
+ "grad_norm": 0.025183985009789467,
105
+ "learning_rate": 0.0009823925488998885,
106
+ "loss": 0.23278099298477173,
107
+ "step": 2304
108
+ },
109
+ {
110
+ "epoch": 0.1182393422936585,
111
+ "grad_norm": 0.025138691067695618,
112
+ "learning_rate": 0.0009770409161149525,
113
+ "loss": 0.21929927170276642,
114
+ "step": 2560
115
+ },
116
+ {
117
+ "epoch": 0.13006327652302435,
118
+ "grad_norm": 0.02561408467590809,
119
+ "learning_rate": 0.0009709988622506973,
120
+ "loss": 0.20501913130283356,
121
+ "step": 2816
122
+ },
123
+ {
124
+ "epoch": 0.1418872107523902,
125
+ "grad_norm": 0.028821036219596863,
126
+ "learning_rate": 0.000964275131968659,
127
+ "loss": 0.19426275789737701,
128
+ "step": 3072
129
+ },
130
+ {
131
+ "epoch": 0.1418872107523902,
132
+ "eval_bleu": 0.19048343372542764,
133
+ "eval_ce_loss": 0.18918462373214226,
134
+ "eval_loss": 0.18918462373214226,
135
+ "step": 3072
136
+ },
137
+ {
138
+ "epoch": 0.1418872107523902,
139
+ "eval_bleu": 0.19048343372542764,
140
+ "eval_ce_loss": 0.18918462373214226,
141
+ "eval_loss": 0.18918462373214226,
142
+ "eval_runtime": 125.1618,
143
+ "eval_samples_per_second": 223.654,
144
+ "eval_steps_per_second": 3.499,
145
+ "step": 3072
146
+ },
147
+ {
148
+ "epoch": 0.15371114498175603,
149
+ "grad_norm": 0.030421536415815353,
150
+ "learning_rate": 0.0009568794565203123,
151
+ "loss": 0.1839032918214798,
152
+ "step": 3328
153
+ },
154
+ {
155
+ "epoch": 0.1655350792111219,
156
+ "grad_norm": 0.032538577914237976,
157
+ "learning_rate": 0.0009488225396630347,
158
+ "loss": 0.17660538852214813,
159
+ "step": 3584
160
+ },
161
+ {
162
+ "epoch": 0.17735901344048774,
163
+ "grad_norm": 0.030282698571681976,
164
+ "learning_rate": 0.0009401160421685646,
165
+ "loss": 0.1695183515548706,
166
+ "step": 3840
167
+ },
168
+ {
169
+ "epoch": 0.18918294766985358,
170
+ "grad_norm": 0.029157549142837524,
171
+ "learning_rate": 0.0009307725649463714,
172
+ "loss": 0.16315339505672455,
173
+ "step": 4096
174
+ },
175
+ {
176
+ "epoch": 0.18918294766985358,
177
+ "eval_bleu": 0.20316795180629035,
178
+ "eval_ce_loss": 0.16114722188885353,
179
+ "eval_loss": 0.16114722188885353,
180
+ "step": 4096
181
+ },
182
+ {
183
+ "epoch": 0.18918294766985358,
184
+ "eval_bleu": 0.20316795180629035,
185
+ "eval_ce_loss": 0.16114722188885353,
186
+ "eval_loss": 0.16114722188885353,
187
+ "eval_runtime": 124.8908,
188
+ "eval_samples_per_second": 224.14,
189
+ "eval_steps_per_second": 3.507,
190
+ "step": 4096
191
+ },
192
+ {
193
+ "epoch": 0.20100688189921945,
194
+ "grad_norm": 0.033927738666534424,
195
+ "learning_rate": 0.0009208056308063659,
196
+ "loss": 0.15908151865005493,
197
+ "step": 4352
198
+ },
199
+ {
200
+ "epoch": 0.2128308161285853,
201
+ "grad_norm": 0.037180621176958084,
202
+ "learning_rate": 0.0009102296648873445,
203
+ "loss": 0.15225762128829956,
204
+ "step": 4608
205
+ },
206
+ {
207
+ "epoch": 0.22465475035795113,
208
+ "grad_norm": 0.037015024572610855,
209
+ "learning_rate": 0.0008990599737794927,
210
+ "loss": 0.1476851999759674,
211
+ "step": 4864
212
+ },
213
+ {
214
+ "epoch": 0.236478684587317,
215
+ "grad_norm": 0.03508929908275604,
216
+ "learning_rate": 0.0008873127233711644,
217
+ "loss": 0.14329124987125397,
218
+ "step": 5120
219
+ },
220
+ {
221
+ "epoch": 0.236478684587317,
222
+ "eval_bleu": 0.21566794230701403,
223
+ "eval_ce_loss": 0.14114715776418987,
224
+ "eval_loss": 0.14114715776418987,
225
+ "step": 5120
226
+ },
227
+ {
228
+ "epoch": 0.236478684587317,
229
+ "eval_bleu": 0.21566794230701403,
230
+ "eval_ce_loss": 0.14114715776418987,
231
+ "eval_loss": 0.14114715776418987,
232
+ "eval_runtime": 125.5711,
233
+ "eval_samples_per_second": 222.926,
234
+ "eval_steps_per_second": 3.488,
235
+ "step": 5120
236
+ },
237
+ {
238
+ "epoch": 0.24830261881668284,
239
+ "grad_norm": 0.03854870796203613,
240
+ "learning_rate": 0.0008750049154520011,
241
+ "loss": 0.14005360007286072,
242
+ "step": 5376
243
+ },
244
+ {
245
+ "epoch": 0.2601265530460487,
246
+ "grad_norm": 0.03188326954841614,
247
+ "learning_rate": 0.0008621543631062487,
248
+ "loss": 0.13561338186264038,
249
+ "step": 5632
250
+ },
251
+ {
252
+ "epoch": 0.27195048727541454,
253
+ "grad_norm": 0.03463684767484665,
254
+ "learning_rate": 0.0008487796649318904,
255
+ "loss": 0.1329711377620697,
256
+ "step": 5888
257
+ },
258
+ {
259
+ "epoch": 0.2837744215047804,
260
+ "grad_norm": 0.0358298160135746,
261
+ "learning_rate": 0.0008349001781229053,
262
+ "loss": 0.1301661729812622,
263
+ "step": 6144
264
+ },
265
+ {
266
+ "epoch": 0.2837744215047804,
267
+ "eval_bleu": 0.22173654323693984,
268
+ "eval_ce_loss": 0.12855397619000852,
269
+ "eval_loss": 0.12855397619000852,
270
+ "step": 6144
271
+ },
272
+ {
273
+ "epoch": 0.2837744215047804,
274
+ "eval_bleu": 0.22173654323693984,
275
+ "eval_ce_loss": 0.12855397619000852,
276
+ "eval_loss": 0.12855397619000852,
277
+ "eval_runtime": 126.9306,
278
+ "eval_samples_per_second": 220.538,
279
+ "eval_steps_per_second": 3.451,
280
+ "step": 6144
281
+ },
282
+ {
283
+ "epoch": 0.2955983557341462,
284
+ "grad_norm": 0.037201233208179474,
285
+ "learning_rate": 0.0008205359904536107,
286
+ "loss": 0.1271919459104538,
287
+ "step": 6400
288
+ },
289
+ {
290
+ "epoch": 0.30742228996351206,
291
+ "grad_norm": 0.0340813584625721,
292
+ "learning_rate": 0.0008057078912056363,
293
+ "loss": 0.12491822242736816,
294
+ "step": 6656
295
+ },
296
+ {
297
+ "epoch": 0.3192462241928779,
298
+ "grad_norm": 0.031119871884584427,
299
+ "learning_rate": 0.0007904373410796086,
300
+ "loss": 0.12252337485551834,
301
+ "step": 6912
302
+ },
303
+ {
304
+ "epoch": 0.3310701584222438,
305
+ "grad_norm": 0.036261022090911865,
306
+ "learning_rate": 0.0007747464411350876,
307
+ "loss": 0.12051380425691605,
308
+ "step": 7168
309
+ },
310
+ {
311
+ "epoch": 0.3310701584222438,
312
+ "eval_bleu": 0.228738826346753,
313
+ "eval_ce_loss": 0.12006643695169933,
314
+ "eval_loss": 0.12006643695169933,
315
+ "step": 7168
316
+ },
317
+ {
318
+ "epoch": 0.3310701584222438,
319
+ "eval_bleu": 0.228738826346753,
320
+ "eval_ce_loss": 0.12006643695169933,
321
+ "eval_loss": 0.12006643695169933,
322
+ "eval_runtime": 128.0461,
323
+ "eval_samples_per_second": 218.617,
324
+ "eval_steps_per_second": 3.421,
325
+ "step": 7168
326
+ },
327
+ {
328
+ "epoch": 0.34289409265160964,
329
+ "grad_norm": 0.03925757110118866,
330
+ "learning_rate": 0.000758657900803716,
331
+ "loss": 0.11830627918243408,
332
+ "step": 7424
333
+ },
334
+ {
335
+ "epoch": 0.3547180268809755,
336
+ "grad_norm": 0.03783626854419708,
337
+ "learning_rate": 0.000742195005021869,
338
+ "loss": 0.11684451997280121,
339
+ "step": 7680
340
+ },
341
+ {
342
+ "epoch": 0.3665419611103413,
343
+ "grad_norm": 0.04456119239330292,
344
+ "learning_rate": 0.0007253815805303786,
345
+ "loss": 0.11479435861110687,
346
+ "step": 7936
347
+ },
348
+ {
349
+ "epoch": 0.37836589533970716,
350
+ "grad_norm": 0.0406075082719326,
351
+ "learning_rate": 0.0007082419613901028,
352
+ "loss": 0.11298805475234985,
353
+ "step": 8192
354
+ },
355
+ {
356
+ "epoch": 0.37836589533970716,
357
+ "eval_bleu": 0.23316787828393487,
358
+ "eval_ce_loss": 0.11301100160501319,
359
+ "eval_loss": 0.11301100160501319,
360
+ "step": 8192
361
+ },
362
+ {
363
+ "epoch": 0.37836589533970716,
364
+ "eval_bleu": 0.23316787828393487,
365
+ "eval_ce_loss": 0.11301100160501319,
366
+ "eval_loss": 0.11301100160501319,
367
+ "eval_runtime": 129.7081,
368
+ "eval_samples_per_second": 215.815,
369
+ "eval_steps_per_second": 3.377,
370
+ "step": 8192
371
+ }
372
+ ],
373
+ "logging_steps": 256,
374
+ "max_steps": 21651,
375
+ "num_input_tokens_seen": 0,
376
+ "num_train_epochs": 1,
377
+ "save_steps": 1024,
378
+ "stateful_callbacks": {
379
+ "TrainerControl": {
380
+ "args": {
381
+ "should_epoch_stop": false,
382
+ "should_evaluate": false,
383
+ "should_log": false,
384
+ "should_save": true,
385
+ "should_training_stop": false
386
+ },
387
+ "attributes": {}
388
+ }
389
+ },
390
+ "total_flos": 0.0,
391
+ "train_batch_size": 64,
392
+ "trial_name": null,
393
+ "trial_params": null
394
+ }
checkpoints-d3.1/checkpoint-8192/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a16bb839f687414b8e48611327c4b9cfddeefe38c031ca70808f9a97c476b7
3
+ size 5137