elsayedissa commited on
Commit
7412490
1 Parent(s): a11d12e

Training in progress, step 2000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49bbe912499e9815db85ebf81eb54a8130fa015e4af147cc5b556b509a4de12b
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cfba6acc2f48d6cbee27062141646fce95fe7c9ec6abcf7a8c2e13168cdb7ed
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e973125a216cde40c6c4f8e3b9480097d49da0899916d84d5d6eaca55778554c
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be6f8e930b75057373efa9b8e571972729960add3845f5586fc078849235cc72
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfaa06738049e37a26018d4bd814954a77d31c74b00db242f54a5bc3a3d9a693
3
- size 14511
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cb4de0fcfb0d5a9b829b897d7551e6c188899b37b17e1c4030b6d4b1d3a6de0
3
+ size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e95ea4d51ce3dbab01e59ff8912c2f7044fb7bd804fa1afc7ea77264e8c7d84
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75b2d5c8b12724e300963a719034bfe44737705d2de06deafe54683c78564862
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ca970d66f7f07c0e8752869b05b946fd6e8bf2f6a38832ab3db1935c1c221fd
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fd6bc27e7186f611a794f2cf9a3fde69378928c584c002486004b9d0cc4bf4e
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.78878255063432,
3
- "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-japanese-24h/checkpoint-1000",
4
- "epoch": 7.633587786259542,
5
- "global_step": 1000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -255,11 +255,260 @@
255
  "eval_steps_per_second": 0.163,
256
  "eval_wer": 0.78878255063432,
257
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
258
  }
259
  ],
260
  "max_steps": 5000,
261
  "num_train_epochs": 39,
262
- "total_flos": 1.05489752168448e+20,
263
  "trial_name": null,
264
  "trial_params": null
265
  }
 
1
  {
2
+ "best_metric": 0.7478299577119965,
3
+ "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-japanese-24h/checkpoint-2000",
4
+ "epoch": 15.267175572519085,
5
+ "global_step": 2000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
255
  "eval_steps_per_second": 0.163,
256
  "eval_wer": 0.78878255063432,
257
  "step": 1000
258
+ },
259
+ {
260
+ "epoch": 7.82,
261
+ "learning_rate": 8.842222222222223e-06,
262
+ "loss": 0.01,
263
+ "step": 1025
264
+ },
265
+ {
266
+ "epoch": 8.02,
267
+ "learning_rate": 8.786666666666668e-06,
268
+ "loss": 0.0094,
269
+ "step": 1050
270
+ },
271
+ {
272
+ "epoch": 8.21,
273
+ "learning_rate": 8.73111111111111e-06,
274
+ "loss": 0.0098,
275
+ "step": 1075
276
+ },
277
+ {
278
+ "epoch": 8.4,
279
+ "learning_rate": 8.675555555555556e-06,
280
+ "loss": 0.0097,
281
+ "step": 1100
282
+ },
283
+ {
284
+ "epoch": 8.59,
285
+ "learning_rate": 8.62e-06,
286
+ "loss": 0.008,
287
+ "step": 1125
288
+ },
289
+ {
290
+ "epoch": 8.78,
291
+ "learning_rate": 8.564444444444445e-06,
292
+ "loss": 0.0085,
293
+ "step": 1150
294
+ },
295
+ {
296
+ "epoch": 8.97,
297
+ "learning_rate": 8.50888888888889e-06,
298
+ "loss": 0.0081,
299
+ "step": 1175
300
+ },
301
+ {
302
+ "epoch": 9.16,
303
+ "learning_rate": 8.453333333333334e-06,
304
+ "loss": 0.0062,
305
+ "step": 1200
306
+ },
307
+ {
308
+ "epoch": 9.35,
309
+ "learning_rate": 8.397777777777778e-06,
310
+ "loss": 0.0063,
311
+ "step": 1225
312
+ },
313
+ {
314
+ "epoch": 9.54,
315
+ "learning_rate": 8.342222222222222e-06,
316
+ "loss": 0.0052,
317
+ "step": 1250
318
+ },
319
+ {
320
+ "epoch": 9.73,
321
+ "learning_rate": 8.286666666666668e-06,
322
+ "loss": 0.0048,
323
+ "step": 1275
324
+ },
325
+ {
326
+ "epoch": 9.92,
327
+ "learning_rate": 8.231111111111112e-06,
328
+ "loss": 0.0053,
329
+ "step": 1300
330
+ },
331
+ {
332
+ "epoch": 10.11,
333
+ "learning_rate": 8.175555555555556e-06,
334
+ "loss": 0.004,
335
+ "step": 1325
336
+ },
337
+ {
338
+ "epoch": 10.31,
339
+ "learning_rate": 8.120000000000002e-06,
340
+ "loss": 0.003,
341
+ "step": 1350
342
+ },
343
+ {
344
+ "epoch": 10.5,
345
+ "learning_rate": 8.064444444444444e-06,
346
+ "loss": 0.0044,
347
+ "step": 1375
348
+ },
349
+ {
350
+ "epoch": 10.69,
351
+ "learning_rate": 8.00888888888889e-06,
352
+ "loss": 0.0039,
353
+ "step": 1400
354
+ },
355
+ {
356
+ "epoch": 10.88,
357
+ "learning_rate": 7.953333333333334e-06,
358
+ "loss": 0.0038,
359
+ "step": 1425
360
+ },
361
+ {
362
+ "epoch": 11.07,
363
+ "learning_rate": 7.897777777777778e-06,
364
+ "loss": 0.0032,
365
+ "step": 1450
366
+ },
367
+ {
368
+ "epoch": 11.26,
369
+ "learning_rate": 7.842222222222224e-06,
370
+ "loss": 0.0017,
371
+ "step": 1475
372
+ },
373
+ {
374
+ "epoch": 11.45,
375
+ "learning_rate": 7.786666666666666e-06,
376
+ "loss": 0.0017,
377
+ "step": 1500
378
+ },
379
+ {
380
+ "epoch": 11.64,
381
+ "learning_rate": 7.731111111111112e-06,
382
+ "loss": 0.0024,
383
+ "step": 1525
384
+ },
385
+ {
386
+ "epoch": 11.83,
387
+ "learning_rate": 7.675555555555556e-06,
388
+ "loss": 0.0019,
389
+ "step": 1550
390
+ },
391
+ {
392
+ "epoch": 12.02,
393
+ "learning_rate": 7.620000000000001e-06,
394
+ "loss": 0.0027,
395
+ "step": 1575
396
+ },
397
+ {
398
+ "epoch": 12.21,
399
+ "learning_rate": 7.564444444444446e-06,
400
+ "loss": 0.0019,
401
+ "step": 1600
402
+ },
403
+ {
404
+ "epoch": 12.4,
405
+ "learning_rate": 7.50888888888889e-06,
406
+ "loss": 0.0027,
407
+ "step": 1625
408
+ },
409
+ {
410
+ "epoch": 12.6,
411
+ "learning_rate": 7.453333333333334e-06,
412
+ "loss": 0.0019,
413
+ "step": 1650
414
+ },
415
+ {
416
+ "epoch": 12.79,
417
+ "learning_rate": 7.3977777777777786e-06,
418
+ "loss": 0.0013,
419
+ "step": 1675
420
+ },
421
+ {
422
+ "epoch": 12.98,
423
+ "learning_rate": 7.342222222222223e-06,
424
+ "loss": 0.0013,
425
+ "step": 1700
426
+ },
427
+ {
428
+ "epoch": 13.17,
429
+ "learning_rate": 7.2866666666666675e-06,
430
+ "loss": 0.0011,
431
+ "step": 1725
432
+ },
433
+ {
434
+ "epoch": 13.36,
435
+ "learning_rate": 7.231111111111112e-06,
436
+ "loss": 0.0011,
437
+ "step": 1750
438
+ },
439
+ {
440
+ "epoch": 13.55,
441
+ "learning_rate": 7.1755555555555556e-06,
442
+ "loss": 0.0009,
443
+ "step": 1775
444
+ },
445
+ {
446
+ "epoch": 13.74,
447
+ "learning_rate": 7.1200000000000004e-06,
448
+ "loss": 0.0009,
449
+ "step": 1800
450
+ },
451
+ {
452
+ "epoch": 13.93,
453
+ "learning_rate": 7.0644444444444445e-06,
454
+ "loss": 0.0013,
455
+ "step": 1825
456
+ },
457
+ {
458
+ "epoch": 14.12,
459
+ "learning_rate": 7.008888888888889e-06,
460
+ "loss": 0.0014,
461
+ "step": 1850
462
+ },
463
+ {
464
+ "epoch": 14.31,
465
+ "learning_rate": 6.953333333333334e-06,
466
+ "loss": 0.0011,
467
+ "step": 1875
468
+ },
469
+ {
470
+ "epoch": 14.5,
471
+ "learning_rate": 6.897777777777779e-06,
472
+ "loss": 0.0007,
473
+ "step": 1900
474
+ },
475
+ {
476
+ "epoch": 14.69,
477
+ "learning_rate": 6.842222222222222e-06,
478
+ "loss": 0.0008,
479
+ "step": 1925
480
+ },
481
+ {
482
+ "epoch": 14.89,
483
+ "learning_rate": 6.786666666666667e-06,
484
+ "loss": 0.0009,
485
+ "step": 1950
486
+ },
487
+ {
488
+ "epoch": 15.08,
489
+ "learning_rate": 6.731111111111111e-06,
490
+ "loss": 0.0006,
491
+ "step": 1975
492
+ },
493
+ {
494
+ "epoch": 15.27,
495
+ "learning_rate": 6.675555555555556e-06,
496
+ "loss": 0.0007,
497
+ "step": 2000
498
+ },
499
+ {
500
+ "epoch": 15.27,
501
+ "eval_loss": 0.35849472880363464,
502
+ "eval_runtime": 1718.5005,
503
+ "eval_samples_per_second": 2.61,
504
+ "eval_steps_per_second": 0.164,
505
+ "eval_wer": 0.7478299577119965,
506
+ "step": 2000
507
  }
508
  ],
509
  "max_steps": 5000,
510
  "num_train_epochs": 39,
511
+ "total_flos": 2.1088396164096e+20,
512
  "trial_name": null,
513
  "trial_params": null
514
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e973125a216cde40c6c4f8e3b9480097d49da0899916d84d5d6eaca55778554c
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be6f8e930b75057373efa9b8e571972729960add3845f5586fc078849235cc72
3
  size 6173655480
runs/Jan28_23-42-14_gpu07.cyverse.org/events.out.tfevents.1674974582.gpu07.cyverse.org.94170.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91c6b9a96115582558449d245016f6bb5797363cc081b0a5f708e7a6764a09e3
3
- size 11002
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ddf710c6446bd8b1abd5e8954c53b0172994cf8f274bf4cf7ece1cd249946a8
3
+ size 17600