ToastyPigeon commited on
Commit
c24944a
·
verified ·
1 Parent(s): 9d3e6c0

Training in progress, step 90, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7e2cd2dbf78c54ea3f0f06d49fd1af5b70a3e03ac9e6c82cb1d62722b2593b6
3
  size 763470136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cd39093da46c760bf0ad6acee7d3afbdc458b320613d02a1b9b87bf1dfd3d57
3
  size 763470136
last-checkpoint/global_step90/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a27a4dddf9f2faeaf94ce4b93f01574aced05b3161789cd7daf355b557ac4f
3
+ size 1152331664
last-checkpoint/global_step90/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bce11d1aa5e058c4096f50eaefdce8b585234e4a66af48e9e3ec4f8497934da3
3
+ size 1152331664
last-checkpoint/global_step90/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29a0796cc9183991e8ebdefaea1a53aa52c5ae33fb8faea94492f5f5b948af0a
3
+ size 348711830
last-checkpoint/global_step90/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:860d64ab323625107a1e3ec5efc33674241ef63698c7df279d798fe80ac1b076
3
+ size 348711830
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step60
 
1
+ global_step90
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13d47ce3a312fdef8e844802a1273de45b2a1adbfb4848f261f3e8b747e28be5
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:402dc8f05e787a8bd1f6017a51a44dd402264604e6e8b23372c047903d00e275
3
  size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df4302bd5b366946b267764b57680418f02385ae358386a07ebe9d879f34434f
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:385e4b7d88dae501cd7416828b65c0f6787ac9f75fe1a28f84bf64ac6675909d
3
  size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aad2d65943628ea6b328be0e7f3056643998d5e6f7a4812061e3d7f5ba16c03e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac3936d03564e869c0f8be0c0a24855f0fbaa9d17a65502bfe26d2b6f28ba7ad
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4006677796327212,
5
  "eval_steps": 30,
6
- "global_step": 60,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -451,6 +451,224 @@
451
  "eval_samples_per_second": 0.314,
452
  "eval_steps_per_second": 0.157,
453
  "step": 60
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
454
  }
455
  ],
456
  "logging_steps": 1,
@@ -470,7 +688,7 @@
470
  "attributes": {}
471
  }
472
  },
473
- "total_flos": 7157458711609344.0,
474
  "train_batch_size": 1,
475
  "trial_name": null,
476
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.6010016694490818,
5
  "eval_steps": 30,
6
+ "global_step": 90,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
451
  "eval_samples_per_second": 0.314,
452
  "eval_steps_per_second": 0.157,
453
  "step": 60
454
+ },
455
+ {
456
+ "epoch": 0.4073455759599332,
457
+ "grad_norm": 0.8826173031830702,
458
+ "learning_rate": 3.6805066361746716e-05,
459
+ "loss": 1.249,
460
+ "step": 61
461
+ },
462
+ {
463
+ "epoch": 0.41402337228714525,
464
+ "grad_norm": 0.7923611594554809,
465
+ "learning_rate": 3.634306321221328e-05,
466
+ "loss": 1.391,
467
+ "step": 62
468
+ },
469
+ {
470
+ "epoch": 0.42070116861435725,
471
+ "grad_norm": 0.761557330941142,
472
+ "learning_rate": 3.587660731404377e-05,
473
+ "loss": 1.219,
474
+ "step": 63
475
+ },
476
+ {
477
+ "epoch": 0.4273789649415693,
478
+ "grad_norm": 0.5702377539149609,
479
+ "learning_rate": 3.5405933541830215e-05,
480
+ "loss": 1.0854,
481
+ "step": 64
482
+ },
483
+ {
484
+ "epoch": 0.4340567612687813,
485
+ "grad_norm": 0.5797801586198355,
486
+ "learning_rate": 3.493127889399126e-05,
487
+ "loss": 1.4702,
488
+ "step": 65
489
+ },
490
+ {
491
+ "epoch": 0.44073455759599334,
492
+ "grad_norm": 0.6326945902799598,
493
+ "learning_rate": 3.4452882373436316e-05,
494
+ "loss": 1.5134,
495
+ "step": 66
496
+ },
497
+ {
498
+ "epoch": 0.44741235392320533,
499
+ "grad_norm": 1.1126743861778277,
500
+ "learning_rate": 3.397098486722039e-05,
501
+ "loss": 1.0709,
502
+ "step": 67
503
+ },
504
+ {
505
+ "epoch": 0.4540901502504174,
506
+ "grad_norm": 0.7069612379078216,
507
+ "learning_rate": 3.348582902525019e-05,
508
+ "loss": 1.3329,
509
+ "step": 68
510
+ },
511
+ {
512
+ "epoch": 0.4607679465776294,
513
+ "grad_norm": 0.611072274240805,
514
+ "learning_rate": 3.2997659138102615e-05,
515
+ "loss": 1.4334,
516
+ "step": 69
517
+ },
518
+ {
519
+ "epoch": 0.4674457429048414,
520
+ "grad_norm": 1.103066910563406,
521
+ "learning_rate": 3.250672101401707e-05,
522
+ "loss": 1.1569,
523
+ "step": 70
524
+ },
525
+ {
526
+ "epoch": 0.4741235392320534,
527
+ "grad_norm": 0.8340276510866299,
528
+ "learning_rate": 3.201326185512364e-05,
529
+ "loss": 1.0757,
530
+ "step": 71
531
+ },
532
+ {
533
+ "epoch": 0.48080133555926546,
534
+ "grad_norm": 0.8188265127022848,
535
+ "learning_rate": 3.1517530132969326e-05,
536
+ "loss": 1.371,
537
+ "step": 72
538
+ },
539
+ {
540
+ "epoch": 0.48747913188647746,
541
+ "grad_norm": 0.5777487141982147,
542
+ "learning_rate": 3.1019775463405195e-05,
543
+ "loss": 1.2921,
544
+ "step": 73
545
+ },
546
+ {
547
+ "epoch": 0.4941569282136895,
548
+ "grad_norm": 0.5750437954715367,
549
+ "learning_rate": 3.052024848089725e-05,
550
+ "loss": 1.3945,
551
+ "step": 74
552
+ },
553
+ {
554
+ "epoch": 0.5008347245409015,
555
+ "grad_norm": 0.7900878178576927,
556
+ "learning_rate": 3.001920071232443e-05,
557
+ "loss": 1.2691,
558
+ "step": 75
559
+ },
560
+ {
561
+ "epoch": 0.5075125208681135,
562
+ "grad_norm": 0.5377760257503877,
563
+ "learning_rate": 2.951688445032726e-05,
564
+ "loss": 1.1235,
565
+ "step": 76
566
+ },
567
+ {
568
+ "epoch": 0.5141903171953256,
569
+ "grad_norm": 0.5732341950023626,
570
+ "learning_rate": 2.9013552626270894e-05,
571
+ "loss": 1.1105,
572
+ "step": 77
573
+ },
574
+ {
575
+ "epoch": 0.5208681135225376,
576
+ "grad_norm": 0.5765972783223694,
577
+ "learning_rate": 2.8509458682886592e-05,
578
+ "loss": 1.5025,
579
+ "step": 78
580
+ },
581
+ {
582
+ "epoch": 0.5275459098497496,
583
+ "grad_norm": 0.5473629053792197,
584
+ "learning_rate": 2.8004856446655614e-05,
585
+ "loss": 0.8961,
586
+ "step": 79
587
+ },
588
+ {
589
+ "epoch": 0.5342237061769616,
590
+ "grad_norm": 0.6328841066421724,
591
+ "learning_rate": 2.7500000000000004e-05,
592
+ "loss": 1.1652,
593
+ "step": 80
594
+ },
595
+ {
596
+ "epoch": 0.5409015025041736,
597
+ "grad_norm": 0.5537252916610816,
598
+ "learning_rate": 2.6995143553344394e-05,
599
+ "loss": 1.5573,
600
+ "step": 81
601
+ },
602
+ {
603
+ "epoch": 0.5475792988313857,
604
+ "grad_norm": 0.670737956141697,
605
+ "learning_rate": 2.6490541317113427e-05,
606
+ "loss": 1.2827,
607
+ "step": 82
608
+ },
609
+ {
610
+ "epoch": 0.5542570951585977,
611
+ "grad_norm": 0.8099350587885803,
612
+ "learning_rate": 2.598644737372911e-05,
613
+ "loss": 1.1264,
614
+ "step": 83
615
+ },
616
+ {
617
+ "epoch": 0.5609348914858097,
618
+ "grad_norm": 0.5790201863930161,
619
+ "learning_rate": 2.548311554967275e-05,
620
+ "loss": 1.25,
621
+ "step": 84
622
+ },
623
+ {
624
+ "epoch": 0.5676126878130217,
625
+ "grad_norm": 0.7651780955233705,
626
+ "learning_rate": 2.4980799287675578e-05,
627
+ "loss": 1.5117,
628
+ "step": 85
629
+ },
630
+ {
631
+ "epoch": 0.5742904841402338,
632
+ "grad_norm": 0.8401680884601399,
633
+ "learning_rate": 2.447975151910276e-05,
634
+ "loss": 1.0176,
635
+ "step": 86
636
+ },
637
+ {
638
+ "epoch": 0.5809682804674458,
639
+ "grad_norm": 0.7026724359120083,
640
+ "learning_rate": 2.3980224536594803e-05,
641
+ "loss": 1.2334,
642
+ "step": 87
643
+ },
644
+ {
645
+ "epoch": 0.5876460767946577,
646
+ "grad_norm": 1.434313668658568,
647
+ "learning_rate": 2.3482469867030676e-05,
648
+ "loss": 1.2272,
649
+ "step": 88
650
+ },
651
+ {
652
+ "epoch": 0.5943238731218697,
653
+ "grad_norm": 0.5970254035001638,
654
+ "learning_rate": 2.298673814487636e-05,
655
+ "loss": 1.2615,
656
+ "step": 89
657
+ },
658
+ {
659
+ "epoch": 0.6010016694490818,
660
+ "grad_norm": 0.5775454892942203,
661
+ "learning_rate": 2.249327898598293e-05,
662
+ "loss": 1.1957,
663
+ "step": 90
664
+ },
665
+ {
666
+ "epoch": 0.6010016694490818,
667
+ "eval_loss": 0.9956896305084229,
668
+ "eval_runtime": 318.961,
669
+ "eval_samples_per_second": 0.314,
670
+ "eval_steps_per_second": 0.157,
671
+ "step": 90
672
  }
673
  ],
674
  "logging_steps": 1,
 
688
  "attributes": {}
689
  }
690
  },
691
+ "total_flos": 1.0736256317128704e+16,
692
  "train_batch_size": 1,
693
  "trial_name": null,
694
  "trial_params": null