nttx commited on
Commit
f93046f
·
verified ·
1 Parent(s): 1c6c43c

Training in progress, step 132, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65c3c008fbebe693d48c3c3c289982803a9089ddc206dd90739ec5a2574c1f40
3
  size 130583912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0477ae16092e920f1a70285ec0f6d2a1ad995f8c873e61b0e023ffc0d47ff57d
3
  size 130583912
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:154095e3f56c856d2019627a7a0e1deb4d5a420097aba2ee5ff882113dacde7f
3
  size 261311226
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a1c59e022210ab0f0eb69ca2730f81f70cad23d48653d5cd87a7b12ca7f7112
3
  size 261311226
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2de528baee32c70568e916c3b8f97af886cc74df910fe8ed9e8f92fd7c304f60
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4628cfb5a20dd1151dbfc415cbd7fc7464cb18b044ebb8faa5231ab7e2521acf
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33fdeac700456f590554babdb1bb06f3350befc1bacd0a19b3d33f64c1a1c32d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86e6e29d81cb95f8c8a14f9b3d672190de4431bdbc9834b879a12c3bb9d7d827
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.39477089047431946,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-100",
4
- "epoch": 1.5173067804646752,
5
  "eval_steps": 25,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -747,6 +747,238 @@
747
  "eval_samples_per_second": 11.898,
748
  "eval_steps_per_second": 11.898,
749
  "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
750
  }
751
  ],
752
  "logging_steps": 1,
@@ -770,12 +1002,12 @@
770
  "should_evaluate": false,
771
  "should_log": false,
772
  "should_save": true,
773
- "should_training_stop": false
774
  },
775
  "attributes": {}
776
  }
777
  },
778
- "total_flos": 1.391015187972096e+17,
779
  "train_batch_size": 1,
780
  "trial_name": null,
781
  "trial_params": null
 
1
  {
2
  "best_metric": 0.39477089047431946,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-100",
4
+ "epoch": 2.002844950213371,
5
  "eval_steps": 25,
6
+ "global_step": 132,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
747
  "eval_samples_per_second": 11.898,
748
  "eval_steps_per_second": 11.898,
749
  "step": 100
750
+ },
751
+ {
752
+ "epoch": 1.5324798482693218,
753
+ "grad_norm": 8.945897102355957,
754
+ "learning_rate": 4.5308376785048434e-05,
755
+ "loss": 11.1062,
756
+ "step": 101
757
+ },
758
+ {
759
+ "epoch": 1.5476529160739687,
760
+ "grad_norm": 9.844161987304688,
761
+ "learning_rate": 4.257717175284103e-05,
762
+ "loss": 10.4609,
763
+ "step": 102
764
+ },
765
+ {
766
+ "epoch": 1.5628259838786156,
767
+ "grad_norm": 9.294745445251465,
768
+ "learning_rate": 3.991719503517014e-05,
769
+ "loss": 10.6322,
770
+ "step": 103
771
+ },
772
+ {
773
+ "epoch": 1.5779990516832623,
774
+ "grad_norm": 9.901968002319336,
775
+ "learning_rate": 3.733021036942205e-05,
776
+ "loss": 11.3513,
777
+ "step": 104
778
+ },
779
+ {
780
+ "epoch": 1.593172119487909,
781
+ "grad_norm": 9.366765975952148,
782
+ "learning_rate": 3.481793309451e-05,
783
+ "loss": 10.2632,
784
+ "step": 105
785
+ },
786
+ {
787
+ "epoch": 1.6083451872925556,
788
+ "grad_norm": 8.39588451385498,
789
+ "learning_rate": 3.238202901349345e-05,
790
+ "loss": 10.2529,
791
+ "step": 106
792
+ },
793
+ {
794
+ "epoch": 1.6235182550972025,
795
+ "grad_norm": 8.285764694213867,
796
+ "learning_rate": 3.0024113289042094e-05,
797
+ "loss": 9.9579,
798
+ "step": 107
799
+ },
800
+ {
801
+ "epoch": 1.6386913229018494,
802
+ "grad_norm": 8.389649391174316,
803
+ "learning_rate": 2.774574937247831e-05,
804
+ "loss": 10.3428,
805
+ "step": 108
806
+ },
807
+ {
808
+ "epoch": 1.653864390706496,
809
+ "grad_norm": 8.863457679748535,
810
+ "learning_rate": 2.554844796710716e-05,
811
+ "loss": 9.0607,
812
+ "step": 109
813
+ },
814
+ {
815
+ "epoch": 1.6690374585111427,
816
+ "grad_norm": 8.510146141052246,
817
+ "learning_rate": 2.3433666026522153e-05,
818
+ "loss": 9.3396,
819
+ "step": 110
820
+ },
821
+ {
822
+ "epoch": 1.6842105263157894,
823
+ "grad_norm": 9.09020709991455,
824
+ "learning_rate": 2.1402805788550138e-05,
825
+ "loss": 9.223,
826
+ "step": 111
827
+ },
828
+ {
829
+ "epoch": 1.6993835941204363,
830
+ "grad_norm": 8.539032936096191,
831
+ "learning_rate": 1.945721384547671e-05,
832
+ "loss": 7.2793,
833
+ "step": 112
834
+ },
835
+ {
836
+ "epoch": 1.714556661925083,
837
+ "grad_norm": 9.459700584411621,
838
+ "learning_rate": 1.759818025116787e-05,
839
+ "loss": 7.6996,
840
+ "step": 113
841
+ },
842
+ {
843
+ "epoch": 1.7297297297297298,
844
+ "grad_norm": 11.580260276794434,
845
+ "learning_rate": 1.5826937665680693e-05,
846
+ "loss": 7.68,
847
+ "step": 114
848
+ },
849
+ {
850
+ "epoch": 1.7449027975343765,
851
+ "grad_norm": 8.748165130615234,
852
+ "learning_rate": 1.4144660537929287e-05,
853
+ "loss": 13.0615,
854
+ "step": 115
855
+ },
856
+ {
857
+ "epoch": 1.7600758653390232,
858
+ "grad_norm": 9.052087783813477,
859
+ "learning_rate": 1.2552464326949302e-05,
860
+ "loss": 13.9056,
861
+ "step": 116
862
+ },
863
+ {
864
+ "epoch": 1.7752489331436698,
865
+ "grad_norm": 9.413413047790527,
866
+ "learning_rate": 1.105140476227575e-05,
867
+ "loss": 13.1337,
868
+ "step": 117
869
+ },
870
+ {
871
+ "epoch": 1.7904220009483167,
872
+ "grad_norm": 8.633859634399414,
873
+ "learning_rate": 9.64247714392597e-06,
874
+ "loss": 11.1511,
875
+ "step": 118
876
+ },
877
+ {
878
+ "epoch": 1.8055950687529636,
879
+ "grad_norm": 7.859529972076416,
880
+ "learning_rate": 8.32661568245081e-06,
881
+ "loss": 10.06,
882
+ "step": 119
883
+ },
884
+ {
885
+ "epoch": 1.8207681365576103,
886
+ "grad_norm": 8.01181697845459,
887
+ "learning_rate": 7.104692879492624e-06,
888
+ "loss": 10.4743,
889
+ "step": 120
890
+ },
891
+ {
892
+ "epoch": 1.835941204362257,
893
+ "grad_norm": 9.660663604736328,
894
+ "learning_rate": 5.977518949259735e-06,
895
+ "loss": 11.0629,
896
+ "step": 121
897
+ },
898
+ {
899
+ "epoch": 1.8511142721669036,
900
+ "grad_norm": 8.536215782165527,
901
+ "learning_rate": 4.945841281301943e-06,
902
+ "loss": 10.3549,
903
+ "step": 122
904
+ },
905
+ {
906
+ "epoch": 1.8662873399715505,
907
+ "grad_norm": 8.264043807983398,
908
+ "learning_rate": 4.010343944942618e-06,
909
+ "loss": 9.8312,
910
+ "step": 123
911
+ },
912
+ {
913
+ "epoch": 1.8814604077761974,
914
+ "grad_norm": 8.763490676879883,
915
+ "learning_rate": 3.1716472356963286e-06,
916
+ "loss": 8.8429,
917
+ "step": 124
918
+ },
919
+ {
920
+ "epoch": 1.896633475580844,
921
+ "grad_norm": 9.253498077392578,
922
+ "learning_rate": 2.430307263972547e-06,
923
+ "loss": 10.0951,
924
+ "step": 125
925
+ },
926
+ {
927
+ "epoch": 1.896633475580844,
928
+ "eval_loss": 0.3835863471031189,
929
+ "eval_runtime": 4.1958,
930
+ "eval_samples_per_second": 11.917,
931
+ "eval_steps_per_second": 11.917,
932
+ "step": 125
933
+ },
934
+ {
935
+ "epoch": 1.9118065433854907,
936
+ "grad_norm": 8.449151039123535,
937
+ "learning_rate": 1.7868155863384415e-06,
938
+ "loss": 9.1299,
939
+ "step": 126
940
+ },
941
+ {
942
+ "epoch": 1.9269796111901374,
943
+ "grad_norm": 8.359270095825195,
944
+ "learning_rate": 1.2415988795847765e-06,
945
+ "loss": 9.0518,
946
+ "step": 127
947
+ },
948
+ {
949
+ "epoch": 1.9421526789947843,
950
+ "grad_norm": 8.54788589477539,
951
+ "learning_rate": 7.950186578116413e-07,
952
+ "loss": 7.7242,
953
+ "step": 128
954
+ },
955
+ {
956
+ "epoch": 1.9573257467994312,
957
+ "grad_norm": 8.631213188171387,
958
+ "learning_rate": 4.473710327209945e-07,
959
+ "loss": 6.4852,
960
+ "step": 129
961
+ },
962
+ {
963
+ "epoch": 1.9724988146040778,
964
+ "grad_norm": 8.945975303649902,
965
+ "learning_rate": 1.988865172754206e-07,
966
+ "loss": 7.753,
967
+ "step": 130
968
+ },
969
+ {
970
+ "epoch": 1.9876718824087245,
971
+ "grad_norm": 8.31573486328125,
972
+ "learning_rate": 4.972987285304375e-08,
973
+ "loss": 11.2559,
974
+ "step": 131
975
+ },
976
+ {
977
+ "epoch": 2.002844950213371,
978
+ "grad_norm": 8.791767120361328,
979
+ "learning_rate": 0.0,
980
+ "loss": 9.4748,
981
+ "step": 132
982
  }
983
  ],
984
  "logging_steps": 1,
 
1002
  "should_evaluate": false,
1003
  "should_log": false,
1004
  "should_save": true,
1005
+ "should_training_stop": true
1006
  },
1007
  "attributes": {}
1008
  }
1009
  },
1010
+ "total_flos": 1.8361400481231667e+17,
1011
  "train_batch_size": 1,
1012
  "trial_name": null,
1013
  "trial_params": null