Check commited on
Commit
d0051dc
Β·
1 Parent(s): 71a1e60

auto-commit

Browse files
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1573691ea7d0fc61236cc2759103b23e73e155f43b48959ab586bdb034de93db
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76e57d6f91737dbbfab290b5599d5b3c4aade067aae28aad47ddc55053256d5c
3
  size 722165393
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ad107c141d3f7809507fa2921dcccdf0788b1defb6ed5e37e0ce2eaa7d50eec
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecd6c18707f6f175faf10b4faeebe9a418239519113272879c515e06333a64f3
3
  size 377909911
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0940cfeead080e2f3ccd94ee15dab324a13e5461273c03d52367d56baa29b599
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e8f08e3f19a28bcef3f6bda763058ad4ec179589a461d0c48f62c72069e50c3
3
  size 14503
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd0c1baf2f5398bf54715f4be73eadcf661402972348e07ee31ecfc7ef64bd75
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9423e2e940526608df5ef6924192b590e1b3251579ffbe9455b67f9f06c55439
3
  size 559
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcf1b5511c9d2f60343a826d533e62e2cd2de37fa4f207684f8625a938a917e1
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01c66c6381dce1adcf2bea3883f64b5ca782f6591999977f33146f32cb6de4b0
3
  size 623
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
- "epoch": 902.995983935743,
5
- "global_step": 112002,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -241866,11 +241866,329 @@
241866
  "eval_steps_per_second": 0.625,
241867
  "eval_wer": 0.1886928860613071,
241868
  "step": 112002
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
241869
  }
241870
  ],
241871
- "max_steps": 620000,
241872
  "num_train_epochs": 5000,
241873
- "total_flos": 3.1520264569981986e+20,
241874
  "trial_name": null,
241875
  "trial_params": null
241876
  }
 
1
  {
2
  "best_metric": 0.1743826049391605,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
4
+ "epoch": 897.9960159362549,
5
+ "global_step": 112251,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
241866
  "eval_steps_per_second": 0.625,
241867
  "eval_wer": 0.1886928860613071,
241868
  "step": 112002
241869
+ },
241870
+ {
241871
+ "epoch": 903.02,
241872
+ "learning_rate": 8.207415185783522e-06,
241873
+ "loss": 0.3015,
241874
+ "step": 112005
241875
+ },
241876
+ {
241877
+ "epoch": 903.06,
241878
+ "learning_rate": 8.207334410339256e-06,
241879
+ "loss": 0.2631,
241880
+ "step": 112010
241881
+ },
241882
+ {
241883
+ "epoch": 903.1,
241884
+ "learning_rate": 8.207253634894992e-06,
241885
+ "loss": 0.2948,
241886
+ "step": 112015
241887
+ },
241888
+ {
241889
+ "epoch": 903.14,
241890
+ "learning_rate": 8.207172859450726e-06,
241891
+ "loss": 0.3397,
241892
+ "step": 112020
241893
+ },
241894
+ {
241895
+ "epoch": 903.18,
241896
+ "learning_rate": 8.207092084006462e-06,
241897
+ "loss": 0.7385,
241898
+ "step": 112025
241899
+ },
241900
+ {
241901
+ "epoch": 903.22,
241902
+ "learning_rate": 8.207011308562198e-06,
241903
+ "loss": 0.9632,
241904
+ "step": 112030
241905
+ },
241906
+ {
241907
+ "epoch": 903.27,
241908
+ "learning_rate": 8.206930533117932e-06,
241909
+ "loss": 0.3019,
241910
+ "step": 112035
241911
+ },
241912
+ {
241913
+ "epoch": 903.31,
241914
+ "learning_rate": 8.206849757673668e-06,
241915
+ "loss": 0.2995,
241916
+ "step": 112040
241917
+ },
241918
+ {
241919
+ "epoch": 903.35,
241920
+ "learning_rate": 8.206768982229402e-06,
241921
+ "loss": 0.3895,
241922
+ "step": 112045
241923
+ },
241924
+ {
241925
+ "epoch": 903.39,
241926
+ "learning_rate": 8.206688206785138e-06,
241927
+ "loss": 0.7798,
241928
+ "step": 112050
241929
+ },
241930
+ {
241931
+ "epoch": 903.43,
241932
+ "learning_rate": 8.206607431340872e-06,
241933
+ "loss": 0.801,
241934
+ "step": 112055
241935
+ },
241936
+ {
241937
+ "epoch": 903.47,
241938
+ "learning_rate": 8.206526655896608e-06,
241939
+ "loss": 0.2984,
241940
+ "step": 112060
241941
+ },
241942
+ {
241943
+ "epoch": 903.51,
241944
+ "learning_rate": 8.206445880452342e-06,
241945
+ "loss": 0.2676,
241946
+ "step": 112065
241947
+ },
241948
+ {
241949
+ "epoch": 903.55,
241950
+ "learning_rate": 8.206365105008078e-06,
241951
+ "loss": 0.3573,
241952
+ "step": 112070
241953
+ },
241954
+ {
241955
+ "epoch": 903.59,
241956
+ "learning_rate": 8.206284329563812e-06,
241957
+ "loss": 0.6617,
241958
+ "step": 112075
241959
+ },
241960
+ {
241961
+ "epoch": 903.63,
241962
+ "learning_rate": 8.206203554119548e-06,
241963
+ "loss": 0.8261,
241964
+ "step": 112080
241965
+ },
241966
+ {
241967
+ "epoch": 903.67,
241968
+ "learning_rate": 8.206122778675282e-06,
241969
+ "loss": 0.2804,
241970
+ "step": 112085
241971
+ },
241972
+ {
241973
+ "epoch": 903.71,
241974
+ "learning_rate": 8.206042003231018e-06,
241975
+ "loss": 0.3004,
241976
+ "step": 112090
241977
+ },
241978
+ {
241979
+ "epoch": 903.75,
241980
+ "learning_rate": 8.205961227786754e-06,
241981
+ "loss": 0.3085,
241982
+ "step": 112095
241983
+ },
241984
+ {
241985
+ "epoch": 903.79,
241986
+ "learning_rate": 8.205880452342488e-06,
241987
+ "loss": 0.6766,
241988
+ "step": 112100
241989
+ },
241990
+ {
241991
+ "epoch": 903.83,
241992
+ "learning_rate": 8.205799676898224e-06,
241993
+ "loss": 0.8378,
241994
+ "step": 112105
241995
+ },
241996
+ {
241997
+ "epoch": 903.87,
241998
+ "learning_rate": 8.205718901453958e-06,
241999
+ "loss": 0.3095,
242000
+ "step": 112110
242001
+ },
242002
+ {
242003
+ "epoch": 903.91,
242004
+ "learning_rate": 8.205638126009694e-06,
242005
+ "loss": 0.2864,
242006
+ "step": 112115
242007
+ },
242008
+ {
242009
+ "epoch": 903.95,
242010
+ "learning_rate": 8.205557350565428e-06,
242011
+ "loss": 0.3575,
242012
+ "step": 112120
242013
+ },
242014
+ {
242015
+ "epoch": 903.99,
242016
+ "learning_rate": 8.205476575121164e-06,
242017
+ "loss": 0.8758,
242018
+ "step": 112125
242019
+ },
242020
+ {
242021
+ "epoch": 904.0,
242022
+ "eval_loss": 0.4589906334877014,
242023
+ "eval_runtime": 43.1012,
242024
+ "eval_samples_per_second": 19.419,
242025
+ "eval_steps_per_second": 0.626,
242026
+ "eval_wer": 0.1792831541218638,
242027
+ "step": 112126
242028
+ },
242029
+ {
242030
+ "epoch": 897.03,
242031
+ "learning_rate": 8.205395799676898e-06,
242032
+ "loss": 0.3184,
242033
+ "step": 112130
242034
+ },
242035
+ {
242036
+ "epoch": 897.07,
242037
+ "learning_rate": 8.205315024232634e-06,
242038
+ "loss": 0.281,
242039
+ "step": 112135
242040
+ },
242041
+ {
242042
+ "epoch": 897.11,
242043
+ "learning_rate": 8.205234248788368e-06,
242044
+ "loss": 0.3728,
242045
+ "step": 112140
242046
+ },
242047
+ {
242048
+ "epoch": 897.15,
242049
+ "learning_rate": 8.205153473344104e-06,
242050
+ "loss": 0.3522,
242051
+ "step": 112145
242052
+ },
242053
+ {
242054
+ "epoch": 897.19,
242055
+ "learning_rate": 8.20507269789984e-06,
242056
+ "loss": 0.8106,
242057
+ "step": 112150
242058
+ },
242059
+ {
242060
+ "epoch": 897.23,
242061
+ "learning_rate": 8.204991922455574e-06,
242062
+ "loss": 0.7244,
242063
+ "step": 112155
242064
+ },
242065
+ {
242066
+ "epoch": 897.27,
242067
+ "learning_rate": 8.20491114701131e-06,
242068
+ "loss": 0.2733,
242069
+ "step": 112160
242070
+ },
242071
+ {
242072
+ "epoch": 897.31,
242073
+ "learning_rate": 8.204830371567044e-06,
242074
+ "loss": 0.3499,
242075
+ "step": 112165
242076
+ },
242077
+ {
242078
+ "epoch": 897.35,
242079
+ "learning_rate": 8.20474959612278e-06,
242080
+ "loss": 0.3831,
242081
+ "step": 112170
242082
+ },
242083
+ {
242084
+ "epoch": 897.39,
242085
+ "learning_rate": 8.204668820678514e-06,
242086
+ "loss": 0.8812,
242087
+ "step": 112175
242088
+ },
242089
+ {
242090
+ "epoch": 897.43,
242091
+ "learning_rate": 8.20458804523425e-06,
242092
+ "loss": 0.5804,
242093
+ "step": 112180
242094
+ },
242095
+ {
242096
+ "epoch": 897.47,
242097
+ "learning_rate": 8.204507269789984e-06,
242098
+ "loss": 0.2352,
242099
+ "step": 112185
242100
+ },
242101
+ {
242102
+ "epoch": 897.51,
242103
+ "learning_rate": 8.20442649434572e-06,
242104
+ "loss": 0.3085,
242105
+ "step": 112190
242106
+ },
242107
+ {
242108
+ "epoch": 897.55,
242109
+ "learning_rate": 8.204345718901454e-06,
242110
+ "loss": 0.4456,
242111
+ "step": 112195
242112
+ },
242113
+ {
242114
+ "epoch": 897.59,
242115
+ "learning_rate": 8.20426494345719e-06,
242116
+ "loss": 1.0291,
242117
+ "step": 112200
242118
+ },
242119
+ {
242120
+ "epoch": 897.63,
242121
+ "learning_rate": 8.204184168012925e-06,
242122
+ "loss": 0.6779,
242123
+ "step": 112205
242124
+ },
242125
+ {
242126
+ "epoch": 897.67,
242127
+ "learning_rate": 8.20410339256866e-06,
242128
+ "loss": 0.3265,
242129
+ "step": 112210
242130
+ },
242131
+ {
242132
+ "epoch": 897.71,
242133
+ "learning_rate": 8.204022617124395e-06,
242134
+ "loss": 0.3081,
242135
+ "step": 112215
242136
+ },
242137
+ {
242138
+ "epoch": 897.75,
242139
+ "learning_rate": 8.20394184168013e-06,
242140
+ "loss": 0.3493,
242141
+ "step": 112220
242142
+ },
242143
+ {
242144
+ "epoch": 897.79,
242145
+ "learning_rate": 8.203861066235865e-06,
242146
+ "loss": 1.0938,
242147
+ "step": 112225
242148
+ },
242149
+ {
242150
+ "epoch": 897.83,
242151
+ "learning_rate": 8.2037802907916e-06,
242152
+ "loss": 0.6269,
242153
+ "step": 112230
242154
+ },
242155
+ {
242156
+ "epoch": 897.87,
242157
+ "learning_rate": 8.203699515347335e-06,
242158
+ "loss": 0.2867,
242159
+ "step": 112235
242160
+ },
242161
+ {
242162
+ "epoch": 897.91,
242163
+ "learning_rate": 8.20361873990307e-06,
242164
+ "loss": 0.2649,
242165
+ "step": 112240
242166
+ },
242167
+ {
242168
+ "epoch": 897.95,
242169
+ "learning_rate": 8.203537964458805e-06,
242170
+ "loss": 0.3792,
242171
+ "step": 112245
242172
+ },
242173
+ {
242174
+ "epoch": 897.99,
242175
+ "learning_rate": 8.20345718901454e-06,
242176
+ "loss": 0.8685,
242177
+ "step": 112250
242178
+ },
242179
+ {
242180
+ "epoch": 898.0,
242181
+ "eval_loss": 0.3989468812942505,
242182
+ "eval_runtime": 41.9774,
242183
+ "eval_samples_per_second": 19.939,
242184
+ "eval_steps_per_second": 0.643,
242185
+ "eval_wer": 0.18449336594766882,
242186
+ "step": 112251
242187
  }
242188
  ],
242189
+ "max_steps": 625000,
242190
  "num_train_epochs": 5000,
242191
+ "total_flos": 3.1590504384111575e+20,
242192
  "trial_name": null,
242193
  "trial_params": null
242194
  }
model-bin/finetune/base/{checkpoint-112002 β†’ checkpoint-112251}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630050545.555799/events.out.tfevents.1630050545.52f5c7e305a3.886.91 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b85c2968aa83ff0f693eff4476fc2b9181761fb526fed2f883a45abf46c87b1
3
+ size 4194
model-bin/finetune/base/log/1630051145.811182/events.out.tfevents.1630051145.52f5c7e305a3.886.93 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:335fda2c5e8f4d514c540425f74d9f5cbabdac3af328383abbdd4d89fc0bc2db
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630050545.52f5c7e305a3.886.90 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77bb63ff76d7de4d2c5a2692cedc707f46a50e1723e17f41e4e37445616375b1
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630051145.52f5c7e305a3.886.92 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a15adaf57895b016132c8c82eefaac064357dffb3d30c30422b3343a5fdf63d2
3
+ size 8622