kavyamanohar commited on
Commit
d92e2e0
1 Parent(s): 03fdd6a

Upload 8 files

Browse files
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-large-v2",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForConditionalGeneration"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": [
11
+ 220,
12
+ 50257
13
+ ],
14
+ "bos_token_id": 50257,
15
+ "classifier_proj_size": 256,
16
+ "d_model": 1280,
17
+ "decoder_attention_heads": 20,
18
+ "decoder_ffn_dim": 5120,
19
+ "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 32,
21
+ "decoder_start_token_id": 50258,
22
+ "dropout": 0.0,
23
+ "encoder_attention_heads": 20,
24
+ "encoder_ffn_dim": 5120,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 32,
27
+ "eos_token_id": 50257,
28
+ "forced_decoder_ids": null,
29
+ "init_std": 0.02,
30
+ "is_encoder_decoder": true,
31
+ "mask_feature_length": 10,
32
+ "mask_feature_min_masks": 0,
33
+ "mask_feature_prob": 0.0,
34
+ "mask_time_length": 10,
35
+ "mask_time_min_masks": 2,
36
+ "mask_time_prob": 0.05,
37
+ "max_length": 448,
38
+ "max_source_positions": 1500,
39
+ "max_target_positions": 448,
40
+ "model_type": "whisper",
41
+ "num_hidden_layers": 32,
42
+ "num_mel_bins": 80,
43
+ "pad_token_id": 50257,
44
+ "scale_embedding": false,
45
+ "suppress_tokens": [],
46
+ "torch_dtype": "float32",
47
+ "transformers_version": "4.28.0",
48
+ "use_cache": true,
49
+ "use_weighted_layer_sum": false,
50
+ "vocab_size": 51865
51
+ }
generation_config.json ADDED
@@ -0,0 +1,221 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "begin_suppress_tokens": [
3
+ 220,
4
+ 50257
5
+ ],
6
+ "bos_token_id": 50257,
7
+ "decoder_start_token_id": 50258,
8
+ "eos_token_id": 50257,
9
+ "forced_decoder_ids": [
10
+ [
11
+ 1,
12
+ null
13
+ ],
14
+ [
15
+ 2,
16
+ 50359
17
+ ]
18
+ ],
19
+ "is_multilingual": true,
20
+ "lang_to_id": {
21
+ "<|af|>": 50327,
22
+ "<|am|>": 50334,
23
+ "<|ar|>": 50272,
24
+ "<|as|>": 50350,
25
+ "<|az|>": 50304,
26
+ "<|ba|>": 50355,
27
+ "<|be|>": 50330,
28
+ "<|bg|>": 50292,
29
+ "<|bn|>": 50302,
30
+ "<|bo|>": 50347,
31
+ "<|br|>": 50309,
32
+ "<|bs|>": 50315,
33
+ "<|ca|>": 50270,
34
+ "<|cs|>": 50283,
35
+ "<|cy|>": 50297,
36
+ "<|da|>": 50285,
37
+ "<|de|>": 50261,
38
+ "<|el|>": 50281,
39
+ "<|en|>": 50259,
40
+ "<|es|>": 50262,
41
+ "<|et|>": 50307,
42
+ "<|eu|>": 50310,
43
+ "<|fa|>": 50300,
44
+ "<|fi|>": 50277,
45
+ "<|fo|>": 50338,
46
+ "<|fr|>": 50265,
47
+ "<|gl|>": 50319,
48
+ "<|gu|>": 50333,
49
+ "<|haw|>": 50352,
50
+ "<|ha|>": 50354,
51
+ "<|he|>": 50279,
52
+ "<|hi|>": 50276,
53
+ "<|hr|>": 50291,
54
+ "<|ht|>": 50339,
55
+ "<|hu|>": 50286,
56
+ "<|hy|>": 50312,
57
+ "<|id|>": 50275,
58
+ "<|is|>": 50311,
59
+ "<|it|>": 50274,
60
+ "<|ja|>": 50266,
61
+ "<|jw|>": 50356,
62
+ "<|ka|>": 50329,
63
+ "<|kk|>": 50316,
64
+ "<|km|>": 50323,
65
+ "<|kn|>": 50306,
66
+ "<|ko|>": 50264,
67
+ "<|la|>": 50294,
68
+ "<|lb|>": 50345,
69
+ "<|ln|>": 50353,
70
+ "<|lo|>": 50336,
71
+ "<|lt|>": 50293,
72
+ "<|lv|>": 50301,
73
+ "<|mg|>": 50349,
74
+ "<|mi|>": 50295,
75
+ "<|mk|>": 50308,
76
+ "<|ml|>": 50296,
77
+ "<|mn|>": 50314,
78
+ "<|mr|>": 50320,
79
+ "<|ms|>": 50282,
80
+ "<|mt|>": 50343,
81
+ "<|my|>": 50346,
82
+ "<|ne|>": 50313,
83
+ "<|nl|>": 50271,
84
+ "<|nn|>": 50342,
85
+ "<|no|>": 50288,
86
+ "<|oc|>": 50328,
87
+ "<|pa|>": 50321,
88
+ "<|pl|>": 50269,
89
+ "<|ps|>": 50340,
90
+ "<|pt|>": 50267,
91
+ "<|ro|>": 50284,
92
+ "<|ru|>": 50263,
93
+ "<|sa|>": 50344,
94
+ "<|sd|>": 50332,
95
+ "<|si|>": 50322,
96
+ "<|sk|>": 50298,
97
+ "<|sl|>": 50305,
98
+ "<|sn|>": 50324,
99
+ "<|so|>": 50326,
100
+ "<|sq|>": 50317,
101
+ "<|sr|>": 50303,
102
+ "<|su|>": 50357,
103
+ "<|sv|>": 50273,
104
+ "<|sw|>": 50318,
105
+ "<|ta|>": 50287,
106
+ "<|te|>": 50299,
107
+ "<|tg|>": 50331,
108
+ "<|th|>": 50289,
109
+ "<|tk|>": 50341,
110
+ "<|tl|>": 50348,
111
+ "<|tr|>": 50268,
112
+ "<|tt|>": 50351,
113
+ "<|uk|>": 50280,
114
+ "<|ur|>": 50290,
115
+ "<|uz|>": 50337,
116
+ "<|vi|>": 50278,
117
+ "<|yi|>": 50335,
118
+ "<|yo|>": 50325,
119
+ "<|zh|>": 50260
120
+ },
121
+ "max_initial_timestamp_index": 1,
122
+ "max_length": 448,
123
+ "no_timestamps_token_id": 50363,
124
+ "pad_token_id": 50257,
125
+ "return_timestamps": false,
126
+ "suppress_tokens": [
127
+ 1,
128
+ 2,
129
+ 7,
130
+ 8,
131
+ 9,
132
+ 10,
133
+ 14,
134
+ 25,
135
+ 26,
136
+ 27,
137
+ 28,
138
+ 29,
139
+ 31,
140
+ 58,
141
+ 59,
142
+ 60,
143
+ 61,
144
+ 62,
145
+ 63,
146
+ 90,
147
+ 91,
148
+ 92,
149
+ 93,
150
+ 359,
151
+ 503,
152
+ 522,
153
+ 542,
154
+ 873,
155
+ 893,
156
+ 902,
157
+ 918,
158
+ 922,
159
+ 931,
160
+ 1350,
161
+ 1853,
162
+ 1982,
163
+ 2460,
164
+ 2627,
165
+ 3246,
166
+ 3253,
167
+ 3268,
168
+ 3536,
169
+ 3846,
170
+ 3961,
171
+ 4183,
172
+ 4667,
173
+ 6585,
174
+ 6647,
175
+ 7273,
176
+ 9061,
177
+ 9383,
178
+ 10428,
179
+ 10929,
180
+ 11938,
181
+ 12033,
182
+ 12331,
183
+ 12562,
184
+ 13793,
185
+ 14157,
186
+ 14635,
187
+ 15265,
188
+ 15618,
189
+ 16553,
190
+ 16604,
191
+ 18362,
192
+ 18956,
193
+ 20075,
194
+ 21675,
195
+ 22520,
196
+ 26130,
197
+ 26161,
198
+ 26435,
199
+ 28279,
200
+ 29464,
201
+ 31650,
202
+ 32302,
203
+ 32470,
204
+ 36865,
205
+ 42863,
206
+ 47425,
207
+ 49870,
208
+ 50254,
209
+ 50258,
210
+ 50358,
211
+ 50359,
212
+ 50360,
213
+ 50361,
214
+ 50362
215
+ ],
216
+ "task_to_id": {
217
+ "transcribe": 50359,
218
+ "translate": 50358
219
+ },
220
+ "transformers_version": "4.28.0"
221
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68548de2ff34a5f9ab8e93fa29c238ffb124578cc90b59b8af3a5a45bb9bf778
3
+ size 14575
scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c44ed699a3e1c5392a880602c263e2c4ec3b6e43c8bd0c330782924af236404
3
+ size 557
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f81cf1cd1cdbdffc3c2356d643cfbae1e251bf74731b75889646ac295271662
3
+ size 627
trainer_state.json ADDED
@@ -0,0 +1,646 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 30.100143061516455,
3
+ "best_model_checkpoint": "./Malasar_50_latest/checkpoint-1000",
4
+ "epoch": 2.846299810246679,
5
+ "global_step": 1500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.05,
12
+ "learning_rate": 4.000000000000001e-06,
13
+ "loss": 2.4065,
14
+ "step": 25
15
+ },
16
+ {
17
+ "epoch": 0.09,
18
+ "learning_rate": 8.8e-06,
19
+ "loss": 0.6768,
20
+ "step": 50
21
+ },
22
+ {
23
+ "epoch": 0.09,
24
+ "eval_loss": 0.15913546085357666,
25
+ "eval_runtime": 1024.1019,
26
+ "eval_samples_per_second": 1.829,
27
+ "eval_steps_per_second": 0.115,
28
+ "eval_wer": 63.91988555078684,
29
+ "step": 50
30
+ },
31
+ {
32
+ "epoch": 0.14,
33
+ "learning_rate": 9.86896551724138e-06,
34
+ "loss": 0.1782,
35
+ "step": 75
36
+ },
37
+ {
38
+ "epoch": 0.19,
39
+ "learning_rate": 9.696551724137932e-06,
40
+ "loss": 0.1711,
41
+ "step": 100
42
+ },
43
+ {
44
+ "epoch": 0.19,
45
+ "eval_loss": 0.11521261185407639,
46
+ "eval_runtime": 1068.1467,
47
+ "eval_samples_per_second": 1.754,
48
+ "eval_steps_per_second": 0.11,
49
+ "eval_wer": 52.989985693848354,
50
+ "step": 100
51
+ },
52
+ {
53
+ "epoch": 0.24,
54
+ "learning_rate": 9.524137931034484e-06,
55
+ "loss": 0.151,
56
+ "step": 125
57
+ },
58
+ {
59
+ "epoch": 0.28,
60
+ "learning_rate": 9.351724137931034e-06,
61
+ "loss": 0.1392,
62
+ "step": 150
63
+ },
64
+ {
65
+ "epoch": 0.28,
66
+ "eval_loss": 0.1063603013753891,
67
+ "eval_runtime": 1067.453,
68
+ "eval_samples_per_second": 1.755,
69
+ "eval_steps_per_second": 0.111,
70
+ "eval_wer": 42.00286123032904,
71
+ "step": 150
72
+ },
73
+ {
74
+ "epoch": 0.33,
75
+ "learning_rate": 9.179310344827587e-06,
76
+ "loss": 0.1149,
77
+ "step": 175
78
+ },
79
+ {
80
+ "epoch": 0.38,
81
+ "learning_rate": 9.006896551724139e-06,
82
+ "loss": 0.1131,
83
+ "step": 200
84
+ },
85
+ {
86
+ "epoch": 0.38,
87
+ "eval_loss": 0.09218524396419525,
88
+ "eval_runtime": 1059.1591,
89
+ "eval_samples_per_second": 1.768,
90
+ "eval_steps_per_second": 0.111,
91
+ "eval_wer": 40.2002861230329,
92
+ "step": 200
93
+ },
94
+ {
95
+ "epoch": 0.43,
96
+ "learning_rate": 8.83448275862069e-06,
97
+ "loss": 0.1052,
98
+ "step": 225
99
+ },
100
+ {
101
+ "epoch": 0.47,
102
+ "learning_rate": 8.662068965517241e-06,
103
+ "loss": 0.1208,
104
+ "step": 250
105
+ },
106
+ {
107
+ "epoch": 0.47,
108
+ "eval_loss": 0.08814000338315964,
109
+ "eval_runtime": 1037.3449,
110
+ "eval_samples_per_second": 1.806,
111
+ "eval_steps_per_second": 0.114,
112
+ "eval_wer": 46.49499284692418,
113
+ "step": 250
114
+ },
115
+ {
116
+ "epoch": 0.52,
117
+ "learning_rate": 8.489655172413795e-06,
118
+ "loss": 0.1107,
119
+ "step": 275
120
+ },
121
+ {
122
+ "epoch": 0.57,
123
+ "learning_rate": 8.317241379310345e-06,
124
+ "loss": 0.1312,
125
+ "step": 300
126
+ },
127
+ {
128
+ "epoch": 0.57,
129
+ "eval_loss": 0.07998558133840561,
130
+ "eval_runtime": 1053.0015,
131
+ "eval_samples_per_second": 1.779,
132
+ "eval_steps_per_second": 0.112,
133
+ "eval_wer": 37.51072961373391,
134
+ "step": 300
135
+ },
136
+ {
137
+ "epoch": 0.62,
138
+ "learning_rate": 8.144827586206897e-06,
139
+ "loss": 0.0947,
140
+ "step": 325
141
+ },
142
+ {
143
+ "epoch": 0.66,
144
+ "learning_rate": 7.972413793103448e-06,
145
+ "loss": 0.1112,
146
+ "step": 350
147
+ },
148
+ {
149
+ "epoch": 0.66,
150
+ "eval_loss": 0.07743828743696213,
151
+ "eval_runtime": 1073.0726,
152
+ "eval_samples_per_second": 1.745,
153
+ "eval_steps_per_second": 0.11,
154
+ "eval_wer": 37.56795422031474,
155
+ "step": 350
156
+ },
157
+ {
158
+ "epoch": 0.71,
159
+ "learning_rate": 7.800000000000002e-06,
160
+ "loss": 0.118,
161
+ "step": 375
162
+ },
163
+ {
164
+ "epoch": 0.76,
165
+ "learning_rate": 7.627586206896552e-06,
166
+ "loss": 0.1,
167
+ "step": 400
168
+ },
169
+ {
170
+ "epoch": 0.76,
171
+ "eval_loss": 0.0833013579249382,
172
+ "eval_runtime": 1078.2864,
173
+ "eval_samples_per_second": 1.737,
174
+ "eval_steps_per_second": 0.109,
175
+ "eval_wer": 33.59084406294707,
176
+ "step": 400
177
+ },
178
+ {
179
+ "epoch": 0.81,
180
+ "learning_rate": 7.455172413793104e-06,
181
+ "loss": 0.1023,
182
+ "step": 425
183
+ },
184
+ {
185
+ "epoch": 0.85,
186
+ "learning_rate": 7.282758620689656e-06,
187
+ "loss": 0.094,
188
+ "step": 450
189
+ },
190
+ {
191
+ "epoch": 0.85,
192
+ "eval_loss": 0.07670725882053375,
193
+ "eval_runtime": 1078.233,
194
+ "eval_samples_per_second": 1.737,
195
+ "eval_steps_per_second": 0.109,
196
+ "eval_wer": 31.702432045779684,
197
+ "step": 450
198
+ },
199
+ {
200
+ "epoch": 0.9,
201
+ "learning_rate": 7.110344827586207e-06,
202
+ "loss": 0.0787,
203
+ "step": 475
204
+ },
205
+ {
206
+ "epoch": 0.95,
207
+ "learning_rate": 6.937931034482759e-06,
208
+ "loss": 0.0758,
209
+ "step": 500
210
+ },
211
+ {
212
+ "epoch": 0.95,
213
+ "eval_loss": 0.07279360294342041,
214
+ "eval_runtime": 1081.1361,
215
+ "eval_samples_per_second": 1.732,
216
+ "eval_steps_per_second": 0.109,
217
+ "eval_wer": 32.33190271816881,
218
+ "step": 500
219
+ },
220
+ {
221
+ "epoch": 1.0,
222
+ "learning_rate": 6.7655172413793116e-06,
223
+ "loss": 0.1144,
224
+ "step": 525
225
+ },
226
+ {
227
+ "epoch": 1.04,
228
+ "learning_rate": 6.593103448275863e-06,
229
+ "loss": 0.0967,
230
+ "step": 550
231
+ },
232
+ {
233
+ "epoch": 1.04,
234
+ "eval_loss": 0.07088593393564224,
235
+ "eval_runtime": 1067.9119,
236
+ "eval_samples_per_second": 1.754,
237
+ "eval_steps_per_second": 0.11,
238
+ "eval_wer": 34.39198855507868,
239
+ "step": 550
240
+ },
241
+ {
242
+ "epoch": 1.09,
243
+ "learning_rate": 6.420689655172414e-06,
244
+ "loss": 0.0497,
245
+ "step": 575
246
+ },
247
+ {
248
+ "epoch": 1.14,
249
+ "learning_rate": 6.248275862068966e-06,
250
+ "loss": 0.0423,
251
+ "step": 600
252
+ },
253
+ {
254
+ "epoch": 1.14,
255
+ "eval_loss": 0.07051743566989899,
256
+ "eval_runtime": 1075.828,
257
+ "eval_samples_per_second": 1.741,
258
+ "eval_steps_per_second": 0.11,
259
+ "eval_wer": 31.04434907010014,
260
+ "step": 600
261
+ },
262
+ {
263
+ "epoch": 1.19,
264
+ "learning_rate": 6.075862068965518e-06,
265
+ "loss": 0.0686,
266
+ "step": 625
267
+ },
268
+ {
269
+ "epoch": 1.23,
270
+ "learning_rate": 5.9034482758620695e-06,
271
+ "loss": 0.0669,
272
+ "step": 650
273
+ },
274
+ {
275
+ "epoch": 1.23,
276
+ "eval_loss": 0.07203543931245804,
277
+ "eval_runtime": 1076.3331,
278
+ "eval_samples_per_second": 1.74,
279
+ "eval_steps_per_second": 0.11,
280
+ "eval_wer": 31.587982832618028,
281
+ "step": 650
282
+ },
283
+ {
284
+ "epoch": 1.28,
285
+ "learning_rate": 5.731034482758621e-06,
286
+ "loss": 0.0615,
287
+ "step": 675
288
+ },
289
+ {
290
+ "epoch": 1.33,
291
+ "learning_rate": 5.558620689655173e-06,
292
+ "loss": 0.0686,
293
+ "step": 700
294
+ },
295
+ {
296
+ "epoch": 1.33,
297
+ "eval_loss": 0.06905751675367355,
298
+ "eval_runtime": 1084.2893,
299
+ "eval_samples_per_second": 1.727,
300
+ "eval_steps_per_second": 0.109,
301
+ "eval_wer": 31.21602288984263,
302
+ "step": 700
303
+ },
304
+ {
305
+ "epoch": 1.38,
306
+ "learning_rate": 5.386206896551725e-06,
307
+ "loss": 0.0704,
308
+ "step": 725
309
+ },
310
+ {
311
+ "epoch": 1.42,
312
+ "learning_rate": 5.213793103448276e-06,
313
+ "loss": 0.0798,
314
+ "step": 750
315
+ },
316
+ {
317
+ "epoch": 1.42,
318
+ "eval_loss": 0.06694240123033524,
319
+ "eval_runtime": 1087.4811,
320
+ "eval_samples_per_second": 1.722,
321
+ "eval_steps_per_second": 0.109,
322
+ "eval_wer": 30.643776824034337,
323
+ "step": 750
324
+ },
325
+ {
326
+ "epoch": 1.47,
327
+ "learning_rate": 5.041379310344828e-06,
328
+ "loss": 0.0756,
329
+ "step": 775
330
+ },
331
+ {
332
+ "epoch": 1.52,
333
+ "learning_rate": 4.8689655172413795e-06,
334
+ "loss": 0.0818,
335
+ "step": 800
336
+ },
337
+ {
338
+ "epoch": 1.52,
339
+ "eval_loss": 0.06710417568683624,
340
+ "eval_runtime": 1073.6009,
341
+ "eval_samples_per_second": 1.745,
342
+ "eval_steps_per_second": 0.11,
343
+ "eval_wer": 32.18884120171674,
344
+ "step": 800
345
+ },
346
+ {
347
+ "epoch": 1.57,
348
+ "learning_rate": 4.6965517241379315e-06,
349
+ "loss": 0.0571,
350
+ "step": 825
351
+ },
352
+ {
353
+ "epoch": 1.61,
354
+ "learning_rate": 4.524137931034483e-06,
355
+ "loss": 0.0841,
356
+ "step": 850
357
+ },
358
+ {
359
+ "epoch": 1.61,
360
+ "eval_loss": 0.06191212683916092,
361
+ "eval_runtime": 1076.9574,
362
+ "eval_samples_per_second": 1.739,
363
+ "eval_steps_per_second": 0.11,
364
+ "eval_wer": 28.898426323319025,
365
+ "step": 850
366
+ },
367
+ {
368
+ "epoch": 1.66,
369
+ "learning_rate": 4.351724137931035e-06,
370
+ "loss": 0.0927,
371
+ "step": 875
372
+ },
373
+ {
374
+ "epoch": 1.71,
375
+ "learning_rate": 4.179310344827587e-06,
376
+ "loss": 0.0566,
377
+ "step": 900
378
+ },
379
+ {
380
+ "epoch": 1.71,
381
+ "eval_loss": 0.06160915642976761,
382
+ "eval_runtime": 1078.2038,
383
+ "eval_samples_per_second": 1.737,
384
+ "eval_steps_per_second": 0.109,
385
+ "eval_wer": 29.127324749642348,
386
+ "step": 900
387
+ },
388
+ {
389
+ "epoch": 1.76,
390
+ "learning_rate": 4.006896551724138e-06,
391
+ "loss": 0.0466,
392
+ "step": 925
393
+ },
394
+ {
395
+ "epoch": 1.8,
396
+ "learning_rate": 3.83448275862069e-06,
397
+ "loss": 0.0542,
398
+ "step": 950
399
+ },
400
+ {
401
+ "epoch": 1.8,
402
+ "eval_loss": 0.06113772094249725,
403
+ "eval_runtime": 1072.2754,
404
+ "eval_samples_per_second": 1.747,
405
+ "eval_steps_per_second": 0.11,
406
+ "eval_wer": 28.841201716738198,
407
+ "step": 950
408
+ },
409
+ {
410
+ "epoch": 1.85,
411
+ "learning_rate": 3.6620689655172415e-06,
412
+ "loss": 0.0536,
413
+ "step": 975
414
+ },
415
+ {
416
+ "epoch": 1.9,
417
+ "learning_rate": 3.489655172413793e-06,
418
+ "loss": 0.0638,
419
+ "step": 1000
420
+ },
421
+ {
422
+ "epoch": 1.9,
423
+ "eval_loss": 0.06121109798550606,
424
+ "eval_runtime": 1073.2525,
425
+ "eval_samples_per_second": 1.745,
426
+ "eval_steps_per_second": 0.11,
427
+ "eval_wer": 30.100143061516455,
428
+ "step": 1000
429
+ },
430
+ {
431
+ "epoch": 1.94,
432
+ "learning_rate": 3.3172413793103453e-06,
433
+ "loss": 0.0559,
434
+ "step": 1025
435
+ },
436
+ {
437
+ "epoch": 1.99,
438
+ "learning_rate": 3.1448275862068965e-06,
439
+ "loss": 0.0503,
440
+ "step": 1050
441
+ },
442
+ {
443
+ "epoch": 1.99,
444
+ "eval_loss": 0.06099317967891693,
445
+ "eval_runtime": 1074.3991,
446
+ "eval_samples_per_second": 1.743,
447
+ "eval_steps_per_second": 0.11,
448
+ "eval_wer": 29.32761087267525,
449
+ "step": 1050
450
+ },
451
+ {
452
+ "epoch": 2.04,
453
+ "learning_rate": 2.9724137931034486e-06,
454
+ "loss": 0.0627,
455
+ "step": 1075
456
+ },
457
+ {
458
+ "epoch": 2.09,
459
+ "learning_rate": 2.8000000000000003e-06,
460
+ "loss": 0.0369,
461
+ "step": 1100
462
+ },
463
+ {
464
+ "epoch": 2.09,
465
+ "eval_loss": 0.06123083084821701,
466
+ "eval_runtime": 1071.2265,
467
+ "eval_samples_per_second": 1.748,
468
+ "eval_steps_per_second": 0.11,
469
+ "eval_wer": 29.32761087267525,
470
+ "step": 1100
471
+ },
472
+ {
473
+ "epoch": 2.13,
474
+ "learning_rate": 2.627586206896552e-06,
475
+ "loss": 0.0395,
476
+ "step": 1125
477
+ },
478
+ {
479
+ "epoch": 2.18,
480
+ "learning_rate": 2.4551724137931036e-06,
481
+ "loss": 0.0273,
482
+ "step": 1150
483
+ },
484
+ {
485
+ "epoch": 2.18,
486
+ "eval_loss": 0.06379574537277222,
487
+ "eval_runtime": 1064.3917,
488
+ "eval_samples_per_second": 1.76,
489
+ "eval_steps_per_second": 0.111,
490
+ "eval_wer": 32.58941344778255,
491
+ "step": 1150
492
+ },
493
+ {
494
+ "epoch": 2.23,
495
+ "learning_rate": 2.2827586206896553e-06,
496
+ "loss": 0.0345,
497
+ "step": 1175
498
+ },
499
+ {
500
+ "epoch": 2.28,
501
+ "learning_rate": 2.110344827586207e-06,
502
+ "loss": 0.0269,
503
+ "step": 1200
504
+ },
505
+ {
506
+ "epoch": 2.28,
507
+ "eval_loss": 0.06369271874427795,
508
+ "eval_runtime": 1147.8053,
509
+ "eval_samples_per_second": 1.632,
510
+ "eval_steps_per_second": 0.103,
511
+ "eval_wer": 31.673819742489272,
512
+ "step": 1200
513
+ },
514
+ {
515
+ "epoch": 2.32,
516
+ "learning_rate": 1.9379310344827586e-06,
517
+ "loss": 0.0446,
518
+ "step": 1225
519
+ },
520
+ {
521
+ "epoch": 2.37,
522
+ "learning_rate": 1.7655172413793103e-06,
523
+ "loss": 0.0384,
524
+ "step": 1250
525
+ },
526
+ {
527
+ "epoch": 2.37,
528
+ "eval_loss": 0.06297700107097626,
529
+ "eval_runtime": 1155.6941,
530
+ "eval_samples_per_second": 1.621,
531
+ "eval_steps_per_second": 0.102,
532
+ "eval_wer": 32.04577968526466,
533
+ "step": 1250
534
+ },
535
+ {
536
+ "epoch": 2.42,
537
+ "learning_rate": 1.5931034482758622e-06,
538
+ "loss": 0.0254,
539
+ "step": 1275
540
+ },
541
+ {
542
+ "epoch": 2.47,
543
+ "learning_rate": 1.4206896551724138e-06,
544
+ "loss": 0.0574,
545
+ "step": 1300
546
+ },
547
+ {
548
+ "epoch": 2.47,
549
+ "eval_loss": 0.062077928334474564,
550
+ "eval_runtime": 1073.2284,
551
+ "eval_samples_per_second": 1.745,
552
+ "eval_steps_per_second": 0.11,
553
+ "eval_wer": 30.014306151645208,
554
+ "step": 1300
555
+ },
556
+ {
557
+ "epoch": 2.51,
558
+ "learning_rate": 1.2482758620689655e-06,
559
+ "loss": 0.0447,
560
+ "step": 1325
561
+ },
562
+ {
563
+ "epoch": 2.56,
564
+ "learning_rate": 1.0758620689655174e-06,
565
+ "loss": 0.053,
566
+ "step": 1350
567
+ },
568
+ {
569
+ "epoch": 2.56,
570
+ "eval_loss": 0.06072333827614784,
571
+ "eval_runtime": 1068.6925,
572
+ "eval_samples_per_second": 1.753,
573
+ "eval_steps_per_second": 0.11,
574
+ "eval_wer": 30.529327610872674,
575
+ "step": 1350
576
+ },
577
+ {
578
+ "epoch": 2.61,
579
+ "learning_rate": 9.034482758620689e-07,
580
+ "loss": 0.0303,
581
+ "step": 1375
582
+ },
583
+ {
584
+ "epoch": 2.66,
585
+ "learning_rate": 7.310344827586207e-07,
586
+ "loss": 0.0217,
587
+ "step": 1400
588
+ },
589
+ {
590
+ "epoch": 2.66,
591
+ "eval_loss": 0.060529597103595734,
592
+ "eval_runtime": 1230.1339,
593
+ "eval_samples_per_second": 1.523,
594
+ "eval_steps_per_second": 0.096,
595
+ "eval_wer": 31.359084406294706,
596
+ "step": 1400
597
+ },
598
+ {
599
+ "epoch": 2.7,
600
+ "learning_rate": 5.586206896551725e-07,
601
+ "loss": 0.0394,
602
+ "step": 1425
603
+ },
604
+ {
605
+ "epoch": 2.75,
606
+ "learning_rate": 3.862068965517242e-07,
607
+ "loss": 0.0327,
608
+ "step": 1450
609
+ },
610
+ {
611
+ "epoch": 2.75,
612
+ "eval_loss": 0.060149554163217545,
613
+ "eval_runtime": 1225.2391,
614
+ "eval_samples_per_second": 1.529,
615
+ "eval_steps_per_second": 0.096,
616
+ "eval_wer": 31.301859799713878,
617
+ "step": 1450
618
+ },
619
+ {
620
+ "epoch": 2.8,
621
+ "learning_rate": 2.1379310344827587e-07,
622
+ "loss": 0.0326,
623
+ "step": 1475
624
+ },
625
+ {
626
+ "epoch": 2.85,
627
+ "learning_rate": 4.137931034482759e-08,
628
+ "loss": 0.0235,
629
+ "step": 1500
630
+ },
631
+ {
632
+ "epoch": 2.85,
633
+ "eval_loss": 0.05990791320800781,
634
+ "eval_runtime": 1212.8468,
635
+ "eval_samples_per_second": 1.544,
636
+ "eval_steps_per_second": 0.097,
637
+ "eval_wer": 30.90128755364807,
638
+ "step": 1500
639
+ }
640
+ ],
641
+ "max_steps": 1500,
642
+ "num_train_epochs": 3,
643
+ "total_flos": 1.018527602098176e+20,
644
+ "trial_name": null,
645
+ "trial_params": null
646
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcbfd556926ab64fec633f1ae6aa2f9095356fe4d7ab23ecb43429525c7a8595
3
+ size 3771