anton-l HF staff commited on
Commit
223df24
1 Parent(s): f69d28b

Training in progress, step 1000

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,315 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/wav2vec2-xls-r-300m",
3
+ "activation_dropout": 0.0,
4
+ "adapter_kernel_size": 3,
5
+ "adapter_stride": 2,
6
+ "add_adapter": false,
7
+ "apply_spec_augment": true,
8
+ "architectures": [
9
+ "Wav2Vec2ForSequenceClassification"
10
+ ],
11
+ "attention_dropout": 0.0,
12
+ "bos_token_id": 1,
13
+ "classifier_proj_size": 256,
14
+ "codevector_dim": 768,
15
+ "contrastive_logits_temperature": 0.1,
16
+ "conv_bias": true,
17
+ "conv_dim": [
18
+ 512,
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512
25
+ ],
26
+ "conv_kernel": [
27
+ 10,
28
+ 3,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 2,
33
+ 2
34
+ ],
35
+ "conv_stride": [
36
+ 5,
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2
43
+ ],
44
+ "ctc_loss_reduction": "mean",
45
+ "ctc_zero_infinity": false,
46
+ "diversity_loss_weight": 0.1,
47
+ "do_stable_layer_norm": true,
48
+ "eos_token_id": 2,
49
+ "feat_extract_activation": "gelu",
50
+ "feat_extract_dropout": 0.0,
51
+ "feat_extract_norm": "layer",
52
+ "feat_proj_dropout": 0.0,
53
+ "feat_quantizer_dropout": 0.0,
54
+ "final_dropout": 0.0,
55
+ "hidden_act": "gelu",
56
+ "hidden_dropout": 0.0,
57
+ "hidden_size": 1024,
58
+ "id2label": {
59
+ "0": "af_za",
60
+ "1": "am_et",
61
+ "2": "ar_eg",
62
+ "3": "as_in",
63
+ "4": "ast_es",
64
+ "5": "az_az",
65
+ "6": "be_by",
66
+ "7": "bn_in",
67
+ "8": "bs_ba",
68
+ "9": "ca_es",
69
+ "10": "ceb_ph",
70
+ "11": "cmn_hans_cn",
71
+ "12": "cs_cz",
72
+ "13": "cy_gb",
73
+ "14": "da_dk",
74
+ "15": "de_de",
75
+ "16": "el_gr",
76
+ "17": "en_us",
77
+ "18": "es_419",
78
+ "19": "et_ee",
79
+ "20": "fa_ir",
80
+ "21": "ff_sn",
81
+ "22": "fi_fi",
82
+ "23": "fil_ph",
83
+ "24": "fr_fr",
84
+ "25": "ga_ie",
85
+ "26": "gl_es",
86
+ "27": "gu_in",
87
+ "28": "ha_ng",
88
+ "29": "he_il",
89
+ "30": "hi_in",
90
+ "31": "hr_hr",
91
+ "32": "hu_hu",
92
+ "33": "hy_am",
93
+ "34": "id_id",
94
+ "35": "ig_ng",
95
+ "36": "is_is",
96
+ "37": "it_it",
97
+ "38": "ja_jp",
98
+ "39": "jv_id",
99
+ "40": "ka_ge",
100
+ "41": "kam_ke",
101
+ "42": "kea_cv",
102
+ "43": "kk_kz",
103
+ "44": "km_kh",
104
+ "45": "kn_in",
105
+ "46": "ko_kr",
106
+ "47": "ku_arab_iq",
107
+ "48": "ky_kg",
108
+ "49": "lb_lu",
109
+ "50": "lg_ug",
110
+ "51": "ln_cd",
111
+ "52": "lo_la",
112
+ "53": "lt_lt",
113
+ "54": "luo_ke",
114
+ "55": "lv_lv",
115
+ "56": "mi_nz",
116
+ "57": "mk_mk",
117
+ "58": "ml_in",
118
+ "59": "mn_mn",
119
+ "60": "mr_in",
120
+ "61": "ms_my",
121
+ "62": "mt_mt",
122
+ "63": "my_mm",
123
+ "64": "nb_no",
124
+ "65": "ne_np",
125
+ "66": "nl_nl",
126
+ "67": "nso_za",
127
+ "68": "ny_mw",
128
+ "69": "oci_fr",
129
+ "70": "om_et",
130
+ "71": "or_in",
131
+ "72": "pa_in",
132
+ "73": "pl_pl",
133
+ "74": "ps_af",
134
+ "75": "pt_br",
135
+ "76": "ro_ro",
136
+ "77": "ru_ru",
137
+ "78": "rup_bg",
138
+ "79": "sd_arab_in",
139
+ "80": "sk_sk",
140
+ "81": "sl_si",
141
+ "82": "sn_zw",
142
+ "83": "so_so",
143
+ "84": "sr_rs",
144
+ "85": "sv_se",
145
+ "86": "sw_ke",
146
+ "87": "ta_in",
147
+ "88": "te_in",
148
+ "89": "tg_tj",
149
+ "90": "th_th",
150
+ "91": "tr_tr",
151
+ "92": "uk_ua",
152
+ "93": "umb_ao",
153
+ "94": "ur_pk",
154
+ "95": "uz_uz",
155
+ "96": "vi_vn",
156
+ "97": "wo_sn",
157
+ "98": "xh_za",
158
+ "99": "yo_ng",
159
+ "100": "yue_hant_hk",
160
+ "101": "zu_za"
161
+ },
162
+ "initializer_range": 0.02,
163
+ "intermediate_size": 4096,
164
+ "label2id": {
165
+ "af_za": 0,
166
+ "am_et": 1,
167
+ "ar_eg": 2,
168
+ "as_in": 3,
169
+ "ast_es": 4,
170
+ "az_az": 5,
171
+ "be_by": 6,
172
+ "bn_in": 7,
173
+ "bs_ba": 8,
174
+ "ca_es": 9,
175
+ "ceb_ph": 10,
176
+ "cmn_hans_cn": 11,
177
+ "cs_cz": 12,
178
+ "cy_gb": 13,
179
+ "da_dk": 14,
180
+ "de_de": 15,
181
+ "el_gr": 16,
182
+ "en_us": 17,
183
+ "es_419": 18,
184
+ "et_ee": 19,
185
+ "fa_ir": 20,
186
+ "ff_sn": 21,
187
+ "fi_fi": 22,
188
+ "fil_ph": 23,
189
+ "fr_fr": 24,
190
+ "ga_ie": 25,
191
+ "gl_es": 26,
192
+ "gu_in": 27,
193
+ "ha_ng": 28,
194
+ "he_il": 29,
195
+ "hi_in": 30,
196
+ "hr_hr": 31,
197
+ "hu_hu": 32,
198
+ "hy_am": 33,
199
+ "id_id": 34,
200
+ "ig_ng": 35,
201
+ "is_is": 36,
202
+ "it_it": 37,
203
+ "ja_jp": 38,
204
+ "jv_id": 39,
205
+ "ka_ge": 40,
206
+ "kam_ke": 41,
207
+ "kea_cv": 42,
208
+ "kk_kz": 43,
209
+ "km_kh": 44,
210
+ "kn_in": 45,
211
+ "ko_kr": 46,
212
+ "ku_arab_iq": 47,
213
+ "ky_kg": 48,
214
+ "lb_lu": 49,
215
+ "lg_ug": 50,
216
+ "ln_cd": 51,
217
+ "lo_la": 52,
218
+ "lt_lt": 53,
219
+ "luo_ke": 54,
220
+ "lv_lv": 55,
221
+ "mi_nz": 56,
222
+ "mk_mk": 57,
223
+ "ml_in": 58,
224
+ "mn_mn": 59,
225
+ "mr_in": 60,
226
+ "ms_my": 61,
227
+ "mt_mt": 62,
228
+ "my_mm": 63,
229
+ "nb_no": 64,
230
+ "ne_np": 65,
231
+ "nl_nl": 66,
232
+ "nso_za": 67,
233
+ "ny_mw": 68,
234
+ "oci_fr": 69,
235
+ "om_et": 70,
236
+ "or_in": 71,
237
+ "pa_in": 72,
238
+ "pl_pl": 73,
239
+ "ps_af": 74,
240
+ "pt_br": 75,
241
+ "ro_ro": 76,
242
+ "ru_ru": 77,
243
+ "rup_bg": 78,
244
+ "sd_arab_in": 79,
245
+ "sk_sk": 80,
246
+ "sl_si": 81,
247
+ "sn_zw": 82,
248
+ "so_so": 83,
249
+ "sr_rs": 84,
250
+ "sv_se": 85,
251
+ "sw_ke": 86,
252
+ "ta_in": 87,
253
+ "te_in": 88,
254
+ "tg_tj": 89,
255
+ "th_th": 90,
256
+ "tr_tr": 91,
257
+ "uk_ua": 92,
258
+ "umb_ao": 93,
259
+ "ur_pk": 94,
260
+ "uz_uz": 95,
261
+ "vi_vn": 96,
262
+ "wo_sn": 97,
263
+ "xh_za": 98,
264
+ "yo_ng": 99,
265
+ "yue_hant_hk": 100,
266
+ "zu_za": 101
267
+ },
268
+ "layer_norm_eps": 1e-05,
269
+ "layerdrop": 0.0,
270
+ "mask_feature_length": 10,
271
+ "mask_feature_min_masks": 0,
272
+ "mask_feature_prob": 0.0,
273
+ "mask_time_length": 10,
274
+ "mask_time_min_masks": 2,
275
+ "mask_time_prob": 0.05,
276
+ "model_type": "wav2vec2",
277
+ "num_adapter_layers": 3,
278
+ "num_attention_heads": 16,
279
+ "num_codevector_groups": 2,
280
+ "num_codevectors_per_group": 320,
281
+ "num_conv_pos_embedding_groups": 16,
282
+ "num_conv_pos_embeddings": 128,
283
+ "num_feat_extract_layers": 7,
284
+ "num_hidden_layers": 24,
285
+ "num_negatives": 100,
286
+ "output_hidden_size": 1024,
287
+ "pad_token_id": 0,
288
+ "proj_codevector_dim": 768,
289
+ "tdnn_dilation": [
290
+ 1,
291
+ 2,
292
+ 3,
293
+ 1,
294
+ 1
295
+ ],
296
+ "tdnn_dim": [
297
+ 512,
298
+ 512,
299
+ 512,
300
+ 512,
301
+ 1500
302
+ ],
303
+ "tdnn_kernel": [
304
+ 5,
305
+ 3,
306
+ 3,
307
+ 1,
308
+ 1
309
+ ],
310
+ "torch_dtype": "float32",
311
+ "transformers_version": "4.18.0.dev0",
312
+ "use_weighted_layer_sum": false,
313
+ "vocab_size": 32,
314
+ "xvector_output_dim": 512
315
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000
9
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7b18c06b982a1048021bf7456f542d17debb886aeb7d13e9c2b85816480760
3
+ size 1263081325
runs/Apr06_17-13-08_anton-xtreme-s/1649265428.053142/events.out.tfevents.1649265428.anton-xtreme-s.4388.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee37d28776ab1f1b4658847cb528261387f1697fad2a1c2c17480dea8b6ca4bf
3
+ size 5092
runs/Apr06_17-13-08_anton-xtreme-s/events.out.tfevents.1649265428.anton-xtreme-s.4388.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:368295e34e505a86e2374336975b3ea086310e549152be72373e3247f2ef67ae
3
+ size 165622
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:066084be4b26a3c14127f115b3eddb8da86c7414bd73b4e6191272353caaed60
3
+ size 3247