Dan Berrebbi commited on
Commit
28f0d7c
1 Parent(s): 7083c60

unziped model

Browse files
data/token_list/bpe_unigram250/bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbd9525abedcbd8ec55ee50fd4a56bc9c01396c8d045038e26271ecb9e88ec00
3
+ size 241098
exp/asr_transformer_baseline/RESULTS.md ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- Generated by scripts/utils/show_asr_result.sh -->
2
+ # RESULTS
3
+ ## Environments
4
+ - date: `Sat Apr 16 14:14:45 EDT 2022`
5
+ - python version: `3.9.12 (main, Apr 5 2022, 06:56:58) [GCC 7.5.0]`
6
+ - espnet version: `espnet 0.10.6a1`
7
+ - pytorch version: `pytorch 1.11.0+cu102`
8
+ - Git hash: `f6cbc61353e0a1cefe81ae596278f7db1f0b7dd9`
9
+ - Commit date: `Fri Apr 15 18:31:26 2022 -0400`
10
+
11
+ ## asr_transformer_baseline
12
+ ### WER
13
+
14
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
15
+ |---|---|---|---|---|---|---|---|---|
16
+ |decode_asr_asr_model_valid.acc.ave_10best/devtest|481|3172|97.4|1.6|1.0|0.2|2.8|15.0|
17
+ |decode_asr_asr_model_valid.acc.ave_10best/test|515|2941|85.2|13.4|1.3|9.1|23.9|58.4|
18
+
19
+ ### CER
20
+
21
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
22
+ |---|---|---|---|---|---|---|---|---|
23
+ |decode_asr_asr_model_valid.acc.ave_10best/devtest|481|16205|98.7|0.2|1.1|0.2|1.5|15.0|
24
+ |decode_asr_asr_model_valid.acc.ave_10best/test|515|16233|95.8|2.0|2.2|2.1|6.3|58.4|
25
+
26
+ ### TER
27
+
28
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
29
+ |---|---|---|---|---|---|---|---|---|
30
+ |decode_asr_asr_model_valid.acc.ave_10best/devtest|481|7555|98.1|0.6|1.4|0.3|2.2|15.0|
31
+ |decode_asr_asr_model_valid.acc.ave_10best/test|515|7998|88.9|6.7|4.5|1.3|12.4|58.4|
32
+
exp/asr_transformer_baseline/config.yaml ADDED
@@ -0,0 +1,428 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config: conf/tuning/train_asr_transformer.yaml
2
+ print_config: false
3
+ log_level: INFO
4
+ dry_run: false
5
+ iterator_type: sequence
6
+ output_dir: exp/asr_transformer_baseline
7
+ ngpu: 1
8
+ seed: 0
9
+ num_workers: 1
10
+ num_att_plot: 3
11
+ dist_backend: nccl
12
+ dist_init_method: env://
13
+ dist_world_size: null
14
+ dist_rank: null
15
+ local_rank: 0
16
+ dist_master_addr: null
17
+ dist_master_port: null
18
+ dist_launcher: null
19
+ multiprocessing_distributed: false
20
+ unused_parameters: false
21
+ sharded_ddp: false
22
+ cudnn_enabled: true
23
+ cudnn_benchmark: false
24
+ cudnn_deterministic: true
25
+ collect_stats: false
26
+ write_collected_feats: false
27
+ max_epoch: 100
28
+ patience: 15
29
+ val_scheduler_criterion:
30
+ - valid
31
+ - loss
32
+ early_stopping_criterion:
33
+ - valid
34
+ - loss
35
+ - min
36
+ best_model_criterion:
37
+ - - valid
38
+ - acc
39
+ - max
40
+ keep_nbest_models: 10
41
+ nbest_averaging_interval: 0
42
+ grad_clip: 5
43
+ grad_clip_type: 2.0
44
+ grad_noise: false
45
+ accum_grad: 4
46
+ no_forward_run: false
47
+ resume: true
48
+ train_dtype: float32
49
+ use_amp: false
50
+ log_interval: null
51
+ use_matplotlib: true
52
+ use_tensorboard: true
53
+ use_wandb: false
54
+ wandb_project: null
55
+ wandb_id: null
56
+ wandb_entity: null
57
+ wandb_name: null
58
+ wandb_model_log_interval: -1
59
+ detect_anomaly: false
60
+ pretrain_path: null
61
+ init_param: []
62
+ ignore_init_mismatch: false
63
+ freeze_param: []
64
+ num_iters_per_epoch: null
65
+ batch_size: 16
66
+ valid_batch_size: null
67
+ batch_bins: 1000000
68
+ valid_batch_bins: null
69
+ train_shape_file:
70
+ - exp/asr_stats_raw_bpe250/train/speech_shape
71
+ - exp/asr_stats_raw_bpe250/train/text_shape.bpe
72
+ valid_shape_file:
73
+ - exp/asr_stats_raw_bpe250/valid/speech_shape
74
+ - exp/asr_stats_raw_bpe250/valid/text_shape.bpe
75
+ batch_type: folded
76
+ valid_batch_type: null
77
+ fold_length:
78
+ - 80000
79
+ - 150
80
+ sort_in_batch: descending
81
+ sort_batch: descending
82
+ multiple_iterator: false
83
+ chunk_length: 500
84
+ chunk_shift_ratio: 0.5
85
+ num_cache_chunks: 1024
86
+ train_data_path_and_name_and_type:
87
+ - - dump/raw/train/wav.scp
88
+ - speech
89
+ - sound
90
+ - - dump/raw/train/text
91
+ - text
92
+ - text
93
+ valid_data_path_and_name_and_type:
94
+ - - dump/raw/dev/wav.scp
95
+ - speech
96
+ - sound
97
+ - - dump/raw/dev/text
98
+ - text
99
+ - text
100
+ allow_variable_data_keys: false
101
+ max_cache_size: 0.0
102
+ max_cache_fd: 32
103
+ valid_max_cache_size: null
104
+ optim: adam
105
+ optim_conf:
106
+ lr: 1
107
+ scheduler: noamlr
108
+ scheduler_conf:
109
+ warmup_steps: 4000
110
+ token_list:
111
+ - <blank>
112
+ - <unk>
113
+ - ▁
114
+ - s
115
+ - u
116
+ - t
117
+ - ''''
118
+ - i
119
+ - r
120
+ - e
121
+ - a
122
+ - ▁est
123
+ - o
124
+ - ▁de
125
+ - l
126
+ - ▁a
127
+ - c
128
+ - é
129
+ - '-'
130
+ - n
131
+ - ▁d
132
+ - re
133
+ - ▁l
134
+ - ▁la
135
+ - m
136
+ - ▁que
137
+ - ▁n
138
+ - ce
139
+ - ▁le
140
+ - d
141
+ - ▁c
142
+ - ▁il
143
+ - 'on'
144
+ - p
145
+ - à
146
+ - ▁qui
147
+ - it
148
+ - ▁f
149
+ - is
150
+ - te
151
+ - ▁qu
152
+ - ▁un
153
+ - in
154
+ - ▁pas
155
+ - ▁ne
156
+ - ▁vous
157
+ - er
158
+ - ▁les
159
+ - ▁et
160
+ - en
161
+ - ▁ma
162
+ - ▁se
163
+ - ▁en
164
+ - ▁on
165
+ - f
166
+ - ent
167
+ - b
168
+ - ▁p
169
+ - ▁t
170
+ - ra
171
+ - ▁b
172
+ - ▁vo
173
+ - che
174
+ - ez
175
+ - ro
176
+ - le
177
+ - eur
178
+ - ne
179
+ - ▁m
180
+ - il
181
+ - or
182
+ - ▁vi
183
+ - vous
184
+ - ▁sa
185
+ - tre
186
+ - es
187
+ - ▁bien
188
+ - ie
189
+ - ▁ou
190
+ - ▁au
191
+ - ▁par
192
+ - ▁pa
193
+ - ▁h
194
+ - ir
195
+ - ▁bon
196
+ - ille
197
+ - me
198
+ - ▁ce
199
+ - ▁y
200
+ - ▁fait
201
+ - ▁des
202
+ - eau
203
+ - ▁avez
204
+ - and
205
+ - ur
206
+ - ant
207
+ - ▁du
208
+ - ▁mo
209
+ - h
210
+ - ▁co
211
+ - ▁plus
212
+ - ▁pour
213
+ - ▁une
214
+ - ▁je
215
+ - ▁faut
216
+ - ier
217
+ - sse
218
+ - ▁é
219
+ - eux
220
+ - nt
221
+ - ▁re
222
+ - ▁cha
223
+ - ▁sont
224
+ - que
225
+ - age
226
+ - ▁tout
227
+ - de
228
+ - y
229
+ - ▁son
230
+ - ▁tou
231
+ - â
232
+ - elle
233
+ - ée
234
+ - ▁dans
235
+ - ▁personne
236
+ - ▁va
237
+ - ▁pr
238
+ - ▁dé
239
+ - ▁con
240
+ - ▁ave
241
+ - ▁si
242
+ - aux
243
+ - ▁mais
244
+ - ▁me
245
+ - ▁peut
246
+ - ▁po
247
+ - nge
248
+ - ▁ba
249
+ - ▁comme
250
+ - ter
251
+ - ▁jamais
252
+ - ine
253
+ - ▁ch
254
+ - ▁quelle
255
+ - ▁j
256
+ - ▁mieux
257
+ - ment
258
+ - ion
259
+ - ette
260
+ - ▁cett
261
+ - ▁faire
262
+ - ▁vaut
263
+ - aire
264
+ - z
265
+ - ▁sur
266
+ - ▁homme
267
+ - ▁soi
268
+ - ▁mon
269
+ - ▁rien
270
+ - ▁nous
271
+ - ▁autre
272
+ - ▁perd
273
+ - ▁bou
274
+ - ▁combien
275
+ - ▁parle
276
+ - ▁donne
277
+ - omp
278
+ - ▁deux
279
+ - oir
280
+ - ▁ici
281
+ - ▁peu
282
+ - ▁grand
283
+ - ▁sou
284
+ - jours
285
+ - ▁pro
286
+ - sans
287
+ - ▁petit
288
+ - ▁femme
289
+ - ard
290
+ - ▁bonne
291
+ - ix
292
+ - use
293
+ - q
294
+ - ▁ami
295
+ - ▁êtes
296
+ - ▁point
297
+ - ▁être
298
+ - ▁prend
299
+ - ▁enfant
300
+ - ▁cour
301
+ - ▁mauvais
302
+ - ▁médecin
303
+ - ement
304
+ - ô
305
+ - û
306
+ - ▁veut
307
+ - ▁trop
308
+ - ation
309
+ - able
310
+ - ▁euh
311
+ - ▁fou
312
+ - jou
313
+ - ▁temps
314
+ - ▁allez
315
+ - ▁app
316
+ - x
317
+ - ▁chien
318
+ - ▁ça
319
+ - ▁doit
320
+ - ▁aller
321
+ - avoir
322
+ - puis
323
+ - ▁plai
324
+ - j
325
+ - ▁dire
326
+ - ▁maître
327
+ - ance
328
+ - éri
329
+ - ▁cheval
330
+ - ▁mort
331
+ - ▁monsieur
332
+ - ▁sui
333
+ - ▁fois
334
+ - ▁porte
335
+ - ▁alors
336
+ - ▁quelqu
337
+ - ▁couleur
338
+ - ▁arrive
339
+ - ▁besoin
340
+ - ▁chose
341
+ - ▁souvent
342
+ - ▁rend
343
+ - ▁plaît
344
+ - ▁bonjour
345
+ - ç
346
+ - ï
347
+ - /
348
+ - w
349
+ - œ
350
+ - k
351
+ - ù
352
+ - î
353
+ - ê
354
+ - è
355
+ - g
356
+ - F
357
+ - P
358
+ - A
359
+ - v
360
+ - <sos/eos>
361
+ init: xavier_uniform
362
+ input_size: null
363
+ ctc_conf:
364
+ dropout_rate: 0.0
365
+ ctc_type: builtin
366
+ reduce: true
367
+ ignore_nan_grad: true
368
+ joint_net_conf: null
369
+ model_conf:
370
+ ctc_weight: 0.3
371
+ lsm_weight: 0.1
372
+ length_normalized_loss: false
373
+ use_preprocessor: true
374
+ token_type: bpe
375
+ bpemodel: data/token_list/bpe_unigram250/bpe.model
376
+ non_linguistic_symbols: null
377
+ cleaner: null
378
+ g2p: null
379
+ speech_volume_normalize: null
380
+ rir_scp: null
381
+ rir_apply_prob: 1.0
382
+ noise_scp: null
383
+ noise_apply_prob: 1.0
384
+ noise_db_range: '13_15'
385
+ frontend: default
386
+ frontend_conf:
387
+ fs: 16k
388
+ specaug: specaug
389
+ specaug_conf:
390
+ apply_time_warp: true
391
+ time_warp_window: 5
392
+ time_warp_mode: bicubic
393
+ apply_freq_mask: true
394
+ freq_mask_width_range:
395
+ - 0
396
+ - 30
397
+ num_freq_mask: 2
398
+ apply_time_mask: true
399
+ time_mask_width_range:
400
+ - 0
401
+ - 40
402
+ num_time_mask: 2
403
+ normalize: utterance_mvn
404
+ normalize_conf: {}
405
+ preencoder: null
406
+ preencoder_conf: {}
407
+ encoder: transformer
408
+ encoder_conf:
409
+ input_layer: conv2d2
410
+ num_blocks: 12
411
+ linear_units: 2048
412
+ dropout_rate: 0.1
413
+ output_size: 256
414
+ attention_heads: 4
415
+ attention_dropout_rate: 0.0
416
+ postencoder: null
417
+ postencoder_conf: {}
418
+ decoder: transformer
419
+ decoder_conf:
420
+ input_layer: embed
421
+ num_blocks: 6
422
+ linear_units: 2048
423
+ dropout_rate: 0.1
424
+ required:
425
+ - output_dir
426
+ - token_list
427
+ version: 0.10.6a1
428
+ distributed: false
exp/asr_transformer_baseline/images/acc.png ADDED
exp/asr_transformer_baseline/images/backward_time.png ADDED
exp/asr_transformer_baseline/images/cer.png ADDED
exp/asr_transformer_baseline/images/cer_ctc.png ADDED
exp/asr_transformer_baseline/images/forward_time.png ADDED
exp/asr_transformer_baseline/images/gpu_max_cached_mem_GB.png ADDED
exp/asr_transformer_baseline/images/iter_time.png ADDED
exp/asr_transformer_baseline/images/loss.png ADDED
exp/asr_transformer_baseline/images/loss_att.png ADDED
exp/asr_transformer_baseline/images/loss_ctc.png ADDED
exp/asr_transformer_baseline/images/optim0_lr0.png ADDED
exp/asr_transformer_baseline/images/optim_step_time.png ADDED
exp/asr_transformer_baseline/images/train_time.png ADDED
exp/asr_transformer_baseline/images/wer.png ADDED
exp/asr_transformer_baseline/valid.acc.ave_10best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:739cb3ace3269c2262846479cc6a548744ad09cc47b25479918208e932aa27c9
3
+ size 114073833
meta.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ espnet: 0.10.6a1
2
+ files:
3
+ asr_model_file: exp/asr_transformer_baseline/valid.acc.ave_10best.pth
4
+ python: "3.9.12 (main, Apr 5 2022, 06:56:58) \n[GCC 7.5.0]"
5
+ timestamp: 1650133017.015989
6
+ torch: 1.11.0+cu102
7
+ yaml_files:
8
+ asr_train_config: exp/asr_transformer_baseline/config.yaml