Siddhant commited on
Commit
7220fc6
1 Parent(s): b53d5d1

import from zenodo

Browse files
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - espnet
4
+ - audio
5
+ - automatic-speech-recognition
6
+ language: en
7
+ datasets:
8
+ - timit
9
+ license: cc-by-4.0
10
+ ---
11
+ ## Example ESPnet2 ASR model
12
+ ### `kamo-naoyuki/timit_asr_train_asr_raw_word_valid.acc.ave`
13
+ ♻️ Imported from https://zenodo.org/record/4284058/
14
+
15
+ This model was trained by kamo-naoyuki using timit/asr1 recipe in [espnet](https://github.com/espnet/espnet/).
16
+ ### Demo: How to use in ESPnet2
17
+ ```python
18
+ # coming soon
19
+ ```
20
+ ### Citing ESPnet
21
+ ```BibTex
22
+ @inproceedings{watanabe2018espnet,
23
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson {Enrique Yalta Soplin} and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
24
+ title={{ESPnet}: End-to-End Speech Processing Toolkit},
25
+ year={2018},
26
+ booktitle={Proceedings of Interspeech},
27
+ pages={2207--2211},
28
+ doi={10.21437/Interspeech.2018-1456},
29
+ url={http://dx.doi.org/10.21437/Interspeech.2018-1456}
30
+ }
31
+ @inproceedings{hayashi2020espnet,
32
+ title={{Espnet-TTS}: Unified, reproducible, and integratable open source end-to-end text-to-speech toolkit},
33
+ author={Hayashi, Tomoki and Yamamoto, Ryuichi and Inoue, Katsuki and Yoshimura, Takenori and Watanabe, Shinji and Toda, Tomoki and Takeda, Kazuya and Zhang, Yu and Tan, Xu},
34
+ booktitle={Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)},
35
+ pages={7654--7658},
36
+ year={2020},
37
+ organization={IEEE}
38
+ }
39
+ ```
40
+ or arXiv:
41
+ ```bibtex
42
+ @misc{watanabe2018espnet,
43
+ title={ESPnet: End-to-End Speech Processing Toolkit},
44
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Enrique Yalta Soplin and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
45
+ year={2018},
46
+ eprint={1804.00015},
47
+ archivePrefix={arXiv},
48
+ primaryClass={cs.CL}
49
+ }
50
+ ```
exp/asr_stats_raw_word/train/feats_stats.npz ADDED
Binary file (1.4 kB). View file
 
exp/asr_train_asr_raw_word/RESULTS.md ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- Generated by scripts/utils/show_asr_result.sh -->
2
+ # RESULTS
3
+ ## Environments
4
+ - date: `Sun Nov 22 05:40:05 JST 2020`
5
+ - python version: `3.8.5 (default, Aug 5 2020, 08:36:46) [GCC 7.3.0]`
6
+ - espnet version: `espnet 0.9.5`
7
+ - pytorch version: `pytorch 1.6.0`
8
+ - Git hash: `f157fcd651df34a714ad2bd5d97e46632e010096`
9
+ - Commit date: `Sun Nov 22 04:29:17 2020 +0900`
10
+
11
+ ## asr_train_asr_raw_word
12
+ ### WER
13
+
14
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
15
+ |---|---|---|---|---|---|---|---|---|
16
+ |decode_asr_asr_model_valid.acc.ave/dev|400|15057|84.4|11.2|4.4|2.3|17.9|99.0|
17
+ |decode_asr_asr_model_valid.acc.ave/test|192|7215|83.1|12.4|4.5|2.6|19.5|99.0|
18
+ |decode_asr_lm_lm_train_lm_rnn_word_valid.loss.ave_asr_model_valid.acc.ave/dev|400|15057|83.0|12.2|4.7|2.8|19.8|99.8|
19
+ |decode_asr_lm_lm_train_lm_rnn_word_valid.loss.ave_asr_model_valid.acc.ave/test|192|7215|81.6|13.5|4.9|3.3|21.7|100.0|
20
+
21
+ ### CER
22
+
23
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
24
+ |---|---|---|---|---|---|---|---|---|
25
+ |decode_asr_asr_model_valid.acc.ave/dev|400|41742|91.0|4.3|4.8|2.8|11.9|99.0|
26
+ |decode_asr_asr_model_valid.acc.ave/test|192|20045|90.4|4.7|4.9|3.2|12.8|99.0|
27
+ |decode_asr_lm_lm_train_lm_rnn_word_valid.loss.ave_asr_model_valid.acc.ave/dev|400|41742|90.2|4.7|5.1|3.3|13.1|99.8|
28
+ |decode_asr_lm_lm_train_lm_rnn_word_valid.loss.ave_asr_model_valid.acc.ave/test|192|20045|89.6|5.3|5.2|3.9|14.4|100.0|
29
+
exp/asr_train_asr_raw_word/config.yaml ADDED
@@ -0,0 +1,192 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config: conf/train_asr.yaml
2
+ print_config: false
3
+ log_level: INFO
4
+ dry_run: false
5
+ iterator_type: sequence
6
+ output_dir: exp/asr_train_asr_raw_word
7
+ ngpu: 1
8
+ seed: 0
9
+ num_workers: 1
10
+ num_att_plot: 3
11
+ dist_backend: nccl
12
+ dist_init_method: env://
13
+ dist_world_size: null
14
+ dist_rank: null
15
+ local_rank: 0
16
+ dist_master_addr: null
17
+ dist_master_port: null
18
+ dist_launcher: null
19
+ multiprocessing_distributed: false
20
+ cudnn_enabled: true
21
+ cudnn_benchmark: false
22
+ cudnn_deterministic: true
23
+ collect_stats: false
24
+ write_collected_feats: false
25
+ max_epoch: 25
26
+ patience: 4
27
+ val_scheduler_criterion:
28
+ - valid
29
+ - loss
30
+ early_stopping_criterion:
31
+ - valid
32
+ - loss
33
+ - min
34
+ best_model_criterion:
35
+ - - valid
36
+ - acc
37
+ - max
38
+ keep_nbest_models: 3
39
+ grad_clip: 5.0
40
+ grad_clip_type: 2.0
41
+ grad_noise: false
42
+ accum_grad: 1
43
+ no_forward_run: false
44
+ resume: true
45
+ train_dtype: float32
46
+ use_amp: false
47
+ log_interval: null
48
+ unused_parameters: false
49
+ use_tensorboard: true
50
+ use_wandb: false
51
+ pretrain_path: null
52
+ init_param: []
53
+ num_iters_per_epoch: null
54
+ batch_size: 30
55
+ valid_batch_size: null
56
+ batch_bins: 1000000
57
+ valid_batch_bins: null
58
+ train_shape_file:
59
+ - exp/asr_stats_raw_word/train/speech_shape
60
+ - exp/asr_stats_raw_word/train/text_shape.word
61
+ valid_shape_file:
62
+ - exp/asr_stats_raw_word/valid/speech_shape
63
+ - exp/asr_stats_raw_word/valid/text_shape.word
64
+ batch_type: folded
65
+ valid_batch_type: null
66
+ fold_length:
67
+ - 80000
68
+ - 150
69
+ sort_in_batch: descending
70
+ sort_batch: descending
71
+ multiple_iterator: false
72
+ chunk_length: 500
73
+ chunk_shift_ratio: 0.5
74
+ num_cache_chunks: 1024
75
+ train_data_path_and_name_and_type:
76
+ - - dump/raw/train/wav.scp
77
+ - speech
78
+ - sound
79
+ - - dump/raw/train/text
80
+ - text
81
+ - text
82
+ valid_data_path_and_name_and_type:
83
+ - - dump/raw/dev/wav.scp
84
+ - speech
85
+ - sound
86
+ - - dump/raw/dev/text
87
+ - text
88
+ - text
89
+ allow_variable_data_keys: false
90
+ max_cache_size: 0.0
91
+ max_cache_fd: 32
92
+ valid_max_cache_size: null
93
+ optim: adadelta
94
+ optim_conf:
95
+ lr: 1.0
96
+ rho: 0.95
97
+ eps: 1.0e-08
98
+ weight_decay: 0
99
+ scheduler: reducelronplateau
100
+ scheduler_conf:
101
+ mode: min
102
+ factor: 0.5
103
+ patience: 1
104
+ token_list:
105
+ - <blank>
106
+ - <unk>
107
+ - sil
108
+ - ih
109
+ - n
110
+ - s
111
+ - ah
112
+ - l
113
+ - r
114
+ - iy
115
+ - er
116
+ - aa
117
+ - t
118
+ - k
119
+ - z
120
+ - m
121
+ - eh
122
+ - p
123
+ - d
124
+ - dh
125
+ - ae
126
+ - ey
127
+ - w
128
+ - f
129
+ - b
130
+ - v
131
+ - uw
132
+ - ay
133
+ - dx
134
+ - hh
135
+ - ow
136
+ - sh
137
+ - ng
138
+ - g
139
+ - jh
140
+ - y
141
+ - ch
142
+ - th
143
+ - aw
144
+ - uh
145
+ - oy
146
+ - <sos/eos>
147
+ init: xavier_uniform
148
+ input_size: null
149
+ ctc_conf:
150
+ dropout_rate: 0.0
151
+ ctc_type: builtin
152
+ reduce: true
153
+ ignore_nan_grad: false
154
+ model_conf:
155
+ ctc_weight: 0.5
156
+ lsm_weight: 0.1
157
+ use_preprocessor: true
158
+ token_type: word
159
+ bpemodel: null
160
+ non_linguistic_symbols: null
161
+ cleaner: null
162
+ g2p: null
163
+ frontend: default
164
+ frontend_conf:
165
+ fs: 16k
166
+ specaug: null
167
+ specaug_conf: {}
168
+ normalize: global_mvn
169
+ normalize_conf:
170
+ stats_file: exp/asr_stats_raw_word/train/feats_stats.npz
171
+ encoder: rnn
172
+ encoder_conf:
173
+ rnn_type: gru
174
+ bidirectional: true
175
+ use_projection: true
176
+ num_layers: 4
177
+ hidden_size: 320
178
+ output_size: 320
179
+ dropout: 0.2
180
+ decoder: rnn
181
+ decoder_conf:
182
+ rnn_type: lstm
183
+ num_layers: 1
184
+ hidden_size: 320
185
+ dropout: 0.0
186
+ att_conf:
187
+ atype: location
188
+ adim: 320
189
+ required:
190
+ - output_dir
191
+ - token_list
192
+ distributed: false
exp/asr_train_asr_raw_word/images/acc.png ADDED
exp/asr_train_asr_raw_word/images/backward_time.png ADDED
exp/asr_train_asr_raw_word/images/cer.png ADDED
exp/asr_train_asr_raw_word/images/cer_ctc.png ADDED
exp/asr_train_asr_raw_word/images/forward_time.png ADDED
exp/asr_train_asr_raw_word/images/iter_time.png ADDED
exp/asr_train_asr_raw_word/images/loss.png ADDED
exp/asr_train_asr_raw_word/images/loss_att.png ADDED
exp/asr_train_asr_raw_word/images/loss_ctc.png ADDED
exp/asr_train_asr_raw_word/images/lr_0.png ADDED
exp/asr_train_asr_raw_word/images/optim_step_time.png ADDED
exp/asr_train_asr_raw_word/images/train_time.png ADDED
exp/asr_train_asr_raw_word/images/wer.png ADDED
exp/asr_train_asr_raw_word/valid.acc.ave_3best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e7240348c40188a88e9171abc4a4420f0e695e12ddffaff134341020f352213
3
+ size 27194178
meta.yaml ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ espnet: 0.9.5
2
+ files:
3
+ asr_model_file: exp/asr_train_asr_raw_word/valid.acc.ave_3best.pth
4
+ python: "3.8.5 (default, Aug 5 2020, 08:36:46) \n[GCC 7.3.0]"
5
+ timestamp: 1605991206.359051
6
+ torch: 1.6.0
7
+ yaml_files:
8
+ asr_train_config: exp/asr_train_asr_raw_word/config.yaml