siuze commited on
Commit
5ab9aa9
1 Parent(s): bf8c8de

Update model

Browse files
README.md ADDED
@@ -0,0 +1,315 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - espnet
4
+ - audio
5
+ - automatic-speech-recognition
6
+ language: can
7
+ datasets:
8
+ - mini_an4
9
+ license: cc-by-4.0
10
+ ---
11
+
12
+ ## ESPnet2 ASR model
13
+
14
+ ### `siuze/Cantonese-MDCC`
15
+
16
+ This model was trained by siuze using mini_an4 recipe in [espnet](https://github.com/espnet/espnet/).
17
+
18
+ ### Demo: How to use in ESPnet2
19
+
20
+ Follow the [ESPnet installation instructions](https://espnet.github.io/espnet/installation.html)
21
+ if you haven't done that already.
22
+
23
+ ```bash
24
+ cd espnet
25
+ git checkout 52160d6ed337e9dec74dd59695fec1548042e0b2
26
+ pip install -e .
27
+ cd egs2/mini_an4/asr1
28
+ ./run.sh --skip_data_prep false --skip_train true --download_model siuze/Cantonese-MDCC
29
+ ```
30
+
31
+ <!-- Generated by scripts/utils/show_asr_result.sh -->
32
+ # RESULTS
33
+ ## Environments
34
+ - date: `Fri Mar 17 23:08:24 CST 2023`
35
+ - python version: `3.8.16 | packaged by conda-forge | (default, Feb 1 2023, 16:01:55) [GCC 11.3.0]`
36
+ - espnet version: `espnet 202301`
37
+ - pytorch version: `pytorch 1.10.0`
38
+ - Git hash: `52160d6ed337e9dec74dd59695fec1548042e0b2`
39
+ - Commit date: `Thu Mar 16 21:37:39 2023 +0000`
40
+
41
+ ## exp/asr_train_asr_transformer_raw_can_char
42
+ ### WER
43
+
44
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
45
+ |---|---|---|---|---|---|---|---|---|
46
+ |inference_asr_model_valid.acc.ave/test|9077|108147|0.0|0.0|100.0|0.0|100.0|100.0|
47
+
48
+ ### CER
49
+
50
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
51
+ |---|---|---|---|---|---|---|---|---|
52
+ |inference_asr_model_valid.acc.ave/test|9077|666586|0.0|0.0|100.0|0.0|100.0|100.0|
53
+
54
+ ### TER
55
+
56
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
57
+ |---|---|---|---|---|---|---|---|---|
58
+
59
+ ## ASR config
60
+
61
+ <details><summary>expand</summary>
62
+
63
+ ```
64
+ config: conf/train_asr_transformer.yaml
65
+ print_config: false
66
+ log_level: INFO
67
+ dry_run: false
68
+ iterator_type: sequence
69
+ output_dir: exp/asr_train_asr_transformer_raw_can_char
70
+ ngpu: 1
71
+ seed: 0
72
+ num_workers: 1
73
+ num_att_plot: 3
74
+ dist_backend: nccl
75
+ dist_init_method: env://
76
+ dist_world_size: null
77
+ dist_rank: null
78
+ local_rank: 0
79
+ dist_master_addr: null
80
+ dist_master_port: null
81
+ dist_launcher: null
82
+ multiprocessing_distributed: false
83
+ unused_parameters: false
84
+ sharded_ddp: false
85
+ cudnn_enabled: true
86
+ cudnn_benchmark: false
87
+ cudnn_deterministic: true
88
+ collect_stats: false
89
+ write_collected_feats: false
90
+ max_epoch: 30
91
+ patience: null
92
+ val_scheduler_criterion:
93
+ - valid
94
+ - loss
95
+ early_stopping_criterion:
96
+ - valid
97
+ - loss
98
+ - min
99
+ best_model_criterion:
100
+ - - valid
101
+ - acc
102
+ - max
103
+ keep_nbest_models: 10
104
+ nbest_averaging_interval: 0
105
+ grad_clip: 5.0
106
+ grad_clip_type: 2.0
107
+ grad_noise: false
108
+ accum_grad: 8
109
+ no_forward_run: false
110
+ resume: true
111
+ train_dtype: float32
112
+ use_amp: false
113
+ log_interval: null
114
+ use_matplotlib: true
115
+ use_tensorboard: true
116
+ create_graph_in_tensorboard: false
117
+ use_wandb: false
118
+ wandb_project: null
119
+ wandb_id: null
120
+ wandb_entity: null
121
+ wandb_name: null
122
+ wandb_model_log_interval: -1
123
+ detect_anomaly: false
124
+ pretrain_path: null
125
+ init_param: []
126
+ ignore_init_mismatch: false
127
+ freeze_param: []
128
+ num_iters_per_epoch: null
129
+ batch_size: 16
130
+ valid_batch_size: null
131
+ batch_bins: 1000000
132
+ valid_batch_bins: null
133
+ train_shape_file:
134
+ - exp/asr_stats_raw_can_char/train/speech_shape
135
+ - exp/asr_stats_raw_can_char/train/text_shape.char
136
+ valid_shape_file:
137
+ - exp/asr_stats_raw_can_char/valid/speech_shape
138
+ - exp/asr_stats_raw_can_char/valid/text_shape.char
139
+ batch_type: folded
140
+ valid_batch_type: null
141
+ fold_length:
142
+ - 80000
143
+ - 150
144
+ sort_in_batch: descending
145
+ sort_batch: descending
146
+ multiple_iterator: false
147
+ chunk_length: 500
148
+ chunk_shift_ratio: 0.5
149
+ num_cache_chunks: 1024
150
+ chunk_excluded_key_prefixes: []
151
+ train_data_path_and_name_and_type:
152
+ - - dump/raw/train/wav.scp
153
+ - speech
154
+ - sound
155
+ - - dump/raw/train/text
156
+ - text
157
+ - text
158
+ valid_data_path_and_name_and_type:
159
+ - - dump/raw/dev/wav.scp
160
+ - speech
161
+ - sound
162
+ - - dump/raw/dev/text
163
+ - text
164
+ - text
165
+ allow_variable_data_keys: false
166
+ max_cache_size: 0.0
167
+ max_cache_fd: 32
168
+ valid_max_cache_size: null
169
+ exclude_weight_decay: false
170
+ exclude_weight_decay_conf: {}
171
+ optim: adam
172
+ optim_conf:
173
+ lr: 0.005
174
+ scheduler: warmuplr
175
+ scheduler_conf:
176
+ warmup_steps: 30000
177
+ token_list:
178
+ - <blank>
179
+ - <unk>
180
+ - <space>
181
+ - '3'
182
+ - '2'
183
+ - '5'
184
+ - g
185
+ - o
186
+ - a
187
+ - n
188
+ - i
189
+ - '4'
190
+ - u
191
+ - e
192
+ - k
193
+ - '1'
194
+ - j
195
+ - y
196
+ - z
197
+ - s
198
+ - h
199
+ - d
200
+ - m
201
+ - l
202
+ - c
203
+ - b
204
+ - f
205
+ - t
206
+ - w
207
+ - p
208
+ - r
209
+ - x
210
+ - v
211
+ - q
212
+ - <sos/eos>
213
+ init: xavier_uniform
214
+ input_size: null
215
+ ctc_conf:
216
+ dropout_rate: 0.0
217
+ ctc_type: builtin
218
+ reduce: true
219
+ ignore_nan_grad: null
220
+ zero_infinity: true
221
+ joint_net_conf: null
222
+ use_preprocessor: true
223
+ token_type: char
224
+ bpemodel: null
225
+ non_linguistic_symbols: null
226
+ cleaner: null
227
+ g2p: null
228
+ speech_volume_normalize: null
229
+ rir_scp: null
230
+ rir_apply_prob: 1.0
231
+ noise_scp: null
232
+ noise_apply_prob: 1.0
233
+ noise_db_range: '13_15'
234
+ short_noise_thres: 0.5
235
+ aux_ctc_tasks: []
236
+ frontend: default
237
+ frontend_conf:
238
+ fs: 16k
239
+ specaug: null
240
+ specaug_conf: {}
241
+ normalize: global_mvn
242
+ normalize_conf:
243
+ stats_file: exp/asr_stats_raw_can_char/train/feats_stats.npz
244
+ model: espnet
245
+ model_conf:
246
+ ctc_weight: 0.3
247
+ lsm_weight: 0.1
248
+ length_normalized_loss: false
249
+ preencoder: null
250
+ preencoder_conf: {}
251
+ encoder: transformer
252
+ encoder_conf:
253
+ output_size: 256
254
+ attention_heads: 4
255
+ linear_units: 2048
256
+ num_blocks: 12
257
+ dropout_rate: 0.1
258
+ positional_dropout_rate: 0.1
259
+ attention_dropout_rate: 0.0
260
+ input_layer: conv2d
261
+ normalize_before: true
262
+ postencoder: null
263
+ postencoder_conf: {}
264
+ decoder: transformer
265
+ decoder_conf:
266
+ attention_heads: 4
267
+ linear_units: 2048
268
+ num_blocks: 6
269
+ dropout_rate: 0.1
270
+ positional_dropout_rate: 0.1
271
+ self_attention_dropout_rate: 0.0
272
+ src_attention_dropout_rate: 0.0
273
+ preprocessor: default
274
+ preprocessor_conf: {}
275
+ required:
276
+ - output_dir
277
+ - token_list
278
+ version: '202301'
279
+ distributed: false
280
+ ```
281
+
282
+ </details>
283
+
284
+
285
+
286
+ ### Citing ESPnet
287
+
288
+ ```BibTex
289
+ @inproceedings{watanabe2018espnet,
290
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
291
+ title={{ESPnet}: End-to-End Speech Processing Toolkit},
292
+ year={2018},
293
+ booktitle={Proceedings of Interspeech},
294
+ pages={2207--2211},
295
+ doi={10.21437/Interspeech.2018-1456},
296
+ url={http://dx.doi.org/10.21437/Interspeech.2018-1456}
297
+ }
298
+
299
+
300
+
301
+
302
+ ```
303
+
304
+ or arXiv:
305
+
306
+ ```bibtex
307
+ @misc{watanabe2018espnet,
308
+ title={ESPnet: End-to-End Speech Processing Toolkit},
309
+ author={Shinji Watanabe and Takaaki Hori and Shigeki Karita and Tomoki Hayashi and Jiro Nishitoba and Yuya Unno and Nelson Yalta and Jahn Heymann and Matthew Wiesner and Nanxin Chen and Adithya Renduchintala and Tsubasa Ochiai},
310
+ year={2018},
311
+ eprint={1804.00015},
312
+ archivePrefix={arXiv},
313
+ primaryClass={cs.CL}
314
+ }
315
+ ```
exp/asr_stats_raw_can_char/train/feats_stats.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b07184dd4b3b1595d7bc7f5cc15447d4ad7715e770a33e0deee3b6e3caca484
3
+ size 1402
exp/asr_train_asr_transformer_raw_can_char/RESULTS.md ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!-- Generated by scripts/utils/show_asr_result.sh -->
2
+ # RESULTS
3
+ ## Environments
4
+ - date: `Fri Mar 17 23:08:24 CST 2023`
5
+ - python version: `3.8.16 | packaged by conda-forge | (default, Feb 1 2023, 16:01:55) [GCC 11.3.0]`
6
+ - espnet version: `espnet 202301`
7
+ - pytorch version: `pytorch 1.10.0`
8
+ - Git hash: `52160d6ed337e9dec74dd59695fec1548042e0b2`
9
+ - Commit date: `Thu Mar 16 21:37:39 2023 +0000`
10
+
11
+ ## exp/asr_train_asr_transformer_raw_can_char
12
+ ### WER
13
+
14
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
15
+ |---|---|---|---|---|---|---|---|---|
16
+ |inference_asr_model_valid.acc.ave/test|9077|108147|0.0|0.0|100.0|0.0|100.0|100.0|
17
+
18
+ ### CER
19
+
20
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
21
+ |---|---|---|---|---|---|---|---|---|
22
+ |inference_asr_model_valid.acc.ave/test|9077|666586|0.0|0.0|100.0|0.0|100.0|100.0|
23
+
24
+ ### TER
25
+
26
+ |dataset|Snt|Wrd|Corr|Sub|Del|Ins|Err|S.Err|
27
+ |---|---|---|---|---|---|---|---|---|
exp/asr_train_asr_transformer_raw_can_char/config.yaml ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config: conf/train_asr_transformer.yaml
2
+ print_config: false
3
+ log_level: INFO
4
+ dry_run: false
5
+ iterator_type: sequence
6
+ output_dir: exp/asr_train_asr_transformer_raw_can_char
7
+ ngpu: 1
8
+ seed: 0
9
+ num_workers: 1
10
+ num_att_plot: 3
11
+ dist_backend: nccl
12
+ dist_init_method: env://
13
+ dist_world_size: null
14
+ dist_rank: null
15
+ local_rank: 0
16
+ dist_master_addr: null
17
+ dist_master_port: null
18
+ dist_launcher: null
19
+ multiprocessing_distributed: false
20
+ unused_parameters: false
21
+ sharded_ddp: false
22
+ cudnn_enabled: true
23
+ cudnn_benchmark: false
24
+ cudnn_deterministic: true
25
+ collect_stats: false
26
+ write_collected_feats: false
27
+ max_epoch: 30
28
+ patience: null
29
+ val_scheduler_criterion:
30
+ - valid
31
+ - loss
32
+ early_stopping_criterion:
33
+ - valid
34
+ - loss
35
+ - min
36
+ best_model_criterion:
37
+ - - valid
38
+ - acc
39
+ - max
40
+ keep_nbest_models: 10
41
+ nbest_averaging_interval: 0
42
+ grad_clip: 5.0
43
+ grad_clip_type: 2.0
44
+ grad_noise: false
45
+ accum_grad: 8
46
+ no_forward_run: false
47
+ resume: true
48
+ train_dtype: float32
49
+ use_amp: false
50
+ log_interval: null
51
+ use_matplotlib: true
52
+ use_tensorboard: true
53
+ create_graph_in_tensorboard: false
54
+ use_wandb: false
55
+ wandb_project: null
56
+ wandb_id: null
57
+ wandb_entity: null
58
+ wandb_name: null
59
+ wandb_model_log_interval: -1
60
+ detect_anomaly: false
61
+ pretrain_path: null
62
+ init_param: []
63
+ ignore_init_mismatch: false
64
+ freeze_param: []
65
+ num_iters_per_epoch: null
66
+ batch_size: 16
67
+ valid_batch_size: null
68
+ batch_bins: 1000000
69
+ valid_batch_bins: null
70
+ train_shape_file:
71
+ - exp/asr_stats_raw_can_char/train/speech_shape
72
+ - exp/asr_stats_raw_can_char/train/text_shape.char
73
+ valid_shape_file:
74
+ - exp/asr_stats_raw_can_char/valid/speech_shape
75
+ - exp/asr_stats_raw_can_char/valid/text_shape.char
76
+ batch_type: folded
77
+ valid_batch_type: null
78
+ fold_length:
79
+ - 80000
80
+ - 150
81
+ sort_in_batch: descending
82
+ sort_batch: descending
83
+ multiple_iterator: false
84
+ chunk_length: 500
85
+ chunk_shift_ratio: 0.5
86
+ num_cache_chunks: 1024
87
+ chunk_excluded_key_prefixes: []
88
+ train_data_path_and_name_and_type:
89
+ - - dump/raw/train/wav.scp
90
+ - speech
91
+ - sound
92
+ - - dump/raw/train/text
93
+ - text
94
+ - text
95
+ valid_data_path_and_name_and_type:
96
+ - - dump/raw/dev/wav.scp
97
+ - speech
98
+ - sound
99
+ - - dump/raw/dev/text
100
+ - text
101
+ - text
102
+ allow_variable_data_keys: false
103
+ max_cache_size: 0.0
104
+ max_cache_fd: 32
105
+ valid_max_cache_size: null
106
+ exclude_weight_decay: false
107
+ exclude_weight_decay_conf: {}
108
+ optim: adam
109
+ optim_conf:
110
+ lr: 0.005
111
+ scheduler: warmuplr
112
+ scheduler_conf:
113
+ warmup_steps: 30000
114
+ token_list:
115
+ - <blank>
116
+ - <unk>
117
+ - <space>
118
+ - '3'
119
+ - '2'
120
+ - '5'
121
+ - g
122
+ - o
123
+ - a
124
+ - n
125
+ - i
126
+ - '4'
127
+ - u
128
+ - e
129
+ - k
130
+ - '1'
131
+ - j
132
+ - y
133
+ - z
134
+ - s
135
+ - h
136
+ - d
137
+ - m
138
+ - l
139
+ - c
140
+ - b
141
+ - f
142
+ - t
143
+ - w
144
+ - p
145
+ - r
146
+ - x
147
+ - v
148
+ - q
149
+ - <sos/eos>
150
+ init: xavier_uniform
151
+ input_size: null
152
+ ctc_conf:
153
+ dropout_rate: 0.0
154
+ ctc_type: builtin
155
+ reduce: true
156
+ ignore_nan_grad: null
157
+ zero_infinity: true
158
+ joint_net_conf: null
159
+ use_preprocessor: true
160
+ token_type: char
161
+ bpemodel: null
162
+ non_linguistic_symbols: null
163
+ cleaner: null
164
+ g2p: null
165
+ speech_volume_normalize: null
166
+ rir_scp: null
167
+ rir_apply_prob: 1.0
168
+ noise_scp: null
169
+ noise_apply_prob: 1.0
170
+ noise_db_range: '13_15'
171
+ short_noise_thres: 0.5
172
+ aux_ctc_tasks: []
173
+ frontend: default
174
+ frontend_conf:
175
+ fs: 16k
176
+ specaug: null
177
+ specaug_conf: {}
178
+ normalize: global_mvn
179
+ normalize_conf:
180
+ stats_file: exp/asr_stats_raw_can_char/train/feats_stats.npz
181
+ model: espnet
182
+ model_conf:
183
+ ctc_weight: 0.3
184
+ lsm_weight: 0.1
185
+ length_normalized_loss: false
186
+ preencoder: null
187
+ preencoder_conf: {}
188
+ encoder: transformer
189
+ encoder_conf:
190
+ output_size: 256
191
+ attention_heads: 4
192
+ linear_units: 2048
193
+ num_blocks: 12
194
+ dropout_rate: 0.1
195
+ positional_dropout_rate: 0.1
196
+ attention_dropout_rate: 0.0
197
+ input_layer: conv2d
198
+ normalize_before: true
199
+ postencoder: null
200
+ postencoder_conf: {}
201
+ decoder: transformer
202
+ decoder_conf:
203
+ attention_heads: 4
204
+ linear_units: 2048
205
+ num_blocks: 6
206
+ dropout_rate: 0.1
207
+ positional_dropout_rate: 0.1
208
+ self_attention_dropout_rate: 0.0
209
+ src_attention_dropout_rate: 0.0
210
+ preprocessor: default
211
+ preprocessor_conf: {}
212
+ required:
213
+ - output_dir
214
+ - token_list
215
+ version: '202301'
216
+ distributed: false
exp/asr_train_asr_transformer_raw_can_char/images/acc.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/backward_time.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/cer.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/cer_ctc.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/forward_time.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/gpu_max_cached_mem_GB.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/iter_time.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/loss.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/loss_att.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/loss_ctc.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/optim0_lr0.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/optim_step_time.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/train_time.png ADDED
exp/asr_train_asr_transformer_raw_can_char/images/wer.png ADDED
exp/asr_train_asr_transformer_raw_can_char/valid.acc.ave_10best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5d6684475e9019eb472b2200dfa896ff578ce446e11eb1515b1982e2159cc71
3
+ size 108694117
meta.yaml ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ espnet: '202301'
2
+ files:
3
+ asr_model_file: exp/asr_train_asr_transformer_raw_can_char/valid.acc.ave_10best.pth
4
+ python: "3.8.16 | packaged by conda-forge | (default, Feb 1 2023, 16:01:55) \n[GCC\
5
+ \ 11.3.0]"
6
+ timestamp: 1679113789.540685
7
+ torch: 1.10.0
8
+ yaml_files:
9
+ asr_train_config: exp/asr_train_asr_transformer_raw_can_char/config.yaml