jinzr commited on
Commit
7c9ac64
1 Parent(s): cc6f62c

init update

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +1 -0
  2. data/lang_bbpe_500/.gitattributes +5 -0
  3. data/lang_bbpe_500/HLG.pt +3 -0
  4. data/lang_bbpe_500/L.pt +3 -0
  5. data/lang_bbpe_500/LG.pt +3 -0
  6. data/lang_bbpe_500/L_disambig.pt +3 -0
  7. data/lang_bbpe_500/Linv.pt +3 -0
  8. data/lang_bbpe_500/bbpe.model +3 -0
  9. data/lang_bbpe_500/lexicon.txt +3 -0
  10. data/lang_bbpe_500/lexicon_disambig.txt +3 -0
  11. data/lang_bbpe_500/text +3 -0
  12. data/lang_bbpe_500/tokens.txt +3 -0
  13. data/lang_bbpe_500/unigram_500.model +3 -0
  14. data/lang_bbpe_500/unigram_500.vocab +500 -0
  15. data/lang_bbpe_500/words.txt +3 -0
  16. decoding_results/fast_beam_search/errs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +0 -0
  17. decoding_results/fast_beam_search/errs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +0 -0
  18. decoding_results/fast_beam_search/log-decode-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model-2024-01-16-10-04-04 +64 -0
  19. decoding_results/fast_beam_search/recogs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +0 -0
  20. decoding_results/fast_beam_search/recogs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +0 -0
  21. decoding_results/fast_beam_search/wer-summary-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +2 -0
  22. decoding_results/fast_beam_search/wer-summary-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt +2 -0
  23. decoding_results/greedy_search/errs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +0 -0
  24. decoding_results/greedy_search/errs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +0 -0
  25. decoding_results/greedy_search/log-decode-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model-2024-01-16-08-52-11 +51 -0
  26. decoding_results/greedy_search/recogs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +0 -0
  27. decoding_results/greedy_search/recogs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +0 -0
  28. decoding_results/greedy_search/wer-summary-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +2 -0
  29. decoding_results/greedy_search/wer-summary-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt +2 -0
  30. decoding_results/modified_beam_search/errs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +0 -0
  31. decoding_results/modified_beam_search/errs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +0 -0
  32. decoding_results/modified_beam_search/log-decode-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model-2024-01-16-10-03-43 +66 -0
  33. decoding_results/modified_beam_search/recogs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +0 -0
  34. decoding_results/modified_beam_search/recogs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +0 -0
  35. decoding_results/modified_beam_search/wer-summary-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +2 -0
  36. decoding_results/modified_beam_search/wer-summary-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt +2 -0
  37. exp/.gitattributes +6 -0
  38. exp/decoder-epoch-40-avg-10.int8.onnx +3 -0
  39. exp/decoder-epoch-40-avg-10.onnx +3 -0
  40. exp/encoder-epoch-40-avg-10.int8.onnx +3 -0
  41. exp/encoder-epoch-40-avg-10.onnx +3 -0
  42. exp/epoch-40.pt +3 -0
  43. exp/jit_script.pt +3 -0
  44. exp/joiner-epoch-40-avg-10.int8.onnx +3 -0
  45. exp/joiner-epoch-40-avg-10.onnx +3 -0
  46. exp/log/log-train-2024-01-15-11-07-41-0 +0 -0
  47. exp/log/log-train-2024-01-15-11-07-41-1 +0 -0
  48. exp/pretrained.pt +3 -0
  49. exp/tensorboard/events.out.tfevents.1705288061.de-74279-k2-train-1-1207150822-75498b8c5f-55j4z.2432177.0 +3 -0
  50. test_wavs/DEV_T0000000000.wav +0 -0
README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ See https://github.com/k2-fsa/icefall/pull/1464
data/lang_bbpe_500/.gitattributes ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ lexicon.txt filter=lfs diff=lfs merge=lfs -text
2
+ lexicon_disambig.txt filter=lfs diff=lfs merge=lfs -text
3
+ tokens.txt filter=lfs diff=lfs merge=lfs -text
4
+ words.txt filter=lfs diff=lfs merge=lfs -text
5
+ text filter=lfs diff=lfs merge=lfs -text
data/lang_bbpe_500/HLG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8687c95087cde6f19eb1a3da6e991cea6421ee6a29e56703e6e77dd362809c95
3
+ size 314849055
data/lang_bbpe_500/L.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9186950ad99a88c69e30f684fc530c7cf19e326b1f6d66f2b965452fe9d2bfc8
3
+ size 3508519
data/lang_bbpe_500/LG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b64480bde684f917dc47355d544cf0db8faf381b2094b4d6e2844f7a9632412
3
+ size 84961187
data/lang_bbpe_500/L_disambig.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62b115aa88c97b09bec8b6de9588d2d1b28c1032ebeccc611c596315ba054eaf
3
+ size 3579367
data/lang_bbpe_500/Linv.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fb66c90e9769439dac4bf6c6f7dc9120160c083f45e4d9d461f41957bf60ad8
3
+ size 3508519
data/lang_bbpe_500/bbpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5964d80e4634fc1d1630c603ba54735e49d7593742879929ab438c57d4114c96
3
+ size 246126
data/lang_bbpe_500/lexicon.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09bd54b1e008c4ed1ed12387e980fe2a7e1d77a8727c8fb911e9f02c984cb858
3
+ size 1292685
data/lang_bbpe_500/lexicon_disambig.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbabfd73b0c8c6f71211c5cec45dc0649a19e0b4901759f0536040d34b5fe611
3
+ size 1303308
data/lang_bbpe_500/text ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccd887d311d3295c7a72b6aa046eb3d58f3585741aff768414ceebda0645569e
3
+ size 7684231
data/lang_bbpe_500/tokens.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99a3d16785340e55786a3d21ba4da4d155d8033b9d351626c7253c7f560d2558
3
+ size 6331
data/lang_bbpe_500/unigram_500.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5964d80e4634fc1d1630c603ba54735e49d7593742879929ab438c57d4114c96
3
+ size 246126
data/lang_bbpe_500/unigram_500.vocab ADDED
@@ -0,0 +1,500 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <blk> 0
2
+ <sos/eos> 0
3
+ <unk> 0
4
+ ▁ƌ -3.19716
5
+ ▁ƍ -3.37846
6
+ ▁Ǝ -3.8666
7
+ ▁ƎĽĥ -3.89993
8
+ ▁Ə -4.09624
9
+ ▁Ɛ -4.35942
10
+ ▁ƋŞġ -4.55055
11
+ ĩ -4.75878
12
+ Ŕ -4.80797
13
+ Ī -4.81606
14
+ Į -4.94232
15
+ Ş -5.00197
16
+ Ő -5.0194
17
+ ł -5.02456
18
+ ť -5.05381
19
+ Ģ -5.05508
20
+ ▁ƌŁŎ -5.07018
21
+ ļ -5.1054
22
+ ń -5.14267
23
+ ĺ -5.15195
24
+ Ť -5.16014
25
+ ŗ -5.16186
26
+ Ŗ -5.18595
27
+ ▁ƌİ -5.2022
28
+ ħ -5.20415
29
+ š -5.23155
30
+ Ħ -5.24698
31
+ ţ -5.25781
32
+ ŝ -5.26309
33
+ œ -5.27034
34
+ ġ -5.27961
35
+ Ł -5.2802
36
+ ņ -5.2807
37
+ ▁ƍĻŕ -5.28214
38
+ ī -5.28272
39
+ ▁ƋŞœ -5.29616
40
+ Š -5.3099
41
+ Ś -5.31238
42
+ ▁ƌĮĢ -5.31462
43
+ ▁ƋŠŠ -5.34844
44
+ Ĵ -5.36336
45
+ ŕ -5.36721
46
+ ▁ƍŁĪ -5.40456
47
+ Ň -5.41222
48
+ ŋ -5.41951
49
+ Ŋ -5.42924
50
+ Ō -5.43171
51
+ ▁ƋŠħ -5.43315
52
+ ▁ƋŠĭ -5.43392
53
+ ▁Ǝš -5.43517
54
+ ř -5.43851
55
+ ▁ƋŞī -5.44169
56
+ ĵ -5.45028
57
+ Ļ -5.46384
58
+ ▁ƋŞĮ -5.46761
59
+ ▁ƌľţ -5.47836
60
+ ĸ -5.47966
61
+ ľ -5.48268
62
+ ō -5.49272
63
+ ŏ -5.49277
64
+ Ĺ -5.50156
65
+ ▁ƌŞģ -5.50943
66
+ ň -5.51563
67
+ Ĥ -5.51939
68
+ Ĭ -5.52276
69
+ į -5.52282
70
+ Ņ -5.52406
71
+ ĭ -5.53405
72
+ ķ -5.5344
73
+ ▁ƌŊō -5.54566
74
+ ģ -5.55137
75
+ Ř -5.56234
76
+ ▁ƋŞĽ -5.5747
77
+ Ľ -5.58446
78
+ Ķ -5.58867
79
+ ▁ƋŞŠ -5.59075
80
+ Ţ -5.5949
81
+ İ -5.60126
82
+ ▁ƌşŚ -5.61297
83
+ ş -5.61799
84
+ ▁ƋŞĪ -5.70592
85
+ ĥ -5.71751
86
+ Ŏ -5.72581
87
+ ▁ƌİĴ -5.734
88
+ ▁Ƌţ -5.73857
89
+ ▁ƌĦ -5.73882
90
+ Ĩ -5.74167
91
+ ▁Ƌš -5.8087
92
+ ▁ƋŞŐ -5.81474
93
+ Π-5.81861
94
+ ▁ƌĨŠ -5.8218
95
+ ▁ƌŁŖ -5.8229
96
+ ▁Əņĭ -5.83288
97
+ ▁ƌĦŒ -5.83919
98
+ ▁ƌĩħ -5.83953
99
+ ▁ƋŢĽ -5.84644
100
+ ▁ƏŚ -5.84917
101
+ ı -5.86467
102
+ ▁ƍĩı -5.88557
103
+ ▁ƌĵĭ -5.90836
104
+ ▁Əťļ -5.90872
105
+ ő -5.92808
106
+ Ŝ -5.94431
107
+ ▁ƌĩŖ -5.95008
108
+ ▁ƋŠō -5.95681
109
+ Ń -5.96487
110
+ ▁ƌŕş -5.96726
111
+ ▁ƋŠķ -5.98296
112
+ ▁ƋŠ -5.98395
113
+ ▁ƍĺŜ -5.98461
114
+ ▁ƌŁŠ -5.9885
115
+ ▁ƏĤţ -5.99015
116
+ ▁ƍĬ -5.99308
117
+ ▁ƍĩť -5.99998
118
+ ▁Ɛġ -6.012
119
+ ▁ƋŞ -6.0229
120
+ ▁ƍłŋ -6.02655
121
+ ▁Əŕ -6.03123
122
+ ▁ƍĹŖ -6.03545
123
+ ▁Ƌšŋ -6.03769
124
+ ▁ƎļŤ -6.04031
125
+ ▁ƍĺŋ -6.07482
126
+ ▁ƋşĬ -6.10443
127
+ ▁Ƌş -6.1052
128
+ ▁ƏġĦ -6.1086
129
+ ▁ƌŖħ -6.11159
130
+ ▁ƏŌĢ -6.11263
131
+ ▁ƎįŖ -6.11464
132
+ ▁ƌŔŜ -6.13157
133
+ ▁ƌľľ -6.14532
134
+ ▁Əśĥ -6.15595
135
+ ▁ƌŊĽ -6.16649
136
+ ▁ƍŁĩ -6.17773
137
+ ▁ƐľŜ -6.19122
138
+ ▁Ƌşń -6.19929
139
+ ▁ƍĹş -6.20167
140
+ ▁ƌıį -6.20384
141
+ ▁ƏĨ -6.20649
142
+ ▁ƍŁŠ -6.21703
143
+ ▁ƋŞĬ -6.22246
144
+ ▁Ǝķń -6.22305
145
+ ▁ƋŠį -6.2236
146
+ ▁ƌĪĮ -6.22885
147
+ ▁ƍŕķ -6.24515
148
+ ▁Ǝģş -6.25851
149
+ ▁ƌŢġ -6.26487
150
+ ▁ƐőĻ -6.28347
151
+ ▁ƌīŎ -6.29021
152
+ ▁Ǝšİ -6.32286
153
+ ▁Əťľ -6.32585
154
+ ▁ƍīŋ -6.3345
155
+ ▁ƎķŎ -6.34218
156
+ ▁Əśľ -6.35403
157
+ ▁ƋţĶ -6.35925
158
+ ▁ƐĨĴ -6.3644
159
+ ś -6.38613
160
+ ▁Ƌšŝ -6.39098
161
+ ▁ƏţŌ -6.40247
162
+ ▁ƌİŕ -6.40269
163
+ ▁ƏŇő -6.40776
164
+ ▁ƏťĨ -6.40784
165
+ ▁ƋŞĨ -6.41243
166
+ ▁ƌœı -6.42182
167
+ ▁ƍŁŒ -6.43345
168
+ ▁ƍĪĬ -6.4338
169
+ ▁ƌĦŎ -6.43499
170
+ ▁ƌİŞ -6.43801
171
+ ▁ƋţŁ -6.44903
172
+ ▁ƍŁġ -6.46051
173
+ ▁Ǝķś -6.46421
174
+ ▁ƎľŔ -6.467
175
+ ▁ƌĦő -6.47982
176
+ ▁ƐĤŎ -6.48043
177
+ ▁ƌĦœ -6.49274
178
+ ▁ƌı -6.4971
179
+ ▁ƍĩĴ -6.50821
180
+ ▁ƌīľ -6.53467
181
+ ▁ƌīŅ -6.53724
182
+ ▁ƐłŇ -6.53788
183
+ ▁ƋŞĤ -6.54633
184
+ ▁ƌŠŌ -6.55184
185
+ ▁ƋŢĢ -6.5633
186
+ ▁ƋŠŊ -6.56532
187
+ ▁ƌŔŃ -6.56756
188
+ ▁ƌńį -6.57743
189
+ ▁ƌĦĤ -6.58198
190
+ ▁Ƌşł -6.58499
191
+ ▁ƌŝŘ -6.5855
192
+ ▁ƌŔĽ -6.58607
193
+ ▁ƌŖİ -6.59051
194
+ ▁ƏťĻ -6.59824
195
+ ▁ƌĦŜ -6.60151
196
+ ▁ƌĶĢ -6.60241
197
+ ▁ƌŝŋ -6.60314
198
+ ▁ƌıĭ -6.60475
199
+ ▁Əġĭ -6.60701
200
+ ▁ƎţĴ -6.60816
201
+ ▁ƌıĩ -6.61271
202
+ ▁ƌħĦ -6.61425
203
+ ▁ƌĦř -6.61917
204
+ ▁ƋšĹ -6.6213
205
+ ▁ƋŞį -6.623
206
+ ▁ƍřĸ -6.62593
207
+ ▁ƌŤĺ -6.62751
208
+ ▁ƎŒ -6.64811
209
+ ▁ƌŖŗ -6.65321
210
+ ▁ƏŔŖ -6.65678
211
+ ▁Ǝıħ -6.66421
212
+ ▁ƌĦŋ -6.66644
213
+ ▁Ɛĸť -6.67939
214
+ ▁ƋŠĬ -6.68516
215
+ ▁ƌıĮ -6.68798
216
+ ▁Əś -6.68867
217
+ ▁ƌĸħ -6.69805
218
+ ▁ƌŋř -6.70211
219
+ ▁ƋšŒ -6.70494
220
+ ▁ƐĤţ -6.70841
221
+ ▁ƋŠŒ -6.70994
222
+ ▁ƋŞŊ -6.7116
223
+ ▁ƌşŜ -6.72638
224
+ ▁ƌĩŜ -6.72911
225
+ ▁ƌĭĹ -6.74188
226
+ ▁ƍķť -6.7514
227
+ ▁ƌşř -6.75474
228
+ ▁ƍŁń -6.76244
229
+ ▁ƏņŎ -6.76775
230
+ ▁ƐĨĮ -6.76956
231
+ ▁ƍŒņ -6.77274
232
+ ▁ƋŞš -6.77455
233
+ ▁ƍĮŔ -6.78155
234
+ ▁ƌīņ -6.78194
235
+ ▁Ƌťł -6.78255
236
+ ▁ƌĩŏ -6.78866
237
+ ▁ƐġĽ -6.79022
238
+ ▁ƎľŞ -6.794
239
+ ▁ƐĢĶ -6.79959
240
+ ▁ƋŤ -6.80011
241
+ ▁ƌŊŏ -6.80894
242
+ ▁ƌŗĸ -6.8179
243
+ ▁ƍİı -6.82488
244
+ ▁ƍīĸ -6.82617
245
+ ▁ƌĭĺ -6.82762
246
+ ▁ƐĨİ -6.82849
247
+ ▁ƐĺŚ -6.83218
248
+ ▁ƌĴĻ -6.83769
249
+ ▁ƌšŠ -6.84596
250
+ ▁ƌľŅ -6.8485
251
+ ▁ƍśŝ -6.85106
252
+ ▁ƌŋţ -6.85296
253
+ ▁ƍĪġ -6.85796
254
+ ▁ƍĸŖ -6.86622
255
+ ▁ƎŔņ -6.87755
256
+ ▁ƌŊĹ -6.87808
257
+ ▁ƎŊŠ -6.88415
258
+ ▁ƌĭŠ -6.88751
259
+ ▁ƍľŚ -6.89993
260
+ ▁ƍġō -6.9102
261
+ ▁ƎœĪ -6.91311
262
+ ▁ƌŠķ -6.92274
263
+ ▁Ƌšį -6.92685
264
+ ▁ƌŇŃ -6.92975
265
+ ▁Əťı -6.93133
266
+ ▁ƏŔņ -6.93208
267
+ ▁ƌţĶ -6.93571
268
+ ▁Ɛņş -6.93584
269
+ ▁ƍķŜ -6.93772
270
+ ▁Ƌťņ -6.94787
271
+ ▁ƍĻį -6.96518
272
+ ▁ƐĻ -6.97996
273
+ ▁Ǝłġ -6.98773
274
+ ▁ƏŔŤ -6.99153
275
+ ▁ƍŖĴ -6.99352
276
+ ▁ƏŖĤ -7.00159
277
+ ▁Əōĥ -7.00256
278
+ ▁ƍĭĢ -7.00855
279
+ ▁ƋţĮ -7.00883
280
+ ▁Əśŝ -7.01024
281
+ ▁ƌœŌ -7.01508
282
+ ▁ƌŤĩ -7.04497
283
+ ▁ƍŅŢ -7.04525
284
+ ▁ƍŘņ -7.04543
285
+ ▁Ƌšī -7.04903
286
+ ▁ƍœň -7.05139
287
+ ▁Əģņ -7.05215
288
+ ▁ƌťĤ -7.05258
289
+ ▁ƌİŖ -7.05538
290
+ ▁ƎĥŜ -7.0732
291
+ ▁ƌįš -7.08255
292
+ ▁ƏŤ -7.08788
293
+ ▁ƌŔĪ -7.08923
294
+ ▁ƍœŊ -7.0894
295
+ ▁ƎŤį -7.09152
296
+ ▁ƍŃŁ -7.0981
297
+ ▁ƍĺŅ -7.1161
298
+ ▁ƏĬ -7.11958
299
+ ▁ƐĻń -7.12495
300
+ ▁ƎōĴ -7.12752
301
+ ▁ƎŠť -7.1287
302
+ ▁ƌŌģ -7.1296
303
+ ▁ƌńŠ -7.13019
304
+ ▁ƏŠő -7.13372
305
+ ▁ƐŁġ -7.14207
306
+ ▁ƍĤĦ -7.14907
307
+ ▁ƍŜĩ -7.15571
308
+ ▁ƏťĴ -7.15889
309
+ ▁ƏŚş -7.16064
310
+ ▁ƏŤŤ -7.1667
311
+ ▁ƌİĺ -7.16705
312
+ ▁ƌŞĤ -7.16808
313
+ ▁ƋŞĹ -7.16887
314
+ ▁ƍŅĨ -7.17123
315
+ ▁ƍŁŐ -7.17708
316
+ ▁ƋŞŁ -7.17868
317
+ ▁ƍŗģ -7.17928
318
+ ▁ƍĢŕ -7.18683
319
+ ▁ƋŠľ -7.18718
320
+ ▁ƍœŋ -7.18878
321
+ ▁ƏŕĢ -7.19008
322
+ ▁ƎŁĬ -7.1915
323
+ ▁ƐĨĭ -7.19848
324
+ ▁ƌıĴ -7.20728
325
+ ▁Ǝķŝ -7.21037
326
+ ▁ƌĮĤ -7.21043
327
+ ▁ƌľŃ -7.21177
328
+ ▁ƐĺŎ -7.2143
329
+ ▁ƍķş -7.21651
330
+ ▁ƌŢİ -7.21941
331
+ ▁Əঠ-7.22107
332
+ ▁ƍĥİ -7.2239
333
+ ▁ƍįŋ -7.22463
334
+ ▁Ǝŗř -7.22499
335
+ ▁ƍīġ -7.25222
336
+ ▁ƌİī -7.2523
337
+ ▁Ɛķġ -7.25232
338
+ ▁ƎľĴ -7.25253
339
+ ▁ƍġš -7.25326
340
+ ▁ƍŖŚ -7.25329
341
+ ▁ƏĢķ -7.25352
342
+ ▁Ƌţİ -7.25534
343
+ ▁Ǝšœ -7.25549
344
+ ▁ƍĪĶ -7.25584
345
+ ▁ƏŔŕ -7.25624
346
+ ▁ƋŠť -7.25755
347
+ ▁ƍĸŚ -7.26318
348
+ ▁ƏĨř -7.26433
349
+ ▁ƋšŜ -7.26677
350
+ ▁Əįŝ -7.27447
351
+ ▁ƐļŇ -7.27894
352
+ ▁ƌŢŅ -7.2813
353
+ ▁ƌŢŠ -7.28651
354
+ ▁Ǝįŕ -7.291
355
+ ▁ƏġĢ -7.29374
356
+ ▁Əŕŋ -7.30514
357
+ ▁ƌŗġ -7.31459
358
+ ▁ƎĪş -7.31631
359
+ ▁ƐĺŔ -7.3168
360
+ ▁ƏŚœ -7.32294
361
+ ▁ƌķŔ -7.32513
362
+ ▁Ǝķŗ -7.32595
363
+ ▁ƍŋŢ -7.32973
364
+ ▁Əŝŕ -7.33158
365
+ ▁ƌŊĥ -7.35735
366
+ ▁Ǝřš -7.36312
367
+ ▁ƐŇĻ -7.36762
368
+ ▁ƎōŖ -7.36854
369
+ ▁ƍįŎ -7.37011
370
+ ▁ƌţŗ -7.38065
371
+ ▁ƍĩĻ -7.38067
372
+ ▁ƎœĹ -7.38105
373
+ ▁ƏŔŏ -7.38235
374
+ ▁ƌşť -7.38297
375
+ ▁Ǝšń -7.38342
376
+ ▁Ƌşı -7.38804
377
+ ▁ƌħř -7.39077
378
+ ▁ƍŜŎ -7.393
379
+ ▁ƌŖĴ -7.39481
380
+ ▁ƍĻŤ -7.39553
381
+ ▁ƌŝŗ -7.40121
382
+ ▁ƌŞŞ -7.40335
383
+ ▁ƐŌĹ -7.40411
384
+ ▁ƐĶŜ -7.40563
385
+ ▁ƐļĮ -7.40689
386
+ ▁ƌŔĭ -7.40758
387
+ ▁ƍĩŝ -7.40847
388
+ ▁ƌĩľ -7.4095
389
+ ▁ƎōĮ -7.41323
390
+ ▁ƎőĬ -7.42286
391
+ ▁ƍńŋ -7.42568
392
+ ▁ƍņĩ -7.43364
393
+ ▁Əıŋ -7.43729
394
+ ▁ƎŎĬ -7.44207
395
+ ▁ƎıĤ -7.44671
396
+ ▁ƌŊĨ -7.45022
397
+ ▁ƍĻĶ -7.45135
398
+ ▁ƌīť -7.45209
399
+ ▁ƌĨħ -7.45284
400
+ ▁Ƌť -7.453
401
+ ▁ƍŒŤ -7.45615
402
+ ▁ƌıĥ -7.45805
403
+ ▁ƎĪĭ -7.4581
404
+ ▁ƏŔŊ -7.46154
405
+ ▁ƏŝĴ -7.4708
406
+ ▁Ǝōŕ -7.4726
407
+ ▁ƐġŅ -7.47302
408
+ ▁Ƌšň -7.48171
409
+ ▁ƌœĻ -7.49111
410
+ ▁ƍķŤ -7.49337
411
+ ▁ƏŜĦ -7.49672
412
+ ▁Ɛļŏ -7.49813
413
+ ▁ƌŝŃ -7.49853
414
+ ▁ƌįń -7.49925
415
+ ▁ƌĢĽ -7.50365
416
+ ▁ƎŠō -7.50453
417
+ ▁ƌħĮ -7.50477
418
+ ▁ƋŢŅ -7.505
419
+ ▁ƌŔŇ -7.5077
420
+ ▁ƎĪŏ -7.50877
421
+ ▁ƍŅŝ -7.51128
422
+ ▁ƏŤĤ -7.51459
423
+ ▁ƍĹĨ -7.51962
424
+ ▁ƌĮĸ -7.5223
425
+ ▁Ɛňį -7.52235
426
+ ▁Ǝĸĭ -7.52834
427
+ ▁ƏţŒ -7.53136
428
+ ▁ƌĴī -7.5314
429
+ ▁ƋŤľ -7.53307
430
+ ▁Əōň -7.53589
431
+ ▁ƏœŌ -7.54836
432
+ ▁ƌŋş -7.55543
433
+ ▁ƎįĨ -7.55666
434
+ ▁ƌĮį -7.55769
435
+ ▁ƐŇĥ -7.56045
436
+ ▁Ɛľħ -7.56263
437
+ ▁ƍįĵ -7.56385
438
+ ▁ƌĴŎ -7.56591
439
+ ▁ƍĭĨ -7.57285
440
+ ▁ƌŕŢ -7.57338
441
+ ▁ƌĮĨ -7.57381
442
+ ▁ƍŕİ -7.57849
443
+ ▁ƎĪĨ -7.5926
444
+ ▁ƌťő -7.5929
445
+ ▁ƌŊŚ -7.59533
446
+ ▁ƍŁŕ -7.59652
447
+ ▁ƍŁĮ -7.59806
448
+ ▁ƌŃĬ -7.60054
449
+ ▁ƌĩļ -7.60532
450
+ ▁Əōħ -7.60735
451
+ ▁ƎľŚ -7.60816
452
+ ▁ƎŅķ -7.61627
453
+ ▁ƐłŃ -7.62093
454
+ ▁ƍŠı -7.62144
455
+ ▁ƍĹŕ -7.6232
456
+ ▁ƎŠŌ -7.62703
457
+ ▁ƍĹţ -7.63111
458
+ ▁ƌĩő -7.63792
459
+ ▁ƎŊŤ -7.6435
460
+ ▁ƍŚš -7.64788
461
+ ▁ƐŇħ -7.64809
462
+ ▁ƍķŕ -7.65523
463
+ ▁ƌħľ -7.6566
464
+ ▁ƌļŎ -7.65925
465
+ ▁ƍŎņ -7.66019
466
+ ▁ƐĽŤ -7.66103
467
+ ▁ƌĭš -7.66153
468
+ ▁ƍĤř -7.66756
469
+ ▁ƎįĬ -7.66772
470
+ ▁ƌľŇ -7.66864
471
+ ▁ƍŞŞ -7.67545
472
+ ▁ƍŃĥ -7.68242
473
+ ▁ƐŏŒ -7.68288
474
+ ▁ƍĩĹ -7.68436
475
+ ▁ƍļŠ -7.68474
476
+ ▁ƌŢĸ -7.68914
477
+ ▁ƌŞŌ -7.69112
478
+ ▁ƏŃĮ -7.69745
479
+ ▁ƎĦō -7.70307
480
+ ▁ƎĬı -7.72098
481
+ ▁ƌĮĺ -7.72345
482
+ ▁ƌŘĢ -7.72504
483
+ ▁ƌīŃ -7.72553
484
+ ▁ƌŤŔ -7.72705
485
+ ▁ƌĥť -7.73721
486
+ ▁ƌōĬ -7.74862
487
+ ▁Ɛļı -7.762
488
+ ▁ƍįō -7.7628
489
+ ▁ƌňő -7.76332
490
+ ▁ƎņŔ -7.76353
491
+ ▁ƌŋŋ -7.76416
492
+ ▁ƍłĤ -7.77251
493
+ Ɩ -8.585
494
+ Ɛ -8.5851
495
+ Ə -8.58521
496
+ Ǝ -8.58531
497
+ Ƌ -8.5854
498
+ ƍ -8.5855
499
+ ƌ -8.5856
500
+ ▁ -8.5857
data/lang_bbpe_500/words.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53b9be0095c5158dae04888f4947511a5c70071977bdf4214b85fc18879f150d
3
+ size 566278
decoding_results/fast_beam_search/errs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/fast_beam_search/errs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/fast_beam_search/log-decode-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model-2024-01-16-10-04-04 ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-16 10:04:04,359 INFO [decode.py:673] Decoding started
2
+ 2024-01-16 10:04:04,359 INFO [decode.py:679] Device: cuda:0
3
+ 2024-01-16 10:04:04,423 INFO [lexicon.py:168] Loading pre-compiled data/lang_bbpe_500/Linv.pt
4
+ 2024-01-16 10:04:04,436 INFO [decode.py:691] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '2989b0b1186fa6022932804f5b39fbb2781ebf42', 'k2-git-date': 'Fri Nov 24 11:34:10 2023', 'lhotse-version': '1.19.0.dev+git.d1ae9c05.dirty', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev/aishell-zipformer-bbpe', 'icefall-git-sha1': 'bce81394-clean', 'icefall-git-date': 'Thu Jan 11 09:56:01 2024', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.4.dev20231207+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.19.0.dev0+git.d1ae9c05.dirty-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1207150822-75498b8c5f-55j4z', 'IP address': '10.177.74.211'}, 'epoch': 40, 'iter': 0, 'avg': 10, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan'), 'bpe_model': 'data/lang_bbpe_500/bbpe.model', 'lang_dir': PosixPath('data/lang_bbpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'ilme_scale': 0.2, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'blank_penalty': 0.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'res_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/fast_beam_search'), 'suffix': 'epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
5
+ 2024-01-16 10:04:04,436 INFO [decode.py:693] About to create model
6
+ 2024-01-16 10:04:04,956 INFO [decode.py:760] Calculating the averaged model over epoch range from 30 (excluded) to 40
7
+ 2024-01-16 10:04:12,532 INFO [decode.py:791] Number of model parameters: 65549011
8
+ 2024-01-16 10:04:12,532 INFO [asr_datamodule.py:371] About to get dev cuts
9
+ 2024-01-16 10:04:12,549 INFO [asr_datamodule.py:312] About to create dev dataset
10
+ 2024-01-16 10:04:13,059 INFO [asr_datamodule.py:329] About to create dev dataloader
11
+ 2024-01-16 10:04:13,059 INFO [asr_datamodule.py:376] About to get test cuts
12
+ 2024-01-16 10:04:13,061 INFO [asr_datamodule.py:341] About to create test dataset
13
+ 2024-01-16 10:04:15,062 INFO [decode.py:564] batch 0/?, cuts processed until now is 35
14
+ 2024-01-16 10:04:27,318 INFO [decode.py:564] batch 20/?, cuts processed until now is 785
15
+ 2024-01-16 10:04:39,148 INFO [decode.py:564] batch 40/?, cuts processed until now is 1572
16
+ 2024-01-16 10:04:50,338 INFO [decode.py:564] batch 60/?, cuts processed until now is 2433
17
+ 2024-01-16 10:05:03,143 INFO [decode.py:564] batch 80/?, cuts processed until now is 3154
18
+ 2024-01-16 10:05:11,212 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.7125, 2.7740, 2.7712, 2.0904], device='cuda:0')
19
+ 2024-01-16 10:05:14,032 INFO [decode.py:564] batch 100/?, cuts processed until now is 4023
20
+ 2024-01-16 10:05:26,372 INFO [decode.py:564] batch 120/?, cuts processed until now is 4776
21
+ 2024-01-16 10:05:37,471 INFO [decode.py:564] batch 140/?, cuts processed until now is 5662
22
+ 2024-01-16 10:05:39,479 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.1403, 1.6074, 1.8219, 3.2897], device='cuda:0')
23
+ 2024-01-16 10:05:45,095 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.1947, 1.4890, 1.7581, 3.0378], device='cuda:0')
24
+ 2024-01-16 10:05:48,552 INFO [decode.py:564] batch 160/?, cuts processed until now is 6552
25
+ 2024-01-16 10:05:59,780 INFO [decode.py:564] batch 180/?, cuts processed until now is 7386
26
+ 2024-01-16 10:06:11,198 INFO [decode.py:564] batch 200/?, cuts processed until now is 8204
27
+ 2024-01-16 10:06:11,753 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.3423, 3.6335, 3.9994, 3.4270], device='cuda:0')
28
+ 2024-01-16 10:06:23,159 INFO [decode.py:564] batch 220/?, cuts processed until now is 8999
29
+ 2024-01-16 10:06:36,041 INFO [decode.py:564] batch 240/?, cuts processed until now is 9811
30
+ 2024-01-16 10:06:47,892 INFO [decode.py:564] batch 260/?, cuts processed until now is 10620
31
+ 2024-01-16 10:06:58,301 INFO [decode.py:564] batch 280/?, cuts processed until now is 11537
32
+ 2024-01-16 10:07:10,090 INFO [decode.py:564] batch 300/?, cuts processed until now is 12391
33
+ 2024-01-16 10:07:21,683 INFO [decode.py:564] batch 320/?, cuts processed until now is 13152
34
+ 2024-01-16 10:07:33,975 INFO [decode.py:564] batch 340/?, cuts processed until now is 13879
35
+ 2024-01-16 10:07:45,332 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/fast_beam_search/recogs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt
36
+ 2024-01-16 10:07:45,825 INFO [utils.py:565] [dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64] %WER 4.17% [8566 / 205341, 197 ins, 500 del, 7869 sub ]
37
+ 2024-01-16 10:07:46,507 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/fast_beam_search/errs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt
38
+ 2024-01-16 10:07:46,527 INFO [decode.py:614]
39
+ For dev, WER of different settings are:
40
+ blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64 4.17 best for dev
41
+
42
+ 2024-01-16 10:07:48,132 INFO [decode.py:564] batch 0/?, cuts processed until now is 31
43
+ 2024-01-16 10:08:04,342 INFO [decode.py:564] batch 20/?, cuts processed until now is 697
44
+ 2024-01-16 10:08:17,927 INFO [decode.py:564] batch 40/?, cuts processed until now is 1395
45
+ 2024-01-16 10:08:29,163 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.4244, 2.1965, 2.9948, 2.2683, 2.6988, 2.7314, 2.9246, 2.6597],
46
+ device='cuda:0')
47
+ 2024-01-16 10:08:29,706 INFO [decode.py:564] batch 60/?, cuts processed until now is 2169
48
+ 2024-01-16 10:08:33,536 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.5973, 3.2926, 2.6284, 2.3168], device='cuda:0')
49
+ 2024-01-16 10:08:43,311 INFO [decode.py:564] batch 80/?, cuts processed until now is 2801
50
+ 2024-01-16 10:08:55,694 INFO [decode.py:564] batch 100/?, cuts processed until now is 3578
51
+ 2024-01-16 10:09:09,070 INFO [decode.py:564] batch 120/?, cuts processed until now is 4249
52
+ 2024-01-16 10:09:20,752 INFO [decode.py:564] batch 140/?, cuts processed until now is 5061
53
+ 2024-01-16 10:09:27,732 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.9109, 5.6864, 5.6041, 5.6176], device='cuda:0')
54
+ 2024-01-16 10:09:32,015 INFO [decode.py:564] batch 160/?, cuts processed until now is 5875
55
+ 2024-01-16 10:09:44,614 INFO [decode.py:564] batch 180/?, cuts processed until now is 6581
56
+ 2024-01-16 10:09:56,882 INFO [decode.py:564] batch 200/?, cuts processed until now is 7122
57
+ 2024-01-16 10:09:58,670 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/fast_beam_search/recogs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt
58
+ 2024-01-16 10:09:58,892 INFO [utils.py:565] [test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64] %WER 4.43% [4646 / 104765, 81 ins, 337 del, 4228 sub ]
59
+ 2024-01-16 10:09:59,240 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/fast_beam_search/errs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt
60
+ 2024-01-16 10:09:59,243 INFO [decode.py:614]
61
+ For test, WER of different settings are:
62
+ blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64 4.43 best for test
63
+
64
+ 2024-01-16 10:09:59,250 INFO [decode.py:832] Done!
decoding_results/fast_beam_search/recogs-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/fast_beam_search/recogs-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/fast_beam_search/wer-summary-dev-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64 4.17
decoding_results/fast_beam_search/wer-summary-test-blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64-epoch-40-avg-10-beam-20.0-max-contexts-8-max-states-64-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ blank_penalty_0.0_beam_20.0_max_contexts_8_max_states_64 4.43
decoding_results/greedy_search/errs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/greedy_search/errs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/greedy_search/log-decode-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model-2024-01-16-08-52-11 ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-16 08:52:11,884 INFO [decode.py:673] Decoding started
2
+ 2024-01-16 08:52:11,884 INFO [decode.py:679] Device: cuda:0
3
+ 2024-01-16 08:52:11,964 INFO [lexicon.py:168] Loading pre-compiled data/lang_bbpe_500/Linv.pt
4
+ 2024-01-16 08:52:11,978 INFO [decode.py:691] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '2989b0b1186fa6022932804f5b39fbb2781ebf42', 'k2-git-date': 'Fri Nov 24 11:34:10 2023', 'lhotse-version': '1.19.0.dev+git.d1ae9c05.dirty', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev/aishell-zipformer-bbpe', 'icefall-git-sha1': 'bce81394-clean', 'icefall-git-date': 'Thu Jan 11 09:56:01 2024', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.4.dev20231207+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.19.0.dev0+git.d1ae9c05.dirty-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-10-0115153933-6959778564-g28xv', 'IP address': '10.177.77.20'}, 'epoch': 40, 'iter': 0, 'avg': 10, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan'), 'bpe_model': 'data/lang_bbpe_500/bbpe.model', 'lang_dir': PosixPath('data/lang_bbpe_500'), 'decoding_method': 'greedy_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'ilme_scale': 0.2, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'blank_penalty': 0.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'res_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/greedy_search'), 'suffix': 'epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
5
+ 2024-01-16 08:52:11,978 INFO [decode.py:693] About to create model
6
+ 2024-01-16 08:52:12,576 INFO [decode.py:760] Calculating the averaged model over epoch range from 30 (excluded) to 40
7
+ 2024-01-16 08:52:22,828 INFO [decode.py:791] Number of model parameters: 65549011
8
+ 2024-01-16 08:52:22,829 INFO [asr_datamodule.py:371] About to get dev cuts
9
+ 2024-01-16 08:52:22,860 INFO [asr_datamodule.py:312] About to create dev dataset
10
+ 2024-01-16 08:52:23,509 INFO [asr_datamodule.py:329] About to create dev dataloader
11
+ 2024-01-16 08:52:23,510 INFO [asr_datamodule.py:376] About to get test cuts
12
+ 2024-01-16 08:52:23,512 INFO [asr_datamodule.py:341] About to create test dataset
13
+ 2024-01-16 08:52:25,546 INFO [decode.py:564] batch 0/?, cuts processed until now is 35
14
+ 2024-01-16 08:52:41,292 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.2616, 1.8791, 2.0020, 3.6743], device='cuda:0')
15
+ 2024-01-16 08:52:42,441 INFO [decode.py:564] batch 50/?, cuts processed until now is 2015
16
+ 2024-01-16 08:52:50,054 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.2876, 3.9804, 3.4153, 3.9264], device='cuda:0')
17
+ 2024-01-16 08:52:59,262 INFO [decode.py:564] batch 100/?, cuts processed until now is 4023
18
+ 2024-01-16 08:53:11,327 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.2526, 3.5149, 3.8873, 3.3463], device='cuda:0')
19
+ 2024-01-16 08:53:13,954 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.6640, 3.5339, 2.7571, 3.2975], device='cuda:0')
20
+ 2024-01-16 08:53:15,340 INFO [decode.py:564] batch 150/?, cuts processed until now is 6112
21
+ 2024-01-16 08:53:21,815 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.6251, 3.5824, 3.8278, 1.6666], device='cuda:0')
22
+ 2024-01-16 08:53:30,311 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2507, 3.2402, 3.4456, 3.2989], device='cuda:0')
23
+ 2024-01-16 08:53:30,910 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.6706, 3.8788, 4.2342, 4.2799], device='cuda:0')
24
+ 2024-01-16 08:53:31,225 INFO [decode.py:564] batch 200/?, cuts processed until now is 8204
25
+ 2024-01-16 08:53:48,005 INFO [decode.py:564] batch 250/?, cuts processed until now is 10182
26
+ 2024-01-16 08:54:06,182 INFO [decode.py:564] batch 300/?, cuts processed until now is 12391
27
+ 2024-01-16 08:54:23,201 INFO [decode.py:564] batch 350/?, cuts processed until now is 14177
28
+ 2024-01-16 08:54:25,461 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/greedy_search/recogs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt
29
+ 2024-01-16 08:54:26,045 INFO [utils.py:565] [dev-greedy_search_blank_penalty_0.0] %WER 4.31% [8846 / 205341, 194 ins, 690 del, 7962 sub ]
30
+ 2024-01-16 08:54:26,887 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/greedy_search/errs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt
31
+ 2024-01-16 08:54:26,891 INFO [decode.py:614]
32
+ For dev, WER of different settings are:
33
+ greedy_search_blank_penalty_0.0 4.31 best for dev
34
+
35
+ 2024-01-16 08:54:28,185 INFO [decode.py:564] batch 0/?, cuts processed until now is 31
36
+ 2024-01-16 08:54:46,025 INFO [decode.py:564] batch 50/?, cuts processed until now is 1797
37
+ 2024-01-16 08:55:03,919 INFO [decode.py:564] batch 100/?, cuts processed until now is 3578
38
+ 2024-01-16 08:55:11,697 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.0320, 1.9381, 2.6667, 1.9219, 2.6101, 2.3684, 2.6248, 2.4339],
39
+ device='cuda:0')
40
+ 2024-01-16 08:55:16,148 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.7322, 2.9529, 2.8860, 2.1513], device='cuda:0')
41
+ 2024-01-16 08:55:20,706 INFO [decode.py:564] batch 150/?, cuts processed until now is 5472
42
+ 2024-01-16 08:55:24,812 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.1918, 3.9201, 3.2397, 3.8215], device='cuda:0')
43
+ 2024-01-16 08:55:38,827 INFO [decode.py:564] batch 200/?, cuts processed until now is 7122
44
+ 2024-01-16 08:55:39,578 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/greedy_search/recogs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt
45
+ 2024-01-16 08:55:39,835 INFO [utils.py:565] [test-greedy_search_blank_penalty_0.0] %WER 4.54% [4759 / 104765, 78 ins, 418 del, 4263 sub ]
46
+ 2024-01-16 08:55:40,266 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/greedy_search/errs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt
47
+ 2024-01-16 08:55:40,269 INFO [decode.py:614]
48
+ For test, WER of different settings are:
49
+ greedy_search_blank_penalty_0.0 4.54 best for test
50
+
51
+ 2024-01-16 08:55:40,280 INFO [decode.py:832] Done!
decoding_results/greedy_search/recogs-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/greedy_search/recogs-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/greedy_search/wer-summary-dev-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ greedy_search_blank_penalty_0.0 4.31
decoding_results/greedy_search/wer-summary-test-greedy_search_blank_penalty_0.0-epoch-40-avg-10-context-2-max-sym-per-frame-1-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ greedy_search_blank_penalty_0.0 4.54
decoding_results/modified_beam_search/errs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/modified_beam_search/errs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/modified_beam_search/log-decode-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model-2024-01-16-10-03-43 ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-01-16 10:03:43,643 INFO [decode.py:673] Decoding started
2
+ 2024-01-16 10:03:43,644 INFO [decode.py:679] Device: cuda:0
3
+ 2024-01-16 10:03:43,716 INFO [lexicon.py:168] Loading pre-compiled data/lang_bbpe_500/Linv.pt
4
+ 2024-01-16 10:03:43,728 INFO [decode.py:691] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '2989b0b1186fa6022932804f5b39fbb2781ebf42', 'k2-git-date': 'Fri Nov 24 11:34:10 2023', 'lhotse-version': '1.19.0.dev+git.d1ae9c05.dirty', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev/aishell-zipformer-bbpe', 'icefall-git-sha1': 'bce81394-clean', 'icefall-git-date': 'Thu Jan 11 09:56:01 2024', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.4.dev20231207+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.19.0.dev0+git.d1ae9c05.dirty-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1207150822-75498b8c5f-55j4z', 'IP address': '10.177.74.211'}, 'epoch': 40, 'iter': 0, 'avg': 10, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan'), 'bpe_model': 'data/lang_bbpe_500/bbpe.model', 'lang_dir': PosixPath('data/lang_bbpe_500'), 'decoding_method': 'modified_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'ilme_scale': 0.2, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'blank_penalty': 0.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'res_dir': PosixPath('zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/modified_beam_search'), 'suffix': 'epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
5
+ 2024-01-16 10:03:43,728 INFO [decode.py:693] About to create model
6
+ 2024-01-16 10:03:44,248 INFO [decode.py:760] Calculating the averaged model over epoch range from 30 (excluded) to 40
7
+ 2024-01-16 10:03:54,889 INFO [decode.py:791] Number of model parameters: 65549011
8
+ 2024-01-16 10:03:54,889 INFO [asr_datamodule.py:371] About to get dev cuts
9
+ 2024-01-16 10:03:54,905 INFO [asr_datamodule.py:312] About to create dev dataset
10
+ 2024-01-16 10:03:55,418 INFO [asr_datamodule.py:329] About to create dev dataloader
11
+ 2024-01-16 10:03:55,419 INFO [asr_datamodule.py:376] About to get test cuts
12
+ 2024-01-16 10:03:55,436 INFO [asr_datamodule.py:341] About to create test dataset
13
+ 2024-01-16 10:03:58,334 INFO [decode.py:564] batch 0/?, cuts processed until now is 35
14
+ 2024-01-16 10:04:28,300 INFO [decode.py:564] batch 20/?, cuts processed until now is 785
15
+ 2024-01-16 10:05:41,488 INFO [decode.py:564] batch 40/?, cuts processed until now is 1572
16
+ 2024-01-16 10:06:59,580 INFO [decode.py:564] batch 60/?, cuts processed until now is 2433
17
+ 2024-01-16 10:07:49,465 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.3113, 4.4908, 4.7379, 2.8855], device='cuda:0')
18
+ 2024-01-16 10:08:27,463 INFO [decode.py:564] batch 80/?, cuts processed until now is 3154
19
+ 2024-01-16 10:09:53,050 INFO [decode.py:564] batch 100/?, cuts processed until now is 4023
20
+ 2024-01-16 10:11:23,163 INFO [decode.py:564] batch 120/?, cuts processed until now is 4776
21
+ 2024-01-16 10:12:52,831 INFO [decode.py:564] batch 140/?, cuts processed until now is 5662
22
+ 2024-01-16 10:14:15,511 INFO [decode.py:564] batch 160/?, cuts processed until now is 6552
23
+ 2024-01-16 10:15:01,893 INFO [decode.py:564] batch 180/?, cuts processed until now is 7386
24
+ 2024-01-16 10:16:36,401 INFO [decode.py:564] batch 200/?, cuts processed until now is 8204
25
+ 2024-01-16 10:17:53,793 INFO [decode.py:564] batch 220/?, cuts processed until now is 8999
26
+ 2024-01-16 10:18:22,416 INFO [decode.py:564] batch 240/?, cuts processed until now is 9811
27
+ 2024-01-16 10:18:28,039 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.0271, 4.7459, 4.7175, 4.5908], device='cuda:0')
28
+ 2024-01-16 10:18:29,565 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.0997, 3.4467, 3.4164, 2.4607], device='cuda:0')
29
+ 2024-01-16 10:18:51,316 INFO [decode.py:564] batch 260/?, cuts processed until now is 10620
30
+ 2024-01-16 10:19:19,767 INFO [decode.py:564] batch 280/?, cuts processed until now is 11537
31
+ 2024-01-16 10:19:27,188 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.6404, 2.3883, 3.2457, 2.5336, 2.9562, 2.8679, 3.1428, 2.8781],
32
+ device='cuda:0')
33
+ 2024-01-16 10:19:48,713 INFO [decode.py:564] batch 300/?, cuts processed until now is 12391
34
+ 2024-01-16 10:20:18,251 INFO [decode.py:564] batch 320/?, cuts processed until now is 13152
35
+ 2024-01-16 10:20:47,711 INFO [decode.py:564] batch 340/?, cuts processed until now is 13879
36
+ 2024-01-16 10:21:08,641 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/modified_beam_search/recogs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt
37
+ 2024-01-16 10:21:09,092 INFO [utils.py:565] [dev-beam_size_4_blank_penalty_0.0] %WER 4.11% [8436 / 205341, 211 ins, 404 del, 7821 sub ]
38
+ 2024-01-16 10:21:09,789 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/modified_beam_search/errs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt
39
+ 2024-01-16 10:21:09,792 INFO [decode.py:614]
40
+ For dev, WER of different settings are:
41
+ beam_size_4_blank_penalty_0.0 4.11 best for dev
42
+
43
+ 2024-01-16 10:21:12,237 INFO [decode.py:564] batch 0/?, cuts processed until now is 31
44
+ 2024-01-16 10:21:42,215 INFO [decode.py:564] batch 20/?, cuts processed until now is 697
45
+ 2024-01-16 10:22:11,516 INFO [decode.py:564] batch 40/?, cuts processed until now is 1395
46
+ 2024-01-16 10:23:08,740 INFO [decode.py:564] batch 60/?, cuts processed until now is 2169
47
+ 2024-01-16 10:23:10,319 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.8814, 4.2700, 4.5619, 3.9857], device='cuda:0')
48
+ 2024-01-16 10:23:38,833 INFO [decode.py:564] batch 80/?, cuts processed until now is 2801
49
+ 2024-01-16 10:24:07,941 INFO [decode.py:564] batch 100/?, cuts processed until now is 3578
50
+ 2024-01-16 10:24:37,343 INFO [decode.py:564] batch 120/?, cuts processed until now is 4249
51
+ 2024-01-16 10:25:01,661 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9999, 1.6139, 2.6933, 2.7164, 2.7188, 2.4630, 2.3008, 2.6660],
52
+ device='cuda:0')
53
+ 2024-01-16 10:25:05,856 INFO [decode.py:564] batch 140/?, cuts processed until now is 5061
54
+ 2024-01-16 10:25:34,588 INFO [decode.py:564] batch 160/?, cuts processed until now is 5875
55
+ 2024-01-16 10:25:36,066 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.7714, 4.1238, 4.3947, 4.4562], device='cuda:0')
56
+ 2024-01-16 10:25:59,475 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.3344, 2.0402, 2.0603, 3.8063], device='cuda:0')
57
+ 2024-01-16 10:26:03,697 INFO [decode.py:564] batch 180/?, cuts processed until now is 6581
58
+ 2024-01-16 10:26:27,635 INFO [decode.py:564] batch 200/?, cuts processed until now is 7122
59
+ 2024-01-16 10:26:30,390 INFO [decode.py:580] The transcripts are stored in zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/modified_beam_search/recogs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt
60
+ 2024-01-16 10:26:30,610 INFO [utils.py:565] [test-beam_size_4_blank_penalty_0.0] %WER 4.37% [4583 / 104765, 96 ins, 262 del, 4225 sub ]
61
+ 2024-01-16 10:26:30,971 INFO [decode.py:598] Wrote detailed error stats to zipformer_bbpe/exp-context-size-2-lr-epochs-10-spec-aug-20-disable-musan/modified_beam_search/errs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt
62
+ 2024-01-16 10:26:30,974 INFO [decode.py:614]
63
+ For test, WER of different settings are:
64
+ beam_size_4_blank_penalty_0.0 4.37 best for test
65
+
66
+ 2024-01-16 10:26:30,982 INFO [decode.py:832] Done!
decoding_results/modified_beam_search/recogs-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/modified_beam_search/recogs-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/modified_beam_search/wer-summary-dev-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_size_4_blank_penalty_0.0 4.11
decoding_results/modified_beam_search/wer-summary-test-beam_size_4_blank_penalty_0.0-epoch-40-avg-10-modified_beam_search-beam-size-4-blank-penalty-0.0-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_size_4_blank_penalty_0.0 4.37
exp/.gitattributes ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ encoder-epoch-40-avg-10.onnx filter=lfs diff=lfs merge=lfs -text
2
+ joiner-epoch-40-avg-10.int8.onnx filter=lfs diff=lfs merge=lfs -text
3
+ joiner-epoch-40-avg-10.onnx filter=lfs diff=lfs merge=lfs -text
4
+ decoder-epoch-40-avg-10.int8.onnx filter=lfs diff=lfs merge=lfs -text
5
+ decoder-epoch-40-avg-10.onnx filter=lfs diff=lfs merge=lfs -text
6
+ encoder-epoch-40-avg-10.int8.onnx filter=lfs diff=lfs merge=lfs -text
exp/decoder-epoch-40-avg-10.int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28af835100c308c054995a2796658ad8ed6688e9d282a66525ad138383f6b3ae
3
+ size 540158
exp/decoder-epoch-40-avg-10.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e13cc842bbe5dd0a78c0a9537a0a1204546a8c402be7542a8d4346b8058c025b
3
+ size 2093097
exp/encoder-epoch-40-avg-10.int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86466a93b311ae97c682714dcc4ad2f950f46868f64be11a23c9a3d0f2364769
3
+ size 69037476
exp/encoder-epoch-40-avg-10.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f74bd4c3c37c4a5e3a23eef5f482594754d72340ea6e6d670e9739955e439e92
3
+ size 260000054
exp/epoch-40.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c3fb2f75b80dbef47d4d352f2d35487e915476fccabbba67842afda99ba812c
3
+ size 1049765648
exp/jit_script.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e9df2da8fee40df1e0161716b63d1736cd9d1447994eea196bbaa47f05cf7cb
3
+ size 264930494
exp/joiner-epoch-40-avg-10.int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dde9191d647df89d65afb68b7d9105d8820dfcfee9583482e1f4d04442cf135
3
+ size 259479
exp/joiner-epoch-40-avg-10.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0af6e5d7b8d589a84c6252c3db33b2e94eb84ba05d81b56f435599a20891a4c8
3
+ size 1026451
exp/log/log-train-2024-01-15-11-07-41-0 ADDED
The diff for this file is too large to render. See raw diff
 
exp/log/log-train-2024-01-15-11-07-41-1 ADDED
The diff for this file is too large to render. See raw diff
 
exp/pretrained.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:890aa98194100fb1edf6f5475c16688c7f6e1a3e059f522d87e221c5bec59976
3
+ size 262602995
exp/tensorboard/events.out.tfevents.1705288061.de-74279-k2-train-1-1207150822-75498b8c5f-55j4z.2432177.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:111733bf4d146be7b42f554de45c111b3281473bfca0404c492c4f41bda10724
3
+ size 332370
test_wavs/DEV_T0000000000.wav ADDED
Binary file (180 kB). View file