yfyeung commited on
Commit
8a15837
1 Parent(s): 23adcb0
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +4 -0
  2. data/lang_bpe_500/HLG.pt +3 -0
  3. data/lang_bpe_500/L.pt +3 -0
  4. data/lang_bpe_500/LG.pt +3 -0
  5. data/lang_bpe_500/Linv.pt +3 -0
  6. data/lang_bpe_500/bpe.model +3 -0
  7. data/lang_bpe_500/lexicon.txt +0 -0
  8. data/lang_bpe_500/tokens.txt +502 -0
  9. data/lang_bpe_500/words.txt +0 -0
  10. data/lm/G_4_gram.pt +3 -0
  11. decoding-results/ctc-decoding/errs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +0 -0
  12. decoding-results/ctc-decoding/errs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +0 -0
  13. decoding-results/ctc-decoding/log-decode-epoch-30-avg-10-use-averaged-model-2022-12-14-15-17-22 +27 -0
  14. decoding-results/ctc-decoding/log-decode-epoch-30-avg-10-use-averaged-model-2022-12-14-15-24-23 +27 -0
  15. decoding-results/ctc-decoding/recogs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +0 -0
  16. decoding-results/ctc-decoding/recogs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +0 -0
  17. decoding-results/ctc-decoding/wer-summary-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +2 -0
  18. decoding-results/ctc-decoding/wer-summary-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt +2 -0
  19. decoding-results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +0 -0
  20. decoding-results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +0 -0
  21. decoding-results/fast_beam_search/log-decode-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model-2022-12-13-15-17-20 +40 -0
  22. decoding-results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +0 -0
  23. decoding-results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +0 -0
  24. decoding-results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +2 -0
  25. decoding-results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt +2 -0
  26. decoding-results/greedy_search/errs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +0 -0
  27. decoding-results/greedy_search/errs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +0 -0
  28. decoding-results/greedy_search/log-decode-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model-2022-12-14-16-40-31 +46 -0
  29. decoding-results/greedy_search/recogs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +0 -0
  30. decoding-results/greedy_search/recogs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +0 -0
  31. decoding-results/greedy_search/wer-summary-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +2 -0
  32. decoding-results/greedy_search/wer-summary-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt +2 -0
  33. decoding-results/modified_beam_search/errs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +0 -0
  34. decoding-results/modified_beam_search/errs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +0 -0
  35. decoding-results/modified_beam_search/log-decode-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model-2022-12-14-16-41-57 +35 -0
  36. decoding-results/modified_beam_search/recogs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +0 -0
  37. decoding-results/modified_beam_search/recogs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +0 -0
  38. decoding-results/modified_beam_search/wer-summary-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +2 -0
  39. decoding-results/modified_beam_search/wer-summary-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt +2 -0
  40. exp/cpu_jit.pt +3 -0
  41. exp/decode.sh +31 -0
  42. exp/epoch-30.pt +3 -0
  43. exp/export.sh +10 -0
  44. exp/log/log-train-2022-12-12-18-45-28-0 +0 -0
  45. exp/log/log-train-2022-12-12-18-45-28-1 +0 -0
  46. exp/log/log-train-2022-12-12-18-45-28-2 +0 -0
  47. exp/log/log-train-2022-12-12-18-45-28-3 +0 -0
  48. exp/pretrained.pt +3 -0
  49. exp/run.sh +11 -0
  50. exp/tensorboard/events.out.tfevents.1670841928.d3-hpc-sjtu-test-004.143030.0 +3 -0
README.md CHANGED
@@ -1,3 +1,7 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ Introduction
5
+ This repo contains pre-trained models, checkpoints, training logs and decoding results of the following pull-request:
6
+
7
+ https://github.com/k2-fsa/icefall/pull/730
data/lang_bpe_500/HLG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ec0554f89485063a4f6fb314843416ee8f7c2c5c71dcb5e375987bc8d50f00a
3
+ size 845007559
data/lang_bpe_500/L.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f513bf1f243d9a9cf40eaeaf871d2cf3c76bcfcf678a1c8297bab1ce00b9e240
3
+ size 19025679
data/lang_bpe_500/LG.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f9b5668d3a11f4bc78044ca8e62a3186b3de5a5c06df56fd582759770515259
3
+ size 249852170
data/lang_bpe_500/Linv.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:806b26522f4558068d45a5ba67cc782da293422c095e5c464957bb2e932f5ed1
3
+ size 19025691
data/lang_bpe_500/bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c53433de083c4a6ad12d034550ef22de68cec62c4f58932a7b6b8b2f1e743fa5
3
+ size 244865
data/lang_bpe_500/lexicon.txt ADDED
The diff for this file is too large to render. See raw diff
 
data/lang_bpe_500/tokens.txt ADDED
@@ -0,0 +1,502 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <blk> 0
2
+ <sos/eos> 1
3
+ <unk> 2
4
+ S 3
5
+ ▁THE 4
6
+ ▁A 5
7
+ T 6
8
+ ▁AND 7
9
+ ED 8
10
+ ▁OF 9
11
+ ▁TO 10
12
+ E 11
13
+ D 12
14
+ N 13
15
+ ING 14
16
+ ▁IN 15
17
+ Y 16
18
+ M 17
19
+ C 18
20
+ ▁I 19
21
+ A 20
22
+ P 21
23
+ ▁HE 22
24
+ R 23
25
+ O 24
26
+ L 25
27
+ RE 26
28
+ I 27
29
+ U 28
30
+ ER 29
31
+ ▁IT 30
32
+ LY 31
33
+ ▁THAT 32
34
+ ▁WAS 33
35
+ ▁ 34
36
+ ▁S 35
37
+ AR 36
38
+ ▁BE 37
39
+ F 38
40
+ ▁C 39
41
+ IN 40
42
+ B 41
43
+ ▁FOR 42
44
+ OR 43
45
+ LE 44
46
+ ' 45
47
+ ▁HIS 46
48
+ ▁YOU 47
49
+ AL 48
50
+ ▁RE 49
51
+ V 50
52
+ ▁B 51
53
+ G 52
54
+ RI 53
55
+ ▁E 54
56
+ ▁WITH 55
57
+ ▁T 56
58
+ ▁AS 57
59
+ LL 58
60
+ ▁P 59
61
+ ▁HER 60
62
+ ST 61
63
+ ▁HAD 62
64
+ ▁SO 63
65
+ ▁F 64
66
+ W 65
67
+ CE 66
68
+ ▁IS 67
69
+ ND 68
70
+ ▁NOT 69
71
+ TH 70
72
+ ▁BUT 71
73
+ EN 72
74
+ ▁SHE 73
75
+ ▁ON 74
76
+ VE 75
77
+ ON 76
78
+ SE 77
79
+ ▁DE 78
80
+ UR 79
81
+ ▁G 80
82
+ CH 81
83
+ K 82
84
+ TER 83
85
+ ▁AT 84
86
+ IT 85
87
+ ▁ME 86
88
+ RO 87
89
+ NE 88
90
+ RA 89
91
+ ES 90
92
+ IL 91
93
+ NG 92
94
+ IC 93
95
+ ▁NO 94
96
+ ▁HIM 95
97
+ ENT 96
98
+ IR 97
99
+ ▁WE 98
100
+ H 99
101
+ ▁DO 100
102
+ ▁ALL 101
103
+ ▁HAVE 102
104
+ LO 103
105
+ ▁BY 104
106
+ ▁MY 105
107
+ ▁MO 106
108
+ ▁THIS 107
109
+ LA 108
110
+ ▁ST 109
111
+ ▁WHICH 110
112
+ ▁CON 111
113
+ ▁THEY 112
114
+ CK 113
115
+ TE 114
116
+ ▁SAID 115
117
+ ▁FROM 116
118
+ ▁GO 117
119
+ ▁WHO 118
120
+ ▁TH 119
121
+ ▁OR 120
122
+ ▁D 121
123
+ ▁W 122
124
+ VER 123
125
+ LI 124
126
+ ▁SE 125
127
+ ▁ONE 126
128
+ ▁CA 127
129
+ ▁AN 128
130
+ ▁LA 129
131
+ ▁WERE 130
132
+ EL 131
133
+ ▁HA 132
134
+ ▁MAN 133
135
+ ▁FA 134
136
+ ▁EX 135
137
+ AD 136
138
+ ▁SU 137
139
+ RY 138
140
+ ▁MI 139
141
+ AT 140
142
+ ▁BO 141
143
+ ▁WHEN 142
144
+ AN 143
145
+ THER 144
146
+ PP 145
147
+ ATION 146
148
+ ▁FI 147
149
+ ▁WOULD 148
150
+ ▁PRO 149
151
+ OW 150
152
+ ET 151
153
+ ▁O 152
154
+ ▁THERE 153
155
+ ▁HO 154
156
+ ION 155
157
+ ▁WHAT 156
158
+ ▁FE 157
159
+ ▁PA 158
160
+ US 159
161
+ MENT 160
162
+ ▁MA 161
163
+ UT 162
164
+ ▁OUT 163
165
+ ▁THEIR 164
166
+ ▁IF 165
167
+ ▁LI 166
168
+ ▁K 167
169
+ ▁WILL 168
170
+ ▁ARE 169
171
+ ID 170
172
+ ▁RO 171
173
+ DE 172
174
+ TION 173
175
+ ▁WA 174
176
+ PE 175
177
+ ▁UP 176
178
+ ▁SP 177
179
+ ▁PO 178
180
+ IGHT 179
181
+ ▁UN 180
182
+ RU 181
183
+ ▁LO 182
184
+ AS 183
185
+ OL 184
186
+ ▁LE 185
187
+ ▁BEEN 186
188
+ ▁SH 187
189
+ ▁RA 188
190
+ ▁SEE 189
191
+ KE 190
192
+ UL 191
193
+ TED 192
194
+ ▁SA 193
195
+ UN 194
196
+ UND 195
197
+ ANT 196
198
+ ▁NE 197
199
+ IS 198
200
+ ▁THEM 199
201
+ CI 200
202
+ GE 201
203
+ ▁COULD 202
204
+ ▁DIS 203
205
+ OM 204
206
+ ISH 205
207
+ HE 206
208
+ EST 207
209
+ ▁SOME 208
210
+ ENCE 209
211
+ ITY 210
212
+ IVE 211
213
+ ▁US 212
214
+ ▁MORE 213
215
+ ▁EN 214
216
+ ARD 215
217
+ ATE 216
218
+ ▁YOUR 217
219
+ ▁INTO 218
220
+ ▁KNOW 219
221
+ ▁CO 220
222
+ ANCE 221
223
+ ▁TIME 222
224
+ ▁WI 223
225
+ ▁YE 224
226
+ AGE 225
227
+ ▁NOW 226
228
+ TI 227
229
+ FF 228
230
+ ABLE 229
231
+ ▁VERY 230
232
+ ▁LIKE 231
233
+ AM 232
234
+ HI 233
235
+ Z 234
236
+ ▁OTHER 235
237
+ ▁THAN 236
238
+ ▁LITTLE 237
239
+ ▁DID 238
240
+ ▁LOOK 239
241
+ TY 240
242
+ ERS 241
243
+ ▁CAN 242
244
+ ▁CHA 243
245
+ ▁AR 244
246
+ X 245
247
+ FUL 246
248
+ UGH 247
249
+ ▁BA 248
250
+ ▁DAY 249
251
+ ▁ABOUT 250
252
+ TEN 251
253
+ IM 252
254
+ ▁ANY 253
255
+ ▁PRE 254
256
+ ▁OVER 255
257
+ IES 256
258
+ NESS 257
259
+ ME 258
260
+ BLE 259
261
+ ▁M 260
262
+ ROW 261
263
+ ▁HAS 262
264
+ ▁GREAT 263
265
+ ▁VI 264
266
+ TA 265
267
+ ▁AFTER 266
268
+ PER 267
269
+ ▁AGAIN 268
270
+ HO 269
271
+ SH 270
272
+ ▁UPON 271
273
+ ▁DI 272
274
+ ▁HAND 273
275
+ ▁COM 274
276
+ IST 275
277
+ TURE 276
278
+ ▁STA 277
279
+ ▁THEN 278
280
+ ▁SHOULD 279
281
+ ▁GA 280
282
+ OUS 281
283
+ OUR 282
284
+ ▁WELL 283
285
+ ▁ONLY 284
286
+ MAN 285
287
+ ▁GOOD 286
288
+ ▁TWO 287
289
+ ▁MAR 288
290
+ ▁SAY 289
291
+ ▁HU 290
292
+ TING 291
293
+ ▁OUR 292
294
+ RESS 293
295
+ ▁DOWN 294
296
+ IOUS 295
297
+ ▁BEFORE 296
298
+ ▁DA 297
299
+ ▁NA 298
300
+ QUI 299
301
+ ▁MADE 300
302
+ ▁EVERY 301
303
+ ▁OLD 302
304
+ ▁EVEN 303
305
+ IG 304
306
+ ▁COME 305
307
+ ▁GRA 306
308
+ ▁RI 307
309
+ ▁LONG 308
310
+ OT 309
311
+ SIDE 310
312
+ WARD 311
313
+ ▁FO 312
314
+ ▁WHERE 313
315
+ MO 314
316
+ LESS 315
317
+ ▁SC 316
318
+ ▁MUST 317
319
+ ▁NEVER 318
320
+ ▁HOW 319
321
+ ▁CAME 320
322
+ ▁SUCH 321
323
+ ▁RU 322
324
+ ▁TAKE 323
325
+ ▁WO 324
326
+ ▁CAR 325
327
+ UM 326
328
+ AK 327
329
+ ▁THINK 328
330
+ ▁MUCH 329
331
+ ▁MISTER 330
332
+ ▁MAY 331
333
+ ▁JO 332
334
+ ▁WAY 333
335
+ ▁COMP 334
336
+ ▁THOUGHT 335
337
+ ▁STO 336
338
+ ▁MEN 337
339
+ ▁BACK 338
340
+ ▁DON 339
341
+ J 340
342
+ ▁LET 341
343
+ ▁TRA 342
344
+ ▁FIRST 343
345
+ ▁JUST 344
346
+ ▁VA 345
347
+ ▁OWN 346
348
+ ▁PLA 347
349
+ ▁MAKE 348
350
+ ATED 349
351
+ ▁HIMSELF 350
352
+ ▁WENT 351
353
+ ▁PI 352
354
+ GG 353
355
+ RING 354
356
+ ▁DU 355
357
+ ▁MIGHT 356
358
+ ▁PART 357
359
+ ▁GIVE 358
360
+ ▁IMP 359
361
+ ▁BU 360
362
+ ▁PER 361
363
+ ▁PLACE 362
364
+ ▁HOUSE 363
365
+ ▁THROUGH 364
366
+ IAN 365
367
+ ▁SW 366
368
+ ▁UNDER 367
369
+ QUE 368
370
+ ▁AWAY 369
371
+ ▁LOVE 370
372
+ QUA 371
373
+ ▁LIFE 372
374
+ ▁GET 373
375
+ ▁WITHOUT 374
376
+ ▁PASS 375
377
+ ▁TURN 376
378
+ IGN 377
379
+ ▁HEAD 378
380
+ ▁MOST 379
381
+ ▁THOSE 380
382
+ ▁SHALL 381
383
+ ▁EYES 382
384
+ ▁COL 383
385
+ ▁STILL 384
386
+ ▁NIGHT 385
387
+ ▁NOTHING 386
388
+ ITION 387
389
+ HA 388
390
+ ▁TELL 389
391
+ ▁WORK 390
392
+ ▁LAST 391
393
+ ▁NEW 392
394
+ ▁FACE 393
395
+ ▁HI 394
396
+ ▁WORD 395
397
+ ▁FOUND 396
398
+ ▁COUNT 397
399
+ ▁OB 398
400
+ ▁WHILE 399
401
+ ▁SHA 400
402
+ ▁MEAN 401
403
+ ▁SAW 402
404
+ ▁PEOPLE 403
405
+ ▁FRIEND 404
406
+ ▁THREE 405
407
+ ▁ROOM 406
408
+ ▁SAME 407
409
+ ▁THOUGH 408
410
+ ▁RIGHT 409
411
+ ▁CHILD 410
412
+ ▁FATHER 411
413
+ ▁ANOTHER 412
414
+ ▁HEART 413
415
+ ▁WANT 414
416
+ ▁TOOK 415
417
+ OOK 416
418
+ ▁LIGHT 417
419
+ ▁MISSUS 418
420
+ ▁OPEN 419
421
+ ▁JU 420
422
+ ▁ASKED 421
423
+ PORT 422
424
+ ▁LEFT 423
425
+ ▁JA 424
426
+ ▁WORLD 425
427
+ ▁HOME 426
428
+ ▁WHY 427
429
+ ▁ALWAYS 428
430
+ ▁ANSWER 429
431
+ ▁SEEMED 430
432
+ ▁SOMETHING 431
433
+ ▁GIRL 432
434
+ ▁BECAUSE 433
435
+ ▁NAME 434
436
+ ▁TOLD 435
437
+ ▁NI 436
438
+ ▁HIGH 437
439
+ IZE 438
440
+ ▁WOMAN 439
441
+ ▁FOLLOW 440
442
+ ▁RETURN 441
443
+ ▁KNEW 442
444
+ ▁EACH 443
445
+ ▁KIND 444
446
+ ▁JE 445
447
+ ▁ACT 446
448
+ ▁LU 447
449
+ ▁CERTAIN 448
450
+ ▁YEARS 449
451
+ ▁QUITE 450
452
+ ▁APPEAR 451
453
+ ▁BETTER 452
454
+ ▁HALF 453
455
+ ▁PRESENT 454
456
+ ▁PRINCE 455
457
+ SHIP 456
458
+ ▁ALSO 457
459
+ ▁BEGAN 458
460
+ ▁HAVING 459
461
+ ▁ENOUGH 460
462
+ ▁PERSON 461
463
+ ▁LADY 462
464
+ ▁WHITE 463
465
+ ▁COURSE 464
466
+ ▁VOICE 465
467
+ ▁SPEAK 466
468
+ ▁POWER 467
469
+ ▁MORNING 468
470
+ ▁BETWEEN 469
471
+ ▁AMONG 470
472
+ ▁KEEP 471
473
+ ▁WALK 472
474
+ ▁MATTER 473
475
+ ▁TEA 474
476
+ ▁BELIEVE 475
477
+ ▁SMALL 476
478
+ ▁TALK 477
479
+ ▁FELT 478
480
+ ▁HORSE 479
481
+ ▁MYSELF 480
482
+ ▁SIX 481
483
+ ▁HOWEVER 482
484
+ ▁FULL 483
485
+ ▁HERSELF 484
486
+ ▁POINT 485
487
+ ▁STOOD 486
488
+ ▁HUNDRED 487
489
+ ▁ALMOST 488
490
+ ▁SINCE 489
491
+ ▁LARGE 490
492
+ ▁LEAVE 491
493
+ ▁PERHAPS 492
494
+ ▁DARK 493
495
+ ▁SUDDEN 494
496
+ ▁REPLIED 495
497
+ ▁ANYTHING 496
498
+ ▁WONDER 497
499
+ ▁UNTIL 498
500
+ Q 499
501
+ #0 500
502
+ #1 501
data/lang_bpe_500/words.txt ADDED
The diff for this file is too large to render. See raw diff
 
data/lm/G_4_gram.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff7d17fc8bf0482b9f0cb9113ea83612ab88df85118aba19c0b81238df80d8ca
3
+ size 3700956590
decoding-results/ctc-decoding/errs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/ctc-decoding/errs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/ctc-decoding/log-decode-epoch-30-avg-10-use-averaged-model-2022-12-14-15-17-22 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-12-14 15:17:22,943 INFO [ctc_decode.py:608] Decoding started
2
+ 2022-12-14 15:17:22,944 INFO [ctc_decode.py:614] Device: cuda:0
3
+ 2022-12-14 15:17:22,944 INFO [ctc_decode.py:615] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'beam_size': 10, 'use_double_scores': True, 'warm_step': 2000, 'env_info': {'k2-version': '1.22', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '6df2d56bd9097bba8d8af12d6c1ef8cb66bf9c17', 'k2-git-date': 'Thu Nov 17 19:06:54 2022', 'lhotse-version': '1.10.0', 'torch-version': '1.13.0', 'torch-cuda-available': True, 'torch-cuda-version': '11.6', 'python-version': '3.1', 'icefall-git-branch': 'blankskip', 'icefall-git-sha1': 'cf69804-dirty', 'icefall-git-date': 'Sat Dec 3 16:30:31 2022', 'icefall-path': '/home/yfy62/icefall', 'k2-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/k2-1.22.dev20221122+cuda11.6.torch1.13.0-py3.10-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/lhotse/__init__.py', 'hostname': 'd3-hpc-sjtu-test-004', 'IP address': '10.11.11.11'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'epoch': 30, 'iter': 0, 'avg': 10, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 0.5, 'hlg_scale': 0.8, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 600, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding'), 'suffix': 'epoch-30-avg-10-use-averaged-model'}
4
+ 2022-12-14 15:17:23,300 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2022-12-14 15:17:24,967 INFO [ctc_decode.py:693] About to create model
6
+ 2022-12-14 15:17:25,311 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
7
+ 2022-12-14 15:17:25,325 INFO [ctc_decode.py:760] Calculating the averaged model over epoch range from 20 (excluded) to 30
8
+ 2022-12-14 15:17:28,896 INFO [ctc_decode.py:777] Number of model parameters: 71164387
9
+ 2022-12-14 15:17:28,896 INFO [asr_datamodule.py:443] About to get test-clean cuts
10
+ 2022-12-14 15:17:28,897 INFO [asr_datamodule.py:450] About to get test-other cuts
11
+ 2022-12-14 15:17:32,819 INFO [ctc_decode.py:526] batch 0/?, cuts processed until now is 43
12
+ 2022-12-14 15:17:53,968 INFO [ctc_decode.py:544] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/recogs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
13
+ 2022-12-14 15:17:54,111 INFO [utils.py:536] [test-clean-ctc-decoding] %WER 6.24% [3280 / 52576, 327 ins, 240 del, 2713 sub ]
14
+ 2022-12-14 15:17:54,339 INFO [ctc_decode.py:555] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/errs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
15
+ 2022-12-14 15:17:54,340 INFO [ctc_decode.py:572]
16
+ For test-clean, WER of different settings are:
17
+ ctc-decoding 6.24 best for test-clean
18
+
19
+ 2022-12-14 15:17:55,518 INFO [ctc_decode.py:526] batch 0/?, cuts processed until now is 52
20
+ 2022-12-14 15:18:17,880 INFO [ctc_decode.py:544] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/recogs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
21
+ 2022-12-14 15:18:17,966 INFO [utils.py:536] [test-other-ctc-decoding] %WER 16.97% [8883 / 52343, 842 ins, 805 del, 7236 sub ]
22
+ 2022-12-14 15:18:18,137 INFO [ctc_decode.py:555] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/errs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
23
+ 2022-12-14 15:18:18,138 INFO [ctc_decode.py:572]
24
+ For test-other, WER of different settings are:
25
+ ctc-decoding 16.97 best for test-other
26
+
27
+ 2022-12-14 15:18:18,138 INFO [ctc_decode.py:810] Done!
decoding-results/ctc-decoding/log-decode-epoch-30-avg-10-use-averaged-model-2022-12-14-15-24-23 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-12-14 15:24:23,674 INFO [ctc_decode.py:608] Decoding started
2
+ 2022-12-14 15:24:23,675 INFO [ctc_decode.py:614] Device: cuda:0
3
+ 2022-12-14 15:24:23,675 INFO [ctc_decode.py:615] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'beam_size': 10, 'use_double_scores': True, 'warm_step': 2000, 'env_info': {'k2-version': '1.22', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '6df2d56bd9097bba8d8af12d6c1ef8cb66bf9c17', 'k2-git-date': 'Thu Nov 17 19:06:54 2022', 'lhotse-version': '1.10.0', 'torch-version': '1.13.0', 'torch-cuda-available': True, 'torch-cuda-version': '11.6', 'python-version': '3.1', 'icefall-git-branch': 'blankskip', 'icefall-git-sha1': 'cf69804-dirty', 'icefall-git-date': 'Sat Dec 3 16:30:31 2022', 'icefall-path': '/home/yfy62/icefall', 'k2-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/k2-1.22.dev20221122+cuda11.6.torch1.13.0-py3.10-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/lhotse/__init__.py', 'hostname': 'd3-hpc-sjtu-test-004', 'IP address': '10.11.11.11'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'epoch': 30, 'iter': 0, 'avg': 10, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 0.5, 'hlg_scale': 0.8, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 600, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding'), 'suffix': 'epoch-30-avg-10-use-averaged-model'}
4
+ 2022-12-14 15:24:24,047 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2022-12-14 15:24:25,648 INFO [ctc_decode.py:693] About to create model
6
+ 2022-12-14 15:24:25,996 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
7
+ 2022-12-14 15:24:26,011 INFO [ctc_decode.py:760] Calculating the averaged model over epoch range from 20 (excluded) to 30
8
+ 2022-12-14 15:24:28,399 INFO [ctc_decode.py:777] Number of model parameters: 71164387
9
+ 2022-12-14 15:24:28,399 INFO [asr_datamodule.py:443] About to get test-clean cuts
10
+ 2022-12-14 15:24:28,400 INFO [asr_datamodule.py:450] About to get test-other cuts
11
+ 2022-12-14 15:24:32,204 INFO [ctc_decode.py:526] batch 0/?, cuts processed until now is 43
12
+ 2022-12-14 15:24:53,572 INFO [ctc_decode.py:544] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/recogs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
13
+ 2022-12-14 15:24:53,690 INFO [utils.py:536] [test-clean-ctc-decoding] %WER 6.24% [3280 / 52576, 327 ins, 240 del, 2713 sub ]
14
+ 2022-12-14 15:24:53,920 INFO [ctc_decode.py:555] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/errs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
15
+ 2022-12-14 15:24:53,920 INFO [ctc_decode.py:572]
16
+ For test-clean, WER of different settings are:
17
+ ctc-decoding 6.24 best for test-clean
18
+
19
+ 2022-12-14 15:24:55,073 INFO [ctc_decode.py:526] batch 0/?, cuts processed until now is 52
20
+ 2022-12-14 15:25:17,528 INFO [ctc_decode.py:544] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/recogs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
21
+ 2022-12-14 15:25:17,661 INFO [utils.py:536] [test-other-ctc-decoding] %WER 16.97% [8883 / 52343, 842 ins, 805 del, 7236 sub ]
22
+ 2022-12-14 15:25:17,856 INFO [ctc_decode.py:555] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/ctc-decoding/errs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt
23
+ 2022-12-14 15:25:17,856 INFO [ctc_decode.py:572]
24
+ For test-other, WER of different settings are:
25
+ ctc-decoding 16.97 best for test-other
26
+
27
+ 2022-12-14 15:25:17,856 INFO [ctc_decode.py:810] Done!
decoding-results/ctc-decoding/recogs-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/ctc-decoding/recogs-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/ctc-decoding/wer-summary-test-clean-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 6.24
decoding-results/ctc-decoding/wer-summary-test-other-ctc-decoding-epoch-30-avg-10-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 16.97
decoding-results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/fast_beam_search/log-decode-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model-2022-12-13-15-17-20 ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-12-13 15:17:20,132 INFO [ctc_guild_decode_bk.py:710] Decoding started
2
+ 2022-12-13 15:17:20,132 INFO [ctc_guild_decode_bk.py:716] Device: cuda:0
3
+ 2022-12-13 15:17:20,134 INFO [ctc_guild_decode_bk.py:731] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'beam_size': 4, 'use_double_scores': True, 'warm_step': 2000, 'env_info': {'k2-version': '1.22', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '6df2d56bd9097bba8d8af12d6c1ef8cb66bf9c17', 'k2-git-date': 'Thu Nov 17 19:06:54 2022', 'lhotse-version': '1.10.0', 'torch-version': '1.13.0', 'torch-cuda-available': True, 'torch-cuda-version': '11.6', 'python-version': '3.1', 'icefall-git-branch': 'blankskip', 'icefall-git-sha1': 'cf69804-dirty', 'icefall-git-date': 'Sat Dec 3 16:30:31 2022', 'icefall-path': '/home/yfy62/icefall', 'k2-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/k2-1.22.dev20221122+cuda11.6.torch1.13.0-py3.10-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/lhotse/__init__.py', 'hostname': 'd3-hpc-sjtu-test-004', 'IP address': '10.11.11.11'}, 'epoch': 30, 'iter': 0, 'avg': 13, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': False, 'decode_chunk_size': 16, 'left_context': 64, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 600, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/fast_beam_search'), 'suffix': 'epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-12-13 15:17:20,134 INFO [ctc_guild_decode_bk.py:733] About to create model
5
+ 2022-12-13 15:17:20,561 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
6
+ 2022-12-13 15:17:20,576 INFO [ctc_guild_decode_bk.py:800] Calculating the averaged model over epoch range from 17 (excluded) to 30
7
+ 2022-12-13 15:17:25,971 INFO [ctc_guild_decode_bk.py:836] Number of model parameters: 71164387
8
+ 2022-12-13 15:17:25,971 INFO [asr_datamodule.py:443] About to get test-clean cuts
9
+ 2022-12-13 15:17:25,972 INFO [asr_datamodule.py:450] About to get test-other cuts
10
+ 2022-12-13 15:17:31,374 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 43
11
+ 2022-12-13 15:17:44,998 INFO [ctc_guild_decode_bk.py:608] batch 20/?, cuts processed until now is 1434
12
+ 2022-12-13 15:17:57,210 INFO [ctc_guild_decode_bk.py:608] batch 40/?, cuts processed until now is 2574
13
+ 2022-12-13 15:17:57,816 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt
14
+ 2022-12-13 15:17:57,952 INFO [utils.py:536] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 7.72% [4061 / 52576, 331 ins, 585 del, 3145 sub ]
15
+ 2022-12-13 15:17:58,178 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt
16
+ 2022-12-13 15:17:58,179 INFO [ctc_guild_decode_bk.py:656]
17
+ For test-clean, WER of different settings are:
18
+ beam_20.0_max_contexts_8_max_states_64 7.72 best for test-clean
19
+
20
+ 2022-12-13 15:17:59,469 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 52
21
+ 2022-12-13 15:18:08,253 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.4268, 2.5728, 2.1033, 2.5628, 1.5892, 2.2766, 2.0166, 2.5159],
22
+ device='cuda:0'), covar=tensor([0.0206, 0.0204, 0.0389, 0.0143, 0.0417, 0.0170, 0.0340, 0.0311],
23
+ device='cuda:0'), in_proj_covar=tensor([0.0058, 0.0043, 0.0064, 0.0047, 0.0061, 0.0046, 0.0059, 0.0047],
24
+ device='cuda:0'), out_proj_covar=tensor([7.3030e-05, 5.2734e-05, 8.0003e-05, 5.6515e-05, 7.6844e-05, 5.6425e-05,
25
+ 7.4780e-05, 6.0481e-05], device='cuda:0')
26
+ 2022-12-13 15:18:12,258 INFO [ctc_guild_decode_bk.py:608] batch 20/?, cuts processed until now is 1646
27
+ 2022-12-13 15:18:23,247 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.6448, 3.1406, 2.3828, 2.9393, 1.9249, 2.7527, 2.2939, 2.9757],
28
+ device='cuda:0'), covar=tensor([0.0206, 0.0196, 0.0335, 0.0147, 0.0424, 0.0154, 0.0334, 0.0352],
29
+ device='cuda:0'), in_proj_covar=tensor([0.0058, 0.0043, 0.0064, 0.0047, 0.0061, 0.0046, 0.0059, 0.0047],
30
+ device='cuda:0'), out_proj_covar=tensor([7.3030e-05, 5.2734e-05, 8.0003e-05, 5.6515e-05, 7.6844e-05, 5.6425e-05,
31
+ 7.4780e-05, 6.0481e-05], device='cuda:0')
32
+ 2022-12-13 15:18:23,820 INFO [ctc_guild_decode_bk.py:608] batch 40/?, cuts processed until now is 2870
33
+ 2022-12-13 15:18:24,826 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt
34
+ 2022-12-13 15:18:24,985 INFO [utils.py:536] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 17.51% [9163 / 52343, 737 ins, 1362 del, 7064 sub ]
35
+ 2022-12-13 15:18:25,235 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt
36
+ 2022-12-13 15:18:25,235 INFO [ctc_guild_decode_bk.py:656]
37
+ For test-other, WER of different settings are:
38
+ beam_20.0_max_contexts_8_max_states_64 17.51 best for test-other
39
+
40
+ 2022-12-13 15:18:25,235 INFO [ctc_guild_decode_bk.py:867] Done!
decoding-results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 7.72
decoding-results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-30-avg-13-beam-20.0-max-contexts-8-max-states-64-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 17.51
decoding-results/greedy_search/errs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/greedy_search/errs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/greedy_search/log-decode-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model-2022-12-14-16-40-31 ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-12-14 16:40:31,275 INFO [ctc_guild_decode_bk.py:710] Decoding started
2
+ 2022-12-14 16:40:31,276 INFO [ctc_guild_decode_bk.py:716] Device: cuda:0
3
+ 2022-12-14 16:40:31,278 INFO [ctc_guild_decode_bk.py:731] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'beam_size': 4, 'use_double_scores': True, 'warm_step': 2000, 'env_info': {'k2-version': '1.22', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '6df2d56bd9097bba8d8af12d6c1ef8cb66bf9c17', 'k2-git-date': 'Thu Nov 17 19:06:54 2022', 'lhotse-version': '1.10.0', 'torch-version': '1.13.0', 'torch-cuda-available': True, 'torch-cuda-version': '11.6', 'python-version': '3.1', 'icefall-git-branch': 'blankskip', 'icefall-git-sha1': 'cf69804-dirty', 'icefall-git-date': 'Sat Dec 3 16:30:31 2022', 'icefall-path': '/home/yfy62/icefall', 'k2-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/k2-1.22.dev20221122+cuda11.6.torch1.13.0-py3.10-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/lhotse/__init__.py', 'hostname': 'd3-hpc-sjtu-test-004', 'IP address': '10.11.11.11'}, 'epoch': 30, 'iter': 0, 'avg': 13, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'greedy_search', 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': False, 'decode_chunk_size': 16, 'left_context': 64, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 600, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/greedy_search'), 'suffix': 'epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-12-14 16:40:31,278 INFO [ctc_guild_decode_bk.py:733] About to create model
5
+ 2022-12-14 16:40:31,774 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
6
+ 2022-12-14 16:40:31,794 INFO [ctc_guild_decode_bk.py:800] Calculating the averaged model over epoch range from 17 (excluded) to 30
7
+ 2022-12-14 16:40:36,459 INFO [ctc_guild_decode_bk.py:836] Number of model parameters: 71164387
8
+ 2022-12-14 16:40:36,459 INFO [asr_datamodule.py:443] About to get test-clean cuts
9
+ 2022-12-14 16:40:36,460 INFO [asr_datamodule.py:450] About to get test-other cuts
10
+ 2022-12-14 16:40:41,112 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 43
11
+ 2022-12-14 16:40:42,842 INFO [zipformer.py:1414] attn_weights_entropy = tensor([5.0845, 5.2382, 5.3612, 5.0017, 5.2092, 4.7563, 4.6855, 4.7861],
12
+ device='cuda:0'), covar=tensor([0.0388, 0.0189, 0.0125, 0.0220, 0.0208, 0.0217, 0.0272, 0.0309],
13
+ device='cuda:0'), in_proj_covar=tensor([0.0182, 0.0146, 0.0126, 0.0149, 0.0135, 0.0156, 0.0173, 0.0173],
14
+ device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0001, 0.0001, 0.0001, 0.0001, 0.0001, 0.0002, 0.0002],
15
+ device='cuda:0')
16
+ 2022-12-14 16:40:45,678 INFO [zipformer.py:1414] attn_weights_entropy = tensor([3.1129, 3.3022, 3.6867, 2.1906, 3.2836, 3.0820, 3.2884, 2.9239],
17
+ device='cuda:0'), covar=tensor([0.0854, 0.0538, 0.0214, 0.2138, 0.0361, 0.0956, 0.0574, 0.1603],
18
+ device='cuda:0'), in_proj_covar=tensor([0.0183, 0.0129, 0.0122, 0.0206, 0.0134, 0.0180, 0.0172, 0.0206],
19
+ device='cuda:0'), out_proj_covar=tensor([1.2879e-04, 9.0075e-05, 8.3072e-05, 1.4434e-04, 9.0510e-05, 1.2637e-04,
20
+ 1.1752e-04, 1.4144e-04], device='cuda:0')
21
+ 2022-12-14 16:40:53,272 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/greedy_search/recogs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt
22
+ 2022-12-14 16:40:53,345 INFO [utils.py:536] [test-clean-greedy_search] %WER 5.74% [3016 / 52576, 326 ins, 267 del, 2423 sub ]
23
+ 2022-12-14 16:40:53,491 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/greedy_search/errs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt
24
+ 2022-12-14 16:40:53,491 INFO [ctc_guild_decode_bk.py:656]
25
+ For test-clean, WER of different settings are:
26
+ greedy_search 5.74 best for test-clean
27
+
28
+ 2022-12-14 16:40:54,313 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 52
29
+ 2022-12-14 16:40:54,451 INFO [zipformer.py:1414] attn_weights_entropy = tensor([4.4679, 4.4290, 4.5958, 5.0097, 4.3190, 4.8124, 4.5159, 4.5238],
30
+ device='cuda:0'), covar=tensor([0.0281, 0.0263, 0.0248, 0.0141, 0.0244, 0.0115, 0.0241, 0.0221],
31
+ device='cuda:0'), in_proj_covar=tensor([0.0103, 0.0086, 0.0093, 0.0085, 0.0073, 0.0079, 0.0079, 0.0085],
32
+ device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0001, 0.0002, 0.0001, 0.0001, 0.0001, 0.0001, 0.0001],
33
+ device='cuda:0')
34
+ 2022-12-14 16:40:55,386 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.8303, 2.5061, 2.4607, 3.5364, 3.2174, 2.9463, 3.1359, 3.0660],
35
+ device='cuda:0'), covar=tensor([0.0377, 0.0609, 0.0640, 0.0145, 0.0443, 0.0493, 0.0152, 0.0411],
36
+ device='cuda:0'), in_proj_covar=tensor([0.0073, 0.0065, 0.0084, 0.0054, 0.0053, 0.0054, 0.0063, 0.0053],
37
+ device='cuda:0'), out_proj_covar=tensor([5.8793e-05, 5.3987e-05, 8.1245e-05, 3.9913e-05, 4.5528e-05, 4.6460e-05,
38
+ 4.7543e-05, 4.2467e-05], device='cuda:0')
39
+ 2022-12-14 16:41:06,368 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/greedy_search/recogs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt
40
+ 2022-12-14 16:41:06,453 INFO [utils.py:536] [test-other-greedy_search] %WER 15.46% [8093 / 52343, 758 ins, 887 del, 6448 sub ]
41
+ 2022-12-14 16:41:06,620 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/greedy_search/errs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt
42
+ 2022-12-14 16:41:06,620 INFO [ctc_guild_decode_bk.py:656]
43
+ For test-other, WER of different settings are:
44
+ greedy_search 15.46 best for test-other
45
+
46
+ 2022-12-14 16:41:06,620 INFO [ctc_guild_decode_bk.py:867] Done!
decoding-results/greedy_search/recogs-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/greedy_search/recogs-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/greedy_search/wer-summary-test-clean-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ greedy_search 5.74
decoding-results/greedy_search/wer-summary-test-other-greedy_search-epoch-30-avg-13-context-2-max-sym-per-frame-1-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ greedy_search 15.46
decoding-results/modified_beam_search/errs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/modified_beam_search/errs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/modified_beam_search/log-decode-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model-2022-12-14-16-41-57 ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-12-14 16:41:57,346 INFO [ctc_guild_decode_bk.py:710] Decoding started
2
+ 2022-12-14 16:41:57,346 INFO [ctc_guild_decode_bk.py:716] Device: cuda:0
3
+ 2022-12-14 16:41:57,348 INFO [ctc_guild_decode_bk.py:731] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'beam_size': 4, 'use_double_scores': True, 'warm_step': 2000, 'env_info': {'k2-version': '1.22', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '6df2d56bd9097bba8d8af12d6c1ef8cb66bf9c17', 'k2-git-date': 'Thu Nov 17 19:06:54 2022', 'lhotse-version': '1.10.0', 'torch-version': '1.13.0', 'torch-cuda-available': True, 'torch-cuda-version': '11.6', 'python-version': '3.1', 'icefall-git-branch': 'blankskip', 'icefall-git-sha1': 'cf69804-dirty', 'icefall-git-date': 'Sat Dec 3 16:30:31 2022', 'icefall-path': '/home/yfy62/icefall', 'k2-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/k2-1.22.dev20221122+cuda11.6.torch1.13.0-py3.10-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/home/yfy62/anaconda3/envs/icefall/lib/python3.10/site-packages/lhotse/__init__.py', 'hostname': 'd3-hpc-sjtu-test-004', 'IP address': '10.11.11.11'}, 'epoch': 30, 'iter': 0, 'avg': 13, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'modified_beam_search', 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': False, 'decode_chunk_size': 16, 'left_context': 64, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'full_libri': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 600, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/modified_beam_search'), 'suffix': 'epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-12-14 16:41:57,348 INFO [ctc_guild_decode_bk.py:733] About to create model
5
+ 2022-12-14 16:41:57,749 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
6
+ 2022-12-14 16:41:57,766 INFO [ctc_guild_decode_bk.py:800] Calculating the averaged model over epoch range from 17 (excluded) to 30
7
+ 2022-12-14 16:42:02,099 INFO [ctc_guild_decode_bk.py:836] Number of model parameters: 71164387
8
+ 2022-12-14 16:42:02,099 INFO [asr_datamodule.py:443] About to get test-clean cuts
9
+ 2022-12-14 16:42:02,100 INFO [asr_datamodule.py:450] About to get test-other cuts
10
+ 2022-12-14 16:42:07,791 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 43
11
+ 2022-12-14 16:42:28,479 INFO [ctc_guild_decode_bk.py:608] batch 20/?, cuts processed until now is 1434
12
+ 2022-12-14 16:42:44,750 INFO [ctc_guild_decode_bk.py:608] batch 40/?, cuts processed until now is 2574
13
+ 2022-12-14 16:42:45,603 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/modified_beam_search/recogs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt
14
+ 2022-12-14 16:42:45,722 INFO [utils.py:536] [test-clean-beam_size_4] %WER 5.66% [2978 / 52576, 335 ins, 241 del, 2402 sub ]
15
+ 2022-12-14 16:42:45,870 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/modified_beam_search/errs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt
16
+ 2022-12-14 16:42:45,870 INFO [ctc_guild_decode_bk.py:656]
17
+ For test-clean, WER of different settings are:
18
+ beam_size_4 5.66 best for test-clean
19
+
20
+ 2022-12-14 16:42:46,534 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.6509, 2.3458, 2.3456, 3.1348, 2.9894, 2.8388, 2.8162, 2.8047],
21
+ device='cuda:0'), covar=tensor([0.0352, 0.0657, 0.0608, 0.0179, 0.0396, 0.0467, 0.0155, 0.0386],
22
+ device='cuda:0'), in_proj_covar=tensor([0.0073, 0.0065, 0.0084, 0.0054, 0.0053, 0.0054, 0.0063, 0.0053],
23
+ device='cuda:0'), out_proj_covar=tensor([5.8793e-05, 5.3987e-05, 8.1245e-05, 3.9913e-05, 4.5528e-05, 4.6460e-05,
24
+ 4.7543e-05, 4.2467e-05], device='cuda:0')
25
+ 2022-12-14 16:42:47,542 INFO [ctc_guild_decode_bk.py:608] batch 0/?, cuts processed until now is 52
26
+ 2022-12-14 16:43:06,573 INFO [ctc_guild_decode_bk.py:608] batch 20/?, cuts processed until now is 1646
27
+ 2022-12-14 16:43:22,122 INFO [ctc_guild_decode_bk.py:608] batch 40/?, cuts processed until now is 2870
28
+ 2022-12-14 16:43:22,829 INFO [ctc_guild_decode_bk.py:626] The transcripts are stored in pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/modified_beam_search/recogs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt
29
+ 2022-12-14 16:43:22,906 INFO [utils.py:536] [test-other-beam_size_4] %WER 15.24% [7975 / 52343, 761 ins, 830 del, 6384 sub ]
30
+ 2022-12-14 16:43:23,086 INFO [ctc_guild_decode_bk.py:639] Wrote detailed error stats to pruned_transducer_stateless7_ctc_bk/exp_lconv_scaling/modified_beam_search/errs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt
31
+ 2022-12-14 16:43:23,086 INFO [ctc_guild_decode_bk.py:656]
32
+ For test-other, WER of different settings are:
33
+ beam_size_4 15.24 best for test-other
34
+
35
+ 2022-12-14 16:43:23,087 INFO [ctc_guild_decode_bk.py:867] Done!
decoding-results/modified_beam_search/recogs-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/modified_beam_search/recogs-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding-results/modified_beam_search/wer-summary-test-clean-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_size_4 5.66
decoding-results/modified_beam_search/wer-summary-test-other-beam_size_4-epoch-30-avg-13-modified_beam_search-beam-size-4-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_size_4 15.24
exp/cpu_jit.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da57e855d2bcd3d3230e3cfb0905d6e0230b616a76c5290db9fad37999ebac45
3
+ size 284936924
exp/decode.sh ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env bash
2
+
3
+ export CUDA_VISIBLE_DEVICES=
4
+
5
+ # greedy search
6
+ ./pruned_transducer_stateless7_ctc_bs/ctc_guild_decode_bs.py \
7
+ --epoch 30 \
8
+ --avg 13 \
9
+ --exp-dir ./pruned_transducer_stateless7_ctc_bs/exp \
10
+ --max-duration 600 \
11
+ --decoding-method greedy_search
12
+
13
+ # modified beam search
14
+ ./pruned_transducer_stateless7_ctc_bs/ctc_guild_decode_bs.py \
15
+ --epoch 30 \
16
+ --avg 13 \
17
+ --exp-dir ./pruned_transducer_stateless7_ctc_bs/exp \
18
+ --max-duration 600 \
19
+ --decoding-method modified_beam_search \
20
+ --beam-size 4
21
+
22
+ # fast beam search (one best)
23
+ ./pruned_transducer_stateless7_ctc_bs/ctc_guild_decode_bs.py \
24
+ --epoch 30 \
25
+ --avg 13 \
26
+ --exp-dir ./pruned_transducer_stateless7_ctc_bs/exp \
27
+ --max-duration 600 \
28
+ --decoding-method fast_beam_search \
29
+ --beam 20.0 \
30
+ --max-contexts 8 \
31
+ --max-states 64
exp/epoch-30.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a51723afbc76d9a3ff9e522f3e3eb385ace7916c2f977ae6fcedd5b3b142d36f
3
+ size 1139299611
exp/export.sh ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env bash
2
+
3
+ export CUDA_VISIBLE_DEVICES=
4
+
5
+ ./pruned_transducer_stateless7_ctc_bs/export.py \
6
+ --exp-dir ./pruned_transducer_stateless7_ctc_bs/exp \
7
+ --bpe-model data/lang_bpe_500/bpe.model \
8
+ --epoch 30 \
9
+ --avg 13 \
10
+ --jit 1
exp/log/log-train-2022-12-12-18-45-28-0 ADDED
The diff for this file is too large to render. See raw diff
 
exp/log/log-train-2022-12-12-18-45-28-1 ADDED
The diff for this file is too large to render. See raw diff
 
exp/log/log-train-2022-12-12-18-45-28-2 ADDED
The diff for this file is too large to render. See raw diff
 
exp/log/log-train-2022-12-12-18-45-28-3 ADDED
The diff for this file is too large to render. See raw diff
 
exp/pretrained.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63d958507e38bd4a627755ae46fd27bd1bbdfc9d4ab10c0903e4d51bae23593a
3
+ size 284952187
exp/run.sh ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env bash
2
+
3
+ export CUDA_VISIBLE_DEVICES="0,1,2,3"
4
+
5
+ ./pruned_transducer_stateless7_ctc_bs/train.py \
6
+ --world-size 4 \
7
+ --num-epochs 30 \
8
+ --start-epoch 1 \
9
+ --exp-dir pruned_transducer_stateless7_ctc_bs/exp \
10
+ --full-libri 0 \
11
+ --max-duration 300
exp/tensorboard/events.out.tfevents.1670841928.d3-hpc-sjtu-test-004.143030.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0db07f460a0ded8858fc69f7d13034073c826f216b137a25ddc8a28504961b57
3
+ size 331437