pkufool commited on
Commit
7de761b
1 Parent(s): 1d7ef3c
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. data/lang_bpe_500/bpe.model +3 -0
  2. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  3. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  4. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  5. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  6. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  7. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  8. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  9. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  10. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  11. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  12. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  13. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  14. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  15. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  16. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  17. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  18. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-40-41 +27 -0
  19. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-42-18 +27 -0
  20. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-30-40 +27 -0
  21. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-32-26 +27 -0
  22. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-34-07 +27 -0
  23. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-35-44 +27 -0
  24. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-37-25 +27 -0
  25. decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-39-02 +27 -0
  26. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  27. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  28. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  29. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  30. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  31. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  32. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  33. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  34. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  35. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  36. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  37. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  38. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  39. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  40. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  41. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  42. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  43. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  44. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  45. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  46. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  47. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  48. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  49. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  50. decode_results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
data/lang_bpe_500/bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c53433de083c4a6ad12d034550ef22de68cec62c4f58932a7b6b8b2f1e743fa5
3
+ size 244865
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-40-41 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:40:41,306 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:40:41,307 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:40:41,309 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 16, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:40:41,310 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:40:41,780 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:40:52,938 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:40:52,938 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:40:52,940 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:40:56,022 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:41:22,984 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:41:34,719 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:41:34,906 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.17% [1666 / 52576, 185 ins, 157 del, 1324 sub ]
13
+ 2022-06-26 17:41:35,136 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:41:35,137 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.17 best for test-clean
17
+
18
+ 2022-06-26 17:41:37,577 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:42:03,106 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:42:14,352 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:42:14,436 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.78% [4073 / 52343, 382 ins, 418 del, 3273 sub ]
22
+ 2022-06-26 17:42:14,668 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:42:14,669 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.78 best for test-other
26
+
27
+ 2022-06-26 17:42:14,669 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-42-18 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:42:18,101 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:42:18,102 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:42:18,105 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 16, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:42:18,105 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:42:18,578 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:42:30,118 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:42:30,118 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:42:30,120 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:42:33,173 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:43:00,974 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:43:12,988 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:43:13,192 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.11% [1637 / 52576, 184 ins, 154 del, 1299 sub ]
13
+ 2022-06-26 17:43:13,419 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:43:13,420 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.11 best for test-clean
17
+
18
+ 2022-06-26 17:43:15,962 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:43:42,754 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:43:54,476 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:43:54,568 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.63% [3996 / 52343, 384 ins, 403 del, 3209 sub ]
22
+ 2022-06-26 17:43:54,801 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:43:54,802 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.63 best for test-other
26
+
27
+ 2022-06-26 17:43:54,802 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-30-40 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:30:40,092 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:30:40,092 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:30:40,095 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 2, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:30:40,095 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:30:40,592 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:30:54,090 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:30:54,091 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:30:54,092 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:30:57,893 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:31:29,164 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:31:42,762 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:31:42,926 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.71% [1951 / 52576, 195 ins, 196 del, 1560 sub ]
13
+ 2022-06-26 17:31:43,154 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:31:43,155 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.71 best for test-clean
17
+
18
+ 2022-06-26 17:31:45,659 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:32:11,724 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:32:23,260 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:32:23,348 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 9.18% [4806 / 52343, 423 ins, 581 del, 3802 sub ]
22
+ 2022-06-26 17:32:23,584 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:32:23,585 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 9.18 best for test-other
26
+
27
+ 2022-06-26 17:32:23,585 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-32-26 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:32:26,847 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:32:26,847 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:32:26,850 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 2, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:32:26,850 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:32:27,328 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:32:38,655 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:32:38,655 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:32:38,658 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:32:42,028 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:33:09,813 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:33:22,157 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:33:22,330 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.61% [1900 / 52576, 187 ins, 177 del, 1536 sub ]
13
+ 2022-06-26 17:33:22,559 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:33:22,560 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.61 best for test-clean
17
+
18
+ 2022-06-26 17:33:25,103 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:33:51,905 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:34:03,692 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:34:03,782 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 9.03% [4724 / 52343, 426 ins, 572 del, 3726 sub ]
22
+ 2022-06-26 17:34:04,019 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:34:04,020 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 9.03 best for test-other
26
+
27
+ 2022-06-26 17:34:04,020 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-34-07 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:34:07,348 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:34:07,348 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:34:07,351 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 4, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:34:07,351 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:34:07,821 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:34:19,068 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:34:19,068 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:34:19,070 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:34:22,097 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:34:48,927 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:35:00,589 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:35:00,759 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.36% [1765 / 52576, 194 ins, 144 del, 1427 sub ]
13
+ 2022-06-26 17:35:00,992 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:35:00,992 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.36 best for test-clean
17
+
18
+ 2022-06-26 17:35:03,438 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:35:29,139 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:35:40,461 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:35:40,550 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.65% [4529 / 52343, 461 ins, 460 del, 3608 sub ]
22
+ 2022-06-26 17:35:40,791 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:35:40,791 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.65 best for test-other
26
+
27
+ 2022-06-26 17:35:40,792 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-35-44 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:35:44,177 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:35:44,178 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:35:44,180 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 4, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:35:44,181 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:35:44,659 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:35:55,802 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:35:55,802 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:35:55,805 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:35:58,954 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:36:27,101 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:36:39,371 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:36:39,543 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.46% [1817 / 52576, 207 ins, 165 del, 1445 sub ]
13
+ 2022-06-26 17:36:39,772 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:36:39,773 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.46 best for test-clean
17
+
18
+ 2022-06-26 17:36:42,347 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:37:09,301 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:37:21,234 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:37:21,321 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.43% [4411 / 52343, 434 ins, 442 del, 3535 sub ]
22
+ 2022-06-26 17:37:21,558 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:37:21,558 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.43 best for test-other
26
+
27
+ 2022-06-26 17:37:21,559 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-37-25 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:37:25,058 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:37:25,058 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:37:25,061 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 8, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:37:25,061 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:37:25,535 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:37:37,009 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:37:37,009 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:37:37,011 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:37:40,062 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:38:07,087 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:38:18,769 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:38:18,936 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.23% [1696 / 52576, 191 ins, 145 del, 1360 sub ]
13
+ 2022-06-26 17:38:19,163 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:38:19,163 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.23 best for test-clean
17
+
18
+ 2022-06-26 17:38:21,627 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:38:47,119 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:38:58,408 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:38:58,498 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.23% [4308 / 52343, 445 ins, 416 del, 3447 sub ]
22
+ 2022-06-26 17:38:58,733 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:38:58,733 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.23 best for test-other
26
+
27
+ 2022-06-26 17:38:58,734 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-39-02 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:39:02,206 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:39:02,207 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:39:02,209 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 36, 'iter': 0, 'avg': 8, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 8, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search'), 'suffix': 'epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:39:02,209 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:39:02,672 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-29.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-30.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-31.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-32.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-33.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-34.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-35.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/epoch-36.pt']
6
+ 2022-06-26 17:39:13,782 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:39:13,782 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:39:13,784 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:39:16,914 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:39:44,621 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:39:56,682 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:39:56,832 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.20% [1682 / 52576, 196 ins, 141 del, 1345 sub ]
13
+ 2022-06-26 17:39:57,055 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:39:57,056 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.2 best for test-clean
17
+
18
+ 2022-06-26 17:39:59,557 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:40:25,921 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:40:37,539 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:40:37,626 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.00% [4190 / 52343, 414 ins, 398 del, 3378 sub ]
22
+ 2022-06-26 17:40:37,864 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:40:37,865 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.0 best for test-other
26
+
27
+ 2022-06-26 17:40:37,865 INFO [decode.py:803] Done!
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.17
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.11
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.71
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.61
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.36
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.46
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.23
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.2
decode_results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-36-avg-8-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 7.78