pkufool commited on
Commit
b7d758b
1 Parent(s): e53936c
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. data/lang_bpe_500/bpe.model +3 -0
  2. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  3. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  4. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  5. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  6. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  7. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  8. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  9. decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  10. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  11. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  12. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  13. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  14. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  15. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  16. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  17. decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  18. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-54-26 +27 -0
  19. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-56-12 +27 -0
  20. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-43-58 +27 -0
  21. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-45-39 +27 -0
  22. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-47-24 +27 -0
  23. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-49-05 +27 -0
  24. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-50-52 +27 -0
  25. decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-52-41 +27 -0
  26. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  27. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  28. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  29. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  30. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  31. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  32. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  33. decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  34. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  35. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  36. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  37. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  38. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  39. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  40. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  41. decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +0 -0
  42. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  43. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  44. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  45. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  46. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  47. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  48. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  49. decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
  50. decode_results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt +2 -0
data/lang_bpe_500/bpe.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c53433de083c4a6ad12d034550ef22de68cec62c4f58932a7b6b8b2f1e743fa5
3
+ size 244865
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-54-26 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:54:26,854 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:54:26,854 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:54:26,857 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 16, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:54:26,857 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:54:27,325 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:54:40,990 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:54:40,990 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:54:40,992 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:54:44,394 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:55:15,594 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:55:28,957 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:55:29,043 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 2.77% [1454 / 52576, 172 ins, 128 del, 1154 sub ]
13
+ 2022-06-26 17:55:29,280 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:55:29,281 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 2.77 best for test-clean
17
+
18
+ 2022-06-26 17:55:31,732 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:55:57,482 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:56:08,617 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:56:08,705 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 6.75% [3532 / 52343, 391 ins, 318 del, 2823 sub ]
22
+ 2022-06-26 17:56:08,940 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:56:08,941 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 6.75 best for test-other
26
+
27
+ 2022-06-26 17:56:08,941 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-56-12 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:56:12,645 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:56:12,645 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:56:12,648 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 16, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:56:12,649 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:56:13,116 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:56:26,442 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:56:26,443 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:56:26,445 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:56:29,529 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:56:57,942 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:57:10,030 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:57:10,112 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 2.73% [1436 / 52576, 186 ins, 126 del, 1124 sub ]
13
+ 2022-06-26 17:57:10,337 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:57:10,338 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 2.73 best for test-clean
17
+
18
+ 2022-06-26 17:57:12,818 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:57:39,395 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:57:50,975 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:57:51,060 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 6.68% [3495 / 52343, 394 ins, 314 del, 2787 sub ]
22
+ 2022-06-26 17:57:51,293 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:57:51,294 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 6.68 best for test-other
26
+
27
+ 2022-06-26 17:57:51,294 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-43-58 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:43:58,473 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:43:58,473 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:43:58,477 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 2, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:43:58,477 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:43:58,942 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:44:12,828 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:44:12,828 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:44:12,830 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:44:15,930 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:44:43,330 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:44:55,207 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:44:55,294 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.26% [1712 / 52576, 193 ins, 174 del, 1345 sub ]
13
+ 2022-06-26 17:44:55,520 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:44:55,521 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.26 best for test-clean
17
+
18
+ 2022-06-26 17:44:58,014 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:45:23,895 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:45:35,079 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:45:35,167 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.20% [4294 / 52343, 418 ins, 555 del, 3321 sub ]
22
+ 2022-06-26 17:45:35,400 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:45:35,401 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.2 best for test-other
26
+
27
+ 2022-06-26 17:45:35,401 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-45-39 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:45:39,091 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:45:39,092 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:45:39,095 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 2, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:45:39,095 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:45:39,558 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:45:53,246 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:45:53,247 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:45:53,249 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:45:56,470 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:46:25,557 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:46:38,236 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:46:38,324 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.24% [1705 / 52576, 180 ins, 186 del, 1339 sub ]
13
+ 2022-06-26 17:46:38,555 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:46:38,555 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.24 best for test-clean
17
+
18
+ 2022-06-26 17:46:41,198 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:47:08,861 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:47:20,790 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:47:20,876 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 8.09% [4237 / 52343, 417 ins, 549 del, 3271 sub ]
22
+ 2022-06-26 17:47:21,108 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:47:21,109 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 8.09 best for test-other
26
+
27
+ 2022-06-26 17:47:21,109 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-47-24 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:47:24,332 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:47:24,333 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:47:24,336 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 4, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:47:24,336 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:47:24,804 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:47:38,564 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:47:38,564 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:47:38,566 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:47:41,619 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:48:09,133 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:48:20,986 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:48:21,070 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.06% [1608 / 52576, 197 ins, 137 del, 1274 sub ]
13
+ 2022-06-26 17:48:21,295 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:48:21,295 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.06 best for test-clean
17
+
18
+ 2022-06-26 17:48:23,743 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:48:49,652 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:49:00,831 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:49:00,919 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.56% [3955 / 52343, 445 ins, 348 del, 3162 sub ]
22
+ 2022-06-26 17:49:01,152 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:49:01,153 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.56 best for test-other
26
+
27
+ 2022-06-26 17:49:01,153 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-49-05 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:49:05,012 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:49:05,012 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:49:05,016 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 4, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:49:05,016 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:49:05,494 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:49:19,223 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:49:19,223 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:49:19,225 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:49:22,771 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:49:54,250 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:50:06,609 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:50:06,693 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 3.06% [1608 / 52576, 204 ins, 142 del, 1262 sub ]
13
+ 2022-06-26 17:50:06,920 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:50:06,920 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 3.06 best for test-clean
17
+
18
+ 2022-06-26 17:50:09,407 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:50:36,375 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:50:48,181 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:50:48,269 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.43% [3887 / 52343, 460 ins, 354 del, 3073 sub ]
22
+ 2022-06-26 17:50:48,502 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:50:48,503 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.43 best for test-other
26
+
27
+ 2022-06-26 17:50:48,503 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-50-52 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:50:52,151 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:50:52,152 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:50:52,155 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 8, 'left_context': 32, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:50:52,156 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:50:52,670 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:51:06,888 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:51:06,888 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:51:06,893 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:51:10,535 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:51:40,745 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:51:54,025 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:51:54,135 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 2.98% [1569 / 52576, 202 ins, 135 del, 1232 sub ]
13
+ 2022-06-26 17:51:54,447 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:51:54,448 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 2.98 best for test-clean
17
+
18
+ 2022-06-26 17:51:57,331 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:52:25,480 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:52:37,924 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:52:38,032 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.08% [3707 / 52343, 397 ins, 347 del, 2963 sub ]
22
+ 2022-06-26 17:52:38,262 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:52:38,263 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.08 best for test-other
26
+
27
+ 2022-06-26 17:52:38,263 INFO [decode.py:803] Done!
decode_results/fast_beam_search/log-decode-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64-2022-06-26-17-52-41 ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-06-26 17:52:41,715 INFO [decode.py:693] Decoding started
2
+ 2022-06-26 17:52:41,715 INFO [decode.py:699] Device: cuda:0
3
+ 2022-06-26 17:52:41,718 INFO [decode.py:714] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'encoder_dim': 512, 'nhead': 8, 'dim_feedforward': 2048, 'num_encoder_layers': 12, 'decoder_dim': 512, 'joiner_dim': 512, 'model_warm_step': 3000, 'env_info': {'k2-version': '1.15.1', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': 'f8d2dba06c000ffee36aab5b66f24e7c9809f116', 'k2-git-date': 'Thu Apr 21 12:20:34 2022', 'lhotse-version': '1.1.0', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'streaming-conformer', 'icefall-git-sha1': 'af80a46-clean', 'icefall-git-date': 'Sat Jun 25 18:24:36 2022', 'icefall-path': '/ceph-kw/kangwei/code/icefall_streaming4', 'k2-path': '/ceph-hw/kangwei/code/k2_release/k2/k2/python/k2/__init__.py', 'lhotse-path': '/ceph-hw/kangwei/dev_tools/anaconda3/envs/rnnt2/lib/python3.8/site-packages/lhotse-1.1.0-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-7-0616225511-78bf4545d8-tv52r', 'IP address': '10.177.77.9'}, 'epoch': 25, 'iter': 0, 'avg': 12, 'exp_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 20.0, 'ngram_lm_scale': 0.01, 'max_contexts': 8, 'max_states': 64, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'simulate_streaming': True, 'decode_chunk_size': 8, 'left_context': 64, 'dynamic_chunk_training': False, 'causal_convolution': True, 'short_chunk_size': 25, 'num_left_chunks': 4, 'max_duration': 1000, 'bucketing_sampler': True, 'num_buckets': 30, 'shuffle': True, 'return_cuts': True, 'num_workers': 2, 'on_the_fly_num_workers': 0, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'manifest_dir': PosixPath('data/fbank'), 'on_the_fly_feats': False, 'res_dir': PosixPath('/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search'), 'suffix': 'epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500}
4
+ 2022-06-26 17:52:41,719 INFO [decode.py:716] About to create model
5
+ 2022-06-26 17:52:42,183 INFO [decode.py:744] averaging ['/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-14.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-15.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-16.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-17.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-18.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-19.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-20.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-21.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-22.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-23.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-24.pt', '/ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/epoch-25.pt']
6
+ 2022-06-26 17:52:55,993 INFO [decode.py:773] Number of model parameters: 80199888
7
+ 2022-06-26 17:52:55,993 INFO [librispeech.py:58] About to get test-clean cuts from data/fbank/librispeech_cuts_test-clean.jsonl.gz
8
+ 2022-06-26 17:52:55,995 INFO [librispeech.py:63] About to get test-other cuts from data/fbank/librispeech_cuts_test-other.jsonl.gz
9
+ 2022-06-26 17:52:59,151 INFO [decode.py:595] batch 0/?, cuts processed until now is 32
10
+ 2022-06-26 17:53:27,812 INFO [decode.py:595] batch 20/?, cuts processed until now is 1760
11
+ 2022-06-26 17:53:40,263 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
12
+ 2022-06-26 17:53:40,354 INFO [utils.py:418] [test-clean-beam_20.0_max_contexts_8_max_states_64] %WER 2.88% [1516 / 52576, 202 ins, 134 del, 1180 sub ]
13
+ 2022-06-26 17:53:40,586 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
14
+ 2022-06-26 17:53:40,587 INFO [decode.py:642]
15
+ For test-clean, WER of different settings are:
16
+ beam_20.0_max_contexts_8_max_states_64 2.88 best for test-clean
17
+
18
+ 2022-06-26 17:53:43,151 INFO [decode.py:595] batch 0/?, cuts processed until now is 35
19
+ 2022-06-26 17:54:10,738 INFO [decode.py:595] batch 20/?, cuts processed until now is 1983
20
+ 2022-06-26 17:54:22,835 INFO [decode.py:612] The transcripts are stored in /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
21
+ 2022-06-26 17:54:22,922 INFO [utils.py:418] [test-other-beam_20.0_max_contexts_8_max_states_64] %WER 7.03% [3681 / 52343, 406 ins, 364 del, 2911 sub ]
22
+ 2022-06-26 17:54:23,159 INFO [decode.py:625] Wrote detailed error stats to /ceph-hw/kangwei/code2/icefall_reworked2/egs/librispeech/ASR/streaming_pruned_transducer_stateless3_fixed_left/exp/fast_beam_search/errs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt
23
+ 2022-06-26 17:54:23,160 INFO [decode.py:642]
24
+ For test-other, WER of different settings are:
25
+ beam_20.0_max_contexts_8_max_states_64 7.03 best for test-other
26
+
27
+ 2022-06-26 17:54:23,160 INFO [decode.py:803] Done!
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/recogs-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
The diff for this file is too large to render. See raw diff
 
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 2.77
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 2.73
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.26
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-2-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.24
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.06
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-4-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 3.06
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 2.98
decode_results/fast_beam_search/wer-summary-test-clean-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-8-left-context-64-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 2.88
decode_results/fast_beam_search/wer-summary-test-other-beam_20.0_max_contexts_8_max_states_64-epoch-25-avg-12-streaming-chunk-size-16-left-context-32-beam-20.0-max-contexts-8-max-states-64.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ beam_20.0_max_contexts_8_max_states_64 6.75