jinzr commited on
Commit
fa1b4b4
1 Parent(s): 38b4e64

updated ctc decoding results

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. decoding_results/ctc-decoding/errs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  2. decoding_results/ctc-decoding/errs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  3. decoding_results/ctc-decoding/errs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  4. decoding_results/ctc-decoding/errs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  5. decoding_results/ctc-decoding/errs-aishell-4-epoch-20-avg-1-use-averaged-model.txt +0 -0
  6. decoding_results/ctc-decoding/errs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  7. decoding_results/ctc-decoding/errs-aishell_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  8. decoding_results/ctc-decoding/errs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt +0 -0
  9. decoding_results/ctc-decoding/errs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  10. decoding_results/ctc-decoding/errs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt +0 -0
  11. decoding_results/ctc-decoding/errs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt +0 -0
  12. decoding_results/ctc-decoding/errs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  13. decoding_results/ctc-decoding/errs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  14. decoding_results/ctc-decoding/errs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  15. decoding_results/ctc-decoding/errs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  16. decoding_results/ctc-decoding/errs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  17. decoding_results/ctc-decoding/errs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  18. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-02-26 +0 -18
  19. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-17-01 +0 -18
  20. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-21-51 +0 -18
  21. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-41-49 +0 -6
  22. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-42-14 +0 -18
  23. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-23-09 +0 -57
  24. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-27-50 +0 -4
  25. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-28-40 +0 -196
  26. decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-25-16-34-00 +266 -0
  27. decoding_results/ctc-decoding/log-decode-epoch-22-avg-1-use-averaged-model-2023-10-17-16-37-56 +0 -3
  28. decoding_results/ctc-decoding/log-decode-epoch-22-avg-1-use-averaged-model-2023-10-17-16-38-08 +0 -18
  29. decoding_results/ctc-decoding/recogs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  30. decoding_results/ctc-decoding/recogs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  31. decoding_results/ctc-decoding/recogs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  32. decoding_results/ctc-decoding/recogs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  33. decoding_results/ctc-decoding/recogs-aishell-4-epoch-20-avg-1-use-averaged-model.txt +0 -0
  34. decoding_results/ctc-decoding/recogs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  35. decoding_results/ctc-decoding/recogs-aishell_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  36. decoding_results/ctc-decoding/recogs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt +0 -0
  37. decoding_results/ctc-decoding/recogs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  38. decoding_results/ctc-decoding/recogs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt +0 -0
  39. decoding_results/ctc-decoding/recogs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt +0 -0
  40. decoding_results/ctc-decoding/recogs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  41. decoding_results/ctc-decoding/recogs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  42. decoding_results/ctc-decoding/recogs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  43. decoding_results/ctc-decoding/recogs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  44. decoding_results/ctc-decoding/recogs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt +0 -0
  45. decoding_results/ctc-decoding/recogs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt +0 -0
  46. decoding_results/ctc-decoding/wer-summary-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt +1 -1
  47. decoding_results/ctc-decoding/wer-summary-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt +1 -1
  48. decoding_results/ctc-decoding/wer-summary-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt +1 -1
  49. decoding_results/ctc-decoding/wer-summary-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt +1 -1
  50. decoding_results/ctc-decoding/wer-summary-aishell-4-epoch-20-avg-1-use-averaged-model.txt +1 -1
decoding_results/ctc-decoding/errs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aishell-4-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-aishell_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-02-26 DELETED
@@ -1,18 +0,0 @@
1
- 2023-10-17 16:02:26,209 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:02:26,210 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:02:26,210 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-3-0423201227-84b4557756-h4fh6', 'IP address': '10.177.77.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 16:02:28,952 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:02:45,974 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:02:47,196 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 16:03:07,159 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 16:03:07,161 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 16:03:07,161 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 16:03:07,256 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 16:03:07,322 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 16:03:07,368 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 16:03:07,389 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 16:03:07,436 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 16:03:07,503 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 16:03:07,587 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 16:03:20,149 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 16:03:22,114 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-17-01 DELETED
@@ -1,18 +0,0 @@
1
- 2023-10-17 16:17:01,559 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:17:01,560 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:17:01,560 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 16:17:03,333 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:17:08,838 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:17:09,434 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 16:17:16,852 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 16:17:16,853 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 16:17:16,853 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 16:17:16,919 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 16:17:16,948 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 16:17:16,993 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 16:17:16,997 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 16:17:17,016 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 16:17:17,073 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 16:17:17,139 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 16:17:23,632 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 16:17:24,681 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-21-51 DELETED
@@ -1,18 +0,0 @@
1
- 2023-10-17 16:21:51,584 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:21:51,584 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:21:51,584 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 16:21:53,329 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:21:58,312 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:21:58,877 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 16:22:02,908 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 16:22:02,908 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 16:22:02,908 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 16:22:02,927 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 16:22:02,931 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 16:22:02,935 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 16:22:02,936 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 16:22:02,940 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 16:22:02,943 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 16:22:02,948 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 16:22:08,929 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 16:22:09,902 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_dev
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-41-49 DELETED
@@ -1,6 +0,0 @@
1
- 2023-10-17 16:41:49,881 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:41:49,881 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:41:49,881 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 16:41:51,627 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:41:56,685 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:41:57,256 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-16-42-14 DELETED
@@ -1,18 +0,0 @@
1
- 2023-10-17 16:42:14,818 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:42:14,818 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:42:14,819 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 16:42:16,549 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:42:21,387 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:42:21,967 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 16:42:25,945 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 16:42:25,945 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 16:42:25,945 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 16:42:25,962 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 16:42:25,966 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 16:42:25,970 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 16:42:25,972 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 16:42:25,975 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 16:42:25,978 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 16:42:25,982 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 16:42:32,311 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 16:42:33,328 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-23-09 DELETED
@@ -1,57 +0,0 @@
1
- 2023-10-17 18:23:09,112 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 18:23:09,112 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 18:23:09,112 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 18:23:10,868 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 18:23:16,848 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 18:23:17,431 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 18:23:22,164 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 18:23:22,164 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 18:23:22,165 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 18:23:22,182 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 18:23:22,186 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 18:23:22,189 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 18:23:22,191 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 18:23:22,194 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 18:23:22,197 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 18:23:22,202 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 18:23:28,526 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 18:23:29,552 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
19
- 2023-10-17 18:23:30,937 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 80
20
- 2023-10-17 18:23:31,149 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([4.3243, 4.1785, 3.8951, 4.4938], device='cuda:0')
21
- 2023-10-17 18:23:32,075 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([4.6184, 3.9753, 4.4056, 4.1082], device='cuda:0')
22
- 2023-10-17 18:23:46,072 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([1.7554, 1.8501, 1.9709, 1.4977], device='cuda:0')
23
- 2023-10-17 18:23:47,852 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.0318, 2.9265, 2.8671, 2.9708], device='cuda:0')
24
- 2023-10-17 18:23:49,153 INFO [ctc_decode.py:485] batch 100/?, cuts processed until now is 9084
25
- 2023-10-17 18:23:53,157 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([1.7971, 2.1446, 2.3575, 2.1475, 2.1401, 2.0719, 2.1833, 2.1629],
26
- device='cuda:0')
27
- 2023-10-17 18:24:07,354 INFO [ctc_decode.py:485] batch 200/?, cuts processed until now is 18516
28
- 2023-10-17 18:24:25,501 INFO [ctc_decode.py:485] batch 300/?, cuts processed until now is 28179
29
- 2023-10-17 18:24:28,319 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.8385, 2.2905, 3.9372, 2.3478], device='cuda:0')
30
- 2023-10-17 18:24:41,962 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([5.3786, 4.9526, 5.2852, 4.9957], device='cuda:0')
31
- 2023-10-17 18:24:43,488 INFO [ctc_decode.py:485] batch 400/?, cuts processed until now is 37667
32
- 2023-10-17 18:24:48,276 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.4047, 2.8120, 1.7099, 2.0791], device='cuda:0')
33
- 2023-10-17 18:24:48,972 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.6806, 3.0397, 2.2822, 2.4676], device='cuda:0')
34
- 2023-10-17 18:25:00,399 INFO [ctc_decode.py:485] batch 500/?, cuts processed until now is 46172
35
- 2023-10-17 18:25:06,922 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
36
- 2023-10-17 18:25:07,811 INFO [utils.py:565] [aidatatang_test-ctc-decoding] %WER 15.26% [43137 / 282666, 7255 ins, 10411 del, 25471 sub ]
37
- 2023-10-17 18:25:09,250 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
38
- 2023-10-17 18:25:09,253 INFO [ctc_decode.py:522]
39
- For aidatatang_test, WER of different settings are:
40
- ctc-decoding 15.26 best for aidatatang_test
41
-
42
- 2023-10-17 18:25:09,254 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_dev
43
- 2023-10-17 18:25:10,768 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 81
44
- 2023-10-17 18:25:21,290 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([1.6683, 3.5095, 3.2214, 3.0099], device='cuda:0')
45
- 2023-10-17 18:25:28,677 INFO [ctc_decode.py:485] batch 100/?, cuts processed until now is 9077
46
- 2023-10-17 18:25:45,744 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([5.3476, 5.0039, 5.2630, 5.0075], device='cuda:0')
47
- 2023-10-17 18:25:46,062 INFO [ctc_decode.py:485] batch 200/?, cuts processed until now is 18432
48
- 2023-10-17 18:25:58,007 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
49
- 2023-10-17 18:25:58,413 INFO [utils.py:565] [aidatatang_dev-ctc-decoding] %WER 14.58% [20723 / 142150, 3444 ins, 5445 del, 11834 sub ]
50
- 2023-10-17 18:25:59,266 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
51
- 2023-10-17 18:25:59,276 INFO [ctc_decode.py:522]
52
- For aidatatang_dev, WER of different settings are:
53
- ctc-decoding 14.58 best for aidatatang_dev
54
-
55
- 2023-10-17 18:25:59,277 INFO [ctc_decode.py:695] Start decoding test set: alimeeting_test
56
- 2023-10-17 18:26:01,217 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 44
57
- 2023-10-17 18:26:10,131 WARNING [ctc_decode.py:683] Excluding cut with ID: R8008_M8016-8062-123 from decoding, num_frames: 6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-27-50 DELETED
@@ -1,4 +0,0 @@
1
- 2023-10-17 18:27:50,959 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 18:27:50,959 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 18:27:50,959 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 18:27:52,674 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-17-18-28-40 DELETED
@@ -1,196 +0,0 @@
1
- 2023-10-17 18:28:40,557 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 18:28:40,557 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 18:28:40,557 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 1200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
- 2023-10-17 18:28:42,268 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 18:28:48,063 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 18:28:48,655 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
- 2023-10-17 18:28:53,461 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 18:28:53,462 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 18:28:53,462 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 18:28:53,480 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 18:28:53,483 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 18:28:53,486 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 18:28:53,488 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 18:28:53,492 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 18:28:53,495 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 18:28:53,500 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 18:28:59,880 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 18:29:00,864 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
19
- 2023-10-17 18:29:03,097 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 321
20
- 2023-10-17 18:29:55,906 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.2702, 2.7870, 3.1113, 2.7685], device='cuda:0')
21
- 2023-10-17 18:30:06,942 INFO [ctc_decode.py:485] batch 100/?, cuts processed until now is 35376
22
- 2023-10-17 18:30:29,144 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
23
- 2023-10-17 18:30:29,895 INFO [utils.py:565] [aidatatang_test-ctc-decoding] %WER 15.26% [43139 / 282666, 7245 ins, 10406 del, 25488 sub ]
24
- 2023-10-17 18:30:31,523 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
25
- 2023-10-17 18:30:31,526 INFO [ctc_decode.py:522]
26
- For aidatatang_test, WER of different settings are:
27
- ctc-decoding 15.26 best for aidatatang_test
28
-
29
- 2023-10-17 18:30:31,527 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_dev
30
- 2023-10-17 18:30:33,562 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 325
31
- 2023-10-17 18:31:02,140 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([4.4563, 4.2981, 3.9703, 4.6275], device='cuda:0')
32
- 2023-10-17 18:31:08,199 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.1070, 2.5969, 1.4553, 1.7578], device='cuda:0')
33
- 2023-10-17 18:31:08,818 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.7113, 3.0239, 2.1108, 2.3879], device='cuda:0')
34
- 2023-10-17 18:31:15,999 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
35
- 2023-10-17 18:31:16,380 INFO [utils.py:565] [aidatatang_dev-ctc-decoding] %WER 14.57% [20711 / 142150, 3442 ins, 5435 del, 11834 sub ]
36
- 2023-10-17 18:31:17,106 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
37
- 2023-10-17 18:31:17,110 INFO [ctc_decode.py:522]
38
- For aidatatang_dev, WER of different settings are:
39
- ctc-decoding 14.57 best for aidatatang_dev
40
-
41
- 2023-10-17 18:31:17,110 INFO [ctc_decode.py:695] Start decoding test set: alimeeting_test
42
- 2023-10-17 18:31:19,674 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 177
43
- 2023-10-17 18:31:25,592 WARNING [ctc_decode.py:683] Excluding cut with ID: R8008_M8016-8062-123 from decoding, num_frames: 6
44
- 2023-10-17 18:31:51,823 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt
45
- 2023-10-17 18:31:51,924 INFO [utils.py:565] [alimeeting_test-ctc-decoding] %WER 69.70% [11401 / 16357, 1 ins, 2019 del, 9381 sub ]
46
- 2023-10-17 18:31:52,197 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt
47
- 2023-10-17 18:31:52,201 INFO [ctc_decode.py:522]
48
- For alimeeting_test, WER of different settings are:
49
- ctc-decoding 69.7 best for alimeeting_test
50
-
51
- 2023-10-17 18:31:52,201 INFO [ctc_decode.py:695] Start decoding test set: alimeeting_eval
52
- 2023-10-17 18:31:54,746 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 140
53
- 2023-10-17 18:32:06,274 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt
54
- 2023-10-17 18:32:06,314 INFO [utils.py:565] [alimeeting_eval-ctc-decoding] %WER 72.85% [4704 / 6457, 0 ins, 930 del, 3774 sub ]
55
- 2023-10-17 18:32:06,496 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt
56
- 2023-10-17 18:32:06,499 INFO [ctc_decode.py:522]
57
- For alimeeting_eval, WER of different settings are:
58
- ctc-decoding 72.85 best for alimeeting_eval
59
-
60
- 2023-10-17 18:32:06,499 INFO [ctc_decode.py:695] Start decoding test set: aishell_test
61
- 2023-10-17 18:32:09,216 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 190
62
- 2023-10-17 18:32:15,951 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.6632, 3.6327, 2.7261, 2.6312], device='cuda:0')
63
- 2023-10-17 18:32:28,588 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell_test-epoch-20-avg-1-use-averaged-model.txt
64
- 2023-10-17 18:32:28,732 INFO [utils.py:565] [aishell_test-ctc-decoding] %WER 13.76% [8867 / 64428, 866 ins, 2091 del, 5910 sub ]
65
- 2023-10-17 18:32:29,114 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell_test-epoch-20-avg-1-use-averaged-model.txt
66
- 2023-10-17 18:32:29,118 INFO [ctc_decode.py:522]
67
- For aishell_test, WER of different settings are:
68
- ctc-decoding 13.76 best for aishell_test
69
-
70
- 2023-10-17 18:32:29,119 INFO [ctc_decode.py:695] Start decoding test set: aishell_dev
71
- 2023-10-17 18:32:32,157 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 215
72
- 2023-10-17 18:33:07,311 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt
73
- 2023-10-17 18:33:07,677 INFO [utils.py:565] [aishell_dev-ctc-decoding] %WER 12.87% [16438 / 127698, 1671 ins, 3693 del, 11074 sub ]
74
- 2023-10-17 18:33:08,275 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt
75
- 2023-10-17 18:33:08,279 INFO [ctc_decode.py:522]
76
- For aishell_dev, WER of different settings are:
77
- ctc-decoding 12.87 best for aishell_dev
78
-
79
- 2023-10-17 18:33:08,279 INFO [ctc_decode.py:695] Start decoding test set: aishell-2_test
80
- 2023-10-17 18:33:10,740 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 334
81
- 2023-10-17 18:33:18,787 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt
82
- 2023-10-17 18:33:18,825 INFO [utils.py:565] [aishell-2_test-ctc-decoding] %WER 25.55% [1278 / 5002, 0 ins, 3 del, 1275 sub ]
83
- 2023-10-17 18:33:18,904 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt
84
- 2023-10-17 18:33:18,907 INFO [ctc_decode.py:522]
85
- For aishell-2_test, WER of different settings are:
86
- ctc-decoding 25.55 best for aishell-2_test
87
-
88
- 2023-10-17 18:33:18,907 INFO [ctc_decode.py:695] Start decoding test set: aishell-2_dev
89
- 2023-10-17 18:33:20,413 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 209
90
- 2023-10-17 18:33:24,585 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt
91
- 2023-10-17 18:33:24,602 INFO [utils.py:565] [aishell-2_dev-ctc-decoding] %WER 23.56% [589 / 2500, 0 ins, 0 del, 589 sub ]
92
- 2023-10-17 18:33:24,638 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt
93
- 2023-10-17 18:33:24,642 INFO [ctc_decode.py:522]
94
- For aishell-2_dev, WER of different settings are:
95
- ctc-decoding 23.56 best for aishell-2_dev
96
-
97
- 2023-10-17 18:33:24,642 INFO [ctc_decode.py:695] Start decoding test set: aishell-4
98
- 2023-10-17 18:33:28,250 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 132
99
- 2023-10-17 18:33:57,329 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-4-epoch-20-avg-1-use-averaged-model.txt
100
- 2023-10-17 18:33:57,399 INFO [utils.py:565] [aishell-4-ctc-decoding] %WER 71.75% [7590 / 10579, 47 ins, 534 del, 7009 sub ]
101
- 2023-10-17 18:33:57,590 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-4-epoch-20-avg-1-use-averaged-model.txt
102
- 2023-10-17 18:33:57,593 INFO [ctc_decode.py:522]
103
- For aishell-4, WER of different settings are:
104
- ctc-decoding 71.75 best for aishell-4
105
-
106
- 2023-10-17 18:33:57,594 INFO [ctc_decode.py:695] Start decoding test set: magicdata_test
107
- 2023-10-17 18:34:00,583 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 230
108
- 2023-10-17 18:35:01,490 INFO [ctc_decode.py:485] batch 100/?, cuts processed until now is 23940
109
- 2023-10-17 18:35:02,615 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt
110
- 2023-10-17 18:35:02,855 INFO [utils.py:565] [magicdata_test-ctc-decoding] %WER 19.34% [4698 / 24286, 284 ins, 7 del, 4407 sub ]
111
- 2023-10-17 18:35:03,201 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt
112
- 2023-10-17 18:35:03,205 INFO [ctc_decode.py:522]
113
- For magicdata_test, WER of different settings are:
114
- ctc-decoding 19.34 best for magicdata_test
115
-
116
- 2023-10-17 18:35:03,205 INFO [ctc_decode.py:695] Start decoding test set: magicdata_dev
117
- 2023-10-17 18:35:06,237 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 211
118
- 2023-10-17 18:35:24,163 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.0037, 2.2850, 2.0049, 3.6487], device='cuda:0')
119
- 2023-10-17 18:35:38,292 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt
120
- 2023-10-17 18:35:38,366 INFO [utils.py:565] [magicdata_dev-ctc-decoding] %WER 22.35% [2653 / 11872, 22 ins, 79 del, 2552 sub ]
121
- 2023-10-17 18:35:38,552 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt
122
- 2023-10-17 18:35:38,556 INFO [ctc_decode.py:522]
123
- For magicdata_dev, WER of different settings are:
124
- ctc-decoding 22.35 best for magicdata_dev
125
-
126
- 2023-10-17 18:35:38,556 INFO [ctc_decode.py:695] Start decoding test set: kespeech-asr_test
127
- 2023-10-17 18:35:41,963 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 179
128
- 2023-10-17 18:35:48,964 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.2457, 2.7709, 3.0159, 2.7526, 2.8671, 2.7205, 3.0313, 2.9588],
129
- device='cuda:0')
130
- 2023-10-17 18:35:52,190 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.1268, 2.6636, 2.8915, 2.6335, 2.7311, 2.5814, 2.9027, 2.8249],
131
- device='cuda:0')
132
- 2023-10-17 18:36:14,760 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([5.0753, 4.3347, 4.8011, 4.4720], device='cuda:0')
133
- 2023-10-17 18:36:18,629 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.0111, 4.3811, 3.5554, 3.9260], device='cuda:0')
134
- 2023-10-17 18:36:19,830 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([5.0782, 4.3275, 4.7918, 4.4625], device='cuda:0')
135
- 2023-10-17 18:36:45,716 INFO [ctc_decode.py:485] batch 100/?, cuts processed until now is 19455
136
- 2023-10-17 18:36:46,846 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt
137
- 2023-10-17 18:36:46,972 INFO [utils.py:565] [kespeech-asr_test-ctc-decoding] %WER 48.71% [9608 / 19723, 0 ins, 3 del, 9605 sub ]
138
- 2023-10-17 18:36:47,307 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt
139
- 2023-10-17 18:36:47,311 INFO [ctc_decode.py:522]
140
- For kespeech-asr_test, WER of different settings are:
141
- ctc-decoding 48.71 best for kespeech-asr_test
142
-
143
- 2023-10-17 18:36:47,312 INFO [ctc_decode.py:695] Start decoding test set: kespeech-asr_dev_phase1
144
- 2023-10-17 18:36:49,451 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 179
145
- 2023-10-17 18:36:56,484 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt
146
- 2023-10-17 18:36:56,499 INFO [utils.py:565] [kespeech-asr_dev_phase1-ctc-decoding] %WER 42.38% [932 / 2199, 0 ins, 0 del, 932 sub ]
147
- 2023-10-17 18:36:56,534 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt
148
- 2023-10-17 18:36:56,537 INFO [ctc_decode.py:522]
149
- For kespeech-asr_dev_phase1, WER of different settings are:
150
- ctc-decoding 42.38 best for kespeech-asr_dev_phase1
151
-
152
- 2023-10-17 18:36:56,537 INFO [ctc_decode.py:695] Start decoding test set: kespeech-asr_dev_phase2
153
- 2023-10-17 18:36:58,442 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 175
154
- 2023-10-17 18:37:05,129 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt
155
- 2023-10-17 18:37:05,145 INFO [utils.py:565] [kespeech-asr_dev_phase2-ctc-decoding] %WER 26.90% [594 / 2208, 0 ins, 0 del, 594 sub ]
156
- 2023-10-17 18:37:05,178 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt
157
- 2023-10-17 18:37:05,181 INFO [ctc_decode.py:522]
158
- For kespeech-asr_dev_phase2, WER of different settings are:
159
- ctc-decoding 26.9 best for kespeech-asr_dev_phase2
160
-
161
- 2023-10-17 18:37:05,182 INFO [ctc_decode.py:695] Start decoding test set: wenetspeech-meeting_test
162
- 2023-10-17 18:37:07,190 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 112
163
- 2023-10-17 18:37:33,122 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([1.8152, 2.0526, 1.9635, 2.3204, 2.4993, 2.4532, 2.2485, 1.9477],
164
- device='cuda:0')
165
- 2023-10-17 18:37:41,448 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt
166
- 2023-10-17 18:37:41,502 INFO [utils.py:565] [wenetspeech-meeting_test-ctc-decoding] %WER 67.29% [5632 / 8370, 2 ins, 0 del, 5630 sub ]
167
- 2023-10-17 18:37:41,673 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt
168
- 2023-10-17 18:37:41,676 INFO [ctc_decode.py:522]
169
- For wenetspeech-meeting_test, WER of different settings are:
170
- ctc-decoding 67.29 best for wenetspeech-meeting_test
171
-
172
- 2023-10-17 18:37:41,676 INFO [ctc_decode.py:695] Start decoding test set: wenetspeech-net_test
173
- 2023-10-17 18:37:41,901 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
174
- 2023-10-17 18:37:43,917 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 175
175
- 2023-10-17 18:38:04,564 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([4.2461, 2.2640, 4.2913, 3.8631], device='cuda:0')
176
- 2023-10-17 18:38:08,841 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([3.7114, 3.6090, 3.4062, 3.7334], device='cuda:0')
177
- 2023-10-17 18:38:38,781 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt
178
- 2023-10-17 18:38:39,042 INFO [utils.py:565] [wenetspeech-net_test-ctc-decoding] %WER 54.24% [13438 / 24773, 2 ins, 19 del, 13417 sub ]
179
- 2023-10-17 18:38:39,477 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt
180
- 2023-10-17 18:38:39,480 INFO [ctc_decode.py:522]
181
- For wenetspeech-net_test, WER of different settings are:
182
- ctc-decoding 54.24 best for wenetspeech-net_test
183
-
184
- 2023-10-17 18:38:39,481 INFO [ctc_decode.py:695] Start decoding test set: wenetspeech_dev
185
- 2023-10-17 18:38:41,576 INFO [ctc_decode.py:485] batch 0/?, cuts processed until now is 157
186
- 2023-10-17 18:38:51,281 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([4.5085, 2.3355, 4.5779, 4.0812], device='cuda:0')
187
- 2023-10-17 18:39:09,310 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([2.2675, 3.6412, 3.7390, 3.8346], device='cuda:0')
188
- 2023-10-17 18:39:20,015 INFO [zipformer.py:1853] name=None, attn_weights_entropy = tensor([1.0711, 2.5839, 2.5771, 3.9200], device='cuda:0')
189
- 2023-10-17 18:39:24,326 INFO [ctc_decode.py:499] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt
190
- 2023-10-17 18:39:24,419 INFO [utils.py:565] [wenetspeech_dev-ctc-decoding] %WER 64.88% [8970 / 13825, 1 ins, 1 del, 8968 sub ]
191
- 2023-10-17 18:39:24,662 INFO [ctc_decode.py:508] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt
192
- 2023-10-17 18:39:24,666 INFO [ctc_decode.py:522]
193
- For wenetspeech_dev, WER of different settings are:
194
- ctc-decoding 64.88 best for wenetspeech_dev
195
-
196
- 2023-10-17 18:39:24,666 INFO [ctc_decode.py:714] Done!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-20-avg-1-use-averaged-model-2023-10-25-16-34-00 ADDED
@@ -0,0 +1,266 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-10-25 16:34:00,598 INFO [ctc_decode.py:562] Decoding started
2
+ 2023-10-25 16:34:00,599 INFO [ctc_decode.py:568] Device: cuda:0
3
+ 2023-10-25 16:34:00,599 INFO [ctc_decode.py:569] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_zipformer_cn', 'icefall-git-sha1': '5b9014f7-clean', 'icefall-git-date': 'Tue Oct 24 16:08:39 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 20, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-20-avg-1-use-averaged-model'}
4
+ 2023-10-25 16:34:02,303 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
+ 2023-10-25 16:34:06,584 INFO [ctc_decode.py:589] About to create model
6
+ 2023-10-25 16:34:07,103 INFO [ctc_decode.py:656] Calculating the averaged model over epoch range from 19 (excluded) to 20
7
+ 2023-10-25 16:34:10,515 INFO [ctc_decode.py:673] Number of model parameters: 69651511
8
+ 2023-10-25 16:34:10,516 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
+ 2023-10-25 16:34:10,516 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
+ 2023-10-25 16:34:10,537 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
+ 2023-10-25 16:34:10,541 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
+ 2023-10-25 16:34:10,544 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
+ 2023-10-25 16:34:10,546 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
+ 2023-10-25 16:34:10,549 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
+ 2023-10-25 16:34:10,553 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
+ 2023-10-25 16:34:10,558 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
+ 2023-10-25 16:34:16,524 WARNING [ctc_decode.py:685] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
+ 2023-10-25 16:34:17,479 INFO [ctc_decode.py:697] Start decoding test set: aidatatang_test
19
+ 2023-10-25 16:34:18,685 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 80
20
+ 2023-10-25 16:34:21,412 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.0520, 3.1377, 2.0728, 2.0816], device='cuda:0')
21
+ 2023-10-25 16:34:30,837 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.5011, 3.0861, 3.4112, 3.0793], device='cuda:0')
22
+ 2023-10-25 16:34:36,258 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 9084
23
+ 2023-10-25 16:34:53,909 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 18516
24
+ 2023-10-25 16:35:05,250 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.4200, 3.4877, 2.4226, 2.5617], device='cuda:0')
25
+ 2023-10-25 16:35:11,730 INFO [ctc_decode.py:487] batch 300/?, cuts processed until now is 28179
26
+ 2023-10-25 16:35:26,845 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9498, 2.2983, 2.4953, 2.2775, 2.2739, 2.2083, 2.3184, 2.3589],
27
+ device='cuda:0')
28
+ 2023-10-25 16:35:29,566 INFO [ctc_decode.py:487] batch 400/?, cuts processed until now is 37667
29
+ 2023-10-25 16:35:46,177 INFO [ctc_decode.py:487] batch 500/?, cuts processed until now is 46172
30
+ 2023-10-25 16:35:52,683 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
31
+ 2023-10-25 16:35:53,533 INFO [utils.py:565] [aidatatang_test-ctc-decoding] %WER 3.36% [15767 / 468933, 2081 ins, 1924 del, 11762 sub ]
32
+ 2023-10-25 16:35:55,168 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt
33
+ 2023-10-25 16:35:55,171 INFO [ctc_decode.py:524]
34
+ For aidatatang_test, WER of different settings are:
35
+ ctc-decoding 3.36 best for aidatatang_test
36
+
37
+ 2023-10-25 16:35:55,172 INFO [ctc_decode.py:697] Start decoding test set: aidatatang_dev
38
+ 2023-10-25 16:35:56,437 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 81
39
+ 2023-10-25 16:36:14,093 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 9077
40
+ 2023-10-25 16:36:14,382 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1796, 2.6064, 2.9965, 2.5505], device='cuda:0')
41
+ 2023-10-25 16:36:16,426 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.6213, 2.0994, 2.2801, 2.0023, 2.2536, 2.3341, 1.9518, 2.1787],
42
+ device='cuda:0')
43
+ 2023-10-25 16:36:19,104 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1558, 2.5520, 2.9477, 2.5092], device='cuda:0')
44
+ 2023-10-25 16:36:24,127 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.8962, 2.4230, 2.6258, 2.3016, 2.5764, 2.6809, 2.2256, 2.4345],
45
+ device='cuda:0')
46
+ 2023-10-25 16:36:31,096 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 18432
47
+ 2023-10-25 16:36:31,516 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.8450, 2.0634, 2.1508, 1.8092], device='cuda:0')
48
+ 2023-10-25 16:36:32,160 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1301, 3.0183, 2.9528, 3.0556], device='cuda:0')
49
+ 2023-10-25 16:36:34,442 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1589, 3.0802, 3.0217, 3.1254], device='cuda:0')
50
+ 2023-10-25 16:36:42,515 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
51
+ 2023-10-25 16:36:42,937 INFO [utils.py:565] [aidatatang_dev-ctc-decoding] %WER 2.86% [6716 / 234524, 886 ins, 1013 del, 4817 sub ]
52
+ 2023-10-25 16:36:43,764 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt
53
+ 2023-10-25 16:36:43,768 INFO [ctc_decode.py:524]
54
+ For aidatatang_dev, WER of different settings are:
55
+ ctc-decoding 2.86 best for aidatatang_dev
56
+
57
+ 2023-10-25 16:36:43,768 INFO [ctc_decode.py:697] Start decoding test set: alimeeting_test
58
+ 2023-10-25 16:36:45,313 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 44
59
+ 2023-10-25 16:36:48,073 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.4251, 3.2546, 3.0462, 3.0440], device='cuda:0')
60
+ 2023-10-25 16:36:50,228 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9540, 1.9129, 2.0041, 3.1405], device='cuda:0')
61
+ 2023-10-25 16:36:53,939 WARNING [ctc_decode.py:685] Excluding cut with ID: R8008_M8016-8062-123 from decoding, num_frames: 6
62
+ 2023-10-25 16:37:06,497 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 7625
63
+ 2023-10-25 16:37:10,965 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.8114, 2.8180, 1.8182, 1.8959], device='cuda:0')
64
+ 2023-10-25 16:37:23,583 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt
65
+ 2023-10-25 16:37:23,991 INFO [utils.py:565] [alimeeting_test-ctc-decoding] %WER 24.28% [50952 / 209845, 4093 ins, 24060 del, 22799 sub ]
66
+ 2023-10-25 16:37:24,829 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt
67
+ 2023-10-25 16:37:24,833 INFO [ctc_decode.py:524]
68
+ For alimeeting_test, WER of different settings are:
69
+ ctc-decoding 24.28 best for alimeeting_test
70
+
71
+ 2023-10-25 16:37:24,833 INFO [ctc_decode.py:697] Start decoding test set: alimeeting_eval
72
+ 2023-10-25 16:37:25,951 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 35
73
+ 2023-10-25 16:37:40,611 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt
74
+ 2023-10-25 16:37:40,758 INFO [utils.py:565] [alimeeting_eval-ctc-decoding] %WER 22.93% [18597 / 81111, 1605 ins, 8426 del, 8566 sub ]
75
+ 2023-10-25 16:37:41,068 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt
76
+ 2023-10-25 16:37:41,071 INFO [ctc_decode.py:524]
77
+ For alimeeting_eval, WER of different settings are:
78
+ ctc-decoding 22.93 best for alimeeting_eval
79
+
80
+ 2023-10-25 16:37:41,072 INFO [ctc_decode.py:697] Start decoding test set: aishell_test
81
+ 2023-10-25 16:37:42,054 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 47
82
+ 2023-10-25 16:37:58,933 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 5468
83
+ 2023-10-25 16:38:03,464 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.8423, 3.3145, 3.6725, 3.2758, 3.5179, 3.1670, 3.0848, 3.5636],
84
+ device='cuda:0')
85
+ 2023-10-25 16:38:04,510 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell_test-epoch-20-avg-1-use-averaged-model.txt
86
+ 2023-10-25 16:38:04,686 INFO [utils.py:565] [aishell_test-ctc-decoding] %WER 2.27% [2380 / 104765, 359 ins, 80 del, 1941 sub ]
87
+ 2023-10-25 16:38:05,034 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell_test-epoch-20-avg-1-use-averaged-model.txt
88
+ 2023-10-25 16:38:05,037 INFO [ctc_decode.py:524]
89
+ For aishell_test, WER of different settings are:
90
+ ctc-decoding 2.27 best for aishell_test
91
+
92
+ 2023-10-25 16:38:05,038 INFO [ctc_decode.py:697] Start decoding test set: aishell_dev
93
+ 2023-10-25 16:38:06,188 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 53
94
+ 2023-10-25 16:38:06,246 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1611, 4.0929, 3.5322, 3.4682], device='cuda:0')
95
+ 2023-10-25 16:38:08,335 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.0768, 2.5535, 2.8456, 2.5765, 2.6620, 2.4640, 2.3513, 2.7931],
96
+ device='cuda:0')
97
+ 2023-10-25 16:38:10,493 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.4007, 4.5441, 5.0637, 4.6880], device='cuda:0')
98
+ 2023-10-25 16:38:11,557 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.7117, 3.1288, 3.5261, 2.2668], device='cuda:0')
99
+ 2023-10-25 16:38:14,764 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1489, 3.2889, 3.1936, 3.3052], device='cuda:0')
100
+ 2023-10-25 16:38:23,271 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 6034
101
+ 2023-10-25 16:38:39,877 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 12198
102
+ 2023-10-25 16:38:44,231 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.0977, 2.4464, 2.9487, 1.8454], device='cuda:0')
103
+ 2023-10-25 16:38:46,332 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt
104
+ 2023-10-25 16:38:46,686 INFO [utils.py:565] [aishell_dev-ctc-decoding] %WER 2.05% [4219 / 205341, 720 ins, 129 del, 3370 sub ]
105
+ 2023-10-25 16:38:47,358 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt
106
+ 2023-10-25 16:38:47,361 INFO [ctc_decode.py:524]
107
+ For aishell_dev, WER of different settings are:
108
+ ctc-decoding 2.05 best for aishell_dev
109
+
110
+ 2023-10-25 16:38:47,361 INFO [ctc_decode.py:697] Start decoding test set: aishell-2_test
111
+ 2023-10-25 16:38:48,202 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 83
112
+ 2023-10-25 16:38:58,110 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt
113
+ 2023-10-25 16:38:58,203 INFO [utils.py:565] [aishell-2_test-ctc-decoding] %WER 3.82% [1891 / 49532, 220 ins, 106 del, 1565 sub ]
114
+ 2023-10-25 16:38:58,387 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt
115
+ 2023-10-25 16:38:58,390 INFO [ctc_decode.py:524]
116
+ For aishell-2_test, WER of different settings are:
117
+ ctc-decoding 3.82 best for aishell-2_test
118
+
119
+ 2023-10-25 16:38:58,390 INFO [ctc_decode.py:697] Start decoding test set: aishell-2_dev
120
+ 2023-10-25 16:38:59,005 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 81
121
+ 2023-10-25 16:39:03,843 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt
122
+ 2023-10-25 16:39:03,888 INFO [utils.py:565] [aishell-2_dev-ctc-decoding] %WER 3.33% [825 / 24802, 68 ins, 34 del, 723 sub ]
123
+ 2023-10-25 16:39:03,981 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt
124
+ 2023-10-25 16:39:03,984 INFO [ctc_decode.py:524]
125
+ For aishell-2_dev, WER of different settings are:
126
+ ctc-decoding 3.33 best for aishell-2_dev
127
+
128
+ 2023-10-25 16:39:03,984 INFO [ctc_decode.py:697] Start decoding test set: aishell-4
129
+ 2023-10-25 16:39:05,396 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 33
130
+ 2023-10-25 16:39:06,514 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.7161, 3.0260, 3.4697, 3.0798, 3.1194, 2.9266, 3.3018, 3.3440],
131
+ device='cuda:0')
132
+ 2023-10-25 16:39:13,020 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.4888, 4.7211, 5.2793, 4.8764], device='cuda:0')
133
+ 2023-10-25 16:39:25,289 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 5000
134
+ 2023-10-25 16:39:37,714 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.4363, 2.2623, 2.1157, 2.1364], device='cuda:0')
135
+ 2023-10-25 16:39:39,959 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-aishell-4-epoch-20-avg-1-use-averaged-model.txt
136
+ 2023-10-25 16:39:40,276 INFO [utils.py:565] [aishell-4-ctc-decoding] %WER 15.45% [27907 / 180665, 4812 ins, 8373 del, 14722 sub ]
137
+ 2023-10-25 16:39:40,915 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-aishell-4-epoch-20-avg-1-use-averaged-model.txt
138
+ 2023-10-25 16:39:40,922 INFO [ctc_decode.py:524]
139
+ For aishell-4, WER of different settings are:
140
+ ctc-decoding 15.45 best for aishell-4
141
+
142
+ 2023-10-25 16:39:40,923 INFO [ctc_decode.py:697] Start decoding test set: magicdata_test
143
+ 2023-10-25 16:39:42,150 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 57
144
+ 2023-10-25 16:39:45,980 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2365, 3.2766, 3.2887, 3.2825], device='cuda:0')
145
+ 2023-10-25 16:39:59,489 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.5007, 3.1130, 3.2352, 2.9490, 3.2145, 2.7877, 2.3506, 2.7924],
146
+ device='cuda:0')
147
+ 2023-10-25 16:40:00,114 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 6425
148
+ 2023-10-25 16:40:01,578 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.3456, 2.5491, 4.3242, 1.8453], device='cuda:0')
149
+ 2023-10-25 16:40:01,887 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.3791, 4.0503, 3.3906, 3.4577], device='cuda:0')
150
+ 2023-10-25 16:40:18,421 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 13211
151
+ 2023-10-25 16:40:20,419 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2894, 3.1911, 3.2014, 3.2303], device='cuda:0')
152
+ 2023-10-25 16:40:35,794 INFO [ctc_decode.py:487] batch 300/?, cuts processed until now is 20136
153
+ 2023-10-25 16:40:52,211 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt
154
+ 2023-10-25 16:40:52,654 INFO [utils.py:565] [magicdata_test-ctc-decoding] %WER 2.77% [6622 / 239091, 1024 ins, 428 del, 5170 sub ]
155
+ 2023-10-25 16:40:53,509 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt
156
+ 2023-10-25 16:40:53,512 INFO [ctc_decode.py:524]
157
+ For magicdata_test, WER of different settings are:
158
+ ctc-decoding 2.77 best for magicdata_test
159
+
160
+ 2023-10-25 16:40:53,513 INFO [ctc_decode.py:697] Start decoding test set: magicdata_dev
161
+ 2023-10-25 16:40:54,770 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 52
162
+ 2023-10-25 16:40:54,808 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.1237, 4.6547, 5.0120, 4.7204], device='cuda:0')
163
+ 2023-10-25 16:41:10,069 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.4620, 2.6096, 4.3213, 1.8959], device='cuda:0')
164
+ 2023-10-25 16:41:11,956 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.0565, 2.5376, 2.6165, 2.4429, 2.4652, 2.3741, 1.8434, 2.0589],
165
+ device='cuda:0')
166
+ 2023-10-25 16:41:12,750 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 5919
167
+ 2023-10-25 16:41:24,277 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.8365, 2.6056, 3.9032, 2.6469], device='cuda:0')
168
+ 2023-10-25 16:41:29,523 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.2491, 3.8004, 4.2328, 4.2613], device='cuda:0')
169
+ 2023-10-25 16:41:30,449 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 11646
170
+ 2023-10-25 16:41:31,119 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.4962, 2.5740, 2.7954, 3.6154], device='cuda:0')
171
+ 2023-10-25 16:41:31,588 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt
172
+ 2023-10-25 16:41:31,804 INFO [utils.py:565] [magicdata_dev-ctc-decoding] %WER 3.49% [4080 / 116800, 645 ins, 274 del, 3161 sub ]
173
+ 2023-10-25 16:41:32,228 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt
174
+ 2023-10-25 16:41:32,231 INFO [ctc_decode.py:524]
175
+ For magicdata_dev, WER of different settings are:
176
+ ctc-decoding 3.49 best for magicdata_dev
177
+
178
+ 2023-10-25 16:41:32,232 INFO [ctc_decode.py:697] Start decoding test set: kespeech-asr_test
179
+ 2023-10-25 16:41:33,489 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 45
180
+ 2023-10-25 16:41:38,619 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1330, 2.6924, 2.8790, 2.6249, 2.7502, 2.6020, 2.9186, 2.8276],
181
+ device='cuda:0')
182
+ 2023-10-25 16:41:52,367 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 4867
183
+ 2023-10-25 16:42:11,157 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 9965
184
+ 2023-10-25 16:42:29,384 INFO [ctc_decode.py:487] batch 300/?, cuts processed until now is 15124
185
+ 2023-10-25 16:42:47,690 INFO [ctc_decode.py:487] batch 400/?, cuts processed until now is 19643
186
+ 2023-10-25 16:42:48,374 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt
187
+ 2023-10-25 16:42:48,854 INFO [utils.py:565] [kespeech-asr_test-ctc-decoding] %WER 8.29% [23525 / 283772, 2738 ins, 1869 del, 18918 sub ]
188
+ 2023-10-25 16:42:49,821 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt
189
+ 2023-10-25 16:42:49,825 INFO [ctc_decode.py:524]
190
+ For kespeech-asr_test, WER of different settings are:
191
+ ctc-decoding 8.29 best for kespeech-asr_test
192
+
193
+ 2023-10-25 16:42:49,825 INFO [ctc_decode.py:697] Start decoding test set: kespeech-asr_dev_phase1
194
+ 2023-10-25 16:42:50,401 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 44
195
+ 2023-10-25 16:42:56,659 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.4923, 4.8152, 5.2681, 4.9011], device='cuda:0')
196
+ 2023-10-25 16:42:58,778 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt
197
+ 2023-10-25 16:42:58,837 INFO [utils.py:565] [kespeech-asr_dev_phase1-ctc-decoding] %WER 6.90% [2182 / 31634, 293 ins, 184 del, 1705 sub ]
198
+ 2023-10-25 16:42:58,950 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt
199
+ 2023-10-25 16:42:58,953 INFO [ctc_decode.py:524]
200
+ For kespeech-asr_dev_phase1, WER of different settings are:
201
+ ctc-decoding 6.9 best for kespeech-asr_dev_phase1
202
+
203
+ 2023-10-25 16:42:58,953 INFO [ctc_decode.py:697] Start decoding test set: kespeech-asr_dev_phase2
204
+ 2023-10-25 16:42:59,835 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 47
205
+ 2023-10-25 16:43:06,820 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.6326, 2.9347, 4.1799, 2.9961], device='cuda:0')
206
+ 2023-10-25 16:43:07,937 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt
207
+ 2023-10-25 16:43:07,995 INFO [utils.py:565] [kespeech-asr_dev_phase2-ctc-decoding] %WER 2.85% [910 / 31928, 133 ins, 58 del, 719 sub ]
208
+ 2023-10-25 16:43:08,106 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt
209
+ 2023-10-25 16:43:08,109 INFO [ctc_decode.py:524]
210
+ For kespeech-asr_dev_phase2, WER of different settings are:
211
+ ctc-decoding 2.85 best for kespeech-asr_dev_phase2
212
+
213
+ 2023-10-25 16:43:08,110 INFO [ctc_decode.py:697] Start decoding test set: wenetspeech-meeting_test
214
+ 2023-10-25 16:43:09,586 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 28
215
+ 2023-10-25 16:43:22,437 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.8702, 3.0656, 1.9713, 2.1329], device='cuda:0')
216
+ 2023-10-25 16:43:23,655 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9881, 2.5628, 4.7751, 1.8426], device='cuda:0')
217
+ 2023-10-25 16:43:25,199 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.5829, 2.5263, 2.3613, 4.8147], device='cuda:0')
218
+ 2023-10-25 16:43:30,212 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 3776
219
+ 2023-10-25 16:43:48,162 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2392, 2.2775, 2.1982, 3.4906], device='cuda:0')
220
+ 2023-10-25 16:43:49,940 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 8092
221
+ 2023-10-25 16:43:51,128 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.1897, 3.7199, 3.5001, 4.8150], device='cuda:0')
222
+ 2023-10-25 16:43:51,827 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt
223
+ 2023-10-25 16:43:52,189 INFO [utils.py:565] [wenetspeech-meeting_test-ctc-decoding] %WER 6.92% [15242 / 220385, 1952 ins, 3482 del, 9808 sub ]
224
+ 2023-10-25 16:43:52,902 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt
225
+ 2023-10-25 16:43:52,906 INFO [ctc_decode.py:524]
226
+ For wenetspeech-meeting_test, WER of different settings are:
227
+ ctc-decoding 6.92 best for wenetspeech-meeting_test
228
+
229
+ 2023-10-25 16:43:52,906 INFO [ctc_decode.py:697] Start decoding test set: wenetspeech-net_test
230
+ 2023-10-25 16:43:53,117 WARNING [ctc_decode.py:685] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
231
+ 2023-10-25 16:43:54,558 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 43
232
+ 2023-10-25 16:44:04,413 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.0308, 2.1754, 2.2245, 3.5140], device='cuda:0')
233
+ 2023-10-25 16:44:11,513 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.8906, 1.8713, 1.7192, 2.6720], device='cuda:0')
234
+ 2023-10-25 16:44:12,503 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([0.9986, 1.9138, 1.9920, 2.9444], device='cuda:0')
235
+ 2023-10-25 16:44:14,225 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 7009
236
+ 2023-10-25 16:44:30,429 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.9908, 2.6090, 2.9691, 2.9051], device='cuda:0')
237
+ 2023-10-25 16:44:31,876 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.7256, 4.8738, 3.7879, 3.5543], device='cuda:0')
238
+ 2023-10-25 16:44:33,966 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 14995
239
+ 2023-10-25 16:44:45,561 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.4122, 2.4497, 1.7039, 1.7159], device='cuda:0')
240
+ 2023-10-25 16:44:52,612 INFO [ctc_decode.py:487] batch 300/?, cuts processed until now is 22693
241
+ 2023-10-25 16:44:58,577 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt
242
+ 2023-10-25 16:44:59,335 INFO [utils.py:565] [wenetspeech-net_test-ctc-decoding] %WER 8.57% [35611 / 415746, 2894 ins, 13613 del, 19104 sub ]
243
+ 2023-10-25 16:45:00,779 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt
244
+ 2023-10-25 16:45:00,782 INFO [ctc_decode.py:524]
245
+ For wenetspeech-net_test, WER of different settings are:
246
+ ctc-decoding 8.57 best for wenetspeech-net_test
247
+
248
+ 2023-10-25 16:45:00,783 INFO [ctc_decode.py:697] Start decoding test set: wenetspeech_dev
249
+ 2023-10-25 16:45:02,490 INFO [ctc_decode.py:487] batch 0/?, cuts processed until now is 39
250
+ 2023-10-25 16:45:07,449 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.3442, 3.0899, 3.7565, 3.7398], device='cuda:0')
251
+ 2023-10-25 16:45:19,403 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.4367, 2.7283, 2.6068, 4.3270], device='cuda:0')
252
+ 2023-10-25 16:45:20,121 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.6771, 3.6862, 3.2837, 3.3905], device='cuda:0')
253
+ 2023-10-25 16:45:21,677 INFO [ctc_decode.py:487] batch 100/?, cuts processed until now is 4983
254
+ 2023-10-25 16:45:25,334 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.9137, 2.1321, 4.0173, 2.2182], device='cuda:0')
255
+ 2023-10-25 16:45:35,271 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([5.7926, 5.6763, 5.3816, 5.9253], device='cuda:0')
256
+ 2023-10-25 16:45:36,045 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2507, 4.5774, 3.8076, 4.2477], device='cuda:0')
257
+ 2023-10-25 16:45:40,086 INFO [ctc_decode.py:487] batch 200/?, cuts processed until now is 10268
258
+ 2023-10-25 16:45:46,450 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.2227, 2.1938, 2.0968, 4.2055], device='cuda:0')
259
+ 2023-10-25 16:45:51,879 INFO [ctc_decode.py:501] The transcripts are stored in zipformer/exp-w-ctc/ctc-decoding/recogs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt
260
+ 2023-10-25 16:45:52,426 INFO [utils.py:565] [wenetspeech_dev-ctc-decoding] %WER 9.41% [31089 / 330498, 1851 ins, 18854 del, 10384 sub ]
261
+ 2023-10-25 16:45:53,512 INFO [ctc_decode.py:510] Wrote detailed error stats to zipformer/exp-w-ctc/ctc-decoding/errs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt
262
+ 2023-10-25 16:45:53,516 INFO [ctc_decode.py:524]
263
+ For wenetspeech_dev, WER of different settings are:
264
+ ctc-decoding 9.41 best for wenetspeech_dev
265
+
266
+ 2023-10-25 16:45:53,516 INFO [ctc_decode.py:716] Done!
decoding_results/ctc-decoding/log-decode-epoch-22-avg-1-use-averaged-model-2023-10-17-16-37-56 DELETED
@@ -1,3 +0,0 @@
1
- 2023-10-17 16:37:56,505 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:37:56,505 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:37:56,505 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 22, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-22-avg-1-use-averaged-model'}
 
 
 
 
decoding_results/ctc-decoding/log-decode-epoch-22-avg-1-use-averaged-model-2023-10-17-16-38-08 DELETED
@@ -1,18 +0,0 @@
1
- 2023-10-17 16:38:08,666 INFO [ctc_decode.py:560] Decoding started
2
- 2023-10-17 16:38:08,666 INFO [ctc_decode.py:566] Device: cuda:0
3
- 2023-10-17 16:38:08,666 INFO [ctc_decode.py:567] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.3', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '821ebc378e7fb99b8adc81950227963332821e01', 'k2-git-date': 'Wed Jul 19 15:38:25 2023', 'lhotse-version': '1.16.0.dev+git.1db4d97a.clean', 'torch-version': '1.11.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.9', 'icefall-git-branch': 'dev_multi_zh-hans', 'icefall-git-sha1': '919793d3-dirty', 'icefall-git-date': 'Thu Sep 7 21:06:37 2023', 'icefall-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/icefall-1.0-py3.9.egg', 'k2-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/k2-1.24.3.dev20230721+cuda10.2.torch1.11.0-py3.9-linux-x86_64.egg/k2/__init__.py', 'lhotse-path': '/star-home/jinzengrui/lib/miniconda3/envs/dev39/lib/python3.9/site-packages/lhotse-1.16.0.dev0+git.1db4d97a.clean-py3.9.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-1-1220091118-57c4d55446-mvd6x', 'IP address': '10.177.22.19'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 22, 'iter': 0, 'avg': 1, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-w-ctc'), 'bpe_model': 'data/lang_bpe_2000/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_2000'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'num_encoder_layers': '2,2,3,4,3,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1024,1536,1024,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,384,512,384,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,256,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': True, 'use_ctc': True, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 300.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-w-ctc/ctc-decoding'), 'suffix': 'epoch-22-avg-1-use-averaged-model'}
4
- 2023-10-17 16:38:10,507 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_2000/Linv.pt
5
- 2023-10-17 16:38:16,142 INFO [ctc_decode.py:587] About to create model
6
- 2023-10-17 16:38:16,733 INFO [ctc_decode.py:654] Calculating the averaged model over epoch range from 21 (excluded) to 22
7
- 2023-10-17 16:38:24,407 INFO [ctc_decode.py:671] Number of model parameters: 69651511
8
- 2023-10-17 16:38:24,407 INFO [multi_dataset.py:221] About to get multidataset test cuts
9
- 2023-10-17 16:38:24,407 INFO [multi_dataset.py:224] Loading Aidatatang_200zh set in lazy mode
10
- 2023-10-17 16:38:24,424 INFO [multi_dataset.py:233] Loading Aishell set in lazy mode
11
- 2023-10-17 16:38:24,428 INFO [multi_dataset.py:242] Loading Aishell-2 set in lazy mode
12
- 2023-10-17 16:38:24,431 INFO [multi_dataset.py:251] Loading Aishell-4 TEST set in lazy mode
13
- 2023-10-17 16:38:24,433 INFO [multi_dataset.py:257] Loading Ali-Meeting set in lazy mode
14
- 2023-10-17 16:38:24,436 INFO [multi_dataset.py:266] Loading MagicData set in lazy mode
15
- 2023-10-17 16:38:24,439 INFO [multi_dataset.py:275] Loading KeSpeech set in lazy mode
16
- 2023-10-17 16:38:24,445 INFO [multi_dataset.py:287] Loading WeNetSpeech set in lazy mode
17
- 2023-10-17 16:38:30,811 WARNING [ctc_decode.py:683] Excluding cut with ID: TEST_NET_Y0000000004_0ub4ZzdHzBc_S00023 from decoding, num_frames: 8
18
- 2023-10-17 16:38:31,821 INFO [ctc_decode.py:695] Start decoding test set: aidatatang_test
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
decoding_results/ctc-decoding/recogs-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aishell-4-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aishell_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-aishell_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-alimeeting_eval-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-alimeeting_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-kespeech-asr_dev_phase1-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-kespeech-asr_dev_phase2-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-kespeech-asr_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-magicdata_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-magicdata_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-wenetspeech-meeting_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-wenetspeech-net_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-wenetspeech_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/wer-summary-aidatatang_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
@@ -1,2 +1,2 @@
1
  settings WER
2
- ctc-decoding 14.57
 
1
  settings WER
2
+ ctc-decoding 2.86
decoding_results/ctc-decoding/wer-summary-aidatatang_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
@@ -1,2 +1,2 @@
1
  settings WER
2
- ctc-decoding 15.26
 
1
  settings WER
2
+ ctc-decoding 3.36
decoding_results/ctc-decoding/wer-summary-aishell-2_dev-epoch-20-avg-1-use-averaged-model.txt CHANGED
@@ -1,2 +1,2 @@
1
  settings WER
2
- ctc-decoding 23.56
 
1
  settings WER
2
+ ctc-decoding 3.33
decoding_results/ctc-decoding/wer-summary-aishell-2_test-epoch-20-avg-1-use-averaged-model.txt CHANGED
@@ -1,2 +1,2 @@
1
  settings WER
2
- ctc-decoding 25.55
 
1
  settings WER
2
+ ctc-decoding 3.82
decoding_results/ctc-decoding/wer-summary-aishell-4-epoch-20-avg-1-use-averaged-model.txt CHANGED
@@ -1,2 +1,2 @@
1
  settings WER
2
- ctc-decoding 71.75
 
1
  settings WER
2
+ ctc-decoding 15.45