pruned-transducer-stateless7-streaming-id / exp /streaming /fast_beam_search /log-decode-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model-2023-06-21-10-04-38
w11wo's picture
Added Model
9a835b2
raw
history blame contribute delete
No virus
15.1 kB
2023-06-21 10:04:38,023 INFO [streaming_decode.py:483] Decoding started
2023-06-21 10:04:38,023 INFO [streaming_decode.py:489] Device: cuda:0
2023-06-21 10:04:38,024 INFO [lexicon.py:168] Loading pre-compiled data/lang_phone/Linv.pt
2023-06-21 10:04:38,027 INFO [streaming_decode.py:497] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.23.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '9426c9f730820d291f5dcb06be337662595fa7b4', 'k2-git-date': 'Sun Feb 5 17:35:01 2023', 'lhotse-version': '1.15.0.dev+git.00d3e36.clean', 'torch-version': '1.13.1+cu117', 'torch-cuda-available': True, 'torch-cuda-version': '11.7', 'python-version': '3.1', 'icefall-git-branch': 'master', 'icefall-git-sha1': 'd3f5d01-dirty', 'icefall-git-date': 'Wed May 31 04:15:45 2023', 'icefall-path': '/root/icefall', 'k2-path': '/usr/local/lib/python3.10/dist-packages/k2/__init__.py', 'lhotse-path': '/root/lhotse/lhotse/__init__.py', 'hostname': 'bookbot-k2', 'IP address': '127.0.0.1'}, 'epoch': 30, 'iter': 0, 'avg': 9, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_streaming/exp'), 'lang_dir': 'data/lang_phone', 'decoding_method': 'fast_beam_search', 'num_active_paths': 4, 'beam': 4, 'max_contexts': 4, 'max_states': 32, 'context_size': 2, 'num_decode_streams': 1500, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'short_chunk_size': 50, 'num_left_chunks': 4, 'decode_chunk_len': 32, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search'), 'suffix': 'epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model', 'blank_id': 0, 'unk_id': 7, 'vocab_size': 33}
2023-06-21 10:04:38,027 INFO [streaming_decode.py:499] About to create model
2023-06-21 10:04:38,604 INFO [zipformer.py:405] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8.
2023-06-21 10:04:38,608 INFO [streaming_decode.py:566] Calculating the averaged model over epoch range from 21 (excluded) to 30
2023-06-21 10:04:42,203 INFO [streaming_decode.py:588] Number of model parameters: 69471350
2023-06-21 10:04:42,204 INFO [multidataset.py:122] About to get LibriVox test cuts
2023-06-21 10:04:42,204 INFO [multidataset.py:124] Loading LibriVox in lazy mode
2023-06-21 10:04:42,204 INFO [multidataset.py:133] About to get FLEURS test cuts
2023-06-21 10:04:42,204 INFO [multidataset.py:135] Loading FLEURS in lazy mode
2023-06-21 10:04:42,205 INFO [multidataset.py:144] About to get Common Voice test cuts
2023-06-21 10:04:42,205 INFO [multidataset.py:146] Loading Common Voice in lazy mode
2023-06-21 10:04:42,471 INFO [streaming_decode.py:380] Cuts processed until now is 0.
2023-06-21 10:04:42,786 INFO [streaming_decode.py:380] Cuts processed until now is 50.
2023-06-21 10:04:43,098 INFO [streaming_decode.py:380] Cuts processed until now is 100.
2023-06-21 10:04:43,444 INFO [streaming_decode.py:380] Cuts processed until now is 150.
2023-06-21 10:04:43,770 INFO [streaming_decode.py:380] Cuts processed until now is 200.
2023-06-21 10:04:44,092 INFO [streaming_decode.py:380] Cuts processed until now is 250.
2023-06-21 10:04:44,416 INFO [streaming_decode.py:380] Cuts processed until now is 300.
2023-06-21 10:04:44,756 INFO [streaming_decode.py:380] Cuts processed until now is 350.
2023-06-21 10:04:45,079 INFO [streaming_decode.py:380] Cuts processed until now is 400.
2023-06-21 10:04:45,405 INFO [streaming_decode.py:380] Cuts processed until now is 450.
2023-06-21 10:04:45,734 INFO [streaming_decode.py:380] Cuts processed until now is 500.
2023-06-21 10:04:46,071 INFO [streaming_decode.py:380] Cuts processed until now is 550.
2023-06-21 10:04:46,405 INFO [streaming_decode.py:380] Cuts processed until now is 600.
2023-06-21 10:04:57,029 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-librivox-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:04:57,063 INFO [utils.py:561] [test-librivox-beam_4_max_contexts_4_max_states_32] %WER 4.81% [1759 / 36594, 280 ins, 892 del, 587 sub ]
2023-06-21 10:04:57,144 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-librivox-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:04:57,145 INFO [streaming_decode.py:450]
For test-librivox, WER of different settings are:
beam_4_max_contexts_4_max_states_32 4.81 best for test-librivox
2023-06-21 10:04:57,149 INFO [streaming_decode.py:380] Cuts processed until now is 0.
2023-06-21 10:04:57,332 INFO [streaming_decode.py:380] Cuts processed until now is 50.
2023-06-21 10:04:57,494 INFO [streaming_decode.py:380] Cuts processed until now is 100.
2023-06-21 10:04:57,663 INFO [streaming_decode.py:380] Cuts processed until now is 150.
2023-06-21 10:04:57,833 INFO [streaming_decode.py:380] Cuts processed until now is 200.
2023-06-21 10:04:58,000 INFO [streaming_decode.py:380] Cuts processed until now is 250.
2023-06-21 10:04:58,161 INFO [streaming_decode.py:380] Cuts processed until now is 300.
2023-06-21 10:04:58,323 INFO [streaming_decode.py:380] Cuts processed until now is 350.
2023-06-21 10:04:58,488 INFO [streaming_decode.py:380] Cuts processed until now is 400.
2023-06-21 10:04:58,656 INFO [streaming_decode.py:380] Cuts processed until now is 450.
2023-06-21 10:04:58,819 INFO [streaming_decode.py:380] Cuts processed until now is 500.
2023-06-21 10:04:58,993 INFO [streaming_decode.py:380] Cuts processed until now is 550.
2023-06-21 10:04:59,176 INFO [streaming_decode.py:380] Cuts processed until now is 600.
2023-06-21 10:04:59,364 INFO [streaming_decode.py:380] Cuts processed until now is 650.
2023-06-21 10:05:34,495 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-fleurs-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:05:34,590 INFO [utils.py:561] [test-fleurs-beam_4_max_contexts_4_max_states_32] %WER 12.93% [12100 / 93580, 1706 ins, 5594 del, 4800 sub ]
2023-06-21 10:05:34,813 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-fleurs-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:05:34,814 INFO [streaming_decode.py:450]
For test-fleurs, WER of different settings are:
beam_4_max_contexts_4_max_states_32 12.93 best for test-fleurs
2023-06-21 10:05:34,820 INFO [streaming_decode.py:380] Cuts processed until now is 0.
2023-06-21 10:05:35,059 INFO [streaming_decode.py:380] Cuts processed until now is 50.
2023-06-21 10:05:35,308 INFO [streaming_decode.py:380] Cuts processed until now is 100.
2023-06-21 10:05:35,583 INFO [streaming_decode.py:380] Cuts processed until now is 150.
2023-06-21 10:05:35,829 INFO [streaming_decode.py:380] Cuts processed until now is 200.
2023-06-21 10:05:36,082 INFO [streaming_decode.py:380] Cuts processed until now is 250.
2023-06-21 10:05:36,315 INFO [streaming_decode.py:380] Cuts processed until now is 300.
2023-06-21 10:05:36,537 INFO [streaming_decode.py:380] Cuts processed until now is 350.
2023-06-21 10:05:36,797 INFO [streaming_decode.py:380] Cuts processed until now is 400.
2023-06-21 10:05:37,028 INFO [streaming_decode.py:380] Cuts processed until now is 450.
2023-06-21 10:05:37,263 INFO [streaming_decode.py:380] Cuts processed until now is 500.
2023-06-21 10:05:37,499 INFO [streaming_decode.py:380] Cuts processed until now is 550.
2023-06-21 10:05:37,720 INFO [streaming_decode.py:380] Cuts processed until now is 600.
2023-06-21 10:05:37,959 INFO [streaming_decode.py:380] Cuts processed until now is 650.
2023-06-21 10:05:38,182 INFO [streaming_decode.py:380] Cuts processed until now is 700.
2023-06-21 10:05:38,406 INFO [streaming_decode.py:380] Cuts processed until now is 750.
2023-06-21 10:05:38,664 INFO [streaming_decode.py:380] Cuts processed until now is 800.
2023-06-21 10:05:38,913 INFO [streaming_decode.py:380] Cuts processed until now is 850.
2023-06-21 10:05:39,251 INFO [streaming_decode.py:380] Cuts processed until now is 900.
2023-06-21 10:05:39,493 INFO [streaming_decode.py:380] Cuts processed until now is 950.
2023-06-21 10:05:39,726 INFO [streaming_decode.py:380] Cuts processed until now is 1000.
2023-06-21 10:05:39,959 INFO [streaming_decode.py:380] Cuts processed until now is 1050.
2023-06-21 10:05:40,192 INFO [streaming_decode.py:380] Cuts processed until now is 1100.
2023-06-21 10:05:40,436 INFO [streaming_decode.py:380] Cuts processed until now is 1150.
2023-06-21 10:05:40,709 INFO [streaming_decode.py:380] Cuts processed until now is 1200.
2023-06-21 10:05:40,959 INFO [streaming_decode.py:380] Cuts processed until now is 1250.
2023-06-21 10:05:41,199 INFO [streaming_decode.py:380] Cuts processed until now is 1300.
2023-06-21 10:05:41,448 INFO [streaming_decode.py:380] Cuts processed until now is 1350.
2023-06-21 10:05:41,697 INFO [streaming_decode.py:380] Cuts processed until now is 1400.
2023-06-21 10:05:41,938 INFO [streaming_decode.py:380] Cuts processed until now is 1450.
2023-06-21 10:05:51,050 INFO [streaming_decode.py:380] Cuts processed until now is 1500.
2023-06-21 10:05:53,941 INFO [streaming_decode.py:380] Cuts processed until now is 1550.
2023-06-21 10:05:55,569 INFO [streaming_decode.py:380] Cuts processed until now is 1600.
2023-06-21 10:05:55,799 INFO [streaming_decode.py:380] Cuts processed until now is 1650.
2023-06-21 10:05:57,493 INFO [streaming_decode.py:380] Cuts processed until now is 1700.
2023-06-21 10:05:57,735 INFO [streaming_decode.py:380] Cuts processed until now is 1750.
2023-06-21 10:05:57,961 INFO [streaming_decode.py:380] Cuts processed until now is 1800.
2023-06-21 10:05:59,694 INFO [streaming_decode.py:380] Cuts processed until now is 1850.
2023-06-21 10:05:59,923 INFO [streaming_decode.py:380] Cuts processed until now is 1900.
2023-06-21 10:06:00,151 INFO [streaming_decode.py:380] Cuts processed until now is 1950.
2023-06-21 10:06:01,771 INFO [streaming_decode.py:380] Cuts processed until now is 2000.
2023-06-21 10:06:01,997 INFO [streaming_decode.py:380] Cuts processed until now is 2050.
2023-06-21 10:06:02,241 INFO [streaming_decode.py:380] Cuts processed until now is 2100.
2023-06-21 10:06:02,465 INFO [streaming_decode.py:380] Cuts processed until now is 2150.
2023-06-21 10:06:04,249 INFO [streaming_decode.py:380] Cuts processed until now is 2200.
2023-06-21 10:06:04,478 INFO [streaming_decode.py:380] Cuts processed until now is 2250.
2023-06-21 10:06:04,710 INFO [streaming_decode.py:380] Cuts processed until now is 2300.
2023-06-21 10:06:06,461 INFO [streaming_decode.py:380] Cuts processed until now is 2350.
2023-06-21 10:06:06,697 INFO [streaming_decode.py:380] Cuts processed until now is 2400.
2023-06-21 10:06:06,931 INFO [streaming_decode.py:380] Cuts processed until now is 2450.
2023-06-21 10:06:08,726 INFO [streaming_decode.py:380] Cuts processed until now is 2500.
2023-06-21 10:06:08,950 INFO [streaming_decode.py:380] Cuts processed until now is 2550.
2023-06-21 10:06:09,187 INFO [streaming_decode.py:380] Cuts processed until now is 2600.
2023-06-21 10:06:10,940 INFO [streaming_decode.py:380] Cuts processed until now is 2650.
2023-06-21 10:06:11,165 INFO [streaming_decode.py:380] Cuts processed until now is 2700.
2023-06-21 10:06:12,942 INFO [streaming_decode.py:380] Cuts processed until now is 2750.
2023-06-21 10:06:13,183 INFO [streaming_decode.py:380] Cuts processed until now is 2800.
2023-06-21 10:06:14,919 INFO [streaming_decode.py:380] Cuts processed until now is 2850.
2023-06-21 10:06:16,667 INFO [streaming_decode.py:380] Cuts processed until now is 2900.
2023-06-21 10:06:18,270 INFO [streaming_decode.py:380] Cuts processed until now is 2950.
2023-06-21 10:06:19,990 INFO [streaming_decode.py:380] Cuts processed until now is 3000.
2023-06-21 10:06:20,222 INFO [streaming_decode.py:380] Cuts processed until now is 3050.
2023-06-21 10:06:21,952 INFO [streaming_decode.py:380] Cuts processed until now is 3100.
2023-06-21 10:06:22,202 INFO [streaming_decode.py:380] Cuts processed until now is 3150.
2023-06-21 10:06:23,959 INFO [streaming_decode.py:380] Cuts processed until now is 3200.
2023-06-21 10:06:24,183 INFO [streaming_decode.py:380] Cuts processed until now is 3250.
2023-06-21 10:06:25,951 INFO [streaming_decode.py:380] Cuts processed until now is 3300.
2023-06-21 10:06:26,203 INFO [streaming_decode.py:380] Cuts processed until now is 3350.
2023-06-21 10:06:27,984 INFO [streaming_decode.py:380] Cuts processed until now is 3400.
2023-06-21 10:06:28,228 INFO [streaming_decode.py:380] Cuts processed until now is 3450.
2023-06-21 10:06:28,468 INFO [streaming_decode.py:380] Cuts processed until now is 3500.
2023-06-21 10:06:30,266 INFO [streaming_decode.py:380] Cuts processed until now is 3550.
2023-06-21 10:06:30,497 INFO [streaming_decode.py:380] Cuts processed until now is 3600.
2023-06-21 10:06:45,693 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-commonvoice-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:06:45,825 INFO [utils.py:561] [test-commonvoice-beam_4_max_contexts_4_max_states_32] %WER 14.96% [19859 / 132787, 3004 ins, 8788 del, 8067 sub ]
2023-06-21 10:06:46,126 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-commonvoice-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt
2023-06-21 10:06:46,126 INFO [streaming_decode.py:450]
For test-commonvoice, WER of different settings are:
beam_4_max_contexts_4_max_states_32 14.96 best for test-commonvoice
2023-06-21 10:06:46,127 INFO [streaming_decode.py:618] Done!