2023-06-21 10:04:38,023 INFO [streaming_decode.py:483] Decoding started 2023-06-21 10:04:38,023 INFO [streaming_decode.py:489] Device: cuda:0 2023-06-21 10:04:38,024 INFO [lexicon.py:168] Loading pre-compiled data/lang_phone/Linv.pt 2023-06-21 10:04:38,027 INFO [streaming_decode.py:497] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.23.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '9426c9f730820d291f5dcb06be337662595fa7b4', 'k2-git-date': 'Sun Feb 5 17:35:01 2023', 'lhotse-version': '1.15.0.dev+git.00d3e36.clean', 'torch-version': '1.13.1+cu117', 'torch-cuda-available': True, 'torch-cuda-version': '11.7', 'python-version': '3.1', 'icefall-git-branch': 'master', 'icefall-git-sha1': 'd3f5d01-dirty', 'icefall-git-date': 'Wed May 31 04:15:45 2023', 'icefall-path': '/root/icefall', 'k2-path': '/usr/local/lib/python3.10/dist-packages/k2/__init__.py', 'lhotse-path': '/root/lhotse/lhotse/__init__.py', 'hostname': 'bookbot-k2', 'IP address': '127.0.0.1'}, 'epoch': 30, 'iter': 0, 'avg': 9, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7_streaming/exp'), 'lang_dir': 'data/lang_phone', 'decoding_method': 'fast_beam_search', 'num_active_paths': 4, 'beam': 4, 'max_contexts': 4, 'max_states': 32, 'context_size': 2, 'num_decode_streams': 1500, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'short_chunk_size': 50, 'num_left_chunks': 4, 'decode_chunk_len': 32, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200.0, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search'), 'suffix': 'epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model', 'blank_id': 0, 'unk_id': 7, 'vocab_size': 33} 2023-06-21 10:04:38,027 INFO [streaming_decode.py:499] About to create model 2023-06-21 10:04:38,604 INFO [zipformer.py:405] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8. 2023-06-21 10:04:38,608 INFO [streaming_decode.py:566] Calculating the averaged model over epoch range from 21 (excluded) to 30 2023-06-21 10:04:42,203 INFO [streaming_decode.py:588] Number of model parameters: 69471350 2023-06-21 10:04:42,204 INFO [multidataset.py:122] About to get LibriVox test cuts 2023-06-21 10:04:42,204 INFO [multidataset.py:124] Loading LibriVox in lazy mode 2023-06-21 10:04:42,204 INFO [multidataset.py:133] About to get FLEURS test cuts 2023-06-21 10:04:42,204 INFO [multidataset.py:135] Loading FLEURS in lazy mode 2023-06-21 10:04:42,205 INFO [multidataset.py:144] About to get Common Voice test cuts 2023-06-21 10:04:42,205 INFO [multidataset.py:146] Loading Common Voice in lazy mode 2023-06-21 10:04:42,471 INFO [streaming_decode.py:380] Cuts processed until now is 0. 2023-06-21 10:04:42,786 INFO [streaming_decode.py:380] Cuts processed until now is 50. 2023-06-21 10:04:43,098 INFO [streaming_decode.py:380] Cuts processed until now is 100. 2023-06-21 10:04:43,444 INFO [streaming_decode.py:380] Cuts processed until now is 150. 2023-06-21 10:04:43,770 INFO [streaming_decode.py:380] Cuts processed until now is 200. 2023-06-21 10:04:44,092 INFO [streaming_decode.py:380] Cuts processed until now is 250. 2023-06-21 10:04:44,416 INFO [streaming_decode.py:380] Cuts processed until now is 300. 2023-06-21 10:04:44,756 INFO [streaming_decode.py:380] Cuts processed until now is 350. 2023-06-21 10:04:45,079 INFO [streaming_decode.py:380] Cuts processed until now is 400. 2023-06-21 10:04:45,405 INFO [streaming_decode.py:380] Cuts processed until now is 450. 2023-06-21 10:04:45,734 INFO [streaming_decode.py:380] Cuts processed until now is 500. 2023-06-21 10:04:46,071 INFO [streaming_decode.py:380] Cuts processed until now is 550. 2023-06-21 10:04:46,405 INFO [streaming_decode.py:380] Cuts processed until now is 600. 2023-06-21 10:04:57,029 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-librivox-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:04:57,063 INFO [utils.py:561] [test-librivox-beam_4_max_contexts_4_max_states_32] %WER 4.81% [1759 / 36594, 280 ins, 892 del, 587 sub ] 2023-06-21 10:04:57,144 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-librivox-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:04:57,145 INFO [streaming_decode.py:450] For test-librivox, WER of different settings are: beam_4_max_contexts_4_max_states_32 4.81 best for test-librivox 2023-06-21 10:04:57,149 INFO [streaming_decode.py:380] Cuts processed until now is 0. 2023-06-21 10:04:57,332 INFO [streaming_decode.py:380] Cuts processed until now is 50. 2023-06-21 10:04:57,494 INFO [streaming_decode.py:380] Cuts processed until now is 100. 2023-06-21 10:04:57,663 INFO [streaming_decode.py:380] Cuts processed until now is 150. 2023-06-21 10:04:57,833 INFO [streaming_decode.py:380] Cuts processed until now is 200. 2023-06-21 10:04:58,000 INFO [streaming_decode.py:380] Cuts processed until now is 250. 2023-06-21 10:04:58,161 INFO [streaming_decode.py:380] Cuts processed until now is 300. 2023-06-21 10:04:58,323 INFO [streaming_decode.py:380] Cuts processed until now is 350. 2023-06-21 10:04:58,488 INFO [streaming_decode.py:380] Cuts processed until now is 400. 2023-06-21 10:04:58,656 INFO [streaming_decode.py:380] Cuts processed until now is 450. 2023-06-21 10:04:58,819 INFO [streaming_decode.py:380] Cuts processed until now is 500. 2023-06-21 10:04:58,993 INFO [streaming_decode.py:380] Cuts processed until now is 550. 2023-06-21 10:04:59,176 INFO [streaming_decode.py:380] Cuts processed until now is 600. 2023-06-21 10:04:59,364 INFO [streaming_decode.py:380] Cuts processed until now is 650. 2023-06-21 10:05:34,495 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-fleurs-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:05:34,590 INFO [utils.py:561] [test-fleurs-beam_4_max_contexts_4_max_states_32] %WER 12.93% [12100 / 93580, 1706 ins, 5594 del, 4800 sub ] 2023-06-21 10:05:34,813 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-fleurs-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:05:34,814 INFO [streaming_decode.py:450] For test-fleurs, WER of different settings are: beam_4_max_contexts_4_max_states_32 12.93 best for test-fleurs 2023-06-21 10:05:34,820 INFO [streaming_decode.py:380] Cuts processed until now is 0. 2023-06-21 10:05:35,059 INFO [streaming_decode.py:380] Cuts processed until now is 50. 2023-06-21 10:05:35,308 INFO [streaming_decode.py:380] Cuts processed until now is 100. 2023-06-21 10:05:35,583 INFO [streaming_decode.py:380] Cuts processed until now is 150. 2023-06-21 10:05:35,829 INFO [streaming_decode.py:380] Cuts processed until now is 200. 2023-06-21 10:05:36,082 INFO [streaming_decode.py:380] Cuts processed until now is 250. 2023-06-21 10:05:36,315 INFO [streaming_decode.py:380] Cuts processed until now is 300. 2023-06-21 10:05:36,537 INFO [streaming_decode.py:380] Cuts processed until now is 350. 2023-06-21 10:05:36,797 INFO [streaming_decode.py:380] Cuts processed until now is 400. 2023-06-21 10:05:37,028 INFO [streaming_decode.py:380] Cuts processed until now is 450. 2023-06-21 10:05:37,263 INFO [streaming_decode.py:380] Cuts processed until now is 500. 2023-06-21 10:05:37,499 INFO [streaming_decode.py:380] Cuts processed until now is 550. 2023-06-21 10:05:37,720 INFO [streaming_decode.py:380] Cuts processed until now is 600. 2023-06-21 10:05:37,959 INFO [streaming_decode.py:380] Cuts processed until now is 650. 2023-06-21 10:05:38,182 INFO [streaming_decode.py:380] Cuts processed until now is 700. 2023-06-21 10:05:38,406 INFO [streaming_decode.py:380] Cuts processed until now is 750. 2023-06-21 10:05:38,664 INFO [streaming_decode.py:380] Cuts processed until now is 800. 2023-06-21 10:05:38,913 INFO [streaming_decode.py:380] Cuts processed until now is 850. 2023-06-21 10:05:39,251 INFO [streaming_decode.py:380] Cuts processed until now is 900. 2023-06-21 10:05:39,493 INFO [streaming_decode.py:380] Cuts processed until now is 950. 2023-06-21 10:05:39,726 INFO [streaming_decode.py:380] Cuts processed until now is 1000. 2023-06-21 10:05:39,959 INFO [streaming_decode.py:380] Cuts processed until now is 1050. 2023-06-21 10:05:40,192 INFO [streaming_decode.py:380] Cuts processed until now is 1100. 2023-06-21 10:05:40,436 INFO [streaming_decode.py:380] Cuts processed until now is 1150. 2023-06-21 10:05:40,709 INFO [streaming_decode.py:380] Cuts processed until now is 1200. 2023-06-21 10:05:40,959 INFO [streaming_decode.py:380] Cuts processed until now is 1250. 2023-06-21 10:05:41,199 INFO [streaming_decode.py:380] Cuts processed until now is 1300. 2023-06-21 10:05:41,448 INFO [streaming_decode.py:380] Cuts processed until now is 1350. 2023-06-21 10:05:41,697 INFO [streaming_decode.py:380] Cuts processed until now is 1400. 2023-06-21 10:05:41,938 INFO [streaming_decode.py:380] Cuts processed until now is 1450. 2023-06-21 10:05:51,050 INFO [streaming_decode.py:380] Cuts processed until now is 1500. 2023-06-21 10:05:53,941 INFO [streaming_decode.py:380] Cuts processed until now is 1550. 2023-06-21 10:05:55,569 INFO [streaming_decode.py:380] Cuts processed until now is 1600. 2023-06-21 10:05:55,799 INFO [streaming_decode.py:380] Cuts processed until now is 1650. 2023-06-21 10:05:57,493 INFO [streaming_decode.py:380] Cuts processed until now is 1700. 2023-06-21 10:05:57,735 INFO [streaming_decode.py:380] Cuts processed until now is 1750. 2023-06-21 10:05:57,961 INFO [streaming_decode.py:380] Cuts processed until now is 1800. 2023-06-21 10:05:59,694 INFO [streaming_decode.py:380] Cuts processed until now is 1850. 2023-06-21 10:05:59,923 INFO [streaming_decode.py:380] Cuts processed until now is 1900. 2023-06-21 10:06:00,151 INFO [streaming_decode.py:380] Cuts processed until now is 1950. 2023-06-21 10:06:01,771 INFO [streaming_decode.py:380] Cuts processed until now is 2000. 2023-06-21 10:06:01,997 INFO [streaming_decode.py:380] Cuts processed until now is 2050. 2023-06-21 10:06:02,241 INFO [streaming_decode.py:380] Cuts processed until now is 2100. 2023-06-21 10:06:02,465 INFO [streaming_decode.py:380] Cuts processed until now is 2150. 2023-06-21 10:06:04,249 INFO [streaming_decode.py:380] Cuts processed until now is 2200. 2023-06-21 10:06:04,478 INFO [streaming_decode.py:380] Cuts processed until now is 2250. 2023-06-21 10:06:04,710 INFO [streaming_decode.py:380] Cuts processed until now is 2300. 2023-06-21 10:06:06,461 INFO [streaming_decode.py:380] Cuts processed until now is 2350. 2023-06-21 10:06:06,697 INFO [streaming_decode.py:380] Cuts processed until now is 2400. 2023-06-21 10:06:06,931 INFO [streaming_decode.py:380] Cuts processed until now is 2450. 2023-06-21 10:06:08,726 INFO [streaming_decode.py:380] Cuts processed until now is 2500. 2023-06-21 10:06:08,950 INFO [streaming_decode.py:380] Cuts processed until now is 2550. 2023-06-21 10:06:09,187 INFO [streaming_decode.py:380] Cuts processed until now is 2600. 2023-06-21 10:06:10,940 INFO [streaming_decode.py:380] Cuts processed until now is 2650. 2023-06-21 10:06:11,165 INFO [streaming_decode.py:380] Cuts processed until now is 2700. 2023-06-21 10:06:12,942 INFO [streaming_decode.py:380] Cuts processed until now is 2750. 2023-06-21 10:06:13,183 INFO [streaming_decode.py:380] Cuts processed until now is 2800. 2023-06-21 10:06:14,919 INFO [streaming_decode.py:380] Cuts processed until now is 2850. 2023-06-21 10:06:16,667 INFO [streaming_decode.py:380] Cuts processed until now is 2900. 2023-06-21 10:06:18,270 INFO [streaming_decode.py:380] Cuts processed until now is 2950. 2023-06-21 10:06:19,990 INFO [streaming_decode.py:380] Cuts processed until now is 3000. 2023-06-21 10:06:20,222 INFO [streaming_decode.py:380] Cuts processed until now is 3050. 2023-06-21 10:06:21,952 INFO [streaming_decode.py:380] Cuts processed until now is 3100. 2023-06-21 10:06:22,202 INFO [streaming_decode.py:380] Cuts processed until now is 3150. 2023-06-21 10:06:23,959 INFO [streaming_decode.py:380] Cuts processed until now is 3200. 2023-06-21 10:06:24,183 INFO [streaming_decode.py:380] Cuts processed until now is 3250. 2023-06-21 10:06:25,951 INFO [streaming_decode.py:380] Cuts processed until now is 3300. 2023-06-21 10:06:26,203 INFO [streaming_decode.py:380] Cuts processed until now is 3350. 2023-06-21 10:06:27,984 INFO [streaming_decode.py:380] Cuts processed until now is 3400. 2023-06-21 10:06:28,228 INFO [streaming_decode.py:380] Cuts processed until now is 3450. 2023-06-21 10:06:28,468 INFO [streaming_decode.py:380] Cuts processed until now is 3500. 2023-06-21 10:06:30,266 INFO [streaming_decode.py:380] Cuts processed until now is 3550. 2023-06-21 10:06:30,497 INFO [streaming_decode.py:380] Cuts processed until now is 3600. 2023-06-21 10:06:45,693 INFO [streaming_decode.py:425] The transcripts are stored in pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/recogs-test-commonvoice-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:06:45,825 INFO [utils.py:561] [test-commonvoice-beam_4_max_contexts_4_max_states_32] %WER 14.96% [19859 / 132787, 3004 ins, 8788 del, 8067 sub ] 2023-06-21 10:06:46,126 INFO [streaming_decode.py:436] Wrote detailed error stats to pruned_transducer_stateless7_streaming/exp/streaming/fast_beam_search/errs-test-commonvoice-epoch-30-avg-9-streaming-chunk-size-32-beam-4-max-contexts-4-max-states-32-use-averaged-model.txt 2023-06-21 10:06:46,126 INFO [streaming_decode.py:450] For test-commonvoice, WER of different settings are: beam_4_max_contexts_4_max_states_32 14.96 best for test-commonvoice 2023-06-21 10:06:46,127 INFO [streaming_decode.py:618] Done!