2022-11-21 09:37:32,455 INFO [decode.py:574] Decoding started 2022-11-21 09:37:32,456 INFO [decode.py:580] Device: cuda:0 2022-11-21 09:37:32,460 INFO [decode.py:590] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 100, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'warm_step': 2000, 'env_info': {'k2-version': '1.21', 'k2-build-type': 'Debug', 'k2-with-cuda': True, 'k2-git-sha1': 'f271e82ef30f75fecbae44b163e1244e53def116', 'k2-git-date': 'Fri Oct 28 05:02:16 2022', 'lhotse-version': '1.9.0.dev+git.97bf4b0.dirty', 'torch-version': '1.10.0+cu111', 'torch-cuda-available': True, 'torch-cuda-version': '11.1', 'python-version': '3.8', 'icefall-git-branch': 'ami_recipe', 'icefall-git-sha1': 'd1b5a16-dirty', 'icefall-git-date': 'Sun Nov 20 22:32:57 2022', 'icefall-path': '/exp/draj/mini_scale_2022/icefall', 'k2-path': '/exp/draj/mini_scale_2022/k2/k2/python/k2/__init__.py', 'lhotse-path': '/exp/draj/mini_scale_2022/lhotse/lhotse/__init__.py', 'hostname': 'r2n06', 'IP address': '10.1.2.6'}, 'epoch': 14, 'iter': 0, 'avg': 8, 'use_averaged_model': True, 'exp_dir': PosixPath('pruned_transducer_stateless7/exp/v2'), 'lang_dir': PosixPath('data/lang_bpe_500'), 'decoding_method': 'fast_beam_search', 'beam_size': 4, 'beam': 4, 'ngram_lm_scale': 0.01, 'max_contexts': 4, 'max_states': 8, 'context_size': 2, 'max_sym_per_frame': 1, 'num_paths': 200, 'nbest_scale': 0.5, 'num_encoder_layers': '2,4,3,2,4', 'feedforward_dims': '1024,1024,2048,2048,1024', 'nhead': '8,8,8,8,8', 'encoder_dims': '384,384,384,384,384', 'attention_dims': '192,192,192,192,192', 'encoder_unmasked_dims': '256,256,256,256,256', 'zipformer_downsampling_factors': '1,2,4,8,2', 'cnn_module_kernels': '31,31,31,31,31', 'decoder_dim': 512, 'joiner_dim': 512, 'manifest_dir': PosixPath('data/manifests'), 'enable_musan': True, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'max_duration': 500, 'max_cuts': None, 'num_buckets': 50, 'on_the_fly_feats': False, 'shuffle': True, 'num_workers': 8, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'ihm_only': False, 'res_dir': PosixPath('pruned_transducer_stateless7/exp/v2/fast_beam_search'), 'suffix': 'epoch-14-avg-8-beam-4-max-contexts-4-max-states-8', 'blank_id': 0, 'unk_id': 2, 'vocab_size': 500} 2022-11-21 09:37:32,460 INFO [decode.py:592] About to create model 2022-11-21 09:37:32,937 INFO [zipformer.py:179] At encoder stack 4, which has downsampling_factor=2, we will combine the outputs of layers 1 and 3, with downsampling_factors=2 and 8. 2022-11-21 09:37:32,947 INFO [decode.py:659] Calculating the averaged model over epoch range from 6 (excluded) to 14 2022-11-21 09:37:39,406 INFO [decode.py:694] Number of model parameters: 70369391 2022-11-21 09:37:39,407 INFO [asr_datamodule.py:392] About to get AMI IHM dev cuts 2022-11-21 09:37:39,409 INFO [asr_datamodule.py:413] About to get AMI IHM test cuts 2022-11-21 09:37:39,411 INFO [asr_datamodule.py:398] About to get AMI SDM dev cuts 2022-11-21 09:37:39,412 INFO [asr_datamodule.py:419] About to get AMI SDM test cuts 2022-11-21 09:37:39,414 INFO [asr_datamodule.py:407] About to get AMI GSS-enhanced dev cuts 2022-11-21 09:37:39,415 INFO [asr_datamodule.py:428] About to get AMI GSS-enhanced test cuts 2022-11-21 09:37:41,443 INFO [decode.py:726] Decoding dev_ihm 2022-11-21 09:37:44,362 INFO [decode.py:469] batch 0/?, cuts processed until now is 72 2022-11-21 09:37:47,217 INFO [decode.py:469] batch 2/?, cuts processed until now is 537 2022-11-21 09:37:50,037 INFO [decode.py:469] batch 4/?, cuts processed until now is 689 2022-11-21 09:37:52,786 INFO [decode.py:469] batch 6/?, cuts processed until now is 823 2022-11-21 09:37:55,427 INFO [decode.py:469] batch 8/?, cuts processed until now is 985 2022-11-21 09:38:00,347 INFO [decode.py:469] batch 10/?, cuts processed until now is 1088 2022-11-21 09:38:03,097 INFO [decode.py:469] batch 12/?, cuts processed until now is 1263 2022-11-21 09:38:05,684 INFO [decode.py:469] batch 14/?, cuts processed until now is 1521 2022-11-21 09:38:07,949 INFO [decode.py:469] batch 16/?, cuts processed until now is 1903 2022-11-21 09:38:11,683 INFO [decode.py:469] batch 18/?, cuts processed until now is 2032 2022-11-21 09:38:15,284 INFO [decode.py:469] batch 20/?, cuts processed until now is 2117 2022-11-21 09:38:17,471 INFO [decode.py:469] batch 22/?, cuts processed until now is 2375 2022-11-21 09:38:19,918 INFO [decode.py:469] batch 24/?, cuts processed until now is 2824 2022-11-21 09:38:22,837 INFO [decode.py:469] batch 26/?, cuts processed until now is 2969 2022-11-21 09:38:25,785 INFO [decode.py:469] batch 28/?, cuts processed until now is 3245 2022-11-21 09:38:26,044 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.5012, 1.5040, 2.2027, 1.7505, 1.7409, 2.1940, 1.7005, 1.7461], device='cuda:0'), covar=tensor([0.0017, 0.0101, 0.0055, 0.0066, 0.0099, 0.0065, 0.0039, 0.0053], device='cuda:0'), in_proj_covar=tensor([0.0018, 0.0018, 0.0018, 0.0025, 0.0021, 0.0019, 0.0024, 0.0024], device='cuda:0'), out_proj_covar=tensor([1.6600e-05, 1.6809e-05, 1.6082e-05, 2.4018e-05, 1.9485e-05, 1.8494e-05, 2.3316e-05, 2.3270e-05], device='cuda:0') 2022-11-21 09:38:28,800 INFO [decode.py:469] batch 30/?, cuts processed until now is 3401 2022-11-21 09:38:32,004 INFO [decode.py:469] batch 32/?, cuts processed until now is 3519 2022-11-21 09:38:34,757 INFO [decode.py:469] batch 34/?, cuts processed until now is 3694 2022-11-21 09:38:37,495 INFO [decode.py:469] batch 36/?, cuts processed until now is 3818 2022-11-21 09:38:40,528 INFO [decode.py:469] batch 38/?, cuts processed until now is 3970 2022-11-21 09:38:42,582 INFO [decode.py:469] batch 40/?, cuts processed until now is 4750 2022-11-21 09:38:45,536 INFO [decode.py:469] batch 42/?, cuts processed until now is 5038 2022-11-21 09:38:49,167 INFO [decode.py:469] batch 44/?, cuts processed until now is 5144 2022-11-21 09:38:52,673 INFO [decode.py:469] batch 46/?, cuts processed until now is 5253 2022-11-21 09:38:55,694 INFO [decode.py:469] batch 48/?, cuts processed until now is 5672 2022-11-21 09:38:58,783 INFO [decode.py:469] batch 50/?, cuts processed until now is 5878 2022-11-21 09:39:01,101 INFO [decode.py:469] batch 52/?, cuts processed until now is 6260 2022-11-21 09:39:03,236 INFO [decode.py:469] batch 54/?, cuts processed until now is 6808 2022-11-21 09:39:05,698 INFO [decode.py:469] batch 56/?, cuts processed until now is 7117 2022-11-21 09:39:08,059 INFO [decode.py:469] batch 58/?, cuts processed until now is 7565 2022-11-21 09:39:10,098 INFO [decode.py:469] batch 60/?, cuts processed until now is 8078 2022-11-21 09:39:12,152 INFO [decode.py:469] batch 62/?, cuts processed until now is 8626 2022-11-21 09:39:14,330 INFO [decode.py:469] batch 64/?, cuts processed until now is 9174 2022-11-21 09:39:16,736 INFO [decode.py:469] batch 66/?, cuts processed until now is 9455 2022-11-21 09:39:18,781 INFO [decode.py:469] batch 68/?, cuts processed until now is 9968 2022-11-21 09:39:20,819 INFO [decode.py:469] batch 70/?, cuts processed until now is 10481 2022-11-21 09:39:22,914 INFO [decode.py:469] batch 72/?, cuts processed until now is 11264 2022-11-21 09:39:25,374 INFO [decode.py:469] batch 74/?, cuts processed until now is 11669 2022-11-21 09:39:27,175 INFO [decode.py:469] batch 76/?, cuts processed until now is 11761 2022-11-21 09:39:28,833 INFO [decode.py:469] batch 78/?, cuts processed until now is 11843 2022-11-21 09:39:30,414 INFO [decode.py:469] batch 80/?, cuts processed until now is 11956 2022-11-21 09:39:31,821 INFO [decode.py:469] batch 82/?, cuts processed until now is 12467 2022-11-21 09:39:35,807 INFO [decode.py:469] batch 84/?, cuts processed until now is 12586 2022-11-21 09:39:37,712 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-dev_ihm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:39:37,868 INFO [utils.py:530] [dev_ihm-beam_4_max_contexts_4_max_states_8] %WER 19.44% [18459 / 94940, 2783 ins, 3992 del, 11684 sub ] 2022-11-21 09:39:38,597 INFO [utils.py:530] [dev_ihm-beam_4_max_contexts_4_max_states_8] %WER 12.30% [45497 / 369873, 10772 ins, 17562 del, 17163 sub ] 2022-11-21 09:39:39,509 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-dev_ihm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:39:39,510 INFO [decode.py:531] For dev_ihm, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 19.44 12.3 best for dev_ihm 2022-11-21 09:39:39,514 INFO [decode.py:726] Decoding test_ihm 2022-11-21 09:39:42,413 INFO [decode.py:469] batch 0/?, cuts processed until now is 69 2022-11-21 09:39:45,118 INFO [decode.py:469] batch 2/?, cuts processed until now is 555 2022-11-21 09:39:48,016 INFO [decode.py:469] batch 4/?, cuts processed until now is 703 2022-11-21 09:39:50,763 INFO [decode.py:469] batch 6/?, cuts processed until now is 830 2022-11-21 09:39:53,368 INFO [decode.py:469] batch 8/?, cuts processed until now is 987 2022-11-21 09:39:57,149 INFO [decode.py:469] batch 10/?, cuts processed until now is 1095 2022-11-21 09:39:59,903 INFO [decode.py:469] batch 12/?, cuts processed until now is 1267 2022-11-21 09:40:02,382 INFO [decode.py:469] batch 14/?, cuts processed until now is 1532 2022-11-21 09:40:04,529 INFO [decode.py:469] batch 16/?, cuts processed until now is 1931 2022-11-21 09:40:08,454 INFO [decode.py:469] batch 18/?, cuts processed until now is 2055 2022-11-21 09:40:12,895 INFO [decode.py:469] batch 20/?, cuts processed until now is 2124 2022-11-21 09:40:15,391 INFO [decode.py:469] batch 22/?, cuts processed until now is 2388 2022-11-21 09:40:17,552 INFO [decode.py:469] batch 24/?, cuts processed until now is 2856 2022-11-21 09:40:20,452 INFO [decode.py:469] batch 26/?, cuts processed until now is 2996 2022-11-21 09:40:23,133 INFO [decode.py:469] batch 28/?, cuts processed until now is 3278 2022-11-21 09:40:25,995 INFO [decode.py:469] batch 30/?, cuts processed until now is 3430 2022-11-21 09:40:29,911 INFO [decode.py:469] batch 32/?, cuts processed until now is 3535 2022-11-21 09:40:32,822 INFO [decode.py:469] batch 34/?, cuts processed until now is 3706 2022-11-21 09:40:35,648 INFO [decode.py:469] batch 36/?, cuts processed until now is 3822 2022-11-21 09:40:38,688 INFO [decode.py:469] batch 38/?, cuts processed until now is 3969 2022-11-21 09:40:41,769 INFO [decode.py:469] batch 40/?, cuts processed until now is 4411 2022-11-21 09:40:43,913 INFO [decode.py:469] batch 42/?, cuts processed until now is 5058 2022-11-21 09:40:46,328 INFO [decode.py:469] batch 44/?, cuts processed until now is 5544 2022-11-21 09:40:49,447 INFO [decode.py:469] batch 46/?, cuts processed until now is 5685 2022-11-21 09:40:51,873 INFO [decode.py:469] batch 48/?, cuts processed until now is 5890 2022-11-21 09:40:54,246 INFO [decode.py:469] batch 50/?, cuts processed until now is 6372 2022-11-21 09:40:56,479 INFO [decode.py:469] batch 52/?, cuts processed until now is 6706 2022-11-21 09:40:58,067 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.4471, 3.4262, 3.4324, 2.9745, 1.8228, 3.2833, 2.1406, 3.1996], device='cuda:0'), covar=tensor([0.0454, 0.0208, 0.0175, 0.0451, 0.0695, 0.0255, 0.0610, 0.0165], device='cuda:0'), in_proj_covar=tensor([0.0176, 0.0146, 0.0155, 0.0177, 0.0173, 0.0154, 0.0169, 0.0152], device='cuda:0'), out_proj_covar=tensor([0.0003, 0.0002, 0.0002, 0.0003, 0.0003, 0.0002, 0.0003, 0.0002], device='cuda:0') 2022-11-21 09:40:58,735 INFO [decode.py:469] batch 54/?, cuts processed until now is 7105 2022-11-21 09:41:02,571 INFO [decode.py:469] batch 56/?, cuts processed until now is 7290 2022-11-21 09:41:04,855 INFO [decode.py:469] batch 58/?, cuts processed until now is 8116 2022-11-21 09:41:08,996 INFO [decode.py:469] batch 60/?, cuts processed until now is 8258 2022-11-21 09:41:11,225 INFO [decode.py:469] batch 62/?, cuts processed until now is 8794 2022-11-21 09:41:13,318 INFO [decode.py:469] batch 64/?, cuts processed until now is 9330 2022-11-21 09:41:16,989 INFO [decode.py:469] batch 66/?, cuts processed until now is 9476 2022-11-21 09:41:20,066 INFO [decode.py:469] batch 68/?, cuts processed until now is 9921 2022-11-21 09:41:22,237 INFO [decode.py:469] batch 70/?, cuts processed until now is 10251 2022-11-21 09:41:25,177 INFO [decode.py:469] batch 72/?, cuts processed until now is 10679 2022-11-21 09:41:27,448 INFO [decode.py:469] batch 74/?, cuts processed until now is 10794 2022-11-21 09:41:29,130 INFO [decode.py:469] batch 76/?, cuts processed until now is 11039 2022-11-21 09:41:29,179 INFO [zipformer.py:1414] attn_weights_entropy = tensor([3.8067, 3.6148, 3.9005, 3.9164, 3.5684, 3.7111, 4.2535, 3.9170], device='cuda:0'), covar=tensor([0.0271, 0.0907, 0.0247, 0.0870, 0.0499, 0.0188, 0.0468, 0.0355], device='cuda:0'), in_proj_covar=tensor([0.0075, 0.0097, 0.0083, 0.0108, 0.0078, 0.0067, 0.0133, 0.0090], device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0002, 0.0002, 0.0002, 0.0002, 0.0001, 0.0003, 0.0002], device='cuda:0') 2022-11-21 09:41:30,374 INFO [decode.py:469] batch 78/?, cuts processed until now is 11155 2022-11-21 09:41:31,933 INFO [decode.py:469] batch 80/?, cuts processed until now is 11600 2022-11-21 09:41:34,284 INFO [decode.py:469] batch 82/?, cuts processed until now is 12041 2022-11-21 09:41:35,705 INFO [decode.py:469] batch 84/?, cuts processed until now is 12110 2022-11-21 09:41:35,872 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-test_ihm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:41:36,025 INFO [utils.py:530] [test_ihm-beam_4_max_contexts_4_max_states_8] %WER 18.04% [16174 / 89659, 1994 ins, 4043 del, 10137 sub ] 2022-11-21 09:41:36,695 INFO [utils.py:530] [test_ihm-beam_4_max_contexts_4_max_states_8] %WER 11.30% [40040 / 354205, 8698 ins, 16856 del, 14486 sub ] 2022-11-21 09:41:37,616 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-test_ihm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:41:37,617 INFO [decode.py:531] For test_ihm, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 18.04 11.3 best for test_ihm 2022-11-21 09:41:37,625 INFO [decode.py:726] Decoding dev_sdm 2022-11-21 09:41:39,203 INFO [zipformer.py:1414] attn_weights_entropy = tensor([4.2901, 4.0348, 3.8759, 3.7100, 4.2236, 3.8320, 1.8017, 4.4545], device='cuda:0'), covar=tensor([0.0167, 0.0258, 0.0231, 0.0258, 0.0227, 0.0300, 0.3030, 0.0169], device='cuda:0'), in_proj_covar=tensor([0.0093, 0.0074, 0.0074, 0.0066, 0.0090, 0.0076, 0.0123, 0.0097], device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0002, 0.0001, 0.0001, 0.0002, 0.0002, 0.0002, 0.0002], device='cuda:0') 2022-11-21 09:41:40,546 INFO [decode.py:469] batch 0/?, cuts processed until now is 71 2022-11-21 09:41:43,132 INFO [decode.py:469] batch 2/?, cuts processed until now is 535 2022-11-21 09:41:45,936 INFO [decode.py:469] batch 4/?, cuts processed until now is 686 2022-11-21 09:41:48,893 INFO [decode.py:469] batch 6/?, cuts processed until now is 819 2022-11-21 09:41:51,541 INFO [decode.py:469] batch 8/?, cuts processed until now is 980 2022-11-21 09:41:55,839 INFO [decode.py:469] batch 10/?, cuts processed until now is 1083 2022-11-21 09:41:58,703 INFO [decode.py:469] batch 12/?, cuts processed until now is 1257 2022-11-21 09:42:01,220 INFO [decode.py:469] batch 14/?, cuts processed until now is 1513 2022-11-21 09:42:02,514 INFO [zipformer.py:1414] attn_weights_entropy = tensor([3.3930, 3.4249, 3.4852, 3.2506, 3.5529, 3.2495, 1.5440, 3.6162], device='cuda:0'), covar=tensor([0.0194, 0.0198, 0.0190, 0.0216, 0.0206, 0.0299, 0.2493, 0.0189], device='cuda:0'), in_proj_covar=tensor([0.0093, 0.0074, 0.0074, 0.0066, 0.0090, 0.0076, 0.0123, 0.0097], device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0002, 0.0001, 0.0001, 0.0002, 0.0002, 0.0002, 0.0002], device='cuda:0') 2022-11-21 09:42:03,569 INFO [decode.py:469] batch 16/?, cuts processed until now is 1892 2022-11-21 09:42:07,050 INFO [decode.py:469] batch 18/?, cuts processed until now is 2020 2022-11-21 09:42:10,600 INFO [decode.py:469] batch 20/?, cuts processed until now is 2106 2022-11-21 09:42:12,923 INFO [decode.py:469] batch 22/?, cuts processed until now is 2362 2022-11-21 09:42:14,974 INFO [decode.py:469] batch 24/?, cuts processed until now is 2807 2022-11-21 09:42:17,760 INFO [decode.py:469] batch 26/?, cuts processed until now is 2952 2022-11-21 09:42:20,532 INFO [decode.py:469] batch 28/?, cuts processed until now is 3226 2022-11-21 09:42:23,370 INFO [decode.py:469] batch 30/?, cuts processed until now is 3381 2022-11-21 09:42:26,622 INFO [decode.py:469] batch 32/?, cuts processed until now is 3499 2022-11-21 09:42:29,823 INFO [decode.py:469] batch 34/?, cuts processed until now is 3673 2022-11-21 09:42:32,820 INFO [decode.py:469] batch 36/?, cuts processed until now is 3797 2022-11-21 09:42:35,553 INFO [decode.py:469] batch 38/?, cuts processed until now is 3948 2022-11-21 09:42:37,578 INFO [decode.py:469] batch 40/?, cuts processed until now is 4722 2022-11-21 09:42:40,099 INFO [decode.py:469] batch 42/?, cuts processed until now is 5007 2022-11-21 09:42:43,349 INFO [decode.py:469] batch 44/?, cuts processed until now is 5112 2022-11-21 09:42:46,682 INFO [decode.py:469] batch 46/?, cuts processed until now is 5219 2022-11-21 09:42:49,708 INFO [decode.py:469] batch 48/?, cuts processed until now is 5636 2022-11-21 09:42:52,248 INFO [decode.py:469] batch 50/?, cuts processed until now is 5842 2022-11-21 09:42:54,302 INFO [decode.py:469] batch 52/?, cuts processed until now is 6222 2022-11-21 09:42:56,475 INFO [decode.py:469] batch 54/?, cuts processed until now is 6766 2022-11-21 09:42:58,796 INFO [decode.py:469] batch 56/?, cuts processed until now is 7072 2022-11-21 09:43:01,027 INFO [decode.py:469] batch 58/?, cuts processed until now is 7518 2022-11-21 09:43:03,490 INFO [decode.py:469] batch 60/?, cuts processed until now is 8027 2022-11-21 09:43:05,674 INFO [decode.py:469] batch 62/?, cuts processed until now is 8571 2022-11-21 09:43:07,764 INFO [decode.py:469] batch 64/?, cuts processed until now is 9115 2022-11-21 09:43:10,242 INFO [decode.py:469] batch 66/?, cuts processed until now is 9395 2022-11-21 09:43:12,522 INFO [decode.py:469] batch 68/?, cuts processed until now is 9904 2022-11-21 09:43:14,750 INFO [decode.py:469] batch 70/?, cuts processed until now is 10413 2022-11-21 09:43:16,714 INFO [decode.py:469] batch 72/?, cuts processed until now is 11190 2022-11-21 09:43:18,837 INFO [decode.py:469] batch 74/?, cuts processed until now is 11589 2022-11-21 09:43:18,995 INFO [zipformer.py:1414] attn_weights_entropy = tensor([1.7028, 1.3026, 1.3419, 1.0055, 1.3474, 1.4107, 0.7656, 1.2962], device='cuda:0'), covar=tensor([0.0270, 0.0298, 0.0359, 0.0717, 0.0552, 0.0613, 0.1124, 0.0303], device='cuda:0'), in_proj_covar=tensor([0.0011, 0.0017, 0.0012, 0.0015, 0.0012, 0.0011, 0.0016, 0.0011], device='cuda:0'), out_proj_covar=tensor([5.8009e-05, 8.0520e-05, 5.9059e-05, 7.3872e-05, 6.4258e-05, 5.7202e-05, 7.6631e-05, 5.7227e-05], device='cuda:0') 2022-11-21 09:43:20,529 INFO [decode.py:469] batch 76/?, cuts processed until now is 11699 2022-11-21 09:43:22,432 INFO [decode.py:469] batch 78/?, cuts processed until now is 11799 2022-11-21 09:43:23,864 INFO [decode.py:469] batch 80/?, cuts processed until now is 11889 2022-11-21 09:43:25,424 INFO [decode.py:469] batch 82/?, cuts processed until now is 12461 2022-11-21 09:43:27,202 INFO [decode.py:469] batch 84/?, cuts processed until now is 12568 2022-11-21 09:43:31,673 INFO [decode.py:469] batch 86/?, cuts processed until now is 12601 2022-11-21 09:43:31,844 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-dev_sdm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:43:32,004 INFO [utils.py:530] [dev_sdm-beam_4_max_contexts_4_max_states_8] %WER 31.11% [29537 / 94940, 4266 ins, 7752 del, 17519 sub ] 2022-11-21 09:43:32,829 INFO [utils.py:530] [dev_sdm-beam_4_max_contexts_4_max_states_8] %WER 22.60% [83608 / 369873, 18843 ins, 33372 del, 31393 sub ] 2022-11-21 09:43:33,804 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-dev_sdm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:43:33,805 INFO [decode.py:531] For dev_sdm, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 31.11 22.6 best for dev_sdm 2022-11-21 09:43:33,810 INFO [decode.py:726] Decoding test_sdm 2022-11-21 09:43:36,347 INFO [decode.py:469] batch 0/?, cuts processed until now is 69 2022-11-21 09:43:39,082 INFO [decode.py:469] batch 2/?, cuts processed until now is 555 2022-11-21 09:43:42,216 INFO [decode.py:469] batch 4/?, cuts processed until now is 703 2022-11-21 09:43:45,306 INFO [decode.py:469] batch 6/?, cuts processed until now is 831 2022-11-21 09:43:48,441 INFO [decode.py:469] batch 8/?, cuts processed until now is 988 2022-11-21 09:43:52,218 INFO [decode.py:469] batch 10/?, cuts processed until now is 1096 2022-11-21 09:43:55,047 INFO [decode.py:469] batch 12/?, cuts processed until now is 1268 2022-11-21 09:43:57,332 INFO [decode.py:469] batch 14/?, cuts processed until now is 1533 2022-11-21 09:43:59,736 INFO [decode.py:469] batch 16/?, cuts processed until now is 1932 2022-11-21 09:44:02,455 INFO [zipformer.py:1414] attn_weights_entropy = tensor([4.5302, 4.5710, 4.4733, 4.5946, 4.3598, 3.8894, 5.0773, 4.4629], device='cuda:0'), covar=tensor([0.0235, 0.0436, 0.0137, 0.0967, 0.0232, 0.0157, 0.0389, 0.0248], device='cuda:0'), in_proj_covar=tensor([0.0075, 0.0097, 0.0083, 0.0108, 0.0078, 0.0067, 0.0133, 0.0090], device='cuda:0'), out_proj_covar=tensor([0.0002, 0.0002, 0.0002, 0.0002, 0.0002, 0.0001, 0.0003, 0.0002], device='cuda:0') 2022-11-21 09:44:03,741 INFO [decode.py:469] batch 18/?, cuts processed until now is 2057 2022-11-21 09:44:08,271 INFO [decode.py:469] batch 20/?, cuts processed until now is 2126 2022-11-21 09:44:10,648 INFO [decode.py:469] batch 22/?, cuts processed until now is 2390 2022-11-21 09:44:12,965 INFO [decode.py:469] batch 24/?, cuts processed until now is 2858 2022-11-21 09:44:16,139 INFO [decode.py:469] batch 26/?, cuts processed until now is 2998 2022-11-21 09:44:18,935 INFO [decode.py:469] batch 28/?, cuts processed until now is 3280 2022-11-21 09:44:21,695 INFO [decode.py:469] batch 30/?, cuts processed until now is 3432 2022-11-21 09:44:25,745 INFO [decode.py:469] batch 32/?, cuts processed until now is 3537 2022-11-21 09:44:28,551 INFO [decode.py:469] batch 34/?, cuts processed until now is 3709 2022-11-21 09:44:31,631 INFO [decode.py:469] batch 36/?, cuts processed until now is 3825 2022-11-21 09:44:34,591 INFO [decode.py:469] batch 38/?, cuts processed until now is 3972 2022-11-21 09:44:37,539 INFO [decode.py:469] batch 40/?, cuts processed until now is 4410 2022-11-21 09:44:39,990 INFO [decode.py:469] batch 42/?, cuts processed until now is 5060 2022-11-21 09:44:42,447 INFO [decode.py:469] batch 44/?, cuts processed until now is 5546 2022-11-21 09:44:45,281 INFO [decode.py:469] batch 46/?, cuts processed until now is 5687 2022-11-21 09:44:47,720 INFO [decode.py:469] batch 48/?, cuts processed until now is 5893 2022-11-21 09:44:50,326 INFO [decode.py:469] batch 50/?, cuts processed until now is 6379 2022-11-21 09:44:52,795 INFO [decode.py:469] batch 52/?, cuts processed until now is 6713 2022-11-21 09:44:55,058 INFO [decode.py:469] batch 54/?, cuts processed until now is 7112 2022-11-21 09:44:59,177 INFO [decode.py:469] batch 56/?, cuts processed until now is 7298 2022-11-21 09:45:01,275 INFO [decode.py:469] batch 58/?, cuts processed until now is 8130 2022-11-21 09:45:05,730 INFO [decode.py:469] batch 60/?, cuts processed until now is 8273 2022-11-21 09:45:07,900 INFO [decode.py:469] batch 62/?, cuts processed until now is 8813 2022-11-21 09:45:10,253 INFO [decode.py:469] batch 64/?, cuts processed until now is 9353 2022-11-21 09:45:14,003 INFO [decode.py:469] batch 66/?, cuts processed until now is 9500 2022-11-21 09:45:17,300 INFO [decode.py:469] batch 68/?, cuts processed until now is 9944 2022-11-21 09:45:19,681 INFO [decode.py:469] batch 70/?, cuts processed until now is 10274 2022-11-21 09:45:22,865 INFO [decode.py:469] batch 72/?, cuts processed until now is 10711 2022-11-21 09:45:25,095 INFO [decode.py:469] batch 74/?, cuts processed until now is 10820 2022-11-21 09:45:26,739 INFO [decode.py:469] batch 76/?, cuts processed until now is 11076 2022-11-21 09:45:27,880 INFO [decode.py:469] batch 78/?, cuts processed until now is 11209 2022-11-21 09:45:29,397 INFO [decode.py:469] batch 80/?, cuts processed until now is 11651 2022-11-21 09:45:31,825 INFO [decode.py:469] batch 82/?, cuts processed until now is 12070 2022-11-21 09:45:33,279 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-test_sdm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:45:33,435 INFO [utils.py:530] [test_sdm-beam_4_max_contexts_4_max_states_8] %WER 32.10% [28784 / 89659, 3596 ins, 8598 del, 16590 sub ] 2022-11-21 09:45:34,315 INFO [utils.py:530] [test_sdm-beam_4_max_contexts_4_max_states_8] %WER 23.50% [83238 / 354205, 17319 ins, 35917 del, 30002 sub ] 2022-11-21 09:45:35,273 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-test_sdm-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:45:35,274 INFO [decode.py:531] For test_sdm, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 32.1 23.5 best for test_sdm 2022-11-21 09:45:35,279 INFO [decode.py:726] Decoding dev_gss 2022-11-21 09:45:37,914 INFO [decode.py:469] batch 0/?, cuts processed until now is 71 2022-11-21 09:45:40,678 INFO [decode.py:469] batch 2/?, cuts processed until now is 535 2022-11-21 09:45:43,449 INFO [decode.py:469] batch 4/?, cuts processed until now is 686 2022-11-21 09:45:46,327 INFO [decode.py:469] batch 6/?, cuts processed until now is 819 2022-11-21 09:45:49,041 INFO [decode.py:469] batch 8/?, cuts processed until now is 980 2022-11-21 09:45:53,390 INFO [decode.py:469] batch 10/?, cuts processed until now is 1083 2022-11-21 09:45:56,306 INFO [decode.py:469] batch 12/?, cuts processed until now is 1257 2022-11-21 09:45:58,535 INFO [decode.py:469] batch 14/?, cuts processed until now is 1513 2022-11-21 09:46:00,627 INFO [decode.py:469] batch 16/?, cuts processed until now is 1892 2022-11-21 09:46:04,010 INFO [decode.py:469] batch 18/?, cuts processed until now is 2020 2022-11-21 09:46:07,474 INFO [decode.py:469] batch 20/?, cuts processed until now is 2106 2022-11-21 09:46:09,689 INFO [decode.py:469] batch 22/?, cuts processed until now is 2362 2022-11-21 09:46:11,708 INFO [decode.py:469] batch 24/?, cuts processed until now is 2807 2022-11-21 09:46:14,435 INFO [decode.py:469] batch 26/?, cuts processed until now is 2952 2022-11-21 09:46:16,945 INFO [decode.py:469] batch 28/?, cuts processed until now is 3226 2022-11-21 09:46:19,604 INFO [decode.py:469] batch 30/?, cuts processed until now is 3381 2022-11-21 09:46:22,979 INFO [decode.py:469] batch 32/?, cuts processed until now is 3499 2022-11-21 09:46:26,083 INFO [decode.py:469] batch 34/?, cuts processed until now is 3673 2022-11-21 09:46:29,165 INFO [decode.py:469] batch 36/?, cuts processed until now is 3797 2022-11-21 09:46:31,904 INFO [decode.py:469] batch 38/?, cuts processed until now is 3948 2022-11-21 09:46:33,984 INFO [decode.py:469] batch 40/?, cuts processed until now is 4722 2022-11-21 09:46:36,610 INFO [decode.py:469] batch 42/?, cuts processed until now is 5007 2022-11-21 09:46:39,891 INFO [decode.py:469] batch 44/?, cuts processed until now is 5112 2022-11-21 09:46:43,039 INFO [decode.py:469] batch 46/?, cuts processed until now is 5219 2022-11-21 09:46:46,406 INFO [decode.py:469] batch 48/?, cuts processed until now is 5636 2022-11-21 09:46:49,197 INFO [decode.py:469] batch 50/?, cuts processed until now is 5842 2022-11-21 09:46:51,301 INFO [decode.py:469] batch 52/?, cuts processed until now is 6222 2022-11-21 09:46:53,536 INFO [decode.py:469] batch 54/?, cuts processed until now is 6766 2022-11-21 09:46:55,864 INFO [decode.py:469] batch 56/?, cuts processed until now is 7072 2022-11-21 09:46:58,017 INFO [decode.py:469] batch 58/?, cuts processed until now is 7518 2022-11-21 09:47:00,197 INFO [decode.py:469] batch 60/?, cuts processed until now is 8027 2022-11-21 09:47:02,408 INFO [decode.py:469] batch 62/?, cuts processed until now is 8571 2022-11-21 09:47:04,619 INFO [decode.py:469] batch 64/?, cuts processed until now is 9115 2022-11-21 09:47:07,066 INFO [decode.py:469] batch 66/?, cuts processed until now is 9395 2022-11-21 09:47:09,317 INFO [decode.py:469] batch 68/?, cuts processed until now is 9904 2022-11-21 09:47:11,531 INFO [decode.py:469] batch 70/?, cuts processed until now is 10413 2022-11-21 09:47:13,766 INFO [decode.py:469] batch 72/?, cuts processed until now is 11190 2022-11-21 09:47:15,970 INFO [decode.py:469] batch 74/?, cuts processed until now is 11589 2022-11-21 09:47:17,655 INFO [decode.py:469] batch 76/?, cuts processed until now is 11699 2022-11-21 09:47:19,467 INFO [decode.py:469] batch 78/?, cuts processed until now is 11799 2022-11-21 09:47:21,008 INFO [decode.py:469] batch 80/?, cuts processed until now is 11889 2022-11-21 09:47:22,661 INFO [decode.py:469] batch 82/?, cuts processed until now is 12461 2022-11-21 09:47:24,541 INFO [decode.py:469] batch 84/?, cuts processed until now is 12568 2022-11-21 09:47:29,330 INFO [decode.py:469] batch 86/?, cuts processed until now is 12601 2022-11-21 09:47:29,498 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-dev_gss-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:47:29,659 INFO [utils.py:530] [dev_gss-beam_4_max_contexts_4_max_states_8] %WER 22.21% [21087 / 94940, 2793 ins, 4898 del, 13396 sub ] 2022-11-21 09:47:30,520 INFO [utils.py:530] [dev_gss-beam_4_max_contexts_4_max_states_8] %WER 14.58% [53945 / 369873, 11680 ins, 21193 del, 21072 sub ] 2022-11-21 09:47:31,464 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-dev_gss-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:47:31,465 INFO [decode.py:531] For dev_gss, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 22.21 14.58 best for dev_gss 2022-11-21 09:47:31,470 INFO [decode.py:726] Decoding test_gss 2022-11-21 09:47:34,030 INFO [decode.py:469] batch 0/?, cuts processed until now is 69 2022-11-21 09:47:36,726 INFO [decode.py:469] batch 2/?, cuts processed until now is 555 2022-11-21 09:47:39,438 INFO [decode.py:469] batch 4/?, cuts processed until now is 703 2022-11-21 09:47:42,207 INFO [decode.py:469] batch 6/?, cuts processed until now is 831 2022-11-21 09:47:45,003 INFO [decode.py:469] batch 8/?, cuts processed until now is 988 2022-11-21 09:47:48,933 INFO [decode.py:469] batch 10/?, cuts processed until now is 1096 2022-11-21 09:47:51,621 INFO [decode.py:469] batch 12/?, cuts processed until now is 1268 2022-11-21 09:47:53,933 INFO [decode.py:469] batch 14/?, cuts processed until now is 1533 2022-11-21 09:47:55,182 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.3347, 3.6062, 2.7166, 1.9545, 3.4071, 1.4218, 3.2604, 2.0932], device='cuda:0'), covar=tensor([0.1072, 0.0145, 0.0907, 0.1570, 0.0184, 0.1695, 0.0277, 0.1152], device='cuda:0'), in_proj_covar=tensor([0.0112, 0.0093, 0.0102, 0.0107, 0.0090, 0.0114, 0.0086, 0.0106], device='cuda:0'), out_proj_covar=tensor([0.0005, 0.0003, 0.0004, 0.0004, 0.0003, 0.0004, 0.0003, 0.0004], device='cuda:0') 2022-11-21 09:47:55,993 INFO [decode.py:469] batch 16/?, cuts processed until now is 1932 2022-11-21 09:47:59,662 INFO [decode.py:469] batch 18/?, cuts processed until now is 2057 2022-11-21 09:48:04,043 INFO [decode.py:469] batch 20/?, cuts processed until now is 2126 2022-11-21 09:48:06,353 INFO [decode.py:469] batch 22/?, cuts processed until now is 2390 2022-11-21 09:48:08,351 INFO [decode.py:469] batch 24/?, cuts processed until now is 2858 2022-11-21 09:48:11,314 INFO [decode.py:469] batch 26/?, cuts processed until now is 2998 2022-11-21 09:48:14,124 INFO [decode.py:469] batch 28/?, cuts processed until now is 3280 2022-11-21 09:48:16,777 INFO [decode.py:469] batch 30/?, cuts processed until now is 3432 2022-11-21 09:48:20,723 INFO [decode.py:469] batch 32/?, cuts processed until now is 3537 2022-11-21 09:48:23,634 INFO [decode.py:469] batch 34/?, cuts processed until now is 3709 2022-11-21 09:48:26,540 INFO [decode.py:469] batch 36/?, cuts processed until now is 3825 2022-11-21 09:48:29,325 INFO [decode.py:469] batch 38/?, cuts processed until now is 3972 2022-11-21 09:48:32,390 INFO [decode.py:469] batch 40/?, cuts processed until now is 4410 2022-11-21 09:48:34,687 INFO [decode.py:469] batch 42/?, cuts processed until now is 5060 2022-11-21 09:48:37,004 INFO [decode.py:469] batch 44/?, cuts processed until now is 5546 2022-11-21 09:48:38,609 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.7190, 4.4780, 3.4362, 1.9964, 4.0394, 1.9545, 3.7939, 2.6835], device='cuda:0'), covar=tensor([0.1413, 0.0111, 0.0604, 0.2627, 0.0186, 0.1757, 0.0252, 0.1546], device='cuda:0'), in_proj_covar=tensor([0.0112, 0.0093, 0.0102, 0.0107, 0.0090, 0.0114, 0.0086, 0.0106], device='cuda:0'), out_proj_covar=tensor([0.0005, 0.0003, 0.0004, 0.0004, 0.0003, 0.0004, 0.0003, 0.0004], device='cuda:0') 2022-11-21 09:48:39,665 INFO [decode.py:469] batch 46/?, cuts processed until now is 5687 2022-11-21 09:48:42,162 INFO [decode.py:469] batch 48/?, cuts processed until now is 5893 2022-11-21 09:48:45,120 INFO [decode.py:469] batch 50/?, cuts processed until now is 6379 2022-11-21 09:48:47,277 INFO [decode.py:469] batch 52/?, cuts processed until now is 6713 2022-11-21 09:48:49,298 INFO [decode.py:469] batch 54/?, cuts processed until now is 7112 2022-11-21 09:48:52,978 INFO [decode.py:469] batch 56/?, cuts processed until now is 7298 2022-11-21 09:48:55,141 INFO [decode.py:469] batch 58/?, cuts processed until now is 8130 2022-11-21 09:48:59,231 INFO [decode.py:469] batch 60/?, cuts processed until now is 8273 2022-11-21 09:49:01,493 INFO [decode.py:469] batch 62/?, cuts processed until now is 8813 2022-11-21 09:49:03,797 INFO [decode.py:469] batch 64/?, cuts processed until now is 9353 2022-11-21 09:49:07,802 INFO [decode.py:469] batch 66/?, cuts processed until now is 9500 2022-11-21 09:49:10,994 INFO [decode.py:469] batch 68/?, cuts processed until now is 9944 2022-11-21 09:49:13,266 INFO [decode.py:469] batch 70/?, cuts processed until now is 10274 2022-11-21 09:49:16,318 INFO [decode.py:469] batch 72/?, cuts processed until now is 10711 2022-11-21 09:49:18,676 INFO [decode.py:469] batch 74/?, cuts processed until now is 10820 2022-11-21 09:49:20,277 INFO [decode.py:469] batch 76/?, cuts processed until now is 11076 2022-11-21 09:49:21,355 INFO [decode.py:469] batch 78/?, cuts processed until now is 11209 2022-11-21 09:49:22,909 INFO [decode.py:469] batch 80/?, cuts processed until now is 11651 2022-11-21 09:49:25,098 INFO [decode.py:469] batch 82/?, cuts processed until now is 12070 2022-11-21 09:49:25,367 INFO [zipformer.py:1414] attn_weights_entropy = tensor([2.8827, 4.1919, 4.0346, 3.3228, 2.0846, 4.1770, 2.3554, 3.6866], device='cuda:0'), covar=tensor([0.0463, 0.0336, 0.0194, 0.0486, 0.0737, 0.0181, 0.0636, 0.0154], device='cuda:0'), in_proj_covar=tensor([0.0176, 0.0146, 0.0155, 0.0177, 0.0173, 0.0154, 0.0169, 0.0152], device='cuda:0'), out_proj_covar=tensor([0.0003, 0.0002, 0.0002, 0.0003, 0.0003, 0.0002, 0.0003, 0.0002], device='cuda:0') 2022-11-21 09:49:26,415 INFO [decode.py:485] The transcripts are stored in pruned_transducer_stateless7/exp/v2/fast_beam_search/recogs-test_gss-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:49:26,555 INFO [utils.py:530] [test_gss-beam_4_max_contexts_4_max_states_8] %WER 22.83% [20466 / 89659, 2179 ins, 5438 del, 12849 sub ] 2022-11-21 09:49:27,307 INFO [utils.py:530] [test_gss-beam_4_max_contexts_4_max_states_8] %WER 15.27% [54095 / 354205, 10381 ins, 23091 del, 20623 sub ] 2022-11-21 09:49:28,185 INFO [decode.py:511] Wrote detailed error stats to pruned_transducer_stateless7/exp/v2/fast_beam_search/wers-test_gss-beam_4_max_contexts_4_max_states_8-epoch-14-avg-8-beam-4-max-contexts-4-max-states-8.txt 2022-11-21 09:49:28,186 INFO [decode.py:531] For test_gss, WER/CER of different settings are: beam_4_max_contexts_4_max_states_8 22.83 15.27 best for test_gss 2022-11-21 09:49:28,190 INFO [decode.py:743] Done!