Zengwei commited on
Commit
3361852
·
1 Parent(s): bb6e3ba

upload files

Browse files
Files changed (27) hide show
  1. README.md +1 -0
  2. decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
  3. decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
  4. decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2023-11-20-11-00-12 +252 -0
  5. decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-24-10-49-48 +247 -0
  6. decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
  7. decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
  8. decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt +29 -0
  9. decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt +29 -0
  10. decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
  11. decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
  12. decoding_results/ctc-decoding/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-26-11-20-47 +29 -0
  13. decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
  14. decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
  15. decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt +2 -0
  16. decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt +2 -0
  17. exp/decode.sh +18 -0
  18. exp/epoch-50.pt +3 -0
  19. exp/export.sh +15 -0
  20. exp/pretrained.pt +3 -0
  21. exp/pretrained_ctc.sh +18 -0
  22. exp/tensorboard/events.out.tfevents.1700103483.de-74279-k2-train-1-1220091118-57c4d55446-mvd6x.1928694.0 +3 -0
  23. exp/train.sh +21 -0
  24. test_wavs/1089-134686-0001.wav +0 -0
  25. test_wavs/1221-135766-0001.wav +0 -0
  26. test_wavs/1221-135766-0002.wav +0 -0
  27. test_wavs/trans.txt +3 -0
README.md ADDED
@@ -0,0 +1 @@
 
 
1
+ See https://github.com/k2-fsa/icefall/pull/1389
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2023-11-20-11-00-12 ADDED
@@ -0,0 +1,252 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2023-11-20 11:00:12,309 INFO [ctc_decode.py:690] Decoding started
2
+ 2023-11-20 11:00:12,310 INFO [ctc_decode.py:696] Device: cuda:0
3
+ 2023-11-20 11:00:12,310 INFO [ctc_decode.py:697] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.clean', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '7886da9b-dirty', 'icefall-git-date': 'Tue Nov 14 20:15:43 2023', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/env/k2_icefall/lib/python3.8/site-packages/lhotse-1.17.0.dev0+git.ccfc5b2c.clean-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-9-0208143539-7dbf569d4f-r7nrb', 'IP address': '10.177.13.150'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
4
+ 2023-11-20 11:00:12,759 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2023-11-20 11:00:18,908 INFO [ctc_decode.py:780] About to create model
6
+ 2023-11-20 11:00:20,576 INFO [ctc_decode.py:847] Calculating the averaged model over epoch range from 21 (excluded) to 50
7
+ 2023-11-20 11:00:35,252 INFO [ctc_decode.py:864] Number of model parameters: 174319650
8
+ 2023-11-20 11:00:35,253 INFO [asr_datamodule.py:465] About to get test-clean cuts
9
+ 2023-11-20 11:00:35,506 INFO [asr_datamodule.py:472] About to get test-other cuts
10
+ 2023-11-20 11:00:38,665 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 14
11
+ 2023-11-20 11:01:45,495 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.9823, 4.1719, 4.5345, 4.4621], device='cuda:0')
12
+ 2023-11-20 11:03:08,514 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.2755, 3.6060, 3.7989, 3.7019], device='cuda:0')
13
+ 2023-11-20 11:03:46,736 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2298
14
+ 2023-11-20 11:03:54,391 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.5861, 4.2275, 4.6746, 4.3422], device='cuda:0')
15
+ 2023-11-20 11:04:10,930 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
16
+ 2023-11-20 11:04:11,092 INFO [utils.py:641] [test-clean-attention_scale_0.01] %WER 2.31% [1215 / 52576, 121 ins, 101 del, 993 sub ]
17
+ 2023-11-20 11:04:11,437 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
18
+ 2023-11-20 11:04:11,477 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
19
+ 2023-11-20 11:04:11,634 INFO [utils.py:641] [test-clean-attention_scale_0.05] %WER 2.26% [1190 / 52576, 118 ins, 100 del, 972 sub ]
20
+ 2023-11-20 11:04:11,977 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
21
+ 2023-11-20 11:04:12,017 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
22
+ 2023-11-20 11:04:12,174 INFO [utils.py:641] [test-clean-attention_scale_0.08] %WER 2.26% [1186 / 52576, 119 ins, 100 del, 967 sub ]
23
+ 2023-11-20 11:04:12,522 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
24
+ 2023-11-20 11:04:12,561 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
25
+ 2023-11-20 11:04:12,719 INFO [utils.py:641] [test-clean-attention_scale_0.1] %WER 2.25% [1184 / 52576, 120 ins, 99 del, 965 sub ]
26
+ 2023-11-20 11:04:13,063 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
27
+ 2023-11-20 11:04:13,103 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
28
+ 2023-11-20 11:04:13,260 INFO [utils.py:641] [test-clean-attention_scale_0.3] %WER 2.19% [1151 / 52576, 114 ins, 101 del, 936 sub ]
29
+ 2023-11-20 11:04:13,618 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
30
+ 2023-11-20 11:04:13,672 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
31
+ 2023-11-20 11:04:13,828 INFO [utils.py:641] [test-clean-attention_scale_0.5] %WER 2.16% [1134 / 52576, 114 ins, 99 del, 921 sub ]
32
+ 2023-11-20 11:04:14,380 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
33
+ 2023-11-20 11:04:14,420 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
34
+ 2023-11-20 11:04:14,571 INFO [utils.py:641] [test-clean-attention_scale_0.6] %WER 2.15% [1132 / 52576, 113 ins, 99 del, 920 sub ]
35
+ 2023-11-20 11:04:14,904 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
36
+ 2023-11-20 11:04:14,942 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
37
+ 2023-11-20 11:04:15,089 INFO [utils.py:641] [test-clean-attention_scale_0.7] %WER 2.15% [1130 / 52576, 113 ins, 99 del, 918 sub ]
38
+ 2023-11-20 11:04:15,423 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
39
+ 2023-11-20 11:04:15,462 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
40
+ 2023-11-20 11:04:15,614 INFO [utils.py:641] [test-clean-attention_scale_0.9] %WER 2.13% [1120 / 52576, 114 ins, 98 del, 908 sub ]
41
+ 2023-11-20 11:04:15,952 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
42
+ 2023-11-20 11:04:15,991 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
43
+ 2023-11-20 11:04:16,144 INFO [utils.py:641] [test-clean-attention_scale_1.0] %WER 2.12% [1115 / 52576, 114 ins, 98 del, 903 sub ]
44
+ 2023-11-20 11:04:16,481 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
45
+ 2023-11-20 11:04:16,519 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
46
+ 2023-11-20 11:04:16,671 INFO [utils.py:641] [test-clean-attention_scale_1.1] %WER 2.13% [1119 / 52576, 116 ins, 98 del, 905 sub ]
47
+ 2023-11-20 11:04:17,017 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
48
+ 2023-11-20 11:04:17,056 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
49
+ 2023-11-20 11:04:17,206 INFO [utils.py:641] [test-clean-attention_scale_1.2] %WER 2.14% [1123 / 52576, 116 ins, 100 del, 907 sub ]
50
+ 2023-11-20 11:04:17,547 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
51
+ 2023-11-20 11:04:17,590 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
52
+ 2023-11-20 11:04:17,740 INFO [utils.py:641] [test-clean-attention_scale_1.3] %WER 2.13% [1119 / 52576, 116 ins, 100 del, 903 sub ]
53
+ 2023-11-20 11:04:18,073 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
54
+ 2023-11-20 11:04:18,111 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
55
+ 2023-11-20 11:04:18,262 INFO [utils.py:641] [test-clean-attention_scale_1.5] %WER 2.12% [1113 / 52576, 117 ins, 99 del, 897 sub ]
56
+ 2023-11-20 11:04:18,806 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
57
+ 2023-11-20 11:04:18,844 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
58
+ 2023-11-20 11:04:18,995 INFO [utils.py:641] [test-clean-attention_scale_1.7] %WER 2.12% [1112 / 52576, 118 ins, 98 del, 896 sub ]
59
+ 2023-11-20 11:04:19,328 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
60
+ 2023-11-20 11:04:19,367 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
61
+ 2023-11-20 11:04:19,518 INFO [utils.py:641] [test-clean-attention_scale_1.9] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
62
+ 2023-11-20 11:04:19,857 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
63
+ 2023-11-20 11:04:19,895 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
64
+ 2023-11-20 11:04:20,046 INFO [utils.py:641] [test-clean-attention_scale_2.0] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
65
+ 2023-11-20 11:04:20,382 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
66
+ 2023-11-20 11:04:20,421 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
67
+ 2023-11-20 11:04:20,571 INFO [utils.py:641] [test-clean-attention_scale_2.1] %WER 2.11% [1111 / 52576, 119 ins, 98 del, 894 sub ]
68
+ 2023-11-20 11:04:20,904 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
69
+ 2023-11-20 11:04:20,942 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
70
+ 2023-11-20 11:04:21,090 INFO [utils.py:641] [test-clean-attention_scale_2.2] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
71
+ 2023-11-20 11:04:21,432 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
72
+ 2023-11-20 11:04:21,471 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
73
+ 2023-11-20 11:04:21,622 INFO [utils.py:641] [test-clean-attention_scale_2.3] %WER 2.12% [1115 / 52576, 120 ins, 98 del, 897 sub ]
74
+ 2023-11-20 11:04:21,957 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
75
+ 2023-11-20 11:04:21,995 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
76
+ 2023-11-20 11:04:22,145 INFO [utils.py:641] [test-clean-attention_scale_2.5] %WER 2.12% [1113 / 52576, 119 ins, 98 del, 896 sub ]
77
+ 2023-11-20 11:04:22,478 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
78
+ 2023-11-20 11:04:22,516 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
79
+ 2023-11-20 11:04:22,666 INFO [utils.py:641] [test-clean-attention_scale_3.0] %WER 2.11% [1107 / 52576, 117 ins, 98 del, 892 sub ]
80
+ 2023-11-20 11:04:23,214 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
81
+ 2023-11-20 11:04:23,254 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
82
+ 2023-11-20 11:04:23,402 INFO [utils.py:641] [test-clean-attention_scale_4.0] %WER 2.11% [1107 / 52576, 118 ins, 98 del, 891 sub ]
83
+ 2023-11-20 11:04:23,736 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
84
+ 2023-11-20 11:04:23,774 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
85
+ 2023-11-20 11:04:23,925 INFO [utils.py:641] [test-clean-attention_scale_5.0] %WER 2.10% [1105 / 52576, 118 ins, 98 del, 889 sub ]
86
+ 2023-11-20 11:04:24,258 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
87
+ 2023-11-20 11:04:24,298 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
88
+ 2023-11-20 11:04:24,447 INFO [utils.py:641] [test-clean-attention_scale_6.0] %WER 2.10% [1105 / 52576, 117 ins, 97 del, 891 sub ]
89
+ 2023-11-20 11:04:24,779 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
90
+ 2023-11-20 11:04:24,818 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
91
+ 2023-11-20 11:04:24,977 INFO [utils.py:641] [test-clean-attention_scale_7.0] %WER 2.10% [1103 / 52576, 117 ins, 96 del, 890 sub ]
92
+ 2023-11-20 11:04:25,310 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
93
+ 2023-11-20 11:04:25,349 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
94
+ 2023-11-20 11:04:25,498 INFO [utils.py:641] [test-clean-attention_scale_8.0] %WER 2.09% [1101 / 52576, 116 ins, 96 del, 889 sub ]
95
+ 2023-11-20 11:04:25,834 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
96
+ 2023-11-20 11:04:25,874 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
97
+ 2023-11-20 11:04:26,024 INFO [utils.py:641] [test-clean-attention_scale_9.0] %WER 2.09% [1101 / 52576, 116 ins, 96 del, 889 sub ]
98
+ 2023-11-20 11:04:26,360 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
99
+ 2023-11-20 11:04:26,363 INFO [ctc_decode.py:642]
100
+ For test-clean, WER of different settings are:
101
+ attention_scale_8.0 2.09 best for test-clean
102
+ attention_scale_9.0 2.09
103
+ attention_scale_5.0 2.1
104
+ attention_scale_6.0 2.1
105
+ attention_scale_7.0 2.1
106
+ attention_scale_2.1 2.11
107
+ attention_scale_3.0 2.11
108
+ attention_scale_4.0 2.11
109
+ attention_scale_1.0 2.12
110
+ attention_scale_1.5 2.12
111
+ attention_scale_1.7 2.12
112
+ attention_scale_1.9 2.12
113
+ attention_scale_2.0 2.12
114
+ attention_scale_2.2 2.12
115
+ attention_scale_2.3 2.12
116
+ attention_scale_2.5 2.12
117
+ attention_scale_0.9 2.13
118
+ attention_scale_1.1 2.13
119
+ attention_scale_1.3 2.13
120
+ attention_scale_1.2 2.14
121
+ attention_scale_0.6 2.15
122
+ attention_scale_0.7 2.15
123
+ attention_scale_0.5 2.16
124
+ attention_scale_0.3 2.19
125
+ attention_scale_0.1 2.25
126
+ attention_scale_0.05 2.26
127
+ attention_scale_0.08 2.26
128
+ attention_scale_0.01 2.31
129
+
130
+ 2023-11-20 11:04:29,382 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 17
131
+ 2023-11-20 11:06:22,319 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.9394, 3.1162, 3.4191, 4.0986, 4.0804, 3.4132, 3.5100, 3.5241],
132
+ device='cuda:0')
133
+ 2023-11-20 11:07:23,774 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.5439, 3.3689, 3.4641, 3.4002], device='cuda:0')
134
+ 2023-11-20 11:07:40,616 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2530
135
+ 2023-11-20 11:07:52,519 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.6239, 2.6018, 2.8474, 2.9662, 2.8372, 2.7533, 2.9036, 2.7962],
136
+ device='cuda:0')
137
+ 2023-11-20 11:08:03,346 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
138
+ 2023-11-20 11:08:03,507 INFO [utils.py:641] [test-other-attention_scale_0.01] %WER 5.11% [2673 / 52343, 275 ins, 211 del, 2187 sub ]
139
+ 2023-11-20 11:08:03,861 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
140
+ 2023-11-20 11:08:03,907 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
141
+ 2023-11-20 11:08:04,066 INFO [utils.py:641] [test-other-attention_scale_0.05] %WER 5.02% [2627 / 52343, 269 ins, 206 del, 2152 sub ]
142
+ 2023-11-20 11:08:04,423 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
143
+ 2023-11-20 11:08:04,469 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
144
+ 2023-11-20 11:08:04,633 INFO [utils.py:641] [test-other-attention_scale_0.08] %WER 4.98% [2607 / 52343, 264 ins, 206 del, 2137 sub ]
145
+ 2023-11-20 11:08:05,283 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
146
+ 2023-11-20 11:08:05,322 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
147
+ 2023-11-20 11:08:05,479 INFO [utils.py:641] [test-other-attention_scale_0.1] %WER 4.96% [2598 / 52343, 265 ins, 204 del, 2129 sub ]
148
+ 2023-11-20 11:08:05,829 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
149
+ 2023-11-20 11:08:05,874 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
150
+ 2023-11-20 11:08:06,028 INFO [utils.py:641] [test-other-attention_scale_0.3] %WER 4.83% [2529 / 52343, 263 ins, 193 del, 2073 sub ]
151
+ 2023-11-20 11:08:06,376 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
152
+ 2023-11-20 11:08:06,418 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
153
+ 2023-11-20 11:08:06,573 INFO [utils.py:641] [test-other-attention_scale_0.5] %WER 4.79% [2507 / 52343, 265 ins, 189 del, 2053 sub ]
154
+ 2023-11-20 11:08:06,922 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
155
+ 2023-11-20 11:08:06,964 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
156
+ 2023-11-20 11:08:07,119 INFO [utils.py:641] [test-other-attention_scale_0.6] %WER 4.77% [2497 / 52343, 265 ins, 187 del, 2045 sub ]
157
+ 2023-11-20 11:08:07,466 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
158
+ 2023-11-20 11:08:07,505 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
159
+ 2023-11-20 11:08:07,663 INFO [utils.py:641] [test-other-attention_scale_0.7] %WER 4.77% [2496 / 52343, 262 ins, 187 del, 2047 sub ]
160
+ 2023-11-20 11:08:08,010 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
161
+ 2023-11-20 11:08:08,052 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
162
+ 2023-11-20 11:08:08,207 INFO [utils.py:641] [test-other-attention_scale_0.9] %WER 4.75% [2485 / 52343, 259 ins, 185 del, 2041 sub ]
163
+ 2023-11-20 11:08:08,555 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
164
+ 2023-11-20 11:08:08,602 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
165
+ 2023-11-20 11:08:08,757 INFO [utils.py:641] [test-other-attention_scale_1.0] %WER 4.73% [2475 / 52343, 257 ins, 183 del, 2035 sub ]
166
+ 2023-11-20 11:08:09,104 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
167
+ 2023-11-20 11:08:09,144 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
168
+ 2023-11-20 11:08:09,533 INFO [utils.py:641] [test-other-attention_scale_1.1] %WER 4.72% [2469 / 52343, 258 ins, 182 del, 2029 sub ]
169
+ 2023-11-20 11:08:09,880 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
170
+ 2023-11-20 11:08:09,920 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
171
+ 2023-11-20 11:08:10,072 INFO [utils.py:641] [test-other-attention_scale_1.2] %WER 4.71% [2463 / 52343, 258 ins, 181 del, 2024 sub ]
172
+ 2023-11-20 11:08:10,411 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
173
+ 2023-11-20 11:08:10,453 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
174
+ 2023-11-20 11:08:10,607 INFO [utils.py:641] [test-other-attention_scale_1.3] %WER 4.70% [2458 / 52343, 258 ins, 179 del, 2021 sub ]
175
+ 2023-11-20 11:08:10,954 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
176
+ 2023-11-20 11:08:10,995 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
177
+ 2023-11-20 11:08:11,152 INFO [utils.py:641] [test-other-attention_scale_1.5] %WER 4.67% [2445 / 52343, 256 ins, 181 del, 2008 sub ]
178
+ 2023-11-20 11:08:11,502 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
179
+ 2023-11-20 11:08:11,541 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
180
+ 2023-11-20 11:08:11,698 INFO [utils.py:641] [test-other-attention_scale_1.7] %WER 4.67% [2443 / 52343, 255 ins, 182 del, 2006 sub ]
181
+ 2023-11-20 11:08:12,043 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
182
+ 2023-11-20 11:08:12,083 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
183
+ 2023-11-20 11:08:12,235 INFO [utils.py:641] [test-other-attention_scale_1.9] %WER 4.66% [2440 / 52343, 256 ins, 181 del, 2003 sub ]
184
+ 2023-11-20 11:08:12,594 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
185
+ 2023-11-20 11:08:12,637 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
186
+ 2023-11-20 11:08:12,791 INFO [utils.py:641] [test-other-attention_scale_2.0] %WER 4.66% [2437 / 52343, 254 ins, 179 del, 2004 sub ]
187
+ 2023-11-20 11:08:13,136 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
188
+ 2023-11-20 11:08:13,177 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
189
+ 2023-11-20 11:08:13,331 INFO [utils.py:641] [test-other-attention_scale_2.1] %WER 4.66% [2438 / 52343, 255 ins, 180 del, 2003 sub ]
190
+ 2023-11-20 11:08:13,948 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
191
+ 2023-11-20 11:08:13,990 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
192
+ 2023-11-20 11:08:14,142 INFO [utils.py:641] [test-other-attention_scale_2.2] %WER 4.66% [2437 / 52343, 255 ins, 180 del, 2002 sub ]
193
+ 2023-11-20 11:08:14,487 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
194
+ 2023-11-20 11:08:14,527 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
195
+ 2023-11-20 11:08:14,683 INFO [utils.py:641] [test-other-attention_scale_2.3] %WER 4.65% [2433 / 52343, 253 ins, 180 del, 2000 sub ]
196
+ 2023-11-20 11:08:15,029 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
197
+ 2023-11-20 11:08:15,069 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
198
+ 2023-11-20 11:08:15,221 INFO [utils.py:641] [test-other-attention_scale_2.5] %WER 4.64% [2430 / 52343, 251 ins, 179 del, 2000 sub ]
199
+ 2023-11-20 11:08:15,567 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
200
+ 2023-11-20 11:08:15,608 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
201
+ 2023-11-20 11:08:15,761 INFO [utils.py:641] [test-other-attention_scale_3.0] %WER 4.63% [2422 / 52343, 253 ins, 176 del, 1993 sub ]
202
+ 2023-11-20 11:08:16,115 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
203
+ 2023-11-20 11:08:16,155 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
204
+ 2023-11-20 11:08:16,308 INFO [utils.py:641] [test-other-attention_scale_4.0] %WER 4.62% [2416 / 52343, 253 ins, 176 del, 1987 sub ]
205
+ 2023-11-20 11:08:16,658 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
206
+ 2023-11-20 11:08:16,698 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
207
+ 2023-11-20 11:08:16,852 INFO [utils.py:641] [test-other-attention_scale_5.0] %WER 4.62% [2418 / 52343, 254 ins, 177 del, 1987 sub ]
208
+ 2023-11-20 11:08:17,199 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
209
+ 2023-11-20 11:08:17,240 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
210
+ 2023-11-20 11:08:17,395 INFO [utils.py:641] [test-other-attention_scale_6.0] %WER 4.62% [2418 / 52343, 254 ins, 178 del, 1986 sub ]
211
+ 2023-11-20 11:08:17,745 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
212
+ 2023-11-20 11:08:17,784 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
213
+ 2023-11-20 11:08:17,939 INFO [utils.py:641] [test-other-attention_scale_7.0] %WER 4.61% [2412 / 52343, 255 ins, 176 del, 1981 sub ]
214
+ 2023-11-20 11:08:18,284 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
215
+ 2023-11-20 11:08:18,324 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
216
+ 2023-11-20 11:08:18,725 INFO [utils.py:641] [test-other-attention_scale_8.0] %WER 4.60% [2407 / 52343, 255 ins, 175 del, 1977 sub ]
217
+ 2023-11-20 11:08:19,071 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
218
+ 2023-11-20 11:08:19,111 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
219
+ 2023-11-20 11:08:19,267 INFO [utils.py:641] [test-other-attention_scale_9.0] %WER 4.59% [2405 / 52343, 255 ins, 175 del, 1975 sub ]
220
+ 2023-11-20 11:08:19,612 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
221
+ 2023-11-20 11:08:19,615 INFO [ctc_decode.py:642]
222
+ For test-other, WER of different settings are:
223
+ attention_scale_9.0 4.59 best for test-other
224
+ attention_scale_8.0 4.6
225
+ attention_scale_7.0 4.61
226
+ attention_scale_4.0 4.62
227
+ attention_scale_5.0 4.62
228
+ attention_scale_6.0 4.62
229
+ attention_scale_3.0 4.63
230
+ attention_scale_2.5 4.64
231
+ attention_scale_2.3 4.65
232
+ attention_scale_1.9 4.66
233
+ attention_scale_2.0 4.66
234
+ attention_scale_2.1 4.66
235
+ attention_scale_2.2 4.66
236
+ attention_scale_1.5 4.67
237
+ attention_scale_1.7 4.67
238
+ attention_scale_1.3 4.7
239
+ attention_scale_1.2 4.71
240
+ attention_scale_1.1 4.72
241
+ attention_scale_1.0 4.73
242
+ attention_scale_0.9 4.75
243
+ attention_scale_0.6 4.77
244
+ attention_scale_0.7 4.77
245
+ attention_scale_0.5 4.79
246
+ attention_scale_0.3 4.83
247
+ attention_scale_0.1 4.96
248
+ attention_scale_0.08 4.98
249
+ attention_scale_0.05 5.02
250
+ attention_scale_0.01 5.11
251
+
252
+ 2023-11-20 11:08:19,616 INFO [ctc_decode.py:897] Done!
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-24-10-49-48 ADDED
@@ -0,0 +1,247 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-24 10:49:48,302 INFO [ctc_decode.py:690] Decoding started
2
+ 2024-05-24 10:49:48,302 INFO [ctc_decode.py:696] Device: cuda:0
3
+ 2024-05-24 10:49:48,302 INFO [ctc_decode.py:697] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '0be32f3d-clean', 'icefall-git-date': 'Fri Feb 23 15:21:01 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-10-0312151423-668f59dc99-b8962', 'IP address': '10.177.6.147'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
4
+ 2024-05-24 10:49:48,578 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2024-05-24 10:49:53,648 INFO [ctc_decode.py:780] About to create model
6
+ 2024-05-24 10:49:54,840 INFO [ctc_decode.py:847] Calculating the averaged model over epoch range from 21 (excluded) to 50
7
+ 2024-05-24 10:50:10,087 INFO [ctc_decode.py:864] Number of model parameters: 174319650
8
+ 2024-05-24 10:50:10,087 INFO [asr_datamodule.py:465] About to get test-clean cuts
9
+ 2024-05-24 10:50:10,239 INFO [asr_datamodule.py:472] About to get test-other cuts
10
+ 2024-05-24 10:50:12,872 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 14
11
+ 2024-05-24 10:50:52,024 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.9550, 4.9446, 4.5893, 4.1813], device='cuda:0')
12
+ 2024-05-24 10:53:16,396 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2298
13
+ 2024-05-24 10:53:39,601 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
14
+ 2024-05-24 10:53:39,698 INFO [utils.py:641] [test-clean-attention_scale_0.01] %WER 2.31% [1215 / 52576, 122 ins, 101 del, 992 sub ]
15
+ 2024-05-24 10:53:39,895 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
16
+ 2024-05-24 10:53:39,923 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
17
+ 2024-05-24 10:53:40,013 INFO [utils.py:641] [test-clean-attention_scale_0.05] %WER 2.26% [1189 / 52576, 119 ins, 101 del, 969 sub ]
18
+ 2024-05-24 10:53:40,205 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
19
+ 2024-05-24 10:53:40,231 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
20
+ 2024-05-24 10:53:40,321 INFO [utils.py:641] [test-clean-attention_scale_0.08] %WER 2.26% [1186 / 52576, 120 ins, 100 del, 966 sub ]
21
+ 2024-05-24 10:53:40,513 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
22
+ 2024-05-24 10:53:40,540 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
23
+ 2024-05-24 10:53:40,638 INFO [utils.py:641] [test-clean-attention_scale_0.1] %WER 2.25% [1183 / 52576, 121 ins, 99 del, 963 sub ]
24
+ 2024-05-24 10:53:40,833 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
25
+ 2024-05-24 10:53:40,861 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
26
+ 2024-05-24 10:53:40,950 INFO [utils.py:641] [test-clean-attention_scale_0.3] %WER 2.19% [1150 / 52576, 116 ins, 99 del, 935 sub ]
27
+ 2024-05-24 10:53:41,145 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
28
+ 2024-05-24 10:53:41,172 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
29
+ 2024-05-24 10:53:41,260 INFO [utils.py:641] [test-clean-attention_scale_0.5] %WER 2.15% [1132 / 52576, 115 ins, 99 del, 918 sub ]
30
+ 2024-05-24 10:53:41,650 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
31
+ 2024-05-24 10:53:41,677 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
32
+ 2024-05-24 10:53:41,767 INFO [utils.py:641] [test-clean-attention_scale_0.6] %WER 2.15% [1131 / 52576, 114 ins, 99 del, 918 sub ]
33
+ 2024-05-24 10:53:41,966 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
34
+ 2024-05-24 10:53:41,992 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
35
+ 2024-05-24 10:53:42,082 INFO [utils.py:641] [test-clean-attention_scale_0.7] %WER 2.14% [1127 / 52576, 114 ins, 99 del, 914 sub ]
36
+ 2024-05-24 10:53:42,274 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
37
+ 2024-05-24 10:53:42,301 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
38
+ 2024-05-24 10:53:42,390 INFO [utils.py:641] [test-clean-attention_scale_0.9] %WER 2.13% [1118 / 52576, 114 ins, 98 del, 906 sub ]
39
+ 2024-05-24 10:53:42,581 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
40
+ 2024-05-24 10:53:42,609 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
41
+ 2024-05-24 10:53:42,698 INFO [utils.py:641] [test-clean-attention_scale_1.0] %WER 2.11% [1110 / 52576, 112 ins, 98 del, 900 sub ]
42
+ 2024-05-24 10:53:42,888 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
43
+ 2024-05-24 10:53:42,914 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
44
+ 2024-05-24 10:53:43,002 INFO [utils.py:641] [test-clean-attention_scale_1.1] %WER 2.12% [1114 / 52576, 114 ins, 98 del, 902 sub ]
45
+ 2024-05-24 10:53:43,200 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
46
+ 2024-05-24 10:53:43,262 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
47
+ 2024-05-24 10:53:43,353 INFO [utils.py:641] [test-clean-attention_scale_1.2] %WER 2.12% [1116 / 52576, 114 ins, 100 del, 902 sub ]
48
+ 2024-05-24 10:53:43,547 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
49
+ 2024-05-24 10:53:43,576 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
50
+ 2024-05-24 10:53:43,666 INFO [utils.py:641] [test-clean-attention_scale_1.3] %WER 2.12% [1114 / 52576, 114 ins, 99 del, 901 sub ]
51
+ 2024-05-24 10:53:43,856 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
52
+ 2024-05-24 10:53:43,881 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
53
+ 2024-05-24 10:53:43,970 INFO [utils.py:641] [test-clean-attention_scale_1.5] %WER 2.11% [1108 / 52576, 117 ins, 98 del, 893 sub ]
54
+ 2024-05-24 10:53:44,384 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
55
+ 2024-05-24 10:53:44,412 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
56
+ 2024-05-24 10:53:44,503 INFO [utils.py:641] [test-clean-attention_scale_1.7] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
57
+ 2024-05-24 10:53:44,695 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
58
+ 2024-05-24 10:53:44,721 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
59
+ 2024-05-24 10:53:44,811 INFO [utils.py:641] [test-clean-attention_scale_1.9] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
60
+ 2024-05-24 10:53:45,005 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
61
+ 2024-05-24 10:53:45,073 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
62
+ 2024-05-24 10:53:45,167 INFO [utils.py:641] [test-clean-attention_scale_2.0] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
63
+ 2024-05-24 10:53:45,359 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
64
+ 2024-05-24 10:53:45,386 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
65
+ 2024-05-24 10:53:45,475 INFO [utils.py:641] [test-clean-attention_scale_2.1] %WER 2.11% [1107 / 52576, 118 ins, 97 del, 892 sub ]
66
+ 2024-05-24 10:53:45,675 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
67
+ 2024-05-24 10:53:45,702 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
68
+ 2024-05-24 10:53:45,792 INFO [utils.py:641] [test-clean-attention_scale_2.2] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
69
+ 2024-05-24 10:53:45,984 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
70
+ 2024-05-24 10:53:46,011 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
71
+ 2024-05-24 10:53:46,101 INFO [utils.py:641] [test-clean-attention_scale_2.3] %WER 2.11% [1111 / 52576, 119 ins, 97 del, 895 sub ]
72
+ 2024-05-24 10:53:46,293 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
73
+ 2024-05-24 10:53:46,339 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
74
+ 2024-05-24 10:53:46,428 INFO [utils.py:641] [test-clean-attention_scale_2.5] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
75
+ 2024-05-24 10:53:46,619 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
76
+ 2024-05-24 10:53:46,646 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
77
+ 2024-05-24 10:53:46,734 INFO [utils.py:641] [test-clean-attention_scale_3.0] %WER 2.10% [1105 / 52576, 116 ins, 98 del, 891 sub ]
78
+ 2024-05-24 10:53:46,965 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
79
+ 2024-05-24 10:53:46,994 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
80
+ 2024-05-24 10:53:47,098 INFO [utils.py:641] [test-clean-attention_scale_4.0] %WER 2.10% [1105 / 52576, 118 ins, 98 del, 889 sub ]
81
+ 2024-05-24 10:53:47,494 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
82
+ 2024-05-24 10:53:47,549 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
83
+ 2024-05-24 10:53:47,638 INFO [utils.py:641] [test-clean-attention_scale_5.0] %WER 2.10% [1106 / 52576, 118 ins, 98 del, 890 sub ]
84
+ 2024-05-24 10:53:47,830 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
85
+ 2024-05-24 10:53:47,857 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
86
+ 2024-05-24 10:53:47,945 INFO [utils.py:641] [test-clean-attention_scale_6.0] %WER 2.10% [1105 / 52576, 117 ins, 97 del, 891 sub ]
87
+ 2024-05-24 10:53:48,161 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
88
+ 2024-05-24 10:53:48,189 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
89
+ 2024-05-24 10:53:48,283 INFO [utils.py:641] [test-clean-attention_scale_7.0] %WER 2.10% [1106 / 52576, 117 ins, 97 del, 892 sub ]
90
+ 2024-05-24 10:53:48,477 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
91
+ 2024-05-24 10:53:48,504 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
92
+ 2024-05-24 10:53:48,594 INFO [utils.py:641] [test-clean-attention_scale_8.0] %WER 2.10% [1105 / 52576, 116 ins, 97 del, 892 sub ]
93
+ 2024-05-24 10:53:48,785 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
94
+ 2024-05-24 10:53:48,811 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
95
+ 2024-05-24 10:53:48,900 INFO [utils.py:641] [test-clean-attention_scale_9.0] %WER 2.10% [1105 / 52576, 116 ins, 97 del, 892 sub ]
96
+ 2024-05-24 10:53:49,098 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
97
+ 2024-05-24 10:53:49,101 INFO [ctc_decode.py:642]
98
+ For test-clean, WER of different settings are:
99
+ attention_scale_3.0 2.1 best for test-clean
100
+ attention_scale_4.0 2.1
101
+ attention_scale_5.0 2.1
102
+ attention_scale_6.0 2.1
103
+ attention_scale_7.0 2.1
104
+ attention_scale_8.0 2.1
105
+ attention_scale_9.0 2.1
106
+ attention_scale_1.0 2.11
107
+ attention_scale_1.5 2.11
108
+ attention_scale_1.7 2.11
109
+ attention_scale_1.9 2.11
110
+ attention_scale_2.0 2.11
111
+ attention_scale_2.1 2.11
112
+ attention_scale_2.2 2.11
113
+ attention_scale_2.3 2.11
114
+ attention_scale_2.5 2.11
115
+ attention_scale_1.1 2.12
116
+ attention_scale_1.2 2.12
117
+ attention_scale_1.3 2.12
118
+ attention_scale_0.9 2.13
119
+ attention_scale_0.7 2.14
120
+ attention_scale_0.5 2.15
121
+ attention_scale_0.6 2.15
122
+ attention_scale_0.3 2.19
123
+ attention_scale_0.1 2.25
124
+ attention_scale_0.05 2.26
125
+ attention_scale_0.08 2.26
126
+ attention_scale_0.01 2.31
127
+
128
+ 2024-05-24 10:53:51,796 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 17
129
+ 2024-05-24 10:54:02,980 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.6454, 2.4206, 3.8113, 4.5973], device='cuda:0')
130
+ 2024-05-24 10:55:02,806 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.3866, 3.0863, 3.1216, 3.0326], device='cuda:0')
131
+ 2024-05-24 10:56:59,598 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2530
132
+ 2024-05-24 10:57:21,501 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
133
+ 2024-05-24 10:57:21,600 INFO [utils.py:641] [test-other-attention_scale_0.01] %WER 5.12% [2678 / 52343, 277 ins, 212 del, 2189 sub ]
134
+ 2024-05-24 10:57:21,801 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
135
+ 2024-05-24 10:57:21,836 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
136
+ 2024-05-24 10:57:21,930 INFO [utils.py:641] [test-other-attention_scale_0.05] %WER 5.02% [2630 / 52343, 270 ins, 206 del, 2154 sub ]
137
+ 2024-05-24 10:57:22,396 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
138
+ 2024-05-24 10:57:22,426 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
139
+ 2024-05-24 10:57:22,519 INFO [utils.py:641] [test-other-attention_scale_0.08] %WER 4.98% [2606 / 52343, 264 ins, 206 del, 2136 sub ]
140
+ 2024-05-24 10:57:22,719 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
141
+ 2024-05-24 10:57:22,747 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
142
+ 2024-05-24 10:57:22,839 INFO [utils.py:641] [test-other-attention_scale_0.1] %WER 4.96% [2597 / 52343, 265 ins, 204 del, 2128 sub ]
143
+ 2024-05-24 10:57:23,039 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
144
+ 2024-05-24 10:57:23,066 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
145
+ 2024-05-24 10:57:23,158 INFO [utils.py:641] [test-other-attention_scale_0.3] %WER 4.83% [2527 / 52343, 261 ins, 194 del, 2072 sub ]
146
+ 2024-05-24 10:57:23,354 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
147
+ 2024-05-24 10:57:23,385 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
148
+ 2024-05-24 10:57:23,485 INFO [utils.py:641] [test-other-attention_scale_0.5] %WER 4.79% [2507 / 52343, 262 ins, 190 del, 2055 sub ]
149
+ 2024-05-24 10:57:23,686 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
150
+ 2024-05-24 10:57:23,714 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
151
+ 2024-05-24 10:57:23,808 INFO [utils.py:641] [test-other-attention_scale_0.6] %WER 4.76% [2493 / 52343, 262 ins, 187 del, 2044 sub ]
152
+ 2024-05-24 10:57:24,010 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
153
+ 2024-05-24 10:57:24,036 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
154
+ 2024-05-24 10:57:24,128 INFO [utils.py:641] [test-other-attention_scale_0.7] %WER 4.76% [2491 / 52343, 260 ins, 186 del, 2045 sub ]
155
+ 2024-05-24 10:57:24,326 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
156
+ 2024-05-24 10:57:24,353 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
157
+ 2024-05-24 10:57:24,444 INFO [utils.py:641] [test-other-attention_scale_0.9] %WER 4.74% [2479 / 52343, 258 ins, 184 del, 2037 sub ]
158
+ 2024-05-24 10:57:24,643 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
159
+ 2024-05-24 10:57:24,672 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
160
+ 2024-05-24 10:57:24,769 INFO [utils.py:641] [test-other-attention_scale_1.0] %WER 4.72% [2468 / 52343, 256 ins, 183 del, 2029 sub ]
161
+ 2024-05-24 10:57:25,183 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
162
+ 2024-05-24 10:57:25,215 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
163
+ 2024-05-24 10:57:25,307 INFO [utils.py:641] [test-other-attention_scale_1.1] %WER 4.70% [2462 / 52343, 257 ins, 182 del, 2023 sub ]
164
+ 2024-05-24 10:57:25,504 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
165
+ 2024-05-24 10:57:25,559 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
166
+ 2024-05-24 10:57:25,651 INFO [utils.py:641] [test-other-attention_scale_1.2] %WER 4.70% [2458 / 52343, 257 ins, 182 del, 2019 sub ]
167
+ 2024-05-24 10:57:25,875 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
168
+ 2024-05-24 10:57:25,936 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
169
+ 2024-05-24 10:57:26,030 INFO [utils.py:641] [test-other-attention_scale_1.3] %WER 4.69% [2454 / 52343, 257 ins, 180 del, 2017 sub ]
170
+ 2024-05-24 10:57:26,228 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
171
+ 2024-05-24 10:57:26,257 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
172
+ 2024-05-24 10:57:26,355 INFO [utils.py:641] [test-other-attention_scale_1.5] %WER 4.66% [2440 / 52343, 254 ins, 181 del, 2005 sub ]
173
+ 2024-05-24 10:57:26,562 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
174
+ 2024-05-24 10:57:26,589 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
175
+ 2024-05-24 10:57:26,681 INFO [utils.py:641] [test-other-attention_scale_1.7] %WER 4.66% [2438 / 52343, 254 ins, 182 del, 2002 sub ]
176
+ 2024-05-24 10:57:26,893 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
177
+ 2024-05-24 10:57:26,921 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
178
+ 2024-05-24 10:57:27,012 INFO [utils.py:641] [test-other-attention_scale_1.9] %WER 4.65% [2436 / 52343, 255 ins, 180 del, 2001 sub ]
179
+ 2024-05-24 10:57:27,219 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
180
+ 2024-05-24 10:57:27,247 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
181
+ 2024-05-24 10:57:27,339 INFO [utils.py:641] [test-other-attention_scale_2.0] %WER 4.65% [2433 / 52343, 253 ins, 178 del, 2002 sub ]
182
+ 2024-05-24 10:57:27,543 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
183
+ 2024-05-24 10:57:27,573 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
184
+ 2024-05-24 10:57:27,673 INFO [utils.py:641] [test-other-attention_scale_2.1] %WER 4.65% [2436 / 52343, 255 ins, 179 del, 2002 sub ]
185
+ 2024-05-24 10:57:27,873 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
186
+ 2024-05-24 10:57:27,900 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
187
+ 2024-05-24 10:57:27,992 INFO [utils.py:641] [test-other-attention_scale_2.2] %WER 4.65% [2435 / 52343, 255 ins, 179 del, 2001 sub ]
188
+ 2024-05-24 10:57:28,396 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
189
+ 2024-05-24 10:57:28,426 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
190
+ 2024-05-24 10:57:28,517 INFO [utils.py:641] [test-other-attention_scale_2.3] %WER 4.64% [2429 / 52343, 253 ins, 179 del, 1997 sub ]
191
+ 2024-05-24 10:57:28,712 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
192
+ 2024-05-24 10:57:28,739 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
193
+ 2024-05-24 10:57:28,834 INFO [utils.py:641] [test-other-attention_scale_2.5] %WER 4.64% [2428 / 52343, 252 ins, 178 del, 1998 sub ]
194
+ 2024-05-24 10:57:29,038 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
195
+ 2024-05-24 10:57:29,100 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
196
+ 2024-05-24 10:57:29,192 INFO [utils.py:641] [test-other-attention_scale_3.0] %WER 4.62% [2417 / 52343, 252 ins, 173 del, 1992 sub ]
197
+ 2024-05-24 10:57:29,392 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
198
+ 2024-05-24 10:57:29,419 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
199
+ 2024-05-24 10:57:29,513 INFO [utils.py:641] [test-other-attention_scale_4.0] %WER 4.60% [2407 / 52343, 251 ins, 173 del, 1983 sub ]
200
+ 2024-05-24 10:57:29,710 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
201
+ 2024-05-24 10:57:29,737 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
202
+ 2024-05-24 10:57:29,829 INFO [utils.py:641] [test-other-attention_scale_5.0] %WER 4.60% [2409 / 52343, 252 ins, 174 del, 1983 sub ]
203
+ 2024-05-24 10:57:30,024 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
204
+ 2024-05-24 10:57:30,051 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
205
+ 2024-05-24 10:57:30,148 INFO [utils.py:641] [test-other-attention_scale_6.0] %WER 4.59% [2405 / 52343, 252 ins, 174 del, 1979 sub ]
206
+ 2024-05-24 10:57:30,352 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
207
+ 2024-05-24 10:57:30,379 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
208
+ 2024-05-24 10:57:30,472 INFO [utils.py:641] [test-other-attention_scale_7.0] %WER 4.59% [2400 / 52343, 253 ins, 172 del, 1975 sub ]
209
+ 2024-05-24 10:57:30,671 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
210
+ 2024-05-24 10:57:30,699 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
211
+ 2024-05-24 10:57:30,791 INFO [utils.py:641] [test-other-attention_scale_8.0] %WER 4.58% [2396 / 52343, 253 ins, 171 del, 1972 sub ]
212
+ 2024-05-24 10:57:31,193 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
213
+ 2024-05-24 10:57:31,219 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
214
+ 2024-05-24 10:57:31,311 INFO [utils.py:641] [test-other-attention_scale_9.0] %WER 4.57% [2394 / 52343, 253 ins, 171 del, 1970 sub ]
215
+ 2024-05-24 10:57:31,514 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
216
+ 2024-05-24 10:57:31,517 INFO [ctc_decode.py:642]
217
+ For test-other, WER of different settings are:
218
+ attention_scale_9.0 4.57 best for test-other
219
+ attention_scale_8.0 4.58
220
+ attention_scale_6.0 4.59
221
+ attention_scale_7.0 4.59
222
+ attention_scale_4.0 4.6
223
+ attention_scale_5.0 4.6
224
+ attention_scale_3.0 4.62
225
+ attention_scale_2.3 4.64
226
+ attention_scale_2.5 4.64
227
+ attention_scale_1.9 4.65
228
+ attention_scale_2.0 4.65
229
+ attention_scale_2.1 4.65
230
+ attention_scale_2.2 4.65
231
+ attention_scale_1.5 4.66
232
+ attention_scale_1.7 4.66
233
+ attention_scale_1.3 4.69
234
+ attention_scale_1.1 4.7
235
+ attention_scale_1.2 4.7
236
+ attention_scale_1.0 4.72
237
+ attention_scale_0.9 4.74
238
+ attention_scale_0.6 4.76
239
+ attention_scale_0.7 4.76
240
+ attention_scale_0.5 4.79
241
+ attention_scale_0.3 4.83
242
+ attention_scale_0.1 4.96
243
+ attention_scale_0.08 4.98
244
+ attention_scale_0.05 5.02
245
+ attention_scale_0.01 5.12
246
+
247
+ 2024-05-24 10:57:31,517 INFO [ctc_decode.py:897] Done!
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ settings WER
2
+ attention_scale_3.0 2.1
3
+ attention_scale_4.0 2.1
4
+ attention_scale_5.0 2.1
5
+ attention_scale_6.0 2.1
6
+ attention_scale_7.0 2.1
7
+ attention_scale_8.0 2.1
8
+ attention_scale_9.0 2.1
9
+ attention_scale_1.0 2.11
10
+ attention_scale_1.5 2.11
11
+ attention_scale_1.7 2.11
12
+ attention_scale_1.9 2.11
13
+ attention_scale_2.0 2.11
14
+ attention_scale_2.1 2.11
15
+ attention_scale_2.2 2.11
16
+ attention_scale_2.3 2.11
17
+ attention_scale_2.5 2.11
18
+ attention_scale_1.1 2.12
19
+ attention_scale_1.2 2.12
20
+ attention_scale_1.3 2.12
21
+ attention_scale_0.9 2.13
22
+ attention_scale_0.7 2.14
23
+ attention_scale_0.5 2.15
24
+ attention_scale_0.6 2.15
25
+ attention_scale_0.3 2.19
26
+ attention_scale_0.1 2.25
27
+ attention_scale_0.05 2.26
28
+ attention_scale_0.08 2.26
29
+ attention_scale_0.01 2.31
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ settings WER
2
+ attention_scale_9.0 4.57
3
+ attention_scale_8.0 4.58
4
+ attention_scale_6.0 4.59
5
+ attention_scale_7.0 4.59
6
+ attention_scale_4.0 4.6
7
+ attention_scale_5.0 4.6
8
+ attention_scale_3.0 4.62
9
+ attention_scale_2.3 4.64
10
+ attention_scale_2.5 4.64
11
+ attention_scale_1.9 4.65
12
+ attention_scale_2.0 4.65
13
+ attention_scale_2.1 4.65
14
+ attention_scale_2.2 4.65
15
+ attention_scale_1.5 4.66
16
+ attention_scale_1.7 4.66
17
+ attention_scale_1.3 4.69
18
+ attention_scale_1.1 4.7
19
+ attention_scale_1.2 4.7
20
+ attention_scale_1.0 4.72
21
+ attention_scale_0.9 4.74
22
+ attention_scale_0.6 4.76
23
+ attention_scale_0.7 4.76
24
+ attention_scale_0.5 4.79
25
+ attention_scale_0.3 4.83
26
+ attention_scale_0.1 4.96
27
+ attention_scale_0.08 4.98
28
+ attention_scale_0.05 5.02
29
+ attention_scale_0.01 5.12
decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-26-11-20-47 ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-26 11:20:47,975 INFO [ctc_decode.py:717] Decoding started
2
+ 2024-05-26 11:20:47,975 INFO [ctc_decode.py:723] Device: cuda:0
3
+ 2024-05-26 11:20:47,975 INFO [ctc_decode.py:724] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '84dfb576-dirty', 'icefall-git-date': 'Sat May 25 17:49:14 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-10-0312151423-668f59dc99-b8962', 'IP address': '10.177.6.147'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
4
+ 2024-05-26 11:20:48,266 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
5
+ 2024-05-26 11:20:53,300 INFO [ctc_decode.py:807] About to create model
6
+ 2024-05-26 11:20:54,579 INFO [ctc_decode.py:874] Calculating the averaged model over epoch range from 21 (excluded) to 50
7
+ 2024-05-26 11:21:03,984 INFO [ctc_decode.py:891] Number of model parameters: 174319650
8
+ 2024-05-26 11:21:03,985 INFO [asr_datamodule.py:467] About to get test-clean cuts
9
+ 2024-05-26 11:21:04,107 INFO [asr_datamodule.py:474] About to get test-other cuts
10
+ 2024-05-26 11:21:05,248 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 14
11
+ 2024-05-26 11:21:40,659 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2298
12
+ 2024-05-26 11:21:43,885 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.6771, 3.2606, 3.3798, 3.6473], device='cuda:0')
13
+ 2024-05-26 11:21:45,487 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
14
+ 2024-05-26 11:21:45,583 INFO [utils.py:657] [test-clean-ctc-decoding] %WER 2.29% [1206 / 52576, 119 ins, 98 del, 989 sub ]
15
+ 2024-05-26 11:21:45,788 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
16
+ 2024-05-26 11:21:45,791 INFO [ctc_decode.py:669]
17
+ For test-clean, WER of different settings are:
18
+ ctc-decoding 2.29 best for test-clean
19
+
20
+ 2024-05-26 11:21:46,547 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 17
21
+ 2024-05-26 11:22:22,471 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2530
22
+ 2024-05-26 11:22:27,123 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
23
+ 2024-05-26 11:22:27,217 INFO [utils.py:657] [test-other-ctc-decoding] %WER 5.14% [2688 / 52343, 277 ins, 209 del, 2202 sub ]
24
+ 2024-05-26 11:22:27,413 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
25
+ 2024-05-26 11:22:27,416 INFO [ctc_decode.py:669]
26
+ For test-other, WER of different settings are:
27
+ ctc-decoding 5.14 best for test-other
28
+
29
+ 2024-05-26 11:22:27,416 INFO [ctc_decode.py:924] Done!
decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
The diff for this file is too large to render. See raw diff
 
decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 2.29
decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ settings WER
2
+ ctc-decoding 5.14
exp/decode.sh ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES="0"
2
+ for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
3
+ ./zipformer/ctc_decode.py \
4
+ --epoch 50 \
5
+ --avg 29 \
6
+ --exp-dir zipformer/exp-large \
7
+ --use-ctc 1 \
8
+ --use-transducer 0 \
9
+ --use-attention-decoder 1 \
10
+ --attention-decoder-loss-scale 0.9 \
11
+ --num-encoder-layers 2,2,4,5,4,2 \
12
+ --feedforward-dim 512,768,1536,2048,1536,768 \
13
+ --encoder-dim 192,256,512,768,512,256 \
14
+ --encoder-unmasked-dim 192,192,256,320,256,192 \
15
+ --max-duration 100 \
16
+ --causal 0 \
17
+ --num-paths 100 \
18
+ --decoding-method $m
exp/epoch-50.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d3956bffeda26a91e5d3f8dbc7ac811bd7332c56467178ce867418646188640
3
+ size 2790381194
exp/export.sh ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ./zipformer/export.py \
2
+ --exp-dir zipformer/exp-large \
3
+ --use-ctc 1 \
4
+ --use-transducer 0 \
5
+ --use-attention-decoder 1 \
6
+ --num-encoder-layers 2,2,4,5,4,2 \
7
+ --feedforward-dim 512,768,1536,2048,1536,768 \
8
+ --encoder-dim 192,256,512,768,512,256 \
9
+ --encoder-unmasked-dim 192,192,256,320,256,192 \
10
+ --tokens data/lang_bpe_500/tokens.txt \
11
+ --epoch 50 \
12
+ --avg 29 \
13
+ --jit 0
14
+
15
+
exp/pretrained.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78d4c46696304b6df6519d9a1355325d1233712e8f7108a97ea5d067b088c637
3
+ size 697818201
exp/pretrained_ctc.sh ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES=7
2
+ for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
3
+ ./zipformer/pretrained_ctc.py \
4
+ --use-ctc 1 \
5
+ --use-transducer 0 \
6
+ --use-attention-decoder 1 \
7
+ --num-encoder-layers 2,2,4,5,4,2 \
8
+ --feedforward-dim 512,768,1536,2048,1536,768 \
9
+ --encoder-dim 192,256,512,768,512,256 \
10
+ --encoder-unmasked-dim 192,192,256,320,256,192 \
11
+ --checkpoint ./zipformer/exp-large/pretrained.pt \
12
+ --tokens data/lang_bpe_500/tokens.txt \
13
+ --method $m \
14
+ --sample-rate 16000 \
15
+ test_wavs/1089-134686-0001.wav \
16
+ test_wavs/1221-135766-0001.wav \
17
+ test_wavs/1221-135766-0002.wav
18
+ done
exp/tensorboard/events.out.tfevents.1700103483.de-74279-k2-train-1-1220091118-57c4d55446-mvd6x.1928694.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24870b4e634398f97a7c5b789d52b2fd07b6b9a0d76b3ef75317e51c8cbb49b0
3
+ size 1608517
exp/train.sh ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ export CUDA_VISIBLE_DEVICES="0,1,2,3"
2
+ # For non-streaming model training:
3
+ ./zipformer/train.py \
4
+ --world-size 4 \
5
+ --num-epochs 50 \
6
+ --start-epoch 1 \
7
+ --use-fp16 1 \
8
+ --exp-dir zipformer/exp-large \
9
+ --full-libri 1 \
10
+ --use-ctc 1 \
11
+ --use-transducer 0 \
12
+ --use-attention-decoder 1 \
13
+ --ctc-loss-scale 0.1 \
14
+ --attention-decoder-loss-scale 0.9 \
15
+ --num-encoder-layers 2,2,4,5,4,2 \
16
+ --feedforward-dim 512,768,1536,2048,1536,768 \
17
+ --encoder-dim 192,256,512,768,512,256 \
18
+ --encoder-unmasked-dim 192,192,256,320,256,192 \
19
+ --max-duration 1200 \
20
+ --master-port 12345
21
+
test_wavs/1089-134686-0001.wav ADDED
Binary file (212 kB). View file
 
test_wavs/1221-135766-0001.wav ADDED
Binary file (535 kB). View file
 
test_wavs/1221-135766-0002.wav ADDED
Binary file (154 kB). View file
 
test_wavs/trans.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ 1089-134686-0001 AFTER EARLY NIGHTFALL THE YELLOW LAMPS WOULD LIGHT UP HERE AND THERE THE SQUALID QUARTER OF THE BROTHELS
2
+ 1221-135766-0001 GOD AS A DIRECT CONSEQUENCE OF THE SIN WHICH MAN THUS PUNISHED HAD GIVEN HER A LOVELY CHILD WHOSE PLACE WAS ON THAT SAME DISHONOURED BOSOM TO CONNECT HER PARENT FOR EVER WITH THE RACE AND DESCENT OF MORTALS AND TO BE FINALLY A BLESSED SOUL IN HEAVEN
3
+ 1221-135766-0002 YET THESE THOUGHTS AFFECTED HESTER PRYNNE LESS WITH HOPE THAN APPREHENSION