upload files
Browse files- README.md +1 -0
- decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2023-11-20-11-00-12 +252 -0
- decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-24-10-49-48 +247 -0
- decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt +29 -0
- decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt +29 -0
- decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/ctc-decoding/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-26-11-20-47 +29 -0
- decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt +0 -0
- decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt +2 -0
- decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt +2 -0
- exp/decode.sh +18 -0
- exp/epoch-50.pt +3 -0
- exp/export.sh +15 -0
- exp/pretrained.pt +3 -0
- exp/pretrained_ctc.sh +18 -0
- exp/tensorboard/events.out.tfevents.1700103483.de-74279-k2-train-1-1220091118-57c4d55446-mvd6x.1928694.0 +3 -0
- exp/train.sh +21 -0
- test_wavs/1089-134686-0001.wav +0 -0
- test_wavs/1221-135766-0001.wav +0 -0
- test_wavs/1221-135766-0002.wav +0 -0
- test_wavs/trans.txt +3 -0
README.md
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
See https://github.com/k2-fsa/icefall/pull/1389
|
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2023-11-20-11-00-12
ADDED
@@ -0,0 +1,252 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2023-11-20 11:00:12,309 INFO [ctc_decode.py:690] Decoding started
|
2 |
+
2023-11-20 11:00:12,310 INFO [ctc_decode.py:696] Device: cuda:0
|
3 |
+
2023-11-20 11:00:12,310 INFO [ctc_decode.py:697] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.clean', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '7886da9b-dirty', 'icefall-git-date': 'Tue Nov 14 20:15:43 2023', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/env/k2_icefall/lib/python3.8/site-packages/lhotse-1.17.0.dev0+git.ccfc5b2c.clean-py3.8.egg/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-9-0208143539-7dbf569d4f-r7nrb', 'IP address': '10.177.13.150'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
|
4 |
+
2023-11-20 11:00:12,759 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
|
5 |
+
2023-11-20 11:00:18,908 INFO [ctc_decode.py:780] About to create model
|
6 |
+
2023-11-20 11:00:20,576 INFO [ctc_decode.py:847] Calculating the averaged model over epoch range from 21 (excluded) to 50
|
7 |
+
2023-11-20 11:00:35,252 INFO [ctc_decode.py:864] Number of model parameters: 174319650
|
8 |
+
2023-11-20 11:00:35,253 INFO [asr_datamodule.py:465] About to get test-clean cuts
|
9 |
+
2023-11-20 11:00:35,506 INFO [asr_datamodule.py:472] About to get test-other cuts
|
10 |
+
2023-11-20 11:00:38,665 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 14
|
11 |
+
2023-11-20 11:01:45,495 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.9823, 4.1719, 4.5345, 4.4621], device='cuda:0')
|
12 |
+
2023-11-20 11:03:08,514 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.2755, 3.6060, 3.7989, 3.7019], device='cuda:0')
|
13 |
+
2023-11-20 11:03:46,736 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2298
|
14 |
+
2023-11-20 11:03:54,391 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.5861, 4.2275, 4.6746, 4.3422], device='cuda:0')
|
15 |
+
2023-11-20 11:04:10,930 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
16 |
+
2023-11-20 11:04:11,092 INFO [utils.py:641] [test-clean-attention_scale_0.01] %WER 2.31% [1215 / 52576, 121 ins, 101 del, 993 sub ]
|
17 |
+
2023-11-20 11:04:11,437 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
18 |
+
2023-11-20 11:04:11,477 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
19 |
+
2023-11-20 11:04:11,634 INFO [utils.py:641] [test-clean-attention_scale_0.05] %WER 2.26% [1190 / 52576, 118 ins, 100 del, 972 sub ]
|
20 |
+
2023-11-20 11:04:11,977 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
21 |
+
2023-11-20 11:04:12,017 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
22 |
+
2023-11-20 11:04:12,174 INFO [utils.py:641] [test-clean-attention_scale_0.08] %WER 2.26% [1186 / 52576, 119 ins, 100 del, 967 sub ]
|
23 |
+
2023-11-20 11:04:12,522 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
24 |
+
2023-11-20 11:04:12,561 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
25 |
+
2023-11-20 11:04:12,719 INFO [utils.py:641] [test-clean-attention_scale_0.1] %WER 2.25% [1184 / 52576, 120 ins, 99 del, 965 sub ]
|
26 |
+
2023-11-20 11:04:13,063 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
27 |
+
2023-11-20 11:04:13,103 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
28 |
+
2023-11-20 11:04:13,260 INFO [utils.py:641] [test-clean-attention_scale_0.3] %WER 2.19% [1151 / 52576, 114 ins, 101 del, 936 sub ]
|
29 |
+
2023-11-20 11:04:13,618 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
30 |
+
2023-11-20 11:04:13,672 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
31 |
+
2023-11-20 11:04:13,828 INFO [utils.py:641] [test-clean-attention_scale_0.5] %WER 2.16% [1134 / 52576, 114 ins, 99 del, 921 sub ]
|
32 |
+
2023-11-20 11:04:14,380 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
33 |
+
2023-11-20 11:04:14,420 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
34 |
+
2023-11-20 11:04:14,571 INFO [utils.py:641] [test-clean-attention_scale_0.6] %WER 2.15% [1132 / 52576, 113 ins, 99 del, 920 sub ]
|
35 |
+
2023-11-20 11:04:14,904 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
36 |
+
2023-11-20 11:04:14,942 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
37 |
+
2023-11-20 11:04:15,089 INFO [utils.py:641] [test-clean-attention_scale_0.7] %WER 2.15% [1130 / 52576, 113 ins, 99 del, 918 sub ]
|
38 |
+
2023-11-20 11:04:15,423 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
39 |
+
2023-11-20 11:04:15,462 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
40 |
+
2023-11-20 11:04:15,614 INFO [utils.py:641] [test-clean-attention_scale_0.9] %WER 2.13% [1120 / 52576, 114 ins, 98 del, 908 sub ]
|
41 |
+
2023-11-20 11:04:15,952 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
42 |
+
2023-11-20 11:04:15,991 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
43 |
+
2023-11-20 11:04:16,144 INFO [utils.py:641] [test-clean-attention_scale_1.0] %WER 2.12% [1115 / 52576, 114 ins, 98 del, 903 sub ]
|
44 |
+
2023-11-20 11:04:16,481 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
45 |
+
2023-11-20 11:04:16,519 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
46 |
+
2023-11-20 11:04:16,671 INFO [utils.py:641] [test-clean-attention_scale_1.1] %WER 2.13% [1119 / 52576, 116 ins, 98 del, 905 sub ]
|
47 |
+
2023-11-20 11:04:17,017 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
48 |
+
2023-11-20 11:04:17,056 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
49 |
+
2023-11-20 11:04:17,206 INFO [utils.py:641] [test-clean-attention_scale_1.2] %WER 2.14% [1123 / 52576, 116 ins, 100 del, 907 sub ]
|
50 |
+
2023-11-20 11:04:17,547 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
51 |
+
2023-11-20 11:04:17,590 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
52 |
+
2023-11-20 11:04:17,740 INFO [utils.py:641] [test-clean-attention_scale_1.3] %WER 2.13% [1119 / 52576, 116 ins, 100 del, 903 sub ]
|
53 |
+
2023-11-20 11:04:18,073 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
54 |
+
2023-11-20 11:04:18,111 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
55 |
+
2023-11-20 11:04:18,262 INFO [utils.py:641] [test-clean-attention_scale_1.5] %WER 2.12% [1113 / 52576, 117 ins, 99 del, 897 sub ]
|
56 |
+
2023-11-20 11:04:18,806 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
57 |
+
2023-11-20 11:04:18,844 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
58 |
+
2023-11-20 11:04:18,995 INFO [utils.py:641] [test-clean-attention_scale_1.7] %WER 2.12% [1112 / 52576, 118 ins, 98 del, 896 sub ]
|
59 |
+
2023-11-20 11:04:19,328 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
60 |
+
2023-11-20 11:04:19,367 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
61 |
+
2023-11-20 11:04:19,518 INFO [utils.py:641] [test-clean-attention_scale_1.9] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
|
62 |
+
2023-11-20 11:04:19,857 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
63 |
+
2023-11-20 11:04:19,895 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
64 |
+
2023-11-20 11:04:20,046 INFO [utils.py:641] [test-clean-attention_scale_2.0] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
|
65 |
+
2023-11-20 11:04:20,382 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
66 |
+
2023-11-20 11:04:20,421 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
67 |
+
2023-11-20 11:04:20,571 INFO [utils.py:641] [test-clean-attention_scale_2.1] %WER 2.11% [1111 / 52576, 119 ins, 98 del, 894 sub ]
|
68 |
+
2023-11-20 11:04:20,904 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
69 |
+
2023-11-20 11:04:20,942 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
70 |
+
2023-11-20 11:04:21,090 INFO [utils.py:641] [test-clean-attention_scale_2.2] %WER 2.12% [1112 / 52576, 119 ins, 98 del, 895 sub ]
|
71 |
+
2023-11-20 11:04:21,432 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
72 |
+
2023-11-20 11:04:21,471 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
73 |
+
2023-11-20 11:04:21,622 INFO [utils.py:641] [test-clean-attention_scale_2.3] %WER 2.12% [1115 / 52576, 120 ins, 98 del, 897 sub ]
|
74 |
+
2023-11-20 11:04:21,957 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
75 |
+
2023-11-20 11:04:21,995 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
76 |
+
2023-11-20 11:04:22,145 INFO [utils.py:641] [test-clean-attention_scale_2.5] %WER 2.12% [1113 / 52576, 119 ins, 98 del, 896 sub ]
|
77 |
+
2023-11-20 11:04:22,478 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
78 |
+
2023-11-20 11:04:22,516 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
79 |
+
2023-11-20 11:04:22,666 INFO [utils.py:641] [test-clean-attention_scale_3.0] %WER 2.11% [1107 / 52576, 117 ins, 98 del, 892 sub ]
|
80 |
+
2023-11-20 11:04:23,214 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
81 |
+
2023-11-20 11:04:23,254 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
82 |
+
2023-11-20 11:04:23,402 INFO [utils.py:641] [test-clean-attention_scale_4.0] %WER 2.11% [1107 / 52576, 118 ins, 98 del, 891 sub ]
|
83 |
+
2023-11-20 11:04:23,736 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
84 |
+
2023-11-20 11:04:23,774 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
85 |
+
2023-11-20 11:04:23,925 INFO [utils.py:641] [test-clean-attention_scale_5.0] %WER 2.10% [1105 / 52576, 118 ins, 98 del, 889 sub ]
|
86 |
+
2023-11-20 11:04:24,258 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
87 |
+
2023-11-20 11:04:24,298 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
88 |
+
2023-11-20 11:04:24,447 INFO [utils.py:641] [test-clean-attention_scale_6.0] %WER 2.10% [1105 / 52576, 117 ins, 97 del, 891 sub ]
|
89 |
+
2023-11-20 11:04:24,779 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
90 |
+
2023-11-20 11:04:24,818 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
91 |
+
2023-11-20 11:04:24,977 INFO [utils.py:641] [test-clean-attention_scale_7.0] %WER 2.10% [1103 / 52576, 117 ins, 96 del, 890 sub ]
|
92 |
+
2023-11-20 11:04:25,310 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
93 |
+
2023-11-20 11:04:25,349 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
94 |
+
2023-11-20 11:04:25,498 INFO [utils.py:641] [test-clean-attention_scale_8.0] %WER 2.09% [1101 / 52576, 116 ins, 96 del, 889 sub ]
|
95 |
+
2023-11-20 11:04:25,834 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
96 |
+
2023-11-20 11:04:25,874 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
97 |
+
2023-11-20 11:04:26,024 INFO [utils.py:641] [test-clean-attention_scale_9.0] %WER 2.09% [1101 / 52576, 116 ins, 96 del, 889 sub ]
|
98 |
+
2023-11-20 11:04:26,360 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
99 |
+
2023-11-20 11:04:26,363 INFO [ctc_decode.py:642]
|
100 |
+
For test-clean, WER of different settings are:
|
101 |
+
attention_scale_8.0 2.09 best for test-clean
|
102 |
+
attention_scale_9.0 2.09
|
103 |
+
attention_scale_5.0 2.1
|
104 |
+
attention_scale_6.0 2.1
|
105 |
+
attention_scale_7.0 2.1
|
106 |
+
attention_scale_2.1 2.11
|
107 |
+
attention_scale_3.0 2.11
|
108 |
+
attention_scale_4.0 2.11
|
109 |
+
attention_scale_1.0 2.12
|
110 |
+
attention_scale_1.5 2.12
|
111 |
+
attention_scale_1.7 2.12
|
112 |
+
attention_scale_1.9 2.12
|
113 |
+
attention_scale_2.0 2.12
|
114 |
+
attention_scale_2.2 2.12
|
115 |
+
attention_scale_2.3 2.12
|
116 |
+
attention_scale_2.5 2.12
|
117 |
+
attention_scale_0.9 2.13
|
118 |
+
attention_scale_1.1 2.13
|
119 |
+
attention_scale_1.3 2.13
|
120 |
+
attention_scale_1.2 2.14
|
121 |
+
attention_scale_0.6 2.15
|
122 |
+
attention_scale_0.7 2.15
|
123 |
+
attention_scale_0.5 2.16
|
124 |
+
attention_scale_0.3 2.19
|
125 |
+
attention_scale_0.1 2.25
|
126 |
+
attention_scale_0.05 2.26
|
127 |
+
attention_scale_0.08 2.26
|
128 |
+
attention_scale_0.01 2.31
|
129 |
+
|
130 |
+
2023-11-20 11:04:29,382 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 17
|
131 |
+
2023-11-20 11:06:22,319 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.9394, 3.1162, 3.4191, 4.0986, 4.0804, 3.4132, 3.5100, 3.5241],
|
132 |
+
device='cuda:0')
|
133 |
+
2023-11-20 11:07:23,774 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.5439, 3.3689, 3.4641, 3.4002], device='cuda:0')
|
134 |
+
2023-11-20 11:07:40,616 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2530
|
135 |
+
2023-11-20 11:07:52,519 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([1.6239, 2.6018, 2.8474, 2.9662, 2.8372, 2.7533, 2.9036, 2.7962],
|
136 |
+
device='cuda:0')
|
137 |
+
2023-11-20 11:08:03,346 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
138 |
+
2023-11-20 11:08:03,507 INFO [utils.py:641] [test-other-attention_scale_0.01] %WER 5.11% [2673 / 52343, 275 ins, 211 del, 2187 sub ]
|
139 |
+
2023-11-20 11:08:03,861 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
140 |
+
2023-11-20 11:08:03,907 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
141 |
+
2023-11-20 11:08:04,066 INFO [utils.py:641] [test-other-attention_scale_0.05] %WER 5.02% [2627 / 52343, 269 ins, 206 del, 2152 sub ]
|
142 |
+
2023-11-20 11:08:04,423 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
143 |
+
2023-11-20 11:08:04,469 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
144 |
+
2023-11-20 11:08:04,633 INFO [utils.py:641] [test-other-attention_scale_0.08] %WER 4.98% [2607 / 52343, 264 ins, 206 del, 2137 sub ]
|
145 |
+
2023-11-20 11:08:05,283 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
146 |
+
2023-11-20 11:08:05,322 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
147 |
+
2023-11-20 11:08:05,479 INFO [utils.py:641] [test-other-attention_scale_0.1] %WER 4.96% [2598 / 52343, 265 ins, 204 del, 2129 sub ]
|
148 |
+
2023-11-20 11:08:05,829 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
149 |
+
2023-11-20 11:08:05,874 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
150 |
+
2023-11-20 11:08:06,028 INFO [utils.py:641] [test-other-attention_scale_0.3] %WER 4.83% [2529 / 52343, 263 ins, 193 del, 2073 sub ]
|
151 |
+
2023-11-20 11:08:06,376 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
152 |
+
2023-11-20 11:08:06,418 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
153 |
+
2023-11-20 11:08:06,573 INFO [utils.py:641] [test-other-attention_scale_0.5] %WER 4.79% [2507 / 52343, 265 ins, 189 del, 2053 sub ]
|
154 |
+
2023-11-20 11:08:06,922 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
155 |
+
2023-11-20 11:08:06,964 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
156 |
+
2023-11-20 11:08:07,119 INFO [utils.py:641] [test-other-attention_scale_0.6] %WER 4.77% [2497 / 52343, 265 ins, 187 del, 2045 sub ]
|
157 |
+
2023-11-20 11:08:07,466 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
158 |
+
2023-11-20 11:08:07,505 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
159 |
+
2023-11-20 11:08:07,663 INFO [utils.py:641] [test-other-attention_scale_0.7] %WER 4.77% [2496 / 52343, 262 ins, 187 del, 2047 sub ]
|
160 |
+
2023-11-20 11:08:08,010 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
161 |
+
2023-11-20 11:08:08,052 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
162 |
+
2023-11-20 11:08:08,207 INFO [utils.py:641] [test-other-attention_scale_0.9] %WER 4.75% [2485 / 52343, 259 ins, 185 del, 2041 sub ]
|
163 |
+
2023-11-20 11:08:08,555 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
164 |
+
2023-11-20 11:08:08,602 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
165 |
+
2023-11-20 11:08:08,757 INFO [utils.py:641] [test-other-attention_scale_1.0] %WER 4.73% [2475 / 52343, 257 ins, 183 del, 2035 sub ]
|
166 |
+
2023-11-20 11:08:09,104 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
167 |
+
2023-11-20 11:08:09,144 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
168 |
+
2023-11-20 11:08:09,533 INFO [utils.py:641] [test-other-attention_scale_1.1] %WER 4.72% [2469 / 52343, 258 ins, 182 del, 2029 sub ]
|
169 |
+
2023-11-20 11:08:09,880 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
170 |
+
2023-11-20 11:08:09,920 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
171 |
+
2023-11-20 11:08:10,072 INFO [utils.py:641] [test-other-attention_scale_1.2] %WER 4.71% [2463 / 52343, 258 ins, 181 del, 2024 sub ]
|
172 |
+
2023-11-20 11:08:10,411 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
173 |
+
2023-11-20 11:08:10,453 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
174 |
+
2023-11-20 11:08:10,607 INFO [utils.py:641] [test-other-attention_scale_1.3] %WER 4.70% [2458 / 52343, 258 ins, 179 del, 2021 sub ]
|
175 |
+
2023-11-20 11:08:10,954 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
176 |
+
2023-11-20 11:08:10,995 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
177 |
+
2023-11-20 11:08:11,152 INFO [utils.py:641] [test-other-attention_scale_1.5] %WER 4.67% [2445 / 52343, 256 ins, 181 del, 2008 sub ]
|
178 |
+
2023-11-20 11:08:11,502 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
179 |
+
2023-11-20 11:08:11,541 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
180 |
+
2023-11-20 11:08:11,698 INFO [utils.py:641] [test-other-attention_scale_1.7] %WER 4.67% [2443 / 52343, 255 ins, 182 del, 2006 sub ]
|
181 |
+
2023-11-20 11:08:12,043 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
182 |
+
2023-11-20 11:08:12,083 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
183 |
+
2023-11-20 11:08:12,235 INFO [utils.py:641] [test-other-attention_scale_1.9] %WER 4.66% [2440 / 52343, 256 ins, 181 del, 2003 sub ]
|
184 |
+
2023-11-20 11:08:12,594 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
185 |
+
2023-11-20 11:08:12,637 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
186 |
+
2023-11-20 11:08:12,791 INFO [utils.py:641] [test-other-attention_scale_2.0] %WER 4.66% [2437 / 52343, 254 ins, 179 del, 2004 sub ]
|
187 |
+
2023-11-20 11:08:13,136 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
188 |
+
2023-11-20 11:08:13,177 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
189 |
+
2023-11-20 11:08:13,331 INFO [utils.py:641] [test-other-attention_scale_2.1] %WER 4.66% [2438 / 52343, 255 ins, 180 del, 2003 sub ]
|
190 |
+
2023-11-20 11:08:13,948 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
191 |
+
2023-11-20 11:08:13,990 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
192 |
+
2023-11-20 11:08:14,142 INFO [utils.py:641] [test-other-attention_scale_2.2] %WER 4.66% [2437 / 52343, 255 ins, 180 del, 2002 sub ]
|
193 |
+
2023-11-20 11:08:14,487 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
194 |
+
2023-11-20 11:08:14,527 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
195 |
+
2023-11-20 11:08:14,683 INFO [utils.py:641] [test-other-attention_scale_2.3] %WER 4.65% [2433 / 52343, 253 ins, 180 del, 2000 sub ]
|
196 |
+
2023-11-20 11:08:15,029 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
197 |
+
2023-11-20 11:08:15,069 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
198 |
+
2023-11-20 11:08:15,221 INFO [utils.py:641] [test-other-attention_scale_2.5] %WER 4.64% [2430 / 52343, 251 ins, 179 del, 2000 sub ]
|
199 |
+
2023-11-20 11:08:15,567 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
200 |
+
2023-11-20 11:08:15,608 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
201 |
+
2023-11-20 11:08:15,761 INFO [utils.py:641] [test-other-attention_scale_3.0] %WER 4.63% [2422 / 52343, 253 ins, 176 del, 1993 sub ]
|
202 |
+
2023-11-20 11:08:16,115 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
203 |
+
2023-11-20 11:08:16,155 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
204 |
+
2023-11-20 11:08:16,308 INFO [utils.py:641] [test-other-attention_scale_4.0] %WER 4.62% [2416 / 52343, 253 ins, 176 del, 1987 sub ]
|
205 |
+
2023-11-20 11:08:16,658 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
206 |
+
2023-11-20 11:08:16,698 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
207 |
+
2023-11-20 11:08:16,852 INFO [utils.py:641] [test-other-attention_scale_5.0] %WER 4.62% [2418 / 52343, 254 ins, 177 del, 1987 sub ]
|
208 |
+
2023-11-20 11:08:17,199 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
209 |
+
2023-11-20 11:08:17,240 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
210 |
+
2023-11-20 11:08:17,395 INFO [utils.py:641] [test-other-attention_scale_6.0] %WER 4.62% [2418 / 52343, 254 ins, 178 del, 1986 sub ]
|
211 |
+
2023-11-20 11:08:17,745 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
212 |
+
2023-11-20 11:08:17,784 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
213 |
+
2023-11-20 11:08:17,939 INFO [utils.py:641] [test-other-attention_scale_7.0] %WER 4.61% [2412 / 52343, 255 ins, 176 del, 1981 sub ]
|
214 |
+
2023-11-20 11:08:18,284 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
215 |
+
2023-11-20 11:08:18,324 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
216 |
+
2023-11-20 11:08:18,725 INFO [utils.py:641] [test-other-attention_scale_8.0] %WER 4.60% [2407 / 52343, 255 ins, 175 del, 1977 sub ]
|
217 |
+
2023-11-20 11:08:19,071 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
218 |
+
2023-11-20 11:08:19,111 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
219 |
+
2023-11-20 11:08:19,267 INFO [utils.py:641] [test-other-attention_scale_9.0] %WER 4.59% [2405 / 52343, 255 ins, 175 del, 1975 sub ]
|
220 |
+
2023-11-20 11:08:19,612 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
221 |
+
2023-11-20 11:08:19,615 INFO [ctc_decode.py:642]
|
222 |
+
For test-other, WER of different settings are:
|
223 |
+
attention_scale_9.0 4.59 best for test-other
|
224 |
+
attention_scale_8.0 4.6
|
225 |
+
attention_scale_7.0 4.61
|
226 |
+
attention_scale_4.0 4.62
|
227 |
+
attention_scale_5.0 4.62
|
228 |
+
attention_scale_6.0 4.62
|
229 |
+
attention_scale_3.0 4.63
|
230 |
+
attention_scale_2.5 4.64
|
231 |
+
attention_scale_2.3 4.65
|
232 |
+
attention_scale_1.9 4.66
|
233 |
+
attention_scale_2.0 4.66
|
234 |
+
attention_scale_2.1 4.66
|
235 |
+
attention_scale_2.2 4.66
|
236 |
+
attention_scale_1.5 4.67
|
237 |
+
attention_scale_1.7 4.67
|
238 |
+
attention_scale_1.3 4.7
|
239 |
+
attention_scale_1.2 4.71
|
240 |
+
attention_scale_1.1 4.72
|
241 |
+
attention_scale_1.0 4.73
|
242 |
+
attention_scale_0.9 4.75
|
243 |
+
attention_scale_0.6 4.77
|
244 |
+
attention_scale_0.7 4.77
|
245 |
+
attention_scale_0.5 4.79
|
246 |
+
attention_scale_0.3 4.83
|
247 |
+
attention_scale_0.1 4.96
|
248 |
+
attention_scale_0.08 4.98
|
249 |
+
attention_scale_0.05 5.02
|
250 |
+
attention_scale_0.01 5.11
|
251 |
+
|
252 |
+
2023-11-20 11:08:19,616 INFO [ctc_decode.py:897] Done!
|
decoding_results/attention-decoder-rescoring-no-ngram/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-24-10-49-48
ADDED
@@ -0,0 +1,247 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-05-24 10:49:48,302 INFO [ctc_decode.py:690] Decoding started
|
2 |
+
2024-05-24 10:49:48,302 INFO [ctc_decode.py:696] Device: cuda:0
|
3 |
+
2024-05-24 10:49:48,302 INFO [ctc_decode.py:697] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '0be32f3d-clean', 'icefall-git-date': 'Fri Feb 23 15:21:01 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-10-0312151423-668f59dc99-b8962', 'IP address': '10.177.6.147'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'attention-decoder-rescoring-no-ngram', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
|
4 |
+
2024-05-24 10:49:48,578 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
|
5 |
+
2024-05-24 10:49:53,648 INFO [ctc_decode.py:780] About to create model
|
6 |
+
2024-05-24 10:49:54,840 INFO [ctc_decode.py:847] Calculating the averaged model over epoch range from 21 (excluded) to 50
|
7 |
+
2024-05-24 10:50:10,087 INFO [ctc_decode.py:864] Number of model parameters: 174319650
|
8 |
+
2024-05-24 10:50:10,087 INFO [asr_datamodule.py:465] About to get test-clean cuts
|
9 |
+
2024-05-24 10:50:10,239 INFO [asr_datamodule.py:472] About to get test-other cuts
|
10 |
+
2024-05-24 10:50:12,872 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 14
|
11 |
+
2024-05-24 10:50:52,024 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.9550, 4.9446, 4.5893, 4.1813], device='cuda:0')
|
12 |
+
2024-05-24 10:53:16,396 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2298
|
13 |
+
2024-05-24 10:53:39,601 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
14 |
+
2024-05-24 10:53:39,698 INFO [utils.py:641] [test-clean-attention_scale_0.01] %WER 2.31% [1215 / 52576, 122 ins, 101 del, 992 sub ]
|
15 |
+
2024-05-24 10:53:39,895 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
16 |
+
2024-05-24 10:53:39,923 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
17 |
+
2024-05-24 10:53:40,013 INFO [utils.py:641] [test-clean-attention_scale_0.05] %WER 2.26% [1189 / 52576, 119 ins, 101 del, 969 sub ]
|
18 |
+
2024-05-24 10:53:40,205 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
19 |
+
2024-05-24 10:53:40,231 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
20 |
+
2024-05-24 10:53:40,321 INFO [utils.py:641] [test-clean-attention_scale_0.08] %WER 2.26% [1186 / 52576, 120 ins, 100 del, 966 sub ]
|
21 |
+
2024-05-24 10:53:40,513 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
22 |
+
2024-05-24 10:53:40,540 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
23 |
+
2024-05-24 10:53:40,638 INFO [utils.py:641] [test-clean-attention_scale_0.1] %WER 2.25% [1183 / 52576, 121 ins, 99 del, 963 sub ]
|
24 |
+
2024-05-24 10:53:40,833 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
25 |
+
2024-05-24 10:53:40,861 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
26 |
+
2024-05-24 10:53:40,950 INFO [utils.py:641] [test-clean-attention_scale_0.3] %WER 2.19% [1150 / 52576, 116 ins, 99 del, 935 sub ]
|
27 |
+
2024-05-24 10:53:41,145 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
28 |
+
2024-05-24 10:53:41,172 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
29 |
+
2024-05-24 10:53:41,260 INFO [utils.py:641] [test-clean-attention_scale_0.5] %WER 2.15% [1132 / 52576, 115 ins, 99 del, 918 sub ]
|
30 |
+
2024-05-24 10:53:41,650 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
31 |
+
2024-05-24 10:53:41,677 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
32 |
+
2024-05-24 10:53:41,767 INFO [utils.py:641] [test-clean-attention_scale_0.6] %WER 2.15% [1131 / 52576, 114 ins, 99 del, 918 sub ]
|
33 |
+
2024-05-24 10:53:41,966 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
34 |
+
2024-05-24 10:53:41,992 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
35 |
+
2024-05-24 10:53:42,082 INFO [utils.py:641] [test-clean-attention_scale_0.7] %WER 2.14% [1127 / 52576, 114 ins, 99 del, 914 sub ]
|
36 |
+
2024-05-24 10:53:42,274 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
37 |
+
2024-05-24 10:53:42,301 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
38 |
+
2024-05-24 10:53:42,390 INFO [utils.py:641] [test-clean-attention_scale_0.9] %WER 2.13% [1118 / 52576, 114 ins, 98 del, 906 sub ]
|
39 |
+
2024-05-24 10:53:42,581 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
40 |
+
2024-05-24 10:53:42,609 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
41 |
+
2024-05-24 10:53:42,698 INFO [utils.py:641] [test-clean-attention_scale_1.0] %WER 2.11% [1110 / 52576, 112 ins, 98 del, 900 sub ]
|
42 |
+
2024-05-24 10:53:42,888 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
43 |
+
2024-05-24 10:53:42,914 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
44 |
+
2024-05-24 10:53:43,002 INFO [utils.py:641] [test-clean-attention_scale_1.1] %WER 2.12% [1114 / 52576, 114 ins, 98 del, 902 sub ]
|
45 |
+
2024-05-24 10:53:43,200 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
46 |
+
2024-05-24 10:53:43,262 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
47 |
+
2024-05-24 10:53:43,353 INFO [utils.py:641] [test-clean-attention_scale_1.2] %WER 2.12% [1116 / 52576, 114 ins, 100 del, 902 sub ]
|
48 |
+
2024-05-24 10:53:43,547 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
49 |
+
2024-05-24 10:53:43,576 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
50 |
+
2024-05-24 10:53:43,666 INFO [utils.py:641] [test-clean-attention_scale_1.3] %WER 2.12% [1114 / 52576, 114 ins, 99 del, 901 sub ]
|
51 |
+
2024-05-24 10:53:43,856 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
52 |
+
2024-05-24 10:53:43,881 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
53 |
+
2024-05-24 10:53:43,970 INFO [utils.py:641] [test-clean-attention_scale_1.5] %WER 2.11% [1108 / 52576, 117 ins, 98 del, 893 sub ]
|
54 |
+
2024-05-24 10:53:44,384 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
55 |
+
2024-05-24 10:53:44,412 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
56 |
+
2024-05-24 10:53:44,503 INFO [utils.py:641] [test-clean-attention_scale_1.7] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
|
57 |
+
2024-05-24 10:53:44,695 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
58 |
+
2024-05-24 10:53:44,721 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
59 |
+
2024-05-24 10:53:44,811 INFO [utils.py:641] [test-clean-attention_scale_1.9] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
|
60 |
+
2024-05-24 10:53:45,005 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
61 |
+
2024-05-24 10:53:45,073 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
62 |
+
2024-05-24 10:53:45,167 INFO [utils.py:641] [test-clean-attention_scale_2.0] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
|
63 |
+
2024-05-24 10:53:45,359 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
64 |
+
2024-05-24 10:53:45,386 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
65 |
+
2024-05-24 10:53:45,475 INFO [utils.py:641] [test-clean-attention_scale_2.1] %WER 2.11% [1107 / 52576, 118 ins, 97 del, 892 sub ]
|
66 |
+
2024-05-24 10:53:45,675 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
67 |
+
2024-05-24 10:53:45,702 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
68 |
+
2024-05-24 10:53:45,792 INFO [utils.py:641] [test-clean-attention_scale_2.2] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
|
69 |
+
2024-05-24 10:53:45,984 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
70 |
+
2024-05-24 10:53:46,011 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
71 |
+
2024-05-24 10:53:46,101 INFO [utils.py:641] [test-clean-attention_scale_2.3] %WER 2.11% [1111 / 52576, 119 ins, 97 del, 895 sub ]
|
72 |
+
2024-05-24 10:53:46,293 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
73 |
+
2024-05-24 10:53:46,339 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
74 |
+
2024-05-24 10:53:46,428 INFO [utils.py:641] [test-clean-attention_scale_2.5] %WER 2.11% [1108 / 52576, 118 ins, 97 del, 893 sub ]
|
75 |
+
2024-05-24 10:53:46,619 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
76 |
+
2024-05-24 10:53:46,646 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
77 |
+
2024-05-24 10:53:46,734 INFO [utils.py:641] [test-clean-attention_scale_3.0] %WER 2.10% [1105 / 52576, 116 ins, 98 del, 891 sub ]
|
78 |
+
2024-05-24 10:53:46,965 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
79 |
+
2024-05-24 10:53:46,994 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
80 |
+
2024-05-24 10:53:47,098 INFO [utils.py:641] [test-clean-attention_scale_4.0] %WER 2.10% [1105 / 52576, 118 ins, 98 del, 889 sub ]
|
81 |
+
2024-05-24 10:53:47,494 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
82 |
+
2024-05-24 10:53:47,549 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
83 |
+
2024-05-24 10:53:47,638 INFO [utils.py:641] [test-clean-attention_scale_5.0] %WER 2.10% [1106 / 52576, 118 ins, 98 del, 890 sub ]
|
84 |
+
2024-05-24 10:53:47,830 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
85 |
+
2024-05-24 10:53:47,857 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
86 |
+
2024-05-24 10:53:47,945 INFO [utils.py:641] [test-clean-attention_scale_6.0] %WER 2.10% [1105 / 52576, 117 ins, 97 del, 891 sub ]
|
87 |
+
2024-05-24 10:53:48,161 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
88 |
+
2024-05-24 10:53:48,189 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
89 |
+
2024-05-24 10:53:48,283 INFO [utils.py:641] [test-clean-attention_scale_7.0] %WER 2.10% [1106 / 52576, 117 ins, 97 del, 892 sub ]
|
90 |
+
2024-05-24 10:53:48,477 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
91 |
+
2024-05-24 10:53:48,504 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
92 |
+
2024-05-24 10:53:48,594 INFO [utils.py:641] [test-clean-attention_scale_8.0] %WER 2.10% [1105 / 52576, 116 ins, 97 del, 892 sub ]
|
93 |
+
2024-05-24 10:53:48,785 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
94 |
+
2024-05-24 10:53:48,811 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
95 |
+
2024-05-24 10:53:48,900 INFO [utils.py:641] [test-clean-attention_scale_9.0] %WER 2.10% [1105 / 52576, 116 ins, 97 del, 892 sub ]
|
96 |
+
2024-05-24 10:53:49,098 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
97 |
+
2024-05-24 10:53:49,101 INFO [ctc_decode.py:642]
|
98 |
+
For test-clean, WER of different settings are:
|
99 |
+
attention_scale_3.0 2.1 best for test-clean
|
100 |
+
attention_scale_4.0 2.1
|
101 |
+
attention_scale_5.0 2.1
|
102 |
+
attention_scale_6.0 2.1
|
103 |
+
attention_scale_7.0 2.1
|
104 |
+
attention_scale_8.0 2.1
|
105 |
+
attention_scale_9.0 2.1
|
106 |
+
attention_scale_1.0 2.11
|
107 |
+
attention_scale_1.5 2.11
|
108 |
+
attention_scale_1.7 2.11
|
109 |
+
attention_scale_1.9 2.11
|
110 |
+
attention_scale_2.0 2.11
|
111 |
+
attention_scale_2.1 2.11
|
112 |
+
attention_scale_2.2 2.11
|
113 |
+
attention_scale_2.3 2.11
|
114 |
+
attention_scale_2.5 2.11
|
115 |
+
attention_scale_1.1 2.12
|
116 |
+
attention_scale_1.2 2.12
|
117 |
+
attention_scale_1.3 2.12
|
118 |
+
attention_scale_0.9 2.13
|
119 |
+
attention_scale_0.7 2.14
|
120 |
+
attention_scale_0.5 2.15
|
121 |
+
attention_scale_0.6 2.15
|
122 |
+
attention_scale_0.3 2.19
|
123 |
+
attention_scale_0.1 2.25
|
124 |
+
attention_scale_0.05 2.26
|
125 |
+
attention_scale_0.08 2.26
|
126 |
+
attention_scale_0.01 2.31
|
127 |
+
|
128 |
+
2024-05-24 10:53:51,796 INFO [ctc_decode.py:596] batch 0/?, cuts processed until now is 17
|
129 |
+
2024-05-24 10:54:02,980 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([4.6454, 2.4206, 3.8113, 4.5973], device='cuda:0')
|
130 |
+
2024-05-24 10:55:02,806 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([3.3866, 3.0863, 3.1216, 3.0326], device='cuda:0')
|
131 |
+
2024-05-24 10:56:59,598 INFO [ctc_decode.py:596] batch 100/?, cuts processed until now is 2530
|
132 |
+
2024-05-24 10:57:21,501 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
133 |
+
2024-05-24 10:57:21,600 INFO [utils.py:641] [test-other-attention_scale_0.01] %WER 5.12% [2678 / 52343, 277 ins, 212 del, 2189 sub ]
|
134 |
+
2024-05-24 10:57:21,801 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
135 |
+
2024-05-24 10:57:21,836 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
136 |
+
2024-05-24 10:57:21,930 INFO [utils.py:641] [test-other-attention_scale_0.05] %WER 5.02% [2630 / 52343, 270 ins, 206 del, 2154 sub ]
|
137 |
+
2024-05-24 10:57:22,396 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
138 |
+
2024-05-24 10:57:22,426 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
139 |
+
2024-05-24 10:57:22,519 INFO [utils.py:641] [test-other-attention_scale_0.08] %WER 4.98% [2606 / 52343, 264 ins, 206 del, 2136 sub ]
|
140 |
+
2024-05-24 10:57:22,719 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
141 |
+
2024-05-24 10:57:22,747 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
142 |
+
2024-05-24 10:57:22,839 INFO [utils.py:641] [test-other-attention_scale_0.1] %WER 4.96% [2597 / 52343, 265 ins, 204 del, 2128 sub ]
|
143 |
+
2024-05-24 10:57:23,039 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
144 |
+
2024-05-24 10:57:23,066 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
145 |
+
2024-05-24 10:57:23,158 INFO [utils.py:641] [test-other-attention_scale_0.3] %WER 4.83% [2527 / 52343, 261 ins, 194 del, 2072 sub ]
|
146 |
+
2024-05-24 10:57:23,354 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
147 |
+
2024-05-24 10:57:23,385 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
148 |
+
2024-05-24 10:57:23,485 INFO [utils.py:641] [test-other-attention_scale_0.5] %WER 4.79% [2507 / 52343, 262 ins, 190 del, 2055 sub ]
|
149 |
+
2024-05-24 10:57:23,686 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
150 |
+
2024-05-24 10:57:23,714 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
151 |
+
2024-05-24 10:57:23,808 INFO [utils.py:641] [test-other-attention_scale_0.6] %WER 4.76% [2493 / 52343, 262 ins, 187 del, 2044 sub ]
|
152 |
+
2024-05-24 10:57:24,010 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
153 |
+
2024-05-24 10:57:24,036 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
154 |
+
2024-05-24 10:57:24,128 INFO [utils.py:641] [test-other-attention_scale_0.7] %WER 4.76% [2491 / 52343, 260 ins, 186 del, 2045 sub ]
|
155 |
+
2024-05-24 10:57:24,326 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
156 |
+
2024-05-24 10:57:24,353 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
157 |
+
2024-05-24 10:57:24,444 INFO [utils.py:641] [test-other-attention_scale_0.9] %WER 4.74% [2479 / 52343, 258 ins, 184 del, 2037 sub ]
|
158 |
+
2024-05-24 10:57:24,643 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
159 |
+
2024-05-24 10:57:24,672 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
160 |
+
2024-05-24 10:57:24,769 INFO [utils.py:641] [test-other-attention_scale_1.0] %WER 4.72% [2468 / 52343, 256 ins, 183 del, 2029 sub ]
|
161 |
+
2024-05-24 10:57:25,183 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
162 |
+
2024-05-24 10:57:25,215 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
163 |
+
2024-05-24 10:57:25,307 INFO [utils.py:641] [test-other-attention_scale_1.1] %WER 4.70% [2462 / 52343, 257 ins, 182 del, 2023 sub ]
|
164 |
+
2024-05-24 10:57:25,504 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
165 |
+
2024-05-24 10:57:25,559 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
166 |
+
2024-05-24 10:57:25,651 INFO [utils.py:641] [test-other-attention_scale_1.2] %WER 4.70% [2458 / 52343, 257 ins, 182 del, 2019 sub ]
|
167 |
+
2024-05-24 10:57:25,875 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
168 |
+
2024-05-24 10:57:25,936 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
169 |
+
2024-05-24 10:57:26,030 INFO [utils.py:641] [test-other-attention_scale_1.3] %WER 4.69% [2454 / 52343, 257 ins, 180 del, 2017 sub ]
|
170 |
+
2024-05-24 10:57:26,228 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
171 |
+
2024-05-24 10:57:26,257 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
172 |
+
2024-05-24 10:57:26,355 INFO [utils.py:641] [test-other-attention_scale_1.5] %WER 4.66% [2440 / 52343, 254 ins, 181 del, 2005 sub ]
|
173 |
+
2024-05-24 10:57:26,562 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
174 |
+
2024-05-24 10:57:26,589 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
175 |
+
2024-05-24 10:57:26,681 INFO [utils.py:641] [test-other-attention_scale_1.7] %WER 4.66% [2438 / 52343, 254 ins, 182 del, 2002 sub ]
|
176 |
+
2024-05-24 10:57:26,893 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
177 |
+
2024-05-24 10:57:26,921 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
178 |
+
2024-05-24 10:57:27,012 INFO [utils.py:641] [test-other-attention_scale_1.9] %WER 4.65% [2436 / 52343, 255 ins, 180 del, 2001 sub ]
|
179 |
+
2024-05-24 10:57:27,219 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
180 |
+
2024-05-24 10:57:27,247 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
181 |
+
2024-05-24 10:57:27,339 INFO [utils.py:641] [test-other-attention_scale_2.0] %WER 4.65% [2433 / 52343, 253 ins, 178 del, 2002 sub ]
|
182 |
+
2024-05-24 10:57:27,543 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
183 |
+
2024-05-24 10:57:27,573 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
184 |
+
2024-05-24 10:57:27,673 INFO [utils.py:641] [test-other-attention_scale_2.1] %WER 4.65% [2436 / 52343, 255 ins, 179 del, 2002 sub ]
|
185 |
+
2024-05-24 10:57:27,873 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
186 |
+
2024-05-24 10:57:27,900 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
187 |
+
2024-05-24 10:57:27,992 INFO [utils.py:641] [test-other-attention_scale_2.2] %WER 4.65% [2435 / 52343, 255 ins, 179 del, 2001 sub ]
|
188 |
+
2024-05-24 10:57:28,396 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
189 |
+
2024-05-24 10:57:28,426 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
190 |
+
2024-05-24 10:57:28,517 INFO [utils.py:641] [test-other-attention_scale_2.3] %WER 4.64% [2429 / 52343, 253 ins, 179 del, 1997 sub ]
|
191 |
+
2024-05-24 10:57:28,712 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
192 |
+
2024-05-24 10:57:28,739 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
193 |
+
2024-05-24 10:57:28,834 INFO [utils.py:641] [test-other-attention_scale_2.5] %WER 4.64% [2428 / 52343, 252 ins, 178 del, 1998 sub ]
|
194 |
+
2024-05-24 10:57:29,038 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
195 |
+
2024-05-24 10:57:29,100 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
196 |
+
2024-05-24 10:57:29,192 INFO [utils.py:641] [test-other-attention_scale_3.0] %WER 4.62% [2417 / 52343, 252 ins, 173 del, 1992 sub ]
|
197 |
+
2024-05-24 10:57:29,392 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
198 |
+
2024-05-24 10:57:29,419 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
199 |
+
2024-05-24 10:57:29,513 INFO [utils.py:641] [test-other-attention_scale_4.0] %WER 4.60% [2407 / 52343, 251 ins, 173 del, 1983 sub ]
|
200 |
+
2024-05-24 10:57:29,710 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
201 |
+
2024-05-24 10:57:29,737 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
202 |
+
2024-05-24 10:57:29,829 INFO [utils.py:641] [test-other-attention_scale_5.0] %WER 4.60% [2409 / 52343, 252 ins, 174 del, 1983 sub ]
|
203 |
+
2024-05-24 10:57:30,024 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
204 |
+
2024-05-24 10:57:30,051 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
205 |
+
2024-05-24 10:57:30,148 INFO [utils.py:641] [test-other-attention_scale_6.0] %WER 4.59% [2405 / 52343, 252 ins, 174 del, 1979 sub ]
|
206 |
+
2024-05-24 10:57:30,352 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
207 |
+
2024-05-24 10:57:30,379 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
208 |
+
2024-05-24 10:57:30,472 INFO [utils.py:641] [test-other-attention_scale_7.0] %WER 4.59% [2400 / 52343, 253 ins, 172 del, 1975 sub ]
|
209 |
+
2024-05-24 10:57:30,671 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
210 |
+
2024-05-24 10:57:30,699 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
211 |
+
2024-05-24 10:57:30,791 INFO [utils.py:641] [test-other-attention_scale_8.0] %WER 4.58% [2396 / 52343, 253 ins, 171 del, 1972 sub ]
|
212 |
+
2024-05-24 10:57:31,193 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
213 |
+
2024-05-24 10:57:31,219 INFO [ctc_decode.py:619] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
214 |
+
2024-05-24 10:57:31,311 INFO [utils.py:641] [test-other-attention_scale_9.0] %WER 4.57% [2394 / 52343, 253 ins, 171 del, 1970 sub ]
|
215 |
+
2024-05-24 10:57:31,514 INFO [ctc_decode.py:628] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/attention-decoder-rescoring-no-ngram/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
216 |
+
2024-05-24 10:57:31,517 INFO [ctc_decode.py:642]
|
217 |
+
For test-other, WER of different settings are:
|
218 |
+
attention_scale_9.0 4.57 best for test-other
|
219 |
+
attention_scale_8.0 4.58
|
220 |
+
attention_scale_6.0 4.59
|
221 |
+
attention_scale_7.0 4.59
|
222 |
+
attention_scale_4.0 4.6
|
223 |
+
attention_scale_5.0 4.6
|
224 |
+
attention_scale_3.0 4.62
|
225 |
+
attention_scale_2.3 4.64
|
226 |
+
attention_scale_2.5 4.64
|
227 |
+
attention_scale_1.9 4.65
|
228 |
+
attention_scale_2.0 4.65
|
229 |
+
attention_scale_2.1 4.65
|
230 |
+
attention_scale_2.2 4.65
|
231 |
+
attention_scale_1.5 4.66
|
232 |
+
attention_scale_1.7 4.66
|
233 |
+
attention_scale_1.3 4.69
|
234 |
+
attention_scale_1.1 4.7
|
235 |
+
attention_scale_1.2 4.7
|
236 |
+
attention_scale_1.0 4.72
|
237 |
+
attention_scale_0.9 4.74
|
238 |
+
attention_scale_0.6 4.76
|
239 |
+
attention_scale_0.7 4.76
|
240 |
+
attention_scale_0.5 4.79
|
241 |
+
attention_scale_0.3 4.83
|
242 |
+
attention_scale_0.1 4.96
|
243 |
+
attention_scale_0.08 4.98
|
244 |
+
attention_scale_0.05 5.02
|
245 |
+
attention_scale_0.01 5.12
|
246 |
+
|
247 |
+
2024-05-24 10:57:31,517 INFO [ctc_decode.py:897] Done!
|
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/attention-decoder-rescoring-no-ngram/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
settings WER
|
2 |
+
attention_scale_3.0 2.1
|
3 |
+
attention_scale_4.0 2.1
|
4 |
+
attention_scale_5.0 2.1
|
5 |
+
attention_scale_6.0 2.1
|
6 |
+
attention_scale_7.0 2.1
|
7 |
+
attention_scale_8.0 2.1
|
8 |
+
attention_scale_9.0 2.1
|
9 |
+
attention_scale_1.0 2.11
|
10 |
+
attention_scale_1.5 2.11
|
11 |
+
attention_scale_1.7 2.11
|
12 |
+
attention_scale_1.9 2.11
|
13 |
+
attention_scale_2.0 2.11
|
14 |
+
attention_scale_2.1 2.11
|
15 |
+
attention_scale_2.2 2.11
|
16 |
+
attention_scale_2.3 2.11
|
17 |
+
attention_scale_2.5 2.11
|
18 |
+
attention_scale_1.1 2.12
|
19 |
+
attention_scale_1.2 2.12
|
20 |
+
attention_scale_1.3 2.12
|
21 |
+
attention_scale_0.9 2.13
|
22 |
+
attention_scale_0.7 2.14
|
23 |
+
attention_scale_0.5 2.15
|
24 |
+
attention_scale_0.6 2.15
|
25 |
+
attention_scale_0.3 2.19
|
26 |
+
attention_scale_0.1 2.25
|
27 |
+
attention_scale_0.05 2.26
|
28 |
+
attention_scale_0.08 2.26
|
29 |
+
attention_scale_0.01 2.31
|
decoding_results/attention-decoder-rescoring-no-ngram/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
settings WER
|
2 |
+
attention_scale_9.0 4.57
|
3 |
+
attention_scale_8.0 4.58
|
4 |
+
attention_scale_6.0 4.59
|
5 |
+
attention_scale_7.0 4.59
|
6 |
+
attention_scale_4.0 4.6
|
7 |
+
attention_scale_5.0 4.6
|
8 |
+
attention_scale_3.0 4.62
|
9 |
+
attention_scale_2.3 4.64
|
10 |
+
attention_scale_2.5 4.64
|
11 |
+
attention_scale_1.9 4.65
|
12 |
+
attention_scale_2.0 4.65
|
13 |
+
attention_scale_2.1 4.65
|
14 |
+
attention_scale_2.2 4.65
|
15 |
+
attention_scale_1.5 4.66
|
16 |
+
attention_scale_1.7 4.66
|
17 |
+
attention_scale_1.3 4.69
|
18 |
+
attention_scale_1.1 4.7
|
19 |
+
attention_scale_1.2 4.7
|
20 |
+
attention_scale_1.0 4.72
|
21 |
+
attention_scale_0.9 4.74
|
22 |
+
attention_scale_0.6 4.76
|
23 |
+
attention_scale_0.7 4.76
|
24 |
+
attention_scale_0.5 4.79
|
25 |
+
attention_scale_0.3 4.83
|
26 |
+
attention_scale_0.1 4.96
|
27 |
+
attention_scale_0.08 4.98
|
28 |
+
attention_scale_0.05 5.02
|
29 |
+
attention_scale_0.01 5.12
|
decoding_results/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/ctc-decoding/log-decode-epoch-50-avg-29-use-averaged-model-2024-05-26-11-20-47
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-05-26 11:20:47,975 INFO [ctc_decode.py:717] Decoding started
|
2 |
+
2024-05-26 11:20:47,975 INFO [ctc_decode.py:723] Device: cuda:0
|
3 |
+
2024-05-26 11:20:47,975 INFO [ctc_decode.py:724] {'best_train_loss': inf, 'best_valid_loss': inf, 'best_train_epoch': -1, 'best_valid_epoch': -1, 'batch_idx_train': 0, 'log_interval': 50, 'reset_interval': 200, 'valid_interval': 3000, 'feature_dim': 80, 'subsampling_factor': 4, 'ignore_id': -1, 'label_smoothing': 0.1, 'warm_step': 2000, 'env_info': {'k2-version': '1.24.4', 'k2-build-type': 'Release', 'k2-with-cuda': True, 'k2-git-sha1': '44a9d5682af9fd3ef77074777e15278ec6d390eb', 'k2-git-date': 'Wed Sep 27 11:22:55 2023', 'lhotse-version': '1.17.0.dev+git.ccfc5b2c.dirty', 'torch-version': '1.10.0+cu102', 'torch-cuda-available': True, 'torch-cuda-version': '10.2', 'python-version': '3.8', 'icefall-git-branch': 'zipformer-ctc-aed', 'icefall-git-sha1': '84dfb576-dirty', 'icefall-git-date': 'Sat May 25 17:49:14 2024', 'icefall-path': '/star-zw/workspace/zipformer/icefall_ctc_aed', 'k2-path': '/star-zw/workspace/k2/k2/k2/python/k2/__init__.py', 'lhotse-path': '/star-zw/workspace/lhotse/lhotse/lhotse/__init__.py', 'hostname': 'de-74279-k2-train-10-0312151423-668f59dc99-b8962', 'IP address': '10.177.6.147'}, 'frame_shift_ms': 10, 'search_beam': 20, 'output_beam': 8, 'min_active_states': 30, 'max_active_states': 10000, 'use_double_scores': True, 'epoch': 50, 'iter': 0, 'avg': 29, 'use_averaged_model': True, 'exp_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large'), 'bpe_model': 'data/lang_bpe_500/bpe.model', 'lang_dir': PosixPath('data/lang_bpe_500'), 'context_size': 2, 'decoding_method': 'ctc-decoding', 'num_paths': 100, 'nbest_scale': 1.0, 'hlg_scale': 0.6, 'lm_dir': PosixPath('data/lm'), 'num_encoder_layers': '2,2,4,5,4,2', 'downsampling_factor': '1,2,4,8,4,2', 'feedforward_dim': '512,768,1536,2048,1536,768', 'num_heads': '4,4,4,8,4,4', 'encoder_dim': '192,256,512,768,512,256', 'query_head_dim': '32', 'value_head_dim': '12', 'pos_head_dim': '4', 'pos_dim': 48, 'encoder_unmasked_dim': '192,192,256,320,256,192', 'cnn_module_kernel': '31,31,15,15,15,31', 'decoder_dim': 512, 'joiner_dim': 512, 'attention_decoder_dim': 512, 'attention_decoder_num_layers': 6, 'attention_decoder_attention_dim': 512, 'attention_decoder_num_heads': 8, 'attention_decoder_feedforward_dim': 2048, 'causal': False, 'chunk_size': '16,32,64,-1', 'left_context_frames': '64,128,256,-1', 'use_transducer': False, 'use_ctc': True, 'use_attention_decoder': True, 'full_libri': True, 'mini_libri': False, 'manifest_dir': PosixPath('data/fbank'), 'max_duration': 200, 'bucketing_sampler': True, 'num_buckets': 30, 'concatenate_cuts': False, 'duration_factor': 1.0, 'gap': 1.0, 'on_the_fly_feats': False, 'shuffle': True, 'drop_last': True, 'return_cuts': True, 'num_workers': 2, 'enable_spec_aug': True, 'spec_aug_time_warp_factor': 80, 'enable_musan': True, 'input_strategy': 'PrecomputedFeatures', 'res_dir': PosixPath('zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding'), 'suffix': 'epoch-50-avg-29-use-averaged-model'}
|
4 |
+
2024-05-26 11:20:48,266 INFO [lexicon.py:168] Loading pre-compiled data/lang_bpe_500/Linv.pt
|
5 |
+
2024-05-26 11:20:53,300 INFO [ctc_decode.py:807] About to create model
|
6 |
+
2024-05-26 11:20:54,579 INFO [ctc_decode.py:874] Calculating the averaged model over epoch range from 21 (excluded) to 50
|
7 |
+
2024-05-26 11:21:03,984 INFO [ctc_decode.py:891] Number of model parameters: 174319650
|
8 |
+
2024-05-26 11:21:03,985 INFO [asr_datamodule.py:467] About to get test-clean cuts
|
9 |
+
2024-05-26 11:21:04,107 INFO [asr_datamodule.py:474] About to get test-other cuts
|
10 |
+
2024-05-26 11:21:05,248 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 14
|
11 |
+
2024-05-26 11:21:40,659 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2298
|
12 |
+
2024-05-26 11:21:43,885 INFO [zipformer.py:1858] name=None, attn_weights_entropy = tensor([2.6771, 3.2606, 3.3798, 3.6473], device='cuda:0')
|
13 |
+
2024-05-26 11:21:45,487 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
14 |
+
2024-05-26 11:21:45,583 INFO [utils.py:657] [test-clean-ctc-decoding] %WER 2.29% [1206 / 52576, 119 ins, 98 del, 989 sub ]
|
15 |
+
2024-05-26 11:21:45,788 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/errs-test-clean-epoch-50-avg-29-use-averaged-model.txt
|
16 |
+
2024-05-26 11:21:45,791 INFO [ctc_decode.py:669]
|
17 |
+
For test-clean, WER of different settings are:
|
18 |
+
ctc-decoding 2.29 best for test-clean
|
19 |
+
|
20 |
+
2024-05-26 11:21:46,547 INFO [ctc_decode.py:623] batch 0/?, cuts processed until now is 17
|
21 |
+
2024-05-26 11:22:22,471 INFO [ctc_decode.py:623] batch 100/?, cuts processed until now is 2530
|
22 |
+
2024-05-26 11:22:27,123 INFO [ctc_decode.py:646] The transcripts are stored in zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
23 |
+
2024-05-26 11:22:27,217 INFO [utils.py:657] [test-other-ctc-decoding] %WER 5.14% [2688 / 52343, 277 ins, 209 del, 2202 sub ]
|
24 |
+
2024-05-26 11:22:27,413 INFO [ctc_decode.py:655] Wrote detailed error stats to zipformer/exp-ctc-0.1-aed-0.9-penalize-attn-large/ctc-decoding/errs-test-other-epoch-50-avg-29-use-averaged-model.txt
|
25 |
+
2024-05-26 11:22:27,416 INFO [ctc_decode.py:669]
|
26 |
+
For test-other, WER of different settings are:
|
27 |
+
ctc-decoding 5.14 best for test-other
|
28 |
+
|
29 |
+
2024-05-26 11:22:27,416 INFO [ctc_decode.py:924] Done!
|
decoding_results/ctc-decoding/recogs-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/ctc-decoding/recogs-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
decoding_results/ctc-decoding/wer-summary-test-clean-epoch-50-avg-29-use-averaged-model.txt
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
settings WER
|
2 |
+
ctc-decoding 2.29
|
decoding_results/ctc-decoding/wer-summary-test-other-epoch-50-avg-29-use-averaged-model.txt
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
settings WER
|
2 |
+
ctc-decoding 5.14
|
exp/decode.sh
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
export CUDA_VISIBLE_DEVICES="0"
|
2 |
+
for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
|
3 |
+
./zipformer/ctc_decode.py \
|
4 |
+
--epoch 50 \
|
5 |
+
--avg 29 \
|
6 |
+
--exp-dir zipformer/exp-large \
|
7 |
+
--use-ctc 1 \
|
8 |
+
--use-transducer 0 \
|
9 |
+
--use-attention-decoder 1 \
|
10 |
+
--attention-decoder-loss-scale 0.9 \
|
11 |
+
--num-encoder-layers 2,2,4,5,4,2 \
|
12 |
+
--feedforward-dim 512,768,1536,2048,1536,768 \
|
13 |
+
--encoder-dim 192,256,512,768,512,256 \
|
14 |
+
--encoder-unmasked-dim 192,192,256,320,256,192 \
|
15 |
+
--max-duration 100 \
|
16 |
+
--causal 0 \
|
17 |
+
--num-paths 100 \
|
18 |
+
--decoding-method $m
|
exp/epoch-50.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d3956bffeda26a91e5d3f8dbc7ac811bd7332c56467178ce867418646188640
|
3 |
+
size 2790381194
|
exp/export.sh
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
./zipformer/export.py \
|
2 |
+
--exp-dir zipformer/exp-large \
|
3 |
+
--use-ctc 1 \
|
4 |
+
--use-transducer 0 \
|
5 |
+
--use-attention-decoder 1 \
|
6 |
+
--num-encoder-layers 2,2,4,5,4,2 \
|
7 |
+
--feedforward-dim 512,768,1536,2048,1536,768 \
|
8 |
+
--encoder-dim 192,256,512,768,512,256 \
|
9 |
+
--encoder-unmasked-dim 192,192,256,320,256,192 \
|
10 |
+
--tokens data/lang_bpe_500/tokens.txt \
|
11 |
+
--epoch 50 \
|
12 |
+
--avg 29 \
|
13 |
+
--jit 0
|
14 |
+
|
15 |
+
|
exp/pretrained.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78d4c46696304b6df6519d9a1355325d1233712e8f7108a97ea5d067b088c637
|
3 |
+
size 697818201
|
exp/pretrained_ctc.sh
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
export CUDA_VISIBLE_DEVICES=7
|
2 |
+
for m in ctc-decoding attention-decoder-rescoring-no-ngram; do
|
3 |
+
./zipformer/pretrained_ctc.py \
|
4 |
+
--use-ctc 1 \
|
5 |
+
--use-transducer 0 \
|
6 |
+
--use-attention-decoder 1 \
|
7 |
+
--num-encoder-layers 2,2,4,5,4,2 \
|
8 |
+
--feedforward-dim 512,768,1536,2048,1536,768 \
|
9 |
+
--encoder-dim 192,256,512,768,512,256 \
|
10 |
+
--encoder-unmasked-dim 192,192,256,320,256,192 \
|
11 |
+
--checkpoint ./zipformer/exp-large/pretrained.pt \
|
12 |
+
--tokens data/lang_bpe_500/tokens.txt \
|
13 |
+
--method $m \
|
14 |
+
--sample-rate 16000 \
|
15 |
+
test_wavs/1089-134686-0001.wav \
|
16 |
+
test_wavs/1221-135766-0001.wav \
|
17 |
+
test_wavs/1221-135766-0002.wav
|
18 |
+
done
|
exp/tensorboard/events.out.tfevents.1700103483.de-74279-k2-train-1-1220091118-57c4d55446-mvd6x.1928694.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24870b4e634398f97a7c5b789d52b2fd07b6b9a0d76b3ef75317e51c8cbb49b0
|
3 |
+
size 1608517
|
exp/train.sh
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
export CUDA_VISIBLE_DEVICES="0,1,2,3"
|
2 |
+
# For non-streaming model training:
|
3 |
+
./zipformer/train.py \
|
4 |
+
--world-size 4 \
|
5 |
+
--num-epochs 50 \
|
6 |
+
--start-epoch 1 \
|
7 |
+
--use-fp16 1 \
|
8 |
+
--exp-dir zipformer/exp-large \
|
9 |
+
--full-libri 1 \
|
10 |
+
--use-ctc 1 \
|
11 |
+
--use-transducer 0 \
|
12 |
+
--use-attention-decoder 1 \
|
13 |
+
--ctc-loss-scale 0.1 \
|
14 |
+
--attention-decoder-loss-scale 0.9 \
|
15 |
+
--num-encoder-layers 2,2,4,5,4,2 \
|
16 |
+
--feedforward-dim 512,768,1536,2048,1536,768 \
|
17 |
+
--encoder-dim 192,256,512,768,512,256 \
|
18 |
+
--encoder-unmasked-dim 192,192,256,320,256,192 \
|
19 |
+
--max-duration 1200 \
|
20 |
+
--master-port 12345
|
21 |
+
|
test_wavs/1089-134686-0001.wav
ADDED
Binary file (212 kB). View file
|
|
test_wavs/1221-135766-0001.wav
ADDED
Binary file (535 kB). View file
|
|
test_wavs/1221-135766-0002.wav
ADDED
Binary file (154 kB). View file
|
|
test_wavs/trans.txt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
1089-134686-0001 AFTER EARLY NIGHTFALL THE YELLOW LAMPS WOULD LIGHT UP HERE AND THERE THE SQUALID QUARTER OF THE BROTHELS
|
2 |
+
1221-135766-0001 GOD AS A DIRECT CONSEQUENCE OF THE SIN WHICH MAN THUS PUNISHED HAD GIVEN HER A LOVELY CHILD WHOSE PLACE WAS ON THAT SAME DISHONOURED BOSOM TO CONNECT HER PARENT FOR EVER WITH THE RACE AND DESCENT OF MORTALS AND TO BE FINALLY A BLESSED SOUL IN HEAVEN
|
3 |
+
1221-135766-0002 YET THESE THOUGHTS AFFECTED HESTER PRYNNE LESS WITH HOPE THAN APPREHENSION
|