Skip to content

csukuangfj/icefall-asr-conformer-ctc-bpe-500

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

How to clone this repo

Caution: git clone https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500.git won't work. If you use it, you will encounter the following error in icefall:

_pickle.UnpicklingError: invalid load key, 'v'

Please use the following commands to clone this repo:

sudo apt-get install git-lfs
git clone https://github.com/csukuangfj/icefall-asr-conformer-ctc-bpe-500.git
cd icefall-asr-conformer-ctc-bpe-500
git lfs pull

Description

This repo contains a pre-trained model using icefall with the following command (it uses 7 GPUs):

cd egs/librispeech/ASR
./conformer_ctc/train.py \
  --world-size 7 \
  --max-duration 150 \
  --bucketing-sampler 1 \
  --full-libri 1 \
  --start-epoch 0 \
  --num-epochs 50 \
  --exp-dir conformer_ctc/exp \
  --lang-dir data/lang_bpe_500

The model exp/pretrained.pt is obtained by the following command:

cd egs/librispeech/ASR
./conformer_ctc/export.py \
  --epoch 49 \
  --avg 15 \
  --exp-dir conformer_ctc/exp \
  --lang-dir data/lang_bpe_500 \
  --jit 0

You can use exp/pretrained.pt to compute the WER for test-clean and test-other:

cd egs/librispeech/ASR
cd conformer_ctc/exp
ln -s pretrained.pt epoch-99.pt
cd ../..
./conformer_ctc/decode.py \
  --exp-dir ./conformer_ctc/exp \
  --lang-dir ./data/lang_bpe_500 \
  --epoch 99 \
  --avg 1 \
  --max-duration 30 \
  --nbest-scale 0.5

You will get the following log:

2021-10-14 20:01:39,246 INFO [decode.py:474] batch 0/804, cuts processed until now is 6
2021-10-14 20:03:19,789 INFO [decode.py:474] batch 100/804, cuts processed until now is 398
2021-10-14 20:04:53,371 INFO [decode.py:474] batch 200/804, cuts processed until now is 792
2021-10-14 20:06:22,611 INFO [decode.py:474] batch 300/804, cuts processed until now is 1160
2021-10-14 20:07:50,482 INFO [decode.py:474] batch 400/804, cuts processed until now is 1574
2021-10-14 20:09:20,129 INFO [decode.py:474] batch 500/804, cuts processed until now is 1965
2021-10-14 20:10:53,932 INFO [decode.py:474] batch 600/804, cuts processed until now is 2308
2021-10-14 20:12:36,852 INFO [decode.py:474] batch 700/804, cuts processed until now is 2498
2021-10-14 20:14:26,477 INFO [decode.py:474] batch 800/804, cuts processed until now is 2614
2021-10-14 20:16:16,519 INFO [decode.py:523]
For test-clean, WER of different settings are:
ngram_lm_scale_1.1_attention_scale_0.9  2.56    best for test-clean
ngram_lm_scale_1.1_attention_scale_1.0  2.56
ngram_lm_scale_1.2_attention_scale_1.0  2.56
ngram_lm_scale_0.9_attention_scale_0.6  2.57
ngram_lm_scale_0.9_attention_scale_0.7  2.57
ngram_lm_scale_1.0_attention_scale_0.6  2.57
ngram_lm_scale_1.0_attention_scale_0.9  2.57
ngram_lm_scale_1.1_attention_scale_0.7  2.57
ngram_lm_scale_1.1_attention_scale_1.1  2.57
ngram_lm_scale_1.2_attention_scale_0.7  2.57
ngram_lm_scale_1.2_attention_scale_0.9  2.57
ngram_lm_scale_1.2_attention_scale_1.1  2.57
ngram_lm_scale_1.2_attention_scale_1.2  2.57
ngram_lm_scale_1.3_attention_scale_0.9  2.57
ngram_lm_scale_1.3_attention_scale_1.1  2.57
ngram_lm_scale_1.5_attention_scale_1.0  2.57
ngram_lm_scale_1.5_attention_scale_1.1  2.57
ngram_lm_scale_1.5_attention_scale_1.2  2.57
ngram_lm_scale_1.5_attention_scale_1.3  2.57
ngram_lm_scale_2.0_attention_scale_1.7  2.57

2021-10-14 20:16:17,979 INFO [decode.py:474] batch 0/782, cuts processed until now is 6
2021-10-14 20:17:54,037 INFO [decode.py:474] batch 100/782, cuts processed until now is 434
2021-10-14 20:19:26,335 INFO [decode.py:474] batch 200/782, cuts processed until now is 885
2021-10-14 20:20:52,910 INFO [decode.py:474] batch 300/782, cuts processed until now is 1327
2021-10-14 20:22:15,968 INFO [decode.py:474] batch 400/782, cuts processed until now is 1807
2021-10-14 20:23:42,595 INFO [decode.py:474] batch 500/782, cuts processed until now is 2238
2021-10-14 20:25:06,502 INFO [decode.py:474] batch 600/782, cuts processed until now is 2584
2021-10-14 20:26:46,868 INFO [decode.py:474] batch 700/782, cuts processed until now is 2785
2021-10-14 20:30:11,556 INFO [decode.py:523]
For test-other, WER of different settings are:
ngram_lm_scale_1.5_attention_scale_1.9  5.8     best for test-other
ngram_lm_scale_1.7_attention_scale_1.9  5.8
ngram_lm_scale_1.7_attention_scale_2.0  5.8
ngram_lm_scale_1.9_attention_scale_2.0  5.8
ngram_lm_scale_1.3_attention_scale_1.5  5.81
ngram_lm_scale_1.5_attention_scale_1.5  5.81
ngram_lm_scale_1.5_attention_scale_1.7  5.81
ngram_lm_scale_1.7_attention_scale_1.7  5.81
ngram_lm_scale_1.5_attention_scale_1.3  5.82
ngram_lm_scale_1.9_attention_scale_1.9  5.82
ngram_lm_scale_1.2_attention_scale_1.2  5.83
ngram_lm_scale_1.3_attention_scale_1.3  5.83
ngram_lm_scale_1.3_attention_scale_1.7  5.83
ngram_lm_scale_1.5_attention_scale_1.2  5.83
ngram_lm_scale_1.5_attention_scale_2.0  5.83
ngram_lm_scale_2.0_attention_scale_2.0  5.83

Note

This repo uses git lfs. See https://git-lfs.github.com/

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published