Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

test #15

Open
wants to merge 77 commits into
base: master
Choose a base branch
from
Open

test #15

Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
77 commits
Select commit Hold shift + click to select a range
3ae47a4
verify have installed ffmpeg (#1117)
SarahSmitho Jun 7, 2023
dca21c2
Fix parameters_names in train.py (#1121)
yfyeung Jun 8, 2023
b4c38d7
Use symlinks for best epochs (#1123)
rouseabout Jun 12, 2023
0cb71ad
add updated zipformer onnx export (#1108)
kakashidan Jun 12, 2023
0ad037d
Add CTC loss option in zipformer recipe (#1111)
yaozengwei Jun 14, 2023
947f061
Fix running exported model on GPU. (#1131)
csukuangfj Jun 15, 2023
0a46579
Fix Zipformer (#1132)
yfyeung Jun 15, 2023
d667dc3
Fix for diagnostic (#1135)
yfyeung Jun 16, 2023
4d5b836
fix small typo (#1144)
frankyoujian Jun 21, 2023
219bba1
zipformer wenetspeech (#1130)
pkufool Jun 26, 2023
968ebd2
Fix ONNX export of the latest streaming zipformer model. (#1148)
csukuangfj Jun 27, 2023
9c2172c
Zipformer for TedLium (#1125)
desh2608 Jun 28, 2023
db71b03
Support int8 quantization in decoder (#1152)
pkufool Jun 29, 2023
c59c89f
Minor fix in tedlium results file (#1153)
desh2608 Jun 29, 2023
ccd8c62
support testing onnx exported model on the test sets (#1150)
yaozengwei Jun 30, 2023
98d8946
zipformer2 logaddexp onnx safe (#1157)
MicKot Jun 30, 2023
c3e23ec
Fix logaddexp for ONNX export (#1158)
csukuangfj Jul 2, 2023
9009d02
Fix ONNX export for the latest non-streaming zipformer. (#1160)
csukuangfj Jul 3, 2023
eca0202
Add start-batch option for RNNLM training (#1161)
nshmyrev Jul 4, 2023
856c0f2
fixed default param for an aishell recipe (#1159)
JinZr Jul 4, 2023
a4402b8
SURT multi-talker ASR recipe (#1126)
desh2608 Jul 4, 2023
b8a1794
Fix zipformer CI test (#1164)
csukuangfj Jul 5, 2023
130ad03
Fix CI test for zipformer CTC (#1165)
csukuangfj Jul 5, 2023
6fd6743
Fix failed CI tests (#1166)
csukuangfj Jul 5, 2023
11523c5
Shallow fusion & LODR documentation (#1142)
marcoyang1998 Jul 6, 2023
ffe816e
Fix blank skip ci test (#1167)
yfyeung Jul 6, 2023
41b16d7
SURT recipe for AMI and ICSI (#1133)
desh2608 Jul 8, 2023
5ed6fc0
add sym link (#1170)
marcoyang1998 Jul 12, 2023
4ab7d61
removed `batch_name` to fix a KeyError with "uttid" (#1172)
JinZr Jul 15, 2023
1dbbd77
Add tests for subsample.py and fix typos (#1180)
csukuangfj Jul 25, 2023
80d922c
Update preprocess_commonvoice.py to fix text normalization bug. (#1181)
kobenaxie Jul 26, 2023
625b33e
Update descriptions for different decoding methods with external LMs …
marcoyang1998 Jul 27, 2023
3fb0a43
Fix conflict (#1187)
marcoyang1998 Jul 27, 2023
19b942c
Update installation doc. (#1188)
csukuangfj Jul 27, 2023
751bb6f
Add docker image for icefall (#1189)
csukuangfj Jul 28, 2023
375520d
Run the yesno recipe with docker in GitHub actions (#1191)
csukuangfj Jul 28, 2023
bcabaf8
Add doc describing how to run icefall within a docker container (#1194)
csukuangfj Aug 1, 2023
1ee251c
Decode zipformer with external LMs (#1193)
marcoyang1998 Aug 3, 2023
00256a7
Fix decode_stream.py (#1208)
yfyeung Aug 9, 2023
74806b7
disable speed perturbation by default (#1176)
JinZr Aug 10, 2023
d6b28a1
Add export script for the yesno recipe. (#1212)
csukuangfj Aug 11, 2023
a81396b
Use tokens.txt to replace bpe.model (#1162)
JinZr Aug 12, 2023
dfccadc
Fix a typo in export_onnx.py for yesno (#1213)
csukuangfj Aug 12, 2023
b0e8a40
Speed up yesno training to finish in ~10s on CPU (#1215)
pzelasko Aug 13, 2023
3b5645f
doc updated (#1214)
JinZr Aug 13, 2023
9a47c08
Update padding modified beam search (#1217)
ezerhouni Aug 14, 2023
fc2df07
Add icefall tutorials for dummies. (#1220)
csukuangfj Aug 16, 2023
4d7f73c
Add context biasing for zipformer recipe (#1204)
pkufool Aug 28, 2023
3a1ce59
Minor fix for documentation (#1229)
marcoyang1998 Aug 29, 2023
8fcadb6
Missing definitions in scaling.py added (#1232)
desh2608 Aug 31, 2023
9ef8145
minor fixes (#1240)
JinZr Sep 4, 2023
d50a9ea
doc str fixes (#1241)
JinZr Sep 7, 2023
c912bd6
Update run-gigaspeech-pruned-transducer-stateless2-2022-05-12.sh (#1242)
JinZr Sep 7, 2023
49a4b67
fixed a CI test issue related to python version (#1243)
JinZr Sep 7, 2023
3199058
enable `sclite_mode` for swbd scoring (#1239)
JinZr Sep 9, 2023
0f1bc6f
Multi_zh-Hans Recipe (#1238)
JinZr Sep 13, 2023
7cc2dae
Fixes to incorporate with the latest Lhotse release (#1249)
JinZr Sep 13, 2023
fba1710
modify tal_csasr recipe (#1252)
Sedrick-Song Sep 14, 2023
565d2c2
Minor fixes to the libricss recipe (#1256)
JinZr Sep 14, 2023
0c564c6
Fix typo in README.md (#1257)
eltociear Sep 17, 2023
7e1288a
fix thchs-30 download command (#1260)
wangtiance Sep 19, 2023
bbb03f7
Update decoder.py (#1262)
JinZr Sep 20, 2023
45d60ef
Update conformer.py (#1200)
l2009312042 Sep 21, 2023
f5dc957
Fix CI tests (#1266)
csukuangfj Sep 21, 2023
34e40a8
Fix exporting decoder model to onnx (#1264)
csukuangfj Sep 22, 2023
ef658d6
fixes for init value of `diagnostics.TensorDiagnosticOptions` (#1269)
JinZr Sep 24, 2023
ef5da48
formatted the entire LibriSpeech recipe (#1270)
JinZr Sep 24, 2023
97f9b9c
Add documentation for RNNLM training (#1267)
marcoyang1998 Sep 25, 2023
e17f884
Fix docs for MVQ (#1272)
marcoyang1998 Sep 25, 2023
1b565dd
added softlinks to local dir (#1273)
JinZr Sep 26, 2023
2318c3f
Support CTC decoding on CPU using OpenFst and kaldi decoders. (#1244)
csukuangfj Sep 26, 2023
8e10ce0
Copy HL decoding script to HLG decoding script
csukuangfj Sep 26, 2023
2fd0673
Support HLG decoding using OpenFst with kaldi decoders
csukuangfj Sep 26, 2023
043aafc
small fixes
csukuangfj Sep 26, 2023
f357d3f
small fixes
csukuangfj Sep 27, 2023
edc37b0
Fix building HLG
csukuangfj Sep 27, 2023
0ab247f
support triggering CI manually
csukuangfj Sep 27, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
1 change: 1 addition & 0 deletions .flake8
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@ exclude =
**/data/**,
icefall/shared/make_kn_lm.py,
icefall/__init__.py
icefall/ctc/__init__.py

ignore =
# E203 white space before ":"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -29,6 +29,9 @@ if [[ x"${GITHUB_EVENT_NAME}" == x"schedule" || x"${GITHUB_EVENT_LABEL_NAME}" ==
ls -lh data/fbank
ls -lh pruned_transducer_stateless2/exp

ln -s data/fbank/cuts_DEV.jsonl.gz data/fbank/gigaspeech_cuts_DEV.jsonl.gz
ln -s data/fbank/cuts_TEST.jsonl.gz data/fbank/gigaspeech_cuts_TEST.jsonl.gz

log "Decoding dev and test"

# use a small value for decoding with CPU
Expand Down
10 changes: 5 additions & 5 deletions .github/scripts/run-librispeech-conformer-ctc3-2022-11-28.sh
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ log "Decode with models exported by torch.jit.trace()"
for m in ctc-decoding 1best; do
./conformer_ctc3/jit_pretrained.py \
--model-filename $repo/exp/jit_trace.pt \
--words-file $repo/data/lang_bpe_500/words.txt \
--words-file $repo/data/lang_bpe_500/words.txt \
--HLG $repo/data/lang_bpe_500/HLG.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--G $repo/data/lm/G_4_gram.pt \
Expand All @@ -53,7 +53,7 @@ log "Export to torchscript model"

./conformer_ctc3/export.py \
--exp-dir $repo/exp \
--lang-dir $repo/data/lang_bpe_500 \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--jit-trace 1 \
--epoch 99 \
--avg 1 \
Expand All @@ -80,9 +80,9 @@ done
for m in ctc-decoding 1best; do
./conformer_ctc3/pretrained.py \
--checkpoint $repo/exp/pretrained.pt \
--words-file $repo/data/lang_bpe_500/words.txt \
--words-file $repo/data/lang_bpe_500/words.txt \
--HLG $repo/data/lang_bpe_500/HLG.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--G $repo/data/lm/G_4_gram.pt \
--method $m \
--sample-rate 16000 \
Expand All @@ -93,7 +93,7 @@ done

echo "GITHUB_EVENT_NAME: ${GITHUB_EVENT_NAME}"
echo "GITHUB_EVENT_LABEL_NAME: ${GITHUB_EVENT_LABEL_NAME}"
if [[ x"${GITHUB_EVENT_NAME}" == x"schedule" || x"${GITHUB_EVENT_LABEL_NAME}" == x"run-decode" ]]; then
if [[ x"${GITHUB_EVENT_NAME}" == x"schedule" || x"${GITHUB_EVENT_LABEL_NAME}" == x"run-decode" ]]; then
mkdir -p conformer_ctc3/exp
ln -s $PWD/$repo/exp/pretrained.pt conformer_ctc3/exp/epoch-999.pt
ln -s $PWD/$repo/data/lang_bpe_500 data/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ log "Test exporting with torch.jit.trace()"

./lstm_transducer_stateless2/export.py \
--exp-dir $repo/exp \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--use-averaged-model 0 \
Expand All @@ -55,7 +55,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -68,7 +68,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -41,7 +41,7 @@ for method in fast_beam_search modified_beam_search beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -49,7 +49,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -48,7 +48,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -30,14 +30,14 @@ popd
log "Export to torchscript model"
./pruned_transducer_stateless3/export.py \
--exp-dir $repo/exp \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--jit 1

./pruned_transducer_stateless3/export.py \
--exp-dir $repo/exp \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--jit-trace 1
Expand Down Expand Up @@ -74,7 +74,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -87,7 +87,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--num-encoder-layers 18 \
--dim-feedforward 2048 \
--nhead 8 \
Expand All @@ -51,7 +51,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ log "Export to torchscript model"
./pruned_transducer_stateless7/export.py \
--exp-dir $repo/exp \
--use-averaged-model false \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--jit 1
Expand All @@ -56,7 +56,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -69,7 +69,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ log "Export to torchscript model"
./pruned_transducer_stateless7_ctc/export.py \
--exp-dir $repo/exp \
--use-averaged-model false \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--jit 1
Expand Down Expand Up @@ -74,7 +74,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -87,7 +87,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,9 +21,9 @@ tree $repo/
ls -lh $repo/test_wavs/*.wav

pushd $repo/exp
git lfs pull --include "data/lang_bpe_500/HLG.pt"
git lfs pull --include "data/lang_bpe_500/L.pt"
git lfs pull --include "data/lang_bpe_500/LG.pt"
git lfs pull --include "data/lang_bpe_500/HLG.pt"
git lfs pull --include "data/lang_bpe_500/Linv.pt"
git lfs pull --include "data/lang_bpe_500/bpe.model"
git lfs pull --include "exp/cpu_jit.pt"
Expand All @@ -36,7 +36,7 @@ log "Export to torchscript model"
./pruned_transducer_stateless7_ctc_bs/export.py \
--exp-dir $repo/exp \
--use-averaged-model false \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--epoch 99 \
--avg 1 \
--jit 1
Expand Down Expand Up @@ -72,7 +72,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -85,7 +85,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,7 @@ log "Export to torchscript model"
./pruned_transducer_stateless7_streaming/export.py \
--exp-dir $repo/exp \
--use-averaged-model false \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--decode-chunk-len 32 \
--epoch 99 \
--avg 1 \
Expand Down Expand Up @@ -81,7 +81,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--decode-chunk-len 32 \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
Expand All @@ -95,7 +95,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--decode-chunk-len 32 \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ log "Decode with models exported by torch.jit.script()"
log "Export to torchscript model"
./pruned_transducer_stateless8/export.py \
--exp-dir $repo/exp \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--use-averaged-model false \
--epoch 99 \
--avg 1 \
Expand All @@ -65,7 +65,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand All @@ -78,7 +78,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ for sym in 1 2 3; do
--method greedy_search \
--max-sym-per-frame $sym \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--simulate-streaming 1 \
--causal-convolution 1 \
$repo/test_wavs/1089-134686-0001.wav \
Expand All @@ -47,7 +47,7 @@ for method in modified_beam_search beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--simulate-streaming 1 \
--causal-convolution 1 \
$repo/test_wavs/1089-134686-0001.wav \
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ ls -lh $repo/test_wavs/*.wav

pushd $repo/exp
git lfs pull --include "data/lang_bpe_500/bpe.model"
git lfs pull --include "data/lang_bpe_500/tokens.txt"
git lfs pull --include "exp/jit_script_chunk_16_left_128.pt"
git lfs pull --include "exp/pretrained.pt"
ln -s pretrained.pt epoch-99.pt
Expand All @@ -33,7 +34,7 @@ log "Export to torchscript model"
./zipformer/export.py \
--exp-dir $repo/exp \
--use-averaged-model false \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--causal 1 \
--chunk-size 16 \
--left-context-frames 128 \
Expand All @@ -46,7 +47,7 @@ ls -lh $repo/exp/*.pt
log "Decode with models exported by torch.jit.script()"

./zipformer/jit_pretrained_streaming.py \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
--nn-model-filename $repo/exp/jit_script_chunk_16_left_128.pt \
$repo/test_wavs/1089-134686-0001.wav

Expand All @@ -60,7 +61,7 @@ for method in greedy_search modified_beam_search fast_beam_search; do
--method $method \
--beam-size 4 \
--checkpoint $repo/exp/pretrained.pt \
--bpe-model $repo/data/lang_bpe_500/bpe.model \
--tokens $repo/data/lang_bpe_500/tokens.txt \
$repo/test_wavs/1089-134686-0001.wav \
$repo/test_wavs/1221-135766-0001.wav \
$repo/test_wavs/1221-135766-0002.wav
Expand Down
Loading