-
Notifications
You must be signed in to change notification settings - Fork 414
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add C++ support for streaming NeMo CTC models. #857
Conversation
Hello, thank you for the speedy response and the model export support! I tried the online-nemo-ctc-decode-files.py with the 480.ms model but the response isn't generated real-time, I assume as it is online the text should be generated as it is decoded. I also noticed a few errors, is it due to the model itself? and is the 1040ms model better or 80ms model better I also wanted to know if streaming transducer can be used with the streaming_server.py file, as it has a separate decoder and joiner. I tried using it but got an error: /Users/runner/work/sherpa-onnx/sherpa-onnx/sherpa-onnx/csrc/online-transducer-model.cc:GetModelType:75 Unsupported model_type: EncDecHybridRNNTCTCBPEModel |
It is decoding files. What do you mean by
Could you tell us what errors you have noticed?
What do you mean by
It has not been implemented yet. Will support it this week. |
Hello,
It has not been implemented yet. Will support it this week. |
That is expected. We are decoding a file and it gives you the result once the file is decoded. Please refer to our microphone examples and you can change them to support NeMo streaming ctc models and then you can see real-time output as you speak.
Yes, I think so.
In terms of accuracy, I think 1040ms is better. In terms of latency, I think 80ms is better. By the way, you can try the Android APK for NeMo streaming CTC models at https://k2-fsa.github.io/sherpa/onnx/android/apk.html APKs for the non-streaming NeMo CTC models can be found at |
Following #843