Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Option to use a language model with Wav2Vec2 transcription #1312

Closed
lfcnassif opened this issue Sep 11, 2022 · 2 comments
Closed

Option to use a language model with Wav2Vec2 transcription #1312

lfcnassif opened this issue Sep 11, 2022 · 2 comments

Comments

@lfcnassif
Copy link
Member

This was left as a future improvement of #1214. This should be investigated:
jonatasgrosman/huggingsound#62

@lfcnassif
Copy link
Member Author

I still wasn't able to use the (suggested) KenshoLMDecoder implementation for a language model from huggingsound library properly to evaluate this. But I managed to use ParlanceLMDecoder implementation together with Jonatas Grosman's fine tuned wav2vec2 large portuguese model, results were not good:
image

@lfcnassif lfcnassif removed their assignment Mar 27, 2024
@lfcnassif
Copy link
Member Author

Since we are migrating to whisper, I'm closing this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant