Skip to content

GPU inference #24

Answered by snakers4
pKrysenko asked this question in Q&A
Discussion options

You must be logged in to vote

Ah, sorry, forgot to mention. "silero-vad"

Well, these models are very small and they were designed to run on CPU
Running them on GPU will not really provide any tangible speed / throughput benefits
We can of course publish the non-quantized versions of these models
But this would make the repo larger and we will have to maintain 2 versions of each model in parallel (so far we tried to keep this repo as minimal as possible)

So the main question is - why?

Replies: 7 comments

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by snakers4
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants
Converted from issue

This discussion was converted from issue #13 on February 03, 2021 05:13.