Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

please include openvino iGPU support for hardware acceleration #25

Closed
jsapede opened this issue Feb 28, 2024 · 6 comments
Closed

please include openvino iGPU support for hardware acceleration #25

jsapede opened this issue Feb 28, 2024 · 6 comments

Comments

@jsapede
Copy link

jsapede commented Feb 28, 2024

as seen on frigate, openvino support highly accelerate ai computing (equivalent to google coral) even on old/small/cheap architectures (> gen6)
combined to faster-whisper or whisper.cpp could higly fluidify the experience of assistant.
iGPu can be easily passed to docker container as done on frigate.

@tannisroot
Copy link

Upstream doesn't support anything but CUDA unfortunately:
OpenNMT/CTranslate2#1072

@synesthesiam
Copy link
Contributor

This is more applicable to this project: https://github.com/rhasspy/wyoming-whisper-cpp

@jsapede
Copy link
Author

jsapede commented Mar 10, 2024

I'm too noob to understand the différence between Whisper/Whisper.cpp except it's a C version.
From what i see it could be possible to compile whisper.cpp with openVINO support and then serve it with Wyoming ?

@Donkey545
Copy link

For those interested in this thread, I made use of a fork of CTranslate2 to build a Wyoming Faster Whisper for ROCm container.

Check it out here if you are interested. I don't have much hardware to test with, so All I have tested is my APU.

This performs about 15x faster than CPU faster-whisper for me, and about 5x faster than the Whisper.cpp implementation that can be found here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants