Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for mixed precision #323

Open
andres-ulloa-de-la-torre opened this issue Oct 24, 2022 · 3 comments
Open

Support for mixed precision #323

andres-ulloa-de-la-torre opened this issue Oct 24, 2022 · 3 comments

Comments

@andres-ulloa-de-la-torre

Have an rx 6900xt which runs inference on stable diffusion in 33s. My 16 GB rtx a4000 does the same in 6.7s. DirectML is not a serious alternative to neither of ROCm and CUDA, without support or emulation for tensor cores. AMD inference times are 6 times slower than the equivalent Nvidia card running CUDA. Even ROCm has massive gains on Radeon cards without any actual matrix cores.

Any chance the plugin gets real mixed precision support? What are your plans going forward with regards to performance?

Thanks in advance for taking your time to address these concerns.

@aliencaocao
Copy link

#315 (comment)

@PatriceVignola
Copy link
Contributor

Hi @andres-ulloa,

As @aliencaocao said, mixed precision is an area that we haven't been focusing on yet but it's on our radar. Is there a particular model that you're looking at?

@cminnoy
Copy link

cminnoy commented Mar 26, 2024

mixed precision float16 on RDNA2 RX 6900 XT would be great for convolutions

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants