-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AMD ROCm support #5
AMD ROCm support #5
Conversation
I have no AMD ROCm GPUs and no way of testing this. Is it correctly understood that it's exclusively ExLlama kernels that will work with AMD GPUs on ROCm 5 (but not 6)? |
I tested on ROCm 5.6.1 |
@IlyasMoutawwakil I just checked AutoGPTQ/AutoGPTQ#515 in rocm5.7 env. It is compiled and there should be no perf issue because they are basically same syntaxes. (Somehow the old syntax is not compatible with rocm6.0 though) |
@casper-hansen for the other part of your question, yes for now Exllama is going to be the only way to run AWQ model on AMD GPUs. Untill we get a hipifiable, ROCm native or performant triton GEMM. |
setup.py
build.yaml
workflow