-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
AMD ROCm support #5
AMD ROCm support #5
Conversation
I have no AMD ROCm GPUs and no way of testing this. Is it correctly understood that it's exclusively ExLlama kernels that will work with AMD GPUs on ROCm 5 (but not 6)? |
I tested on ROCm 5.6.1 |
@IlyasMoutawwakil I just checked AutoGPTQ/AutoGPTQ#515 in rocm5.7 env. It is compiled and there should be no perf issue because they are basically same syntaxes. (Somehow the old syntax is not compatible with rocm6.0 though) |
@casper-hansen for the other part of your question, yes for now Exllama is going to be the only way to run AWQ model on AMD GPUs. Untill we get a hipifiable, ROCm native or performant triton GEMM. |
setup.py
build.yaml
workflow