Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix(llama-cpp-server): fix rocm build by setting GGML_HIPBLAS #2835

Merged
merged 1 commit into from
Aug 12, 2024

Conversation

richard-jfc
Copy link
Contributor

@richard-jfc richard-jfc commented Aug 12, 2024

Various LLAMA_ build flags have been renamed to GGML_ in llama-cpp-server (See ggerganov/llama.cpp@f3f6542) and need to be renamed in tabby to allow correctly targeting Metal/Rocm/Vulkan.

It appears that this was already done for Cuda.

May be related: #2811

@wsxiaoys
Copy link
Member

Have you verified if the PR fixes your issue?

cmake handles the migration with macros defined in https://demo.tabbyml.com/files/git/llama.cpp/-/blob/7a221b672e49dfae459b1af27210ba3f2b5419b6/CMakeLists.txt?plain=1#L101

So my guess is it won't fix any build issues you are currently encountering.

@richard-jfc
Copy link
Contributor Author

I built rocm v0.15.0 tag using docker. And have been running the built Rocm version via docker.

Without my change I was seeing a CPU spike for every chat request and no VRAM or GPU usage.

After my change I was seeing roughly 6GB VRAM and a GPU usage spike.

@wsxiaoys
Copy link
Member

It does seem that GGML_HIPBLAS is the flag not handled by the llama_option_depr migration. In that case, would you consider updating the PR to contain only the GGML_HIPBLAS change, since it's the only one you've verified?

@richard-jfc
Copy link
Contributor Author

Thanks, Done

@wsxiaoys wsxiaoys changed the title Fixing renamed llama-server build flags fix(llama-cpp-server): fix rocm build by setting GGML_HIPBLAS Aug 12, 2024
@wsxiaoys wsxiaoys enabled auto-merge (squash) August 12, 2024 01:18
@wsxiaoys wsxiaoys disabled auto-merge August 12, 2024 01:32
@wsxiaoys wsxiaoys merged commit 48dba77 into TabbyML:main Aug 12, 2024
3 of 5 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants