Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ggml-cuda.cu:3211: ERROR: CUDA kernel vec_dot_q5_K_q8_1_impl_vmmq #5686

Closed
DanCard opened this issue Feb 23, 2024 Discussed in #5685 · 2 comments
Closed

ggml-cuda.cu:3211: ERROR: CUDA kernel vec_dot_q5_K_q8_1_impl_vmmq #5686

DanCard opened this issue Feb 23, 2024 Discussed in #5685 · 2 comments

Comments

@DanCard
Copy link

DanCard commented Feb 23, 2024

Discussed in #5685

Originally posted by DanCard February 23, 2024
ggml-cuda.cu:3211: ERROR: CUDA kernel vec_dot_q5_K_q8_1_impl_vmmq has no device code compatible with CUDA arch 520. ggml-cuda.cu was compiled for: 520
This worked yesterday. I did a git pull, make clean, make and then get this error today.
nvidia rtx 3090
System: debian testing
Command line:
~/github/llama.cpp/main -m ~/models/miqu-1-70b.q5_K_M.gguf -c 0 -i --color -t 16 --n-gpu-layers 24 --temp 0.8 -p "bob"

I reverted previous two commits and issue went away.
~/github/llama.cpp$ git reset --hard HEAD~2
HEAD is now at 334f76f sync : ggml

@DanCard
Copy link
Author

DanCard commented Feb 24, 2024

This is no longer an issue with latest update.

@DanCard DanCard closed this as completed Feb 24, 2024
@szymonrucinski
Copy link

Hey I still keep getting this error for the server. I am on the main branch as well.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants