-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: AssertionError: libcuda.so cannot found with vllm/vllm-openai:v0.4.0 #3808
Comments
Check / set your CUDA_PATH environment variable |
I have same issue, how you to fix it? |
You may want to run your job on docker with specified cuda 11.8 or 12.1 |
same issue with me. I ran the mixtral-8x7B model using vllm/vllm-openai:v0.4.0 on k8s with H100. |
I've tested multiple versions. This issue didn't occur until version 0.2.7, but it started happening from version 0.3.0 onwards. |
This is not a problem of In general, |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
Your current environment
Running in Kubernetes on H100 in vllm/vllm-openai:v0.4.0
🐛 Describe the bug
Seems like there have been some weird dependency issues since v0.2.7. We would love to use flash attention when you are able to fix this! Thank you! 🙇
The text was updated successfully, but these errors were encountered: