-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Bus error (core dumped) #8974
Comments
Which model are you using? @youkaichao any thoughts on this? |
the information is not very helpful. maybe follow https://docs.vllm.ai/en/latest/getting_started/debugging.html and provide more information? |
Hi, Thanks for your reply. I've followed the instructions from the debugging page. I write a minimal script, vllm_debug.py
I've run it with all four flags on
But no additional error message is given except for the 'Bus Error'. I've also run the test.py file to check GPU communications, it returns the following error
I've followed #8420, but it's not really the case. As I created a new conda environment and directly run the command 'pip install vllm'. I am downgrading to python 3.9.19 with vllm 0.5.4, which works perfectly fine on my other machine. I really wonder why. Everything worked before, and my whole machine collapses after I upgrade my vllm to 0.6.2, and it keeps saying Bus Error even after I, reinstalled nccl, create a separate new conda environment, and downgrade packages. Thank you very much for your support. Here is the bt from the latest core dump, it shows something different.
|
I am using model Llama-3-8B-Instruct, this machine is my dev machine with RTX Quadro 8000 |
you need to paste the output log of this command |
Thank you for your help. However, the file is way too large (278MB). Here is the last 500 lines using tail command
|
from the log, it seems the last time you see vllm code, is did you try to remove the guided decoding?
to
it is just to identify the root cause. |
Ah! You are absolutely correct, by removing the guided_choice, my local version can generate utterances normally. But, why? What could cause my local machine fail to do the guided_choice constrained generation? |
can be an outlier issue. see #7831 for example. and #7831 (comment) for changing the guided decoding backend. |
Thank you very much. Having guided_decoding_backend='lm-format-enforcer' in the engine solved my problem. But I do wonder why this doesn't work on my Quadro RTX 8000 machine but works on my A100 machine. |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
Greetings, I was using vllm 0.5.2 just fine, but I want to do some constrained decoding and therefore upgraded to 0.6.2. Then I always encouter this Bus error when I attempt to inference.
I've tried to delete and re-create conda environment,
downgrade to 0.5.4
uninstall/reinstall nccl from source/apt/conda multiple times (according to #3916)
but non of this helped. However, my other machine works perfectly fine. My code is very standard
Could someone help me out?
Much appreciated
bt from gdp core file
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: