-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Building VLLM from source and running inference: No module named 'vllm._C' #3061
Comments
I have the same problem (vllm built from source): Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/workspace/vllm/vllm/entrypoints/llm.py", line 109, in __init__
self.llm_engine = LLMEngine.from_engine_args(engine_args)
File "/workspace/vllm/vllm/engine/llm_engine.py", line 371, in from_engine_args
engine = cls(*engine_configs,
File "/workspace/vllm/vllm/engine/llm_engine.py", line 120, in __init__
self._init_workers()
File "/workspace/vllm/vllm/engine/llm_engine.py", line 143, in _init_workers
from vllm.worker.worker import Worker
File "/workspace/vllm/vllm/worker/worker.py", line 11, in <module>
from vllm.model_executor import set_random_seed
File "/workspace/vllm/vllm/model_executor/__init__.py", line 2, in <module>
from vllm.model_executor.model_loader import get_model
File "/workspace/vllm/vllm/model_executor/model_loader.py", line 10, in <module>
from vllm.model_executor.weight_utils import (get_quant_config,
File "/workspace/vllm/vllm/model_executor/weight_utils.py", line 18, in <module>
from vllm.model_executor.layers.quantization import (get_quantization_config,
File "/workspace/vllm/vllm/model_executor/layers/quantization/__init__.py", line 4, in <module>
from vllm.model_executor.layers.quantization.awq import AWQConfig
File "/workspace/vllm/vllm/model_executor/layers/quantization/awq.py", line 6, in <module>
from vllm._C import ops
ModuleNotFoundError: No module named 'vllm._C' |
Had the same issue too, it turned out because i had a folder named vllm in my working directory. Unfortunately I had another issue, so now gave up. |
Installing vllm==0.2.6 solves at least the When building vllm without precompiled (VLLM_USE_PRECOMPILED=false), it leads to the |
I installed using "python setup.py install" and got this error. I fixed it with "python setup.py develop" |
@Lena-Jurkschat @george-kuanli-peng |
Unfortunately, not. "python setup.py develop" did not work either in combination with VLLM_USE_PRECOMPILED=false. |
I successfully installed In
and
So, How to fix it? |
I just solved it this way. The problem is with So, what we have to do is to change |
I too am getting this error and not in a position to find and replace all the instances of 'vllm._C' in the code. Cc @liangfu Hardware: inf2.8xlarge
|
Was also seeing the same error on an inf2 instance with the latest release. Running the following in the vLLM directory before installing with pip solved the issue for me.
I'm not sure if this is a solution for all distributions.
|
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
Hi,
after building vllm from source, the following error occures when running a multi-gpu inference using a local ray instance:
I already checked Issue #1814, which does not help. So there is no additional vllm folder to delete, which could lead to confusion.
I run the following to build vllm:
I run the inference using
However, building vllm via pip instead leads to an MPI error when running multi-gpu inference (probably due to version incompatiablity of MPI on my System and the prebuild vllm things?), so I wanted to build it from source.
Some Specs:
The text was updated successfully, but these errors were encountered: