-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to run on dGPU #10515
Comments
NO,it still have new problem.
|
I also find ipex-llm can't support streamlit now.notice: this code still need replace new api error
|
We haven't been able to reproduce this issue yet on our Arc A770. Would you mind running the |
NO problem. env-check.sh
|
Based on the provided environment information, it seems that PyTorch and IPEX are not installed. Could you please set up the correct environment and then run the shell script? |
You can follow the guide below to set up the environment or check if the environment is correct: https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html#linux |
I forgot to source oneAPI environment,so the bash don't check truely. env-check.sh
|
Sorry, we still can't reproduce the error you encountered while running chatglm3/streamchat.py. The error you mentioned is OUT_OF_HOST_MEMORY, indicating a memory overflow on the CPU, but you are actually running model inference on XPU. Therefore, could you please provide further details on the input parameters you used when running chatglm3/streamchat.py, including the question prompt, max_new_token, etc., so that we can further replicate the issue? |
My config is default,if you can‘t reproduce the error,I can provide ssh information to you? |
Sure, you could leave your email address and I'll contact you. |
|
Please ensure that the modeling file for chatglm3 is downloaded from the official repository. You could go to ModelScope to download the corresponding file. |
Thanks, maybe I downloaded the basic model instead of the chat model |
hello, I tried to run the code from
https://gitee.com/Pauntech/chat-glm3/blob/master/chatglm3_web_demo.py
,but I face a problem.
you can see that it run on the cpu? but the code clearly offload to the xpu.
the result of
sycl-ls
:I used the method from https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/KeyFeatures/multi_gpus_selection.html,but it still fail.
I also run the code from
https://github.com/intel-analytics/BigDL/blob/main/python/llm/example/GPU/HF-Transformers-AutoModels/Model/chatglm3/streamchat.py
.but I face the problem:
so,how to run the code in dGPU.?
The text was updated successfully, but these errors were encountered: