-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
modification on llamacpp readme after Ipex-llm latest update #11971
modification on llamacpp readme after Ipex-llm latest update #11971
Conversation
llama_new_context_with_model: SYCL_Host compute buffer size = 288.02 MiB | ||
llama_new_context_with_model: graph nodes = 1062 | ||
llama_new_context_with_model: graph splits = 2 | ||
Native API failed. Native API returns: -5 (PI_ERROR_OUT_OF_RESOURCES) -5 (PI_ERROR_OUT_OF_RESOURCES)Exception caught at file:C:/Users/Administrator/actions-runner/cpp-release/_work/llm.cpp/llm.cpp/llama-cpp-bigdl/ggml/src/ggml-sycl.cpp, line:2856 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
just keep latest
@@ -127,16 +126,12 @@ To use GPU acceleration, several environment variables are required or recommend | |||
|
|||
```cmd | |||
set SYCL_CACHE_PERSISTENT=1 | |||
rem under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
revert it
|
||
#### 14. `Native API failed` error | ||
On latest version of `ipex-llm`, you might come across `native API failed` error with certain models without the `-c` parameter. Simply adding `-c 1024` would resolve this problem. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
On latest version of ipex-llm's llama.cpp, you might come across native API failed
error with certain models without the -c
parameter. Simply adding -c xx
would resolve this problem.
|
||
#### 13. Core dump when having both integrated and dedicated graphics | ||
If you have both integrated and dedicated graphics on your computer and didn't specify which device to use, it will cause a core dump. Therefore, you need to run `export ONEAPI_DEVICE_SELECTOR=level_zero:0` before running `llama-cli`. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If you have both integrated and dedicated graphics displayed in your llama.cpp's device log and don't specify which device to use, it will cause a core dump. In such case, you may need to specify export ONEAPI_DEVICE_SELECTOR=level_zero:0
before running llama-cli
.
Please also add number for ollama's troubleshooting. |
``` | ||
|
||
Under your current directory, you can also execute below command to have interactive chat with Llama3: | ||
|
||
- For **Linux users**: | ||
|
||
```bash | ||
./main -ngl 33 --interactive-first --color -e --in-prefix '<|start_header_id|>user<|end_header_id|>\n\n' --in-suffix '<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n' -r '<|eot_id|>' -m <model_dir>/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf | ||
./llama-cli -ngl 33 --interactive-first --color -e --in-prefix '<|start_header_id|>user<|end_header_id|>\n\n' --in-suffix '<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n' -r '<|eot_id|>' -m <model_dir>/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf -c 1024 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please confirm if it still works for new llama.cpp
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
…nalytics#11971) * update on readme after ipex-llm update * update on readme after ipex-llm update * rebase & delete redundancy * revise * add numbers for troubleshooting
@rnwang04