Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

modification on llamacpp readme after Ipex-llm latest update #11971

Merged
merged 5 commits into from
Aug 30, 2024

Conversation

JinheTang
Copy link
Contributor

  1. main -> llama-cli
  2. -c 1024
  3. 2 new troubleshoots
    @rnwang04

llama_new_context_with_model: SYCL_Host compute buffer size = 288.02 MiB
llama_new_context_with_model: graph nodes = 1062
llama_new_context_with_model: graph splits = 2
Native API failed. Native API returns: -5 (PI_ERROR_OUT_OF_RESOURCES) -5 (PI_ERROR_OUT_OF_RESOURCES)Exception caught at file:C:/Users/Administrator/actions-runner/cpp-release/_work/llm.cpp/llm.cpp/llama-cpp-bigdl/ggml/src/ggml-sycl.cpp, line:2856
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

just keep latest

@@ -127,16 +126,12 @@ To use GPU acceleration, several environment variables are required or recommend

```cmd
set SYCL_CACHE_PERSISTENT=1
rem under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

revert it


#### 14. `Native API failed` error
On latest version of `ipex-llm`, you might come across `native API failed` error with certain models without the `-c` parameter. Simply adding `-c 1024` would resolve this problem.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

On latest version of ipex-llm's llama.cpp, you might come across native API failed error with certain models without the -c parameter. Simply adding -c xx would resolve this problem.


#### 13. Core dump when having both integrated and dedicated graphics
If you have both integrated and dedicated graphics on your computer and didn't specify which device to use, it will cause a core dump. Therefore, you need to run `export ONEAPI_DEVICE_SELECTOR=level_zero:0` before running `llama-cli`.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If you have both integrated and dedicated graphics displayed in your llama.cpp's device log and don't specify which device to use, it will cause a core dump. In such case, you may need to specify export ONEAPI_DEVICE_SELECTOR=level_zero:0 before running llama-cli.

@rnwang04
Copy link
Contributor

rnwang04 commented Aug 30, 2024

Please also add number for ollama's troubleshooting.

```

Under your current directory, you can also execute below command to have interactive chat with Llama3:

- For **Linux users**:

```bash
./main -ngl 33 --interactive-first --color -e --in-prefix '<|start_header_id|>user<|end_header_id|>\n\n' --in-suffix '<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n' -r '<|eot_id|>' -m <model_dir>/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf
./llama-cli -ngl 33 --interactive-first --color -e --in-prefix '<|start_header_id|>user<|end_header_id|>\n\n' --in-suffix '<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n' -r '<|eot_id|>' -m <model_dir>/Meta-Llama-3-8B-Instruct-Q4_K_M.gguf -c 1024
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please confirm if it still works for new llama.cpp

Copy link
Contributor

@rnwang04 rnwang04 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@rnwang04 rnwang04 merged commit e895e1b into intel-analytics:main Aug 30, 2024
@JinheTang JinheTang deleted the ipex-llm-latest-update branch August 30, 2024 05:34
cranechu0131 pushed a commit to cranechu0131/ipex-llm that referenced this pull request Sep 9, 2024
…nalytics#11971)

* update on readme after ipex-llm update

* update on readme after ipex-llm update

* rebase & delete redundancy

* revise

* add numbers for troubleshooting
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants