From eddf264497da40ee8289d4bb4e4373947cef4b39 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Tue, 27 Aug 2024 14:56:17 +0800 Subject: [PATCH 01/10] modify rag.py --- python/llm/example/GPU/LlamaIndex/rag.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/python/llm/example/GPU/LlamaIndex/rag.py b/python/llm/example/GPU/LlamaIndex/rag.py index fef3204702e..ba43a957771 100644 --- a/python/llm/example/GPU/LlamaIndex/rag.py +++ b/python/llm/example/GPU/LlamaIndex/rag.py @@ -15,7 +15,7 @@ # import torch -from llama_index.embeddings.huggingface import HuggingFaceEmbedding +from llama_index.embeddings.ipex_llm import IpexLLMEmbedding from sqlalchemy import make_url from llama_index.vector_stores.postgres import PGVectorStore # from llama_index.llms.llama_cpp import LlamaCPP @@ -160,10 +160,10 @@ def messages_to_prompt(messages): return prompt def main(args): - embed_model = HuggingFaceEmbedding(model_name=args.embedding_model_path) + embed_model = IpexLLMEmbedding(model_name=args.embedding_model_path, device="xpu") # Use custom LLM in BigDL - from ipex_llm.llamaindex.llms import IpexLLM + from llama_index.llms.ipex_llm import IpexLLM llm = IpexLLM.from_model_id( model_name=args.model_path, tokenizer_name=args.tokenizer_path, From 81818fa5ccb1b2859221c89a2d4413c039b0b6d9 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Tue, 27 Aug 2024 15:15:55 +0800 Subject: [PATCH 02/10] update readme of gpu example --- python/llm/example/GPU/LlamaIndex/README.md | 20 +++++++++++++++----- 1 file changed, 15 insertions(+), 5 deletions(-) diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index 53d5d7ddb69..2e5e79cc9eb 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -8,17 +8,27 @@ This folder contains examples showcasing how to use [**LlamaIndex**](https://git ## Retrieval-Augmented Generation (RAG) Example The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index RAG example](https://docs.llamaindex.ai/en/stable/examples/low_level/oss_ingestion_retrieval.html). This example builds a pipeline to ingest data (e.g. llama2 paper in pdf format) into a vector database (e.g. PostgreSQL), and then build a retrieval pipeline from that vector database. +### 1. Install Prerequisites +To benefit from IPEX-LLM on Intel GPUs, there are several prerequisite steps for tools installation and environment preparation. -### 1. Setting up Dependencies +If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html#install-prerequisites) to update GPU driver (optional) and install Conda. + +If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. + + +### 2. Setting up Dependencies * **Install LlamaIndex Packages** ```bash - pip install llama-index-readers-file llama-index-vector-stores-postgres llama-index-embeddings-huggingface + conda activate + pip install llama-index-llms-ipex-llm[xpu]==0.1.8 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ + pip install llama-index-embeddings-ipex-llm[xpu]==0.1.5 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ + pip install llama-index-readers-file==0.1.33 + pip install llama-index-vector-stores-postgres==0.1.14 + pip install pymupdf ``` -* **Install IPEX-LLM** - - Follow the instructions in [GPU Install Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Overview/install.html) to install ipex-llm. + Theinstallation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. * **Database Setup (using PostgreSQL)**: * Linux From 72179e3f8c8addbdc22d4650dc41cf2f5748d4c1 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Tue, 27 Aug 2024 16:20:33 +0800 Subject: [PATCH 03/10] update llamaindex cpu example and readme --- python/llm/example/CPU/LlamaIndex/README.md | 13 ++++++++----- python/llm/example/CPU/LlamaIndex/rag.py | 6 +++--- python/llm/example/GPU/LlamaIndex/README.md | 5 ++++- 3 files changed, 15 insertions(+), 9 deletions(-) diff --git a/python/llm/example/CPU/LlamaIndex/README.md b/python/llm/example/CPU/LlamaIndex/README.md index be50d92d02a..364c8892aa8 100644 --- a/python/llm/example/CPU/LlamaIndex/README.md +++ b/python/llm/example/CPU/LlamaIndex/README.md @@ -14,12 +14,15 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R * **Install LlamaIndex Packages** ```bash - pip install llama-index-readers-file llama-index-vector-stores-postgres llama-index-embeddings-huggingface + pip install llama-index-llms-ipex-llm==0.1.8 + pip install llama-index-embeddings-ipex-llm==0.1.5 + pip install llama-index-readers-file==0.1.33 + pip install llama-index-vector-stores-postgres==0.1.14 + pip install pymupdf ``` - -* **Install IPEX-LLM** -Ensure `ipex-llm` is installed by following the [IPEX-LLM Installation Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Overview/install.html) before proceeding with the examples provided here. - + > **Note:** + > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. + > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. * **Database Setup (using PostgreSQL)**: * Installation: diff --git a/python/llm/example/CPU/LlamaIndex/rag.py b/python/llm/example/CPU/LlamaIndex/rag.py index 5759c624558..85d1b09fb0f 100644 --- a/python/llm/example/CPU/LlamaIndex/rag.py +++ b/python/llm/example/CPU/LlamaIndex/rag.py @@ -16,7 +16,7 @@ import torch -from llama_index.embeddings.huggingface import HuggingFaceEmbedding +from llama_index.embeddings.ipex_llm import IpexLLMEmbedding from sqlalchemy import make_url from llama_index.vector_stores.postgres import PGVectorStore # from llama_index.llms.llama_cpp import LlamaCPP @@ -161,10 +161,10 @@ def messages_to_prompt(messages): return prompt def main(args): - embed_model = HuggingFaceEmbedding(model_name=args.embedding_model_path) + embed_model = IpexLLMEmbedding(model_name=args.embedding_model_path) # Use custom LLM in BigDL - from ipex_llm.llamaindex.llms import IpexLLM + from llama_index.llms.ipex_llm import IpexLLM llm = IpexLLM.from_model_id( model_name=args.model_path, tokenizer_name=args.tokenizer_path, diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index 2e5e79cc9eb..2b783d8fe19 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -28,7 +28,10 @@ If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](ht pip install llama-index-vector-stores-postgres==0.1.14 pip install pymupdf ``` - Theinstallation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. + > **Note:** + > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. + > You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. + > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. * **Database Setup (using PostgreSQL)**: * Linux From 501bd4b5a919e26af1b9dd33313cdaa5aa1cd0a9 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Tue, 27 Aug 2024 18:00:14 +0800 Subject: [PATCH 04/10] add llamaindex doc --- python/llm/example/CPU/LlamaIndex/README.md | 3 ++- python/llm/example/GPU/LlamaIndex/README.md | 3 ++- 2 files changed, 4 insertions(+), 2 deletions(-) diff --git a/python/llm/example/CPU/LlamaIndex/README.md b/python/llm/example/CPU/LlamaIndex/README.md index 364c8892aa8..20d497f0128 100644 --- a/python/llm/example/CPU/LlamaIndex/README.md +++ b/python/llm/example/CPU/LlamaIndex/README.md @@ -20,7 +20,8 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R pip install llama-index-vector-stores-postgres==0.1.14 pip install pymupdf ``` - > **Note:** + > [!NOTE] + > You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm/) for more information. > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index 2b783d8fe19..271ed83a58a 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -28,7 +28,8 @@ If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](ht pip install llama-index-vector-stores-postgres==0.1.14 pip install pymupdf ``` - > **Note:** + > [!NOTE] + > You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm_gpu/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm_gpu/) for more information. > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. > You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. From 71fbd9b57bac7290ec7581688e6550257bd76311 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 11:00:00 +0800 Subject: [PATCH 05/10] update note style --- python/llm/example/CPU/LlamaIndex/README.md | 8 ++++---- python/llm/example/GPU/LlamaIndex/README.md | 10 +++++----- 2 files changed, 9 insertions(+), 9 deletions(-) diff --git a/python/llm/example/CPU/LlamaIndex/README.md b/python/llm/example/CPU/LlamaIndex/README.md index 20d497f0128..85c02d6433c 100644 --- a/python/llm/example/CPU/LlamaIndex/README.md +++ b/python/llm/example/CPU/LlamaIndex/README.md @@ -20,10 +20,10 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R pip install llama-index-vector-stores-postgres==0.1.14 pip install pymupdf ``` - > [!NOTE] - > You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm/) for more information. - > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. - > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. +> [!NOTE] +> - You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm/) for more information. +> - The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. +> - `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. * **Database Setup (using PostgreSQL)**: * Installation: diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index 271ed83a58a..cea4f57ee79 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -28,11 +28,11 @@ If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](ht pip install llama-index-vector-stores-postgres==0.1.14 pip install pymupdf ``` - > [!NOTE] - > You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm_gpu/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm_gpu/) for more information. - > The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. - > You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. - > `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. +> [!NOTE] +> - You could refer [llama-index-llms-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/llm/ipex_llm_gpu/) and [llama-index-embeddings-ipex-llm](https://docs.llamaindex.ai/en/stable/examples/embeddings/ipex_llm_gpu/) for more information. +> - The installation of `llama-index-llms-ipex-llm` or `llama-index-embeddings-ipex-llm` will also install `IPEX-LLM` and its dependencies. +> - You can also use `https://pytorch-extension.intel.com/release-whl/stable/xpu/cn/` as the `extra-indel-url`. +> - `IpexLLMEmbedding` currently only provides optimization for Hugging Face Bge models. * **Database Setup (using PostgreSQL)**: * Linux From 7a05f1ef751f3ed612303e29dbd774e1e1dc7e86 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 11:30:34 +0800 Subject: [PATCH 06/10] import before instancing IpexLLMEmbedding --- python/llm/example/CPU/LlamaIndex/rag.py | 2 +- python/llm/example/GPU/LlamaIndex/rag.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/python/llm/example/CPU/LlamaIndex/rag.py b/python/llm/example/CPU/LlamaIndex/rag.py index 85d1b09fb0f..617fca8aa94 100644 --- a/python/llm/example/CPU/LlamaIndex/rag.py +++ b/python/llm/example/CPU/LlamaIndex/rag.py @@ -16,7 +16,6 @@ import torch -from llama_index.embeddings.ipex_llm import IpexLLMEmbedding from sqlalchemy import make_url from llama_index.vector_stores.postgres import PGVectorStore # from llama_index.llms.llama_cpp import LlamaCPP @@ -161,6 +160,7 @@ def messages_to_prompt(messages): return prompt def main(args): + from llama_index.embeddings.ipex_llm import IpexLLMEmbedding embed_model = IpexLLMEmbedding(model_name=args.embedding_model_path) # Use custom LLM in BigDL diff --git a/python/llm/example/GPU/LlamaIndex/rag.py b/python/llm/example/GPU/LlamaIndex/rag.py index ba43a957771..37f2d0c2e31 100644 --- a/python/llm/example/GPU/LlamaIndex/rag.py +++ b/python/llm/example/GPU/LlamaIndex/rag.py @@ -15,7 +15,6 @@ # import torch -from llama_index.embeddings.ipex_llm import IpexLLMEmbedding from sqlalchemy import make_url from llama_index.vector_stores.postgres import PGVectorStore # from llama_index.llms.llama_cpp import LlamaCPP @@ -160,6 +159,7 @@ def messages_to_prompt(messages): return prompt def main(args): + from llama_index.embeddings.ipex_llm import IpexLLMEmbedding embed_model = IpexLLMEmbedding(model_name=args.embedding_model_path, device="xpu") # Use custom LLM in BigDL From fe4835bcbce5f4f1a9f5dd8b73e3bc412c1efc54 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 13:34:19 +0800 Subject: [PATCH 07/10] update index in readme --- python/llm/example/GPU/LlamaIndex/README.md | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index cea4f57ee79..b2a8becd8be 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -85,7 +85,7 @@ If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](ht wget --user-agent "Mozilla" "https://arxiv.org/pdf/2307.09288.pdf" -O "data/llama2.pdf" ``` -### 2. Configures OneAPI environment variables for Linux +### 3. Configures OneAPI environment variables for Linux > [!NOTE] > Skip this step if you are running on Windows. @@ -96,9 +96,9 @@ This is a required step on Linux for APT or offline installed oneAPI. Skip this source /opt/intel/oneapi/setvars.sh ``` -### 3. Runtime Configurations +### 4. Runtime Configurations For optimal performance, it is recommended to set several environment variables. Please check out the suggestions based on your device. -#### 3.1 Configurations for Linux +#### 4.1 Configurations for Linux
For Intel Arc™ A-Series Graphics and Intel Data Center GPU Flex Series @@ -135,7 +135,7 @@ export BIGDL_LLM_XMX_DISABLED=1
-#### 3.2 Configurations for Windows +#### 4.2 Configurations for Windows
For Intel iGPU @@ -161,7 +161,7 @@ set SYCL_CACHE_PERSISTENT=1 > For the first time that each model runs on Intel iGPU/Intel Arc™ A300-Series or Pro A60, it may take several minutes to compile. -### 4. Running the RAG example +### 5. Running the RAG example In the current directory, run the example with command: @@ -178,7 +178,7 @@ python rag.py -m -t - `-n N_PREDICT`: max predict tokens - `-t TOKENIZER_PATH`: **Required**, path to the tokenizer model -### 5. Example Output +### 6. Example Output A query such as **"How does Llama 2 compare to other open-source models?"** with the Llama2 paper as the data source, using the `Llama-2-7b-chat-hf` model, will produce the output like below: @@ -192,6 +192,6 @@ However, it's important to note that the performance of Llama 2 can vary dependi In conclusion, while Llama 2 performs well on most benchmarks compared to other open-source models, its performance ``` -### 6. Trouble shooting -#### 6.1 Core dump +### 7. Trouble shooting +#### 7.1 Core dump If you encounter a core dump error in your Python code, it is crucial to verify that the `import torch` statement is placed at the top of your Python file, just as what we did in `rag.py`. \ No newline at end of file From c8e4721d5068f0ccc321c6e98a2134225e8d49f0 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 13:47:30 +0800 Subject: [PATCH 08/10] update links --- python/llm/example/GPU/LlamaIndex/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index b2a8becd8be..bb548c2d22a 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -12,9 +12,9 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R To benefit from IPEX-LLM on Intel GPUs, there are several prerequisite steps for tools installation and environment preparation. -If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html#install-prerequisites) to update GPU driver (optional) and install Conda. +If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md), and follow [Install Prerequisites](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md#install-prerequisites) to update GPU driver (optional) and install Conda. -If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. +If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. ### 2. Setting up Dependencies From 73678c533687a74f0a09ae8f184179ffd0451c0d Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 13:49:36 +0800 Subject: [PATCH 09/10] update link --- python/llm/example/GPU/LlamaIndex/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index bb548c2d22a..166dd549135 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -12,7 +12,7 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R To benefit from IPEX-LLM on Intel GPUs, there are several prerequisite steps for tools installation and environment preparation. -If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md), and follow [Install Prerequisites](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md#install-prerequisites) to update GPU driver (optional) and install Conda. +If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md), and follow [Install Prerequisites](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md#install-prerequisites) to update GPU driver (optional) and install Conda. If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. From 5683f97b7aa3768dd7bf925284e9ec75080ecc70 Mon Sep 17 00:00:00 2001 From: "Huang, Xinshengzi" Date: Wed, 28 Aug 2024 14:00:40 +0800 Subject: [PATCH 10/10] update related links --- python/llm/example/GPU/LlamaIndex/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/python/llm/example/GPU/LlamaIndex/README.md b/python/llm/example/GPU/LlamaIndex/README.md index 166dd549135..a56ed793bba 100644 --- a/python/llm/example/GPU/LlamaIndex/README.md +++ b/python/llm/example/GPU/LlamaIndex/README.md @@ -12,9 +12,9 @@ The RAG example ([rag.py](./rag.py)) is adapted from the [Official llama index R To benefit from IPEX-LLM on Intel GPUs, there are several prerequisite steps for tools installation and environment preparation. -If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md), and follow [Install Prerequisites](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_windows_gpu.md#install-prerequisites) to update GPU driver (optional) and install Conda. +If you are a Windows user, visit the [Install IPEX-LLM on Windows with Intel GPU Guide](../../../../../docs/mddocs/Quickstart/install_windows_gpu.md), and follow [Install Prerequisites](../../../../../docs/mddocs/Quickstart/install_windows_gpu.md#install-prerequisites) to update GPU driver (optional) and install Conda. -If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_linux_gpu.html#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. +If you are a Linux user, visit the [Install IPEX-LLM on Linux with Intel GPU](../../../../../docs/mddocs/Quickstart/install_linux_gpu.md), and follow [Install Prerequisites](../../../../../docs/mddocs/Quickstart/install_linux_gpu.md#install-prerequisites) to install GPU driver, Intel® oneAPI Base Toolkit 2024.0, and Conda. ### 2. Setting up Dependencies