Skip to content

Commit

Permalink
Update fastchat demo script (#12367)
Browse files Browse the repository at this point in the history
* Update README.md

* Update vllm_docker_quickstart.md
  • Loading branch information
liu-shaojun authored Nov 8, 2024
1 parent 51f7f87 commit fad15c8
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 2 deletions.
4 changes: 3 additions & 1 deletion docker/llm/serving/xpu/docker/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -102,7 +102,9 @@ To set up model serving using `IPEX-LLM` as backend using FastChat, you can refe
# start controller
python -m fastchat.serve.controller &
export USE_XETLA=OFF
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
export TORCH_LLM_ALLREDUCE=0
export CCL_DG2_ALLREDUCE=1
Expand Down
4 changes: 3 additions & 1 deletion docs/mddocs/DockerGuides/vllm_docker_quickstart.md
Original file line number Diff line number Diff line change
Expand Up @@ -831,7 +831,9 @@ We can set up model serving using `IPEX-LLM` as backend using FastChat, the foll
# start controller
python -m fastchat.serve.controller &
export USE_XETLA=OFF
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=2
export TORCH_LLM_ALLREDUCE=0
export CCL_DG2_ALLREDUCE=1
Expand Down

0 comments on commit fad15c8

Please sign in to comment.