Skip to content

Commit

Permalink
Browse files Browse the repository at this point in the history
  • Loading branch information
gc-fu authored Nov 27, 2024
1 parent cb7b089 commit 8331875
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions docker/llm/serving/xpu/docker/vllm_offline_inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,6 +54,8 @@
disable_async_output_proc=True,
distributed_executor_backend="ray",
max_model_len=2000,
trust_remote_code=True,
block_size=8,
max_num_batched_tokens=2000)
# Generate texts from the prompts. The output is a list of RequestOutput objects
# that contain the prompt, generated text, and other information.
Expand Down

0 comments on commit 8331875

Please sign in to comment.