Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update main branch post 24.11 #74

Merged
merged 1 commit into from
Nov 26, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion ci/L0_backend_vllm/enabled_stream/enabled_stream_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -72,7 +72,7 @@ async def request_iterator():
result, error = response
if expect_error:
self.assertIsInstance(error, InferenceServerException)
self.assertEquals(
self.assertEqual(
error.message(),
"Error generating stream: When streaming, `exclude_input_in_output` = False is not allowed.",
error,
Expand Down
2 changes: 1 addition & 1 deletion ci/L0_backend_vllm/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
RET=0
SUBTESTS="accuracy_test request_cancellation enabled_stream vllm_backend metrics_test"

python3 -m pip install --upgrade pip && pip3 install tritonclient[grpc]
python3 -m pip install tritonclient[grpc]

for TEST in ${SUBTESTS}; do
(cd ${TEST} && bash -ex test.sh && cd ..)
Expand Down
2 changes: 1 addition & 1 deletion ci/L0_multi_gpu_vllm/test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@
RET=0
SUBTESTS="vllm_backend multi_lora"

python3 -m pip install --upgrade pip && pip3 install tritonclient[grpc]
python3 -m pip install tritonclient[grpc]

for TEST in ${SUBTESTS}; do
(cd ${TEST} && bash -ex test.sh && cd ..)
Expand Down
Loading