From 10585e035ec564cd376146c3fe5ffe427a43c92c Mon Sep 17 00:00:00 2001 From: Robert Shaw <114415538+robertgshaw2-neuralmagic@users.noreply.github.com> Date: Fri, 15 Mar 2024 19:35:36 -0500 Subject: [PATCH] Removed Extraneous Print Message From OAI Server (#3440) --- vllm/entrypoints/openai/serving_completion.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/vllm/entrypoints/openai/serving_completion.py b/vllm/entrypoints/openai/serving_completion.py index bfd7c9b50cf32..5f2be878a7b76 100644 --- a/vllm/entrypoints/openai/serving_completion.py +++ b/vllm/entrypoints/openai/serving_completion.py @@ -309,10 +309,7 @@ async def completion_stream_generator( except ValueError as e: # TODO: Use a vllm-specific Validation Error data = self.create_streaming_error_response(str(e)) - print("yield", f"data: {data}\n\n") yield f"data: {data}\n\n" - - print("yield", "data: [DONE]\n\n") yield "data: [DONE]\n\n" def request_output_to_completion_response(