-
-
Notifications
You must be signed in to change notification settings - Fork 5.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
early_stopping
potentially not working via api request
#2938
Comments
Additionally, I can set the value of |
Upon further inspection, it looks like it has just been left out of the completion/chat-completion apis, possibly an oversight?: https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/protocol.py#L56 |
Opened a PR with a potential quick 4-line fix, let me know if it looks like anything is missing! #2939 |
It's not an oversight originally because it is not part of the official API https://platform.openai.com/docs/api-reference/chat/create But it does seem needed. Thank you for your PR |
|
Closed by #2939 |
While using
v0.3.1
,early_stopping
will not toggle toTrue
due to an omission in the protocol definition (see below comments). I am prompting like this:and this is what I get on server side:
Everything else I set is there, but the early stopping isn't, even though none of my other options should be incompatible:
vllm/vllm/sampling_params.py
Line 106 in 264017a
The text was updated successfully, but these errors were encountered: