-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Misc]: Can we remove vllm/entrypoints/api_server.py
?
#3852
Comments
It's probably a good idea. However, it would probably be better to move it into a subfolder like vllm_example. It's possible that in the future, it could be useful to create servers implementation of Mistral AI API or Anthropic. |
We can leave the openai server where it is so that in future if other proprietary APIs are added they can live adjacent to the openai directory. |
vllm.entrypoints/api_server.py
?vllm/entrypoints/api_server.py
?
There exist outstanding issues which suggest that there are some bugs in that file.
Yet, #2858 clearly mentions that the file is not supposed to be further updated. @simon-mo what are the reasons behind this? If we want this file to serve as an example, IMO we should at least make sure that it is bug-free. |
We can fix bug. We just shouldn't add more features to keep it with parity for OpenAI compatibility. |
Does |
Perhaps we should move it to |
Maybe this could be kept for supporting features in beta/experimental or something that's not covered by open ai api spec. For example this was the only way I found to send For example mistral v3 function calling requires custom tokenization that isn't covered by chat template is one such example and I think it could be useful debugging other OSS models that come with their own prompt templates that sometimes may not be optimal for whatever reason. Note: I might be missing the docs regarding sending tokens directly to open ai api spec but I didn't find anything regarding that. |
This is a valid point.
We kept the file there for backward compatibility as some people are still using it. |
Correct me if I'm wrong, but it looks like we can't use this file directly to test out those features. If we have to copy this file and modify it elsewhere anyway, I don't see the harm in moving it to the examples and thus adding it to the official documentation. |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
Anything you want to discuss about vllm.
While gardening GitHub issues I've seen many issues where the user is using
-m vllm.entrypoints.api_server
, which indicates that users are either not aware of ignoring the note at the top of the file:vllm/vllm/entrypoints/api_server.py
Lines 1 to 7 in b778200
Can we remove it to avoid future confusion?
The text was updated successfully, but these errors were encountered: