-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for vLLM and TRT-LLM running in OpenAI compatible mode #6583
Comments
Thanks for submitting feature request! CC @nnshah1 on the request for starting server with the OpenAI compatible API. For the client side, we have introduced the generate endpoint. which is an OpenAI-like endpoint support and can improve Triton adoption for LLM use cases. |
@vecorro It's a bit challenging as the OpenAI API is a moving target and already the completion api is legacy and out of date with many applications. As triton as an inference server serves many types of inference we've focused on providing ways to customize endpoints also with the expectation that many actual deployments would use an additional api gateway to translate from a service api to the inference api (for example langchain). that being said - your perspective is important. Is the legacy completion api enough or would we need to support the newer apis as well? Do you think integration with a project like langchain or other llm projects would be a viable approach or would triton need to provide the interface directly? |
The key here, to be usefull to the community, is actuality & compatibility. every other often used tool, like LiteLLM (middlware/API-Proxy) or loaders, like ooba's TGI, vLLM, FastChat are on the run to implement IMHO: I have been testing all these and just had a peek into triton, but already assume, it would be the much supirior solution of running multiple LLMs locally - IF there would be an out of the box compatibility with the openAI-API. |
I agree with @nnshah1; the key is actuality and compatibility, and I would add ease of use. For experimentation/ded purposes, vLLM is excellent as it allows you to try an LLM without the complexity and time required by TensorRT-LLM. Making the existing vLLM implementation of the legacy OpenAI API can be good enough; meanwhile, vLLM evolves to support newer versions of the OpenAI API. I'm on the VMware team working with NVIDIA on the private AI initiative. From conversations with customers looking for ways to run LLMs on-premises, they are getting started on that journey, so a quick drop-in replacement of OpenAI's models would facilitate things for them. Thanks all! |
@npuichigo has mentioned an integration that could be useful here: |
https://github.com/npuichigo/openai_trtllm provides an OpenAI-like API for trtllm triton backend, but I think vllm in triton would be something alike. |
Closing issue due to inactivity. Please reopen if you would like to follow up with this issue. |
Hello, Seems it's supporting vllm now: But can we use the vllm OpenAI APIs ? Thanks a lot |
We don't currently support it directly - but are still thinking through the best ways to add a compatible API in a way that we can maintain. |
Thank you for the clearance~~ |
is there any workaround so far ?
|
Is this planned for any specific release? |
@anubhav-agrawal-mu-sigma - we are currently planning a tutorial showcasing how to create an open ai compatible triton server using triton's in-process python api - that tutorial is planned for September time frame. |
@nnshah1 : I am interested in the mentioned tutorial, please let me know when you have it online. Many thanks ! |
@nnshah1 I’m excited to see the tutorial! Would you please to let me know when it will be available? Thanks in advance! |
@nnshah1 I'm interested too! |
You can follow along the current preview PR here: Some things may change- so please consider this as BETA until we finalize a few of the internal structure - but this will be the basis. |
Is your feature request related to a problem? Please describe.
I'd like to be able to run vLLM emulating the OpenAI compatible API to use vLLM as a drop-in replacement of ChatGPT.
Describe the solution you'd like
I'd like Triton allow me run vLLM as indicated in vLLM documentation
Example:
python -m vllm.entrypoints.openai.api_server --model facebook/opt-125m
Describe alternatives you've considered
It is possible to use the REST API, however, for developers already leveraging OpenAI and serving open-source LLMs using the OpenAI API would allow a faster path for replacement of OpenAI
The text was updated successfully, but these errors were encountered: