You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
vLLM backend works well and is easy to set up, compared to TensorRT which had me pulling my hair.
However it lacks the OpenAI compatible endpoint that ships with vLLM itself.
The /generate endpoint on its own requires work to setup for chat applications (that I honestly don't know how to do).
In essence, just by adopting vLLM triton instead of vLLM, you have to develop classes and interfaces for all these things.
Not to mention that LangChain has no LLM implementation and LlamaIndex's is a bit primitive, undocumented and bugs out.
Describe the solution you'd like
Include vLLM's OpenAI compatible endpoint as an endpoint while using Triton.
Additional context
Pros:
Better integration with Langchain (through ChatOpenAI) and LlamaIndex
Triton becomes orders of magnitude easier to setup, run and migrate to (i.e you don't have to rebuild your whole toolset to accommodate Triton)
Better out-of-the-box integration with a ton of tools in the market that integrate with OpenAI compatible endpoints (eg. Langfuse, Langsmith)
It would be wonderful if it existed as a feature for all backends, but for now, with vLLM's implementation as reference, maybe that is the best starting point.
Is your feature request related to a problem? Please describe.
vLLM backend works well and is easy to set up, compared to TensorRT which had me pulling my hair.
However it lacks the OpenAI compatible endpoint that ships with vLLM itself.
The
/generate
endpoint on its own requires work to setup for chat applications (that I honestly don't know how to do).In essence, just by adopting vLLM triton instead of vLLM, you have to develop classes and interfaces for all these things.
Not to mention that LangChain has no LLM implementation and LlamaIndex's is a bit primitive, undocumented and bugs out.
Describe the solution you'd like
Include vLLM's OpenAI compatible endpoint as an endpoint while using Triton.
Additional context
Pros:
ChatOpenAI
) and LlamaIndexIt would be wonderful if it existed as a feature for all backends, but for now, with vLLM's implementation as reference, maybe that is the best starting point.
https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/serving_chat.py
https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/api_server.py
https://github.com/npuichigo/openai_trtllm/tree/main
The text was updated successfully, but these errors were encountered: