-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Triton LangChain Provider #6655
Comments
Thank you @peterableda for submitting a feature request! I have filed a ticket for us to discuss and investigate further. |
@npuichigo has posted this on the TRT-LLM discussion. |
Welcome to have a try with https://github.com/npuichigo/openai_trtllm if anyone want an openai-like API to use with LangChain. |
This seems to be the same/v similar request: langchain-ai/langchain#13975 |
Closing issue due to inactivity. Please reopen if you would like to follow up with this issue. |
Is your feature request related to a problem? Please describe.
LangChain is one of the most widely used libraries to build LLM based applications with a wide range of integrations to LLM providers. Triton Inference Server should be supported within that community.
I would like to host open source LLMs from HuggingFace in Triton as a Coding Assistant for JupyterLab. I would use the Jupyter-AI extension that already integrates with a wide range of LLM Providers, it relies on LangChain to abstract the interactions.
Describe the solution you'd like
Build an NVIDIA Triton Inference Server LangChain Provider.
Additional context
A similar request was raised in the LangChain community: langchain-ai/langchain#12474
I also saw LangChain coming up as an alternative solution in the #6583 request.
The text was updated successfully, but these errors were encountered: