-
Notifications
You must be signed in to change notification settings - Fork 402
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Docs/examples nv ai foundation models #396
Conversation
Signed-off-by: Patricia Pampanelli <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The README for the bot is good 👍 .
However, to avoid duplicating the ABC bot config, you should restructure it as a user guide in the /docs/user_guides/llm/nvidia_ai_endpoints
folder.
The user guide should be written in a notebook and then coverted to .md (https://github.com/NVIDIA/NeMo-Guardrails/blob/develop/CONTRIBUTING.md#jupyter-notebook-documentation).
The user guide should copy the abc
configuration and only change the engine/model part and show that it works (e.g., similar to how this guide uses the step 6 from the getting started guide and makes changes: https://github.com/NVIDIA/NeMo-Guardrails/tree/develop/docs/user_guides/jailbreak_detection_heuristics).
Last but not least, if enough time, we should also evaluate the llama2_70b
and playground_mixtral_8x7b
models. See #387 for more context.
Remove nvidia ai endpoints reference Signed-off-by: Patricia Pampanelli <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should also add the config
folder for the user guide to .gitignore
so that it does not get added. I know that for other configurations we've let it in, but since the ABC bot is a bigger one, we don't want to duplicate.
from nemoguardrails.llm.providers import register_llm_provider | ||
from langchain_nvidia_ai_endpoints import ChatNVIDIA | ||
|
||
register_llm_provider("nvidia_ai_endpoints", ChatNVIDIA) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This should be registered by default, the user should not have to do this.
See: #383 for how support for Generative AI Hub is added.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done here: #403
"role": "context", | ||
"content": { | ||
"relevant_chunks": """ | ||
Employees are eligible for the following time off: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The ABC bot has the knowledge base configured, so providing explicit relevant chunks should not be needed. It should work without.
Thanks @patriciapampanelli ! Looks good. I've made a few tweaks to make this similar to the Vertex AI notebook. We'll do the evaluation in a future PR. |
No description provided.