Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[META] Add new guardrail type based on LLM #2463

Closed
ylwu-amzn opened this issue May 21, 2024 · 0 comments
Closed

[META] Add new guardrail type based on LLM #2463

ylwu-amzn opened this issue May 21, 2024 · 0 comments
Assignees
Labels
enhancement New feature or request v2.15.0

Comments

@ylwu-amzn
Copy link
Collaborator

Currently we have only one type of guardrail which check model input/output is toxic with stop words and regex. This is limited as it can only do exact match with stop words and regex. We plan to build a new type of guardrail which can use LLM to detect if input/output is toxic.

@ylwu-amzn ylwu-amzn added enhancement New feature or request untriaged and removed untriaged labels May 21, 2024
@dhrubo-os dhrubo-os moved this to In Progress in ml-commons projects May 21, 2024
@jngz-es jngz-es mentioned this issue Jun 2, 2024
5 tasks
@github-project-automation github-project-automation bot moved this from In Progress to Done in ml-commons projects Jun 14, 2024
@github-project-automation github-project-automation bot moved this to 2.15.0 (Release window opens on June 10th, 2024 and closes on June 25th, 2024) in OpenSearch Project Roadmap Aug 30, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request v2.15.0
Projects
Status: 2.15.0 (Release window opens on June 10th, 2024 and closes on June 25th, 2024)
Development

No branches or pull requests

2 participants