You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently we have only one type of guardrail which check model input/output is toxic with stop words and regex. This is limited as it can only do exact match with stop words and regex. We plan to build a new type of guardrail which can use LLM to detect if input/output is toxic.
The text was updated successfully, but these errors were encountered:
Currently we have only one type of guardrail which check model input/output is toxic with stop words and regex. This is limited as it can only do exact match with stop words and regex. We plan to build a new type of guardrail which can use LLM to detect if input/output is toxic.
The text was updated successfully, but these errors were encountered: