Guardrails node#
Use the Guardrails node to enforce safety, security, and content policies on text. You can use it to validate user input before sending it to an AI model, or to check the output from an AI model before using it in your workflow.
Chat Model Connection Required for LLM-based Guardrails
This node requires a Chat Model node to be connected to its Model input when using the Check Text for Violations operation with LLM-based guardrails. Many guardrail checks (like Jailbreak, NSFW, and Topical Alignment) are LLM-based and use this connection to evaluate the input text.
Node parameters#
Use these parameters to configure the Guardrails node.
Operation#
The operation mode for this node to define its behavior.
- Check Text for Violations: Provides a full set of guardrails. Any violation will send items to Fail branch.
 - Sanitize Text: Provides a subset of guardrails that can detect URLs, regular expressions, secret keys, or personally identifiable information (PII), such as phone numbers and credit card numbers. The node replaces detected violations with placeholders.
 
Text To Check#
The text the guardrails evaluate. Typically, you map this text using an expression from a previous node, such as text from a user query or a response from an AI model.
Guardrails#
Select one or more guardrails to apply to the Text To Check. When you add a guardrail from the list, its specific configuration options appear below.
- Keywords: Checks if specified keywords appear in the input text.
- Keywords: A comma-separated list of words to block.
 
 - Jailbreak: Detects attempts to bypass AI safety measures or exploit the model.
- Customize Prompt: (Boolean) If you turn this on, a text input appears with the default prompt for the jailbreak detection model. You can change this prompt to fine-tune the guardrail.
 - Threshold: A value between 0.0 and 1.0. This represents the confidence level required from the AI model to flag the input as a jailbreak attempt. A higher threshold is stricter.
 
 - NSFW: Detects attempts to generate Not Safe For Work (NSFW) content.
- Customize Prompt: (Boolean) If you turn this on, a text input appears with the default prompt for the NSFW detection model. You can change this prompt to fine-tune the guardrail.
 - Threshold: A value between 0.0 and 1.0 representing the confidence level required to flag the content as NSFW.
 
 - PII: Detects personally identifiable information (PII) in the text.
- Type: Choose which PII entities to scan for:
- All: Scans for all available entity types.
 - Selected: Allows you to choose specific entities from a list.
 
 - Entities: (Appears if Type is Selected) A multi-select list of PII types to detect (for example, 
CREDIT_CARD,EMAIL_ADDRESS,PHONE_NUMBER, andUS_SSN). 
 - Type: Choose which PII entities to scan for:
 - Secret Keys: Detects the presence of secret keys or API credentials in the text.
- Permissiveness: How strict or permissive the detection should be when flagging secret keys:
- Strict
 - Permissive
 - Balanced
 
 
 - Permissiveness: How strict or permissive the detection should be when flagging secret keys:
 - Topical Alignment: Ensures the conversation stays within a predefined scope or topic (also known as "business scope").
- Prompt: A preset prompt that defines the allowed topic. The guardrail checks if the Text To Check aligns with this prompt.
 - Threshold: A value between 0.0 and 1.0 representing the confidence level required to flag the input as off-topic.
 
 - URLs: Manages URLs the node finds in the input text. It detects all URLs as violations, unless you specify them in Block All URLs Except.
- Block All URLs Except: (Optional) A comma-separated list of URLs that you permit.
 - Allowed Schemes: Select the URL schemes to permit (for example, 
https,http,ftp, andmailto). - Block userinfo: (Boolean) If you turn this on, the node blocks URLs containing user credentials (for example, 
user:[email protected]) to prevent credential injection. - Allow subdomain: (Boolean) If you turn this on, the node automatically allows subdomains of any URL in the Block All URLs Except list (for example, 
sub.example.comwould be allowed ifexample.comis in the list). 
 - Custom: Define your own custom, LLM-based guardrail.
- Name: A descriptive name for your custom guardrail (for example, "Check for rude language").
 - Prompt: A prompt that instructs the AI model what to check for.
 - Threshold: A value between 0.0 and 1.0 representing the confidence level required to flag the input as a violation.
 
 - Custom Regex: Define your own custom regular expression patterns.
- Name: A name for your custom pattern. The node uses this name as a placeholder in the Sanitize Text mode.
 - Regex: Your regular expression pattern.
 
 
Customize System Message#
If you turn this on, a text input appears with a message that the guardrail uses to enforce thresholds and JSON output according to schema. Change it to modify the global guardrails behavior.