Prompt guards
Prompt guards are security policies that inspect LLM requests and responses to detect and block harmful, policy-violating, or inappropriate content before it reaches the model or the user. You can apply prompt guards to the request phase, the response phase, or both.
Agentgateway supports the following prompt guard options:
- Regex filters: Use custom regex patterns or built-in PII detectors to reject requests or mask responses that contain sensitive data such as SSNs, email addresses, or credentials.
- AWS Bedrock Guardrails: Use AWS-managed guardrail policies to filter content based on topics, words, PII, and other safety criteria.
- Google Model Armor: Use Google Cloud’s Model Armor service to sanitize user prompts and model responses against configurable safety templates.