Skip to main content
BETAThis is a new design — give feedback

Guardrails

Also known as: Safety Guardrails, AI Guardrails, Policy Guardrails

In one sentence

Rules and filters that prevent AI from generating harmful, biased, or inappropriate content. They act as safety boundaries that keep AI systems operating within acceptable limits.

Explain like I'm 12

Guardrails are like the bumpers at a bowling alley—they stop the ball from going into the gutter. For AI, they stop it from saying dangerous, offensive, or completely wrong things.

In context

When you use ChatGPT and it refuses to explain how to make something dangerous, that's a guardrail in action. Companies add their own guardrails too—a customer service bot might be programmed to never discuss competitor products, always escalate suicide mentions to a human, and avoid giving medical or legal advice. Guardrails can work at the input level (blocking certain requests), the output level (filtering responses), or both.

See also

Related Guides

Learn more about Guardrails in these guides: