Implementing Safety Guardrails

Learn how to use Llama Stack’s Safety API to filter potentially harmful content. Register and apply safety shields to agents, protecting both input and output through a structured, provider-based moderation system.

Generative AI models are powerful, but not without risk. They may produce harmful, offensive, biased, or unsafe content, especially when prompted with adversarial or ambiguous inputs. In many production applications, this is unacceptable.