Guardian Agents monitor and protect AI agents to prevent errors and unauthorized actions in multi-agent systems.

Who watches the AI? Other AI.
What It Is
When AI agents work independently, someone needs to watch. And that someone is increasingly AI itself. Gartner predicts that Guardian Agents will capture 10-15% of the entire agentic AI market by 2030.
Three types:
- Reviewer: Checks the output of other agents - are the results correct?
- Monitor: Observes the behavior of other agents in real-time - is everything running as planned?
- Protector: Blocks unwanted or unauthorized actions - stop, you're not allowed to do that
Avivah Litan from Gartner puts it plainly: "Humans cannot keep up with the potential for errors and malicious activities" in multi-agent systems. The speed and volume of agent work simply exceeds human control capacity.
How To Spot It
- AI agents occasionally produce results that nobody has reviewed - and nobody notices
- An agent makes decisions outside its intended scope
- The team relies on manual spot checks instead of systematic control
What To Do (FL2 - Coordination Level)
- Triple-secure: Don't rely on just one control mechanism - think Review + Monitoring + Protection together
- Ensure independence: Guardian agents must not share the same weaknesses as the agents they oversee
- Connect with [[Humans Above the Loop]]: Guardian agents are the FL2 layer between strategic steering (FL3) and operational execution (FL1)
- Connect with [[Autonomy Gradient]]: The higher an agent's autonomy level, the more important the guardian layer becomes
The Trap
Who watches the watchers? Guardian agents can themselves hallucinate, miss errors, or be biased in their evaluation. The answer is not infinite nesting (guards guarding guards), but deliberate diversity of control mechanisms - and at the right points: humans.