Agent Evaluation provides three safety checkpoints for agentic AI systems: evaluate tool calls before execution, scan tool results before passing them back to the agent, and detect prompt injection in any text the agent is about to process.Documentation Index
Fetch the complete documentation index at: https://docs.responsibleailabs.ai/llms.txt
Use this file to discover all available pages before exploring further.
The three checkpoints
Tool Call Evaluation
ALLOW / FLAG / BLOCK before a tool runs. 1.5–3.0 credits.
Tool Result Scanning
PII detection + injection check on tool output. 0.5–1.0 credits.
Prompt Injection Detection
Fast injection scan on any input text. 0.5 credits.
Quick example
What’s next
API Reference: Tool Call
Full specification for tool call evaluation.
Python SDK: Agent Evaluation
Python SDK reference for all three agent endpoints.