For Coding Agents
AGNT Evaluations
Coming soon. This product is under active development and not yet available.
Test your prompts before they hit production. Run evaluation suites against prompt changes, catch regressions automatically, and maintain quality as your agents evolve.
What's planned
Test suites with cases and assertions. Define expected behaviors, output constraints, and quality thresholds. Group related test cases into suites that run together.
Automated regression testing on publish. Before a version goes live, run your test suite against it. Catch regressions before your users do. Block production deploys that fail critical assertions.
Production eval configs for continuous monitoring. Keep evaluating after deploy. Sample live traffic, run assertions against real responses, and track quality metrics over time.
Self-healing loops. When quality degrades, get specific feedback about what changed and suggested fixes. Not just "it broke" -- actionable guidance to get back on track.
Current state
Some test infrastructure already exists in Studio -- test presets, test suites, test cases, and assertions are defined in the data model. The plumbing is there. What's missing is the evaluation runtime that actually executes tests and reports results.
We're building this. If you want early access, reach out.