
AI Agent Testing Reveals Critical Gaps: Why 90% Success Isn't Good Enough
May 23, 2025
A recent WIRED investigation into AI agent testing has exposed an uncomfortable truth that every company racing to deploy autonomous AI should read twice. When veteran software engineer Jay Prakash Thakur conducted AI agent evaluations on his prototypes, 9 out of 10 times everything worked perfectly. But that 10% failure rate in AI agent testing? It's exactly why we're not ready for the autonomous future we're promising.
The Stakes Are Higher Than "Extra Onions"
The article's examples might sound almost comical at first—"onion rings" becoming "extra onions," agents linking to the wrong product pages, or HR bots approving leave requests they should deny. But strip away the seemingly benign scenarios and you're left with a sobering reality: we're building systems that fail in unpredictable ways, and we have no systematic way to catch these failures before they reach customers.
Consider what Thakur's AI agent testing revealed when his restaurant ordering system processed complex requests:
Orders with more than five items regularly failed
Food allergies could be mishandled
Price comparisons linked to wrong products
Now scale this to Gartner's prediction that agentic AI will handle 80% of customer service queries by 2029. Without proper voice and chat AI evals, we're not just talking about frustrated customers—we're talking about potential financial damage, safety risks, and legal liability on an unprecedented scale.
The Legal Powder Keg
OpenAI's senior legal counsel Joseph Fireman made a telling observation at a recent conference: when AI agents cause harm, aggrieved parties will "go after those with the deepest pockets." This isn't speculation—it's already happening. Airlines are being held legally responsible for coupons their chatbots invented. Legal firms are apologizing to judges for AI-generated citations.
But here's the terrifying part: as agent systems become more complex, pinpointing responsibility becomes nearly impossible. Thakur compared debugging multi-agent failures to "reconstructing a conversation based on different people's notes." When multiple agents from different companies interact in a single system, who's liable when things go wrong?
Why Current AI Agent Testing Approaches Are Insufficient
The industry's response to these AI agent evaluation challenges has been predictably superficial:
Add more human oversight (defeating the purpose of automation)
Create "judge" agents to monitor other agents (adding complexity, not solving reliability)
Hope that insurance will cover the inevitable mistakes
This is backwards thinking. We're trying to fix reliability problems after deployment instead of ensuring reliability through comprehensive AI agent testing before deployment.
The AI Agent Evals, Simulation, and Monitoring Imperative
The path forward requires a fundamentally different approach; one that treats AI agent testing as an engineering discipline, not an afterthought. This means three critical capabilities for comprehensive AI agent evaluations:
1. Advanced AI Agent Testing Through Simulation
Before any agent system touches a real customer, it needs comprehensive AI agent testing against thousands of edge cases in controlled environments. Thakur's restaurant prototype should have undergone extensive AI agent evals with complex orders, allergy scenarios, and adversarial inputs long before he discovered the 10% failure rate.
2. Rigorous AI Agent Evaluation Frameworks
We need standardized AI agent evals that go beyond "does it work most of the time?" Effective voice and chat AI evals must measure:
Error propagation in multi-agent systems
Graceful degradation under stress
Boundary condition handling
Cross-agent communication fidelity
Performance across different conversation types and user intents
3. Real-Time Monitoring and Circuit Breakers
Once deployed, agent systems need continuous monitoring with automatic safeguards. When an agent starts making unusual decisions or error rates spike, the system should gracefully degrade or halt operations rather than continue causing damage.
Why Coval Is Leading AI Agent Testing Innovation
At Coval, we've seen this AI agent evaluation challenge coming from miles away. While the industry races to build more complex agent systems, we've been focused on the critical AI agent testing infrastructure needed to make them reliable.
Our team combines deep expertise in:
AI safety and alignment research - understanding how and why AI systems fail in AI agent evals
Large-scale AI agent testing - building simulation environments that catch edge cases before they become customer problems costing you millions of dollars
Real-time voice and chat AI evals - creating monitoring systems that track agent performance across different interaction types
Advanced AI agent evaluation frameworks - developing comprehensive testing protocols that ensure reliability at scale
We're not building more agents, we're building the AI agent testing and evaluation infrastructure that makes agents trustworthy for production use.
The Future of AI Agent Testing
Attorney Dazza Greenwood put it perfectly: "If you have a 10 percent error rate with 'add onions,' that to me is nowhere near release. Work your systems out so that you're not inflicting harm on people to start with."
The AI agent revolution is inevitable, but it doesn't have to be reckless. Companies that invest in comprehensive AI agent testing, robust AI agent evaluations, and continuous voice and chat AI evals today will be the ones customers trust tomorrow. Those that don't will be the ones writing apology letters to judges.
The choice is simple: we can either engineer reliability into AI agents through rigorous AI agent testing from the ground up, or we can wait for the inevitable failures to teach us why we should have.
At Coval, we're betting on engineering excellence in AI agent evals. Because when AI agents are handling 80% of customer interactions by 2029, "sorry, the AI made a mistake" won't be good enough.
Ready to build AI agents that customers can actually trust? Let's talk about how Coval's AI agent testing and evaluation platform can help you catch the 10% before your customers do. Contact us to learn more about our comprehensive voice and chat AI evals solutions.