AiPhreaks ← Back to News Feed

A New Framework for Evaluating Voice Agents (EVA)

By Jakub Antkiewicz

2026-03-31T08:58:46Z

A team of researchers from ServiceNow-AI has released EVA, an open-source framework for evaluating conversational voice agents that simultaneously measures task accuracy and conversational experience. The new benchmark addresses a critical gap in existing evaluation methods, which typically assess an agent's ability to complete a task or the quality of its interaction, but not both at the same time. By integrating these two dimensions, EVA provides a more holistic view of an agent's real-world performance, where flawed speech recognition or unnatural pacing can render an otherwise capable system unusable.

EVA functions using a bot-to-bot architecture where a user simulator, configured with a specific goal and persona, interacts with the voice agent over a live audio stream. The framework produces two primary scores: EVA-A (Accuracy) and EVA-X (Experience). Accuracy is measured by assessing task completion against a ground truth, faithfulness to provided information to check for hallucinations, and the fidelity of the agent's spoken audio for critical details like confirmation codes. Experience is evaluated based on the conciseness of responses, the agent's ability to advance the conversation, and its turn-taking dynamics. The evaluation uses a mix of deterministic code-based checks and LLM-as-Judge models to score these qualitative and quantitative aspects.

The most significant finding from initial tests on 20 proprietary and open-source systems is a consistent tradeoff between accuracy and experience. Agents that performed well on task completion tended to provide a worse user experience, and vice versa. This suggests developers face a difficult optimization challenge that is invisible to benchmarks focused solely on task success. The results also pinpointed named entity transcription and complex, multi-step workflows as common failure points across systems. The findings indicate that building effective voice agents requires a deliberate balance between functional correctness and conversational quality, a nuance that will likely influence future development priorities in the industry.

The discovery of an 'Accuracy-Experience' tradeoff fundamentally reframes what a 'good' voice agent is. Success is no longer just about task completion, but about navigating a complex optimization problem where improving technical accuracy can directly degrade the user's conversational experience, forcing product teams to make more deliberate design choices.