VoiceAssistant-Eval: Benchmarking AI Assistants across Listening, Speaking, and Viewing

10,497 examples, 13 tasks: a holistic yardstick for voice-first multimodal assistants.

Voice assistants are rapidly evolving into multimodal agents that must hear, speak, and see. Yet evaluation has lagged behind capability. VoiceAssistant-Eval fills this gap with a comprehensive benchmark of 10,497 curated examples across 13 task categories, spanning natural sounds, music, spoken dialogue (listening); multi-turn and role-play imitation (speaking); and heterogeneous images (viewing).

Twenty-one open-source models and GPT-4o-Audio are assessed for response content, speech quality, and cross-modal consistency. Three key findings emerge: (1) proprietary models do not universally dominate; (2) most models speak well but struggle with audio understanding; and (3) carefully designed smaller models can rival much larger ones. Notably, Step-Audio-2-mini (7B) more than doubles the listening accuracy of LLaMA-Omni2-32B-Bilingual.

The benchmark also surfaces hard problems: audio+visual joint reasoning and role-play voice imitation remain challenging. Robustness and safety alignment gaps persist, underscoring the need for evaluation that captures real-world edge cases and user expectations.

Why it matters: developers can finally compare systems apples-to-apples across modalities, pinpoint failure modes, and prioritize training investments. For product teams, VoiceAssistant-Eval provides measurable targets for improvements in listening comprehension, speech naturalness, and multimodal grounding.

Code and data will be released, creating a shared platform to drive the next generation of voice-first AI.

Paper: http://arxiv.org/abs/2509.22651v1

Register: https://www.AiFeta.com

#AI #Speech #Multimodal #Benchmark #LLM #Audio #ComputerVision

Read more