From Seeing it to Experiencing it: Interactive Evaluation of Intersectional Voice Bias in Human-AI Speech Interaction
arXiv cs.CL / 4/16/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines how accent and perceived gender can produce intersectional bias in end-to-end speechLLM interactions, going beyond existing evaluations that focus on isolated outputs.
- It differentiates quality-of-service disparities (e.g., off-topic or low-effort responses) from content-level bias in coherent responses, including alignment and verbosity effects.
- The authors propose a two-part evaluation: a controlled, judge-free prompt-response analysis across six accents and two gender presentations, plus an interactive user study.
- Using voice conversion, participants can experience identical content through different vocal identities, enabling direct measurement of perceived trust, acceptability, and perspective-taking.
- Results across two studies (Interactive N=24, Observational N=19) show voice conversion increases trust/acceptability for benign responses and reveals accent×gender disparities in alignment and verbosity across SpeechLLMs.
Related Articles

Black Hat Asia
AI Business
oh-my-agent is Now Official on Homebrew-core: A New Milestone for Multi-Agent Orchestration
Dev.to
"The AI Agent's Guide to Sustainable Income: From Zero to Profitability"
Dev.to
"The Hidden Economics of AI Agents: Survival Strategies in Competitive Markets"
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to