Rethinking Entropy Allocation in LLM-based ASR: Understanding the Dynamics between Speech Encoders and LLMs

arXiv cs.CL / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes LLM-based ASR using an “entropy allocation” lens, proposing three metrics to quantify how training reduces uncertainty across the speech encoder versus the LLM.
  • It identifies inefficiencies in current training paradigms as a key driver of tradeoffs among recognition quality, latency/overhead, and hallucination rates.
  • The authors propose a capability-boundary-aware multi-stage training strategy that (a) redesigns pretraining to reduce the speech–text modality gap and (b) uses iterative asynchronous SFT between alignment and joint SFT to prevent excessive encoder representation drift.
  • Experiments on Mandarin and English benchmarks indicate competitive performance with state-of-the-art systems while using only 2.3B parameters, alongside improved hallucination mitigation via encoder–LLM decoupling.
  • Overall, the work presents a principled training framework aimed at making LLM-based ASR more efficient and robust for real-world deployment constraints.

Abstract

Integrating large language models (LLMs) into automatic speech recognition (ASR) has become a dominant paradigm. Although recent LLM-based ASR models have shown promising performance on public benchmarks, it remains challenging to balance recognition quality with latency and overhead, while hallucinations further limit real-world deployment. In this study, we revisit LLM-based ASR from an entropy allocation perspective and introduce three metrics to characterize how training paradigms allocate entropy reduction between the speech encoder and the LLM. To remedy entropy-allocation inefficiencies in prevailing approaches, we propose a principled multi-stage training strategy grounded in capability-boundary awareness, optimizing parameter efficiency and hallucination robustness. Specifically, we redesign the pretraining strategy to alleviate the speech-text modality gap, and further introduce an iterative asynchronous SFT stage between alignment and joint SFT to preserve functional decoupling and constrain encoder representation drift. Experiments on Mandarin and English benchmarks show that our method achieves competitive performance with state-of-the-art models using only 2.3B parameters, while also effectively mitigating hallucinations through our decoupling-oriented design.