From Text to Talk: Audio-Language Model Needs Non-Autoregressive Joint Training

arXiv cs.CL / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current text-audio multimodal models often use autoregressive (AR) approaches even though text and audio should be modeled differently due to their distinct dependency structures (target-target vs source-target).
  • It proposes Text-to-Talk (TtT), a unified Transformer framework that combines AR text generation with non-autoregressive (NAR) audio diffusion to enable joint training under a single objective.
  • The method leverages “absorbing discrete diffusion” and introduces a modality-aware attention mechanism that enforces causal decoding for text while allowing bidirectional modeling within audio spans.
  • To reduce train-test discrepancies, the authors add three training strategies and use block-wise parallel diffusion at inference to synthesize audio efficiently for variable-length outputs.
  • Experiments across Audio-QA, ASR, AAC, and speech-to-speech benchmarks reportedly outperform strong AR and NAR baselines, with ablations supporting the contributions of each component.

Abstract

Recent advances in large language models (LLMs) have attracted significant interest in extending their capabilities to multimodal scenarios, particularly for speech-to-speech conversational systems. However, existing multimodal models handling interleaved audio and text rely on autoregressive (AR) methods, overlooking that text depends on target-target relations whereas audio depends mainly on source-target relations. In this work, we propose Text-to-Talk (TtT), a unified audio-text framework that integrates AR text generation with non-autoregressive (NAR) audio diffusion in a single Transformer. By leveraging the any-order AR property of absorbing discrete diffusion, our approach provides a unified training objective for text and audio. To support this hybrid generation paradigm, we design a modality-aware attention mechanism that enforces causal decoding for text while allowing bidirectional modeling within audio spans, and further introduce three training strategies that reduce train-test discrepancies. During inference, TtT employs block-wise diffusion to synthesize audio in parallel while flexibly handling variable-length outputs. Comprehensive experiments on Audio-QA, ASR, AAC and speech-to-speech benchmarks show that TtT consistently surpasses strong AR and NAR baselines, with additional ablation and training-strategy analyses confirming the contribution of each component. We will open-source our models, data and code to facilitate future research in this direction.