JAL-Turn: Joint Acoustic-Linguistic Modeling for Real-Time and Robust Turn-Taking Detection in Full-Duplex Spoken Dialogue Systems

arXiv cs.CL / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces JAL-Turn, a lightweight speech-only turn-taking detection framework designed for industrial-grade full-duplex spoken dialogue systems where robustness and low latency are difficult to achieve.
  • JAL-Turn uses a joint acoustic-linguistic modeling approach with a cross-attention module to integrate pre-trained acoustic representations with linguistic features for fast hold-vs-shift prediction.
  • By sharing a frozen ASR encoder, the method runs turn-taking prediction fully in parallel with speech recognition, aiming to add no extra end-to-end latency or computational cost.
  • The authors also propose an automated, scalable data construction pipeline that derives turn-taking labels from large real-world dialogue corpora.
  • Experiments on multilingual public benchmarks and an in-house Japanese customer-service dataset show JAL-Turn improves turn-taking detection accuracy over strong baselines while preserving real-time performance.

Abstract

Despite recent advances, efficient and robust turn-taking detection remains a significant challenge in industrial-grade Voice AI agent deployments. Many existing systems rely solely on acoustic or semantic cues, leading to suboptimal accuracy and stability, while recent attempts to endow large language models with full-duplex capabilities require costly full-duplex data and incur substantial training and deployment overheads, limiting real-time performance. In this paper, we propose JAL-Turn, a lightweight and efficient speech-only turn-taking framework that adopts a joint acoustic-linguistic modeling paradigm, in which a cross-attention module adaptively integrates pre-trained acoustic representations with linguistic features to support low-latency prediction of hold vs shift states. By sharing a frozen ASR encoder, JAL-Turn enables turn-taking prediction to run fully in parallel with speech recognition, introducing no additional end-to-end latency or computational overhead. In addition, we introduce a scalable data construction pipeline that automatically derives reliable turn-taking labels from large-scale real-world dialogue corpora. Extensive experiments on public multilingual benchmarks and an in-house Japanese customer-service dataset show that JAL-Turn consistently outperforms strong state-of-the-art baselines in detection accuracy while maintaining superior real-time performance.