Aligning Backchannel and Dialogue Context Representations via Contrastive LLM Fine-Tuning

arXiv cs.CL / 4/21/2026

📰 NewsModels & Research

Key Points

  • The paper studies how backchannel meaning is conveyed jointly by lexical form and prosody, beyond prior work focused mainly on backchannel timing.
  • It introduces a two-stage method that first fine-tunes large language models on dialogue transcripts to obtain contextual representations, then learns a joint embedding space linking dialogue contexts with backchannel realizations.
  • The authors evaluate learned alignment against human perception using triadic similarity judgments (including prosodic and cross-lexical similarity) and a context–backchannel suitability/fit task.
  • Results show improved context-to-backchannel retrieval over prior approaches and suggest that backchannel form is strongly influenced by extended conversational context.
  • The learned embeddings match human judgments better than using raw WavLM features, indicating the benefit of LLM-based context modeling plus contrastive fine-tuning.

Abstract

Backchannels (e.g., `yeah', `mhm', and `right') are short, non-interruptive feedback signals whose lexical form and prosody jointly convey pragmatic meaning. While prior computational research has largely focused on predicting backchannel timing, the relationship between lexico-prosodic form and meaning remains underexplored. We propose a two-stage framework: first, fine-tuning large language models on dialogue transcripts to derive rich contextual representations; and second, learning a joint embedding space for dialogue contexts and backchannel realizations. We evaluate alignment with human perception via triadic similarity judgments (prosodic and cross-lexical) and a context-backchannel suitability task. Our results demonstrate that the learned projections substantially improve context-backchannel retrieval compared to previous methods. In addition, they reveal that backchannel form is highly sensitive to extended conversational context and that the learned embeddings align more closely with human judgments than raw WavLM features.