Learning to Interrupt in Language-based Multi-agent Communication

arXiv cs.CL / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how LLM-based multi-agent communication can be made less verbose and more cost-effective by enabling listeners to interrupt speakers when clarification or opinions are needed.
  • It argues that prior message-compression approaches often fail to adapt to different listeners and to identify what information is actually relevant in context.
  • The authors propose an interruptible communication framework (HANDRAISER) that learns when to interrupt by predicting appropriate interruption points using estimated future reward and communication cost.
  • Experiments across text pictionary (2 agents), meeting scheduling (3 agents), and debate (3 agents) show a 32.2% reduction in communication cost versus a baseline while maintaining comparable or better task performance.
  • The learned interruption policy generalizes across different agent configurations and tasks, indicating the approach can transfer beyond a single setup.

Abstract

Multi-agent systems using large language models (LLMs) have demonstrated impressive capabilities across various domains. However, current agent communication suffers from verbose output that overload context and increase computational costs. Although existing approaches focus on compressing the message from the speaker side, they struggle to adapt to different listeners and identify relevant information. An effective way in human communication is to allow the listener to interrupt and express their opinion or ask for clarification. Motivated by this, we propose an interruptible communication framework that allows the agent who is listening to interrupt the current speaker. Through prompting experiments, we find that current LLMs are often overconfident and interrupt before receiving enough information. Therefore, we propose a learning method that predicts the appropriate interruption points based on the estimated future reward and cost. We evaluate our framework across various multi-agent scenarios, including 2-agent text pictionary games, 3-agent meeting scheduling, and 3-agent debate. The results of the experiment show that our HANDRAISER can reduce the communication cost by 32.2% compared to the baseline with comparable or superior task performance. This learned interruption behavior can also be generalized to different agents and tasks.