Empirical Comparison of Agent Communication Protocols for Task Orchestration

arXiv cs.AI / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the lack of empirical evidence comparing two major AI agent communication approaches—tool-integration-only versus inter-agent delegation—despite growing enterprise adoption.
  • It proposes the first systematic benchmark that evaluates tool-only, delegation-only, and hybrid multi-agent architectures using standardized queries at multiple complexity levels.
  • The benchmark measures practical system trade-offs including response time, context-window usage, monetary cost, error recovery behavior, and implementation complexity.
  • The study is positioned as a foundation for selecting and designing agent orchestration architectures based on quantitative performance rather than anecdotal reports.

Abstract

Context. Nowadays, artificial intelligence agent systems are transforming from single-tool interactions to complex multi-agent orchestrations. As a result, two competing communication protocols have emerged: a tool integration protocol that standardizes how agents invoke external tools, and an inter-agent delegation protocol that enables autonomous agents to discover and delegate tasks to one another. Despite widespread industry adoption by dozens of enterprise partners, no empirical comparison of these protocols exists in the literature. Objective. The goal of this work is to develop the first systematic benchmark comparing tool-integration-only, multi-agent delegation, and hybrid architectures across standardized queries at three complexity levels, and to quantify the trade-offs in response time, context window consumption, monetary cost, error recovery, and implementation complexity.