A Delta-Aware Orchestration Framework for Scalable Multi-Agent Edge Computing

arXiv cs.LG / 4/23/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper identifies a “Synergistic Collapse” in large-scale multi-agent edge deployments where scaling beyond 100 agents triggers superlinear performance degradation that individual optimizations can’t stop.
  • Using a Smart City case with 150 cameras and MADDPG, it reports Deadline Satisfaction falling from 78% to 34%, leading to roughly $180,000 in annual cost overruns.
  • It proposes DAOEF (Delta-Aware Orchestration for Edge Federations), which combines Differential Neural Caching, Criticality-Based Action Space Pruning, and Learned Hardware Affinity Matching to address exponential action-space growth, redundant computation, and task-agnostic scheduling together.
  • Experiments show that each component is necessary but not sufficient alone, while the full DAOEF framework delivers a multiplicative improvement (1.45x) and significant latency reductions (62% in a 200-agent cloud deployment, with sub-linear growth up to 250 agents).
  • The authors also validate the approach on both datasets (100–250 agents) and a 20-device physical testbed, demonstrating robustness beyond purely simulated settings.

Abstract

The Synergistic Collapse occurs when scaling beyond 100 agents causes superlinear performance degradation that individual optimizations cannot prevent. We observe this collapse with 150 cameras in Smart City deployment using MADDPG, where Deadline Satisfaction drops from 78% to 34%, producing approximately $180,000 in annual cost overruns. Prior work has addressed each contributing factor in isolation: exponential action-space growth, computational redundancy among spatially adjacent agents, and task-agnostic hardware scheduling. None has examined how these three factors interact and amplify each other. We present DAOEF (Delta-Aware Orchestration for Edge Federations), a framework that addresses all three simultaneously through: (1) Differential Neural Caching, which stores intermediate layer activations and computes only the input deltas, achieving 2.1x higher hit ratios (72% vs. 35%) than output-level caching while staying within 2% accuracy loss through empirically calibrated similarity thresholds; (2) Criticality-Based Action Space Pruning, which organizes agents into priority tiers and reduces coordination complexity from O(n2) to O(n log n) with less than 6% optimality loss; and (3) Learned Hardware Affinity Matching, which assigns tasks to their optimal accelerator (GPU, CPU, NPU, or FPGA) to prevent compounding mismatch penalties. Controlled factor-isolation experiments confirm that each mechanism is necessary but insufficient on its own: removing any single mechanism increases latency by more than 40%, validating that the gains are interdependent rather than additive. Across four datasets (100-250 agents) and a 20-device physical testbed, DAOEF achieves a 1.45x multiplicative gain over applying the three mechanisms independently. A 200-agent cloud deployment yields 62% latency reduction (280 ms vs. 735 ms), sub-linear latency growth up to 250 agents.