Distributed Online Convex Optimization with Compressed Communication: Optimal Regret and Applications

arXiv cs.LG / 4/13/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies distributed online convex optimization (D-OCO) under compressed communication between local learners and a central server to reduce communication bottlenecks in large-scale streaming settings.
  • It derives lower bounds showing that compression can degrade regret, with different rates for convex versus strongly convex losses in terms of the compression ratio δ and time horizon T.
  • It introduces an “optimal” algorithm that achieves regret bounds of O(δ^{-1/2}√T) for convex losses and O(δ^{-1}log T) for strongly convex losses.
  • The method combines an error-feedback mechanism with Follow-the-Regularized-Leader to control interaction between compression error and projection error, and uses online compression to prevent error accumulation from bidirectional compression.
  • The work extends to an offline stochastic setting via online-to-batch conversion and provides convergence rates for non-smooth optimization with compressed communication and domain constraints.

Abstract

Distributed online convex optimization (D-OCO) is a powerful paradigm for modeling distributed scenarios with streaming data. However, the communication cost between local learners and the central server is substantial in large-scale applications. To alleviate this bottleneck, we initiate the study of D-OCO with compressed communication. Firstly, to quantify the compression impact, we establish the \Omega(\delta^{-1/2}\sqrt{T}) and \Omega(\delta^{-1}\log{T}) lower bounds for convex and strongly convex loss functions, respectively, where \delta \in (0,1] is the compression ratio. Secondly, we propose an optimal algorithm, which enjoys regret bounds of O(\delta^{-1/2}\sqrt{T}) and O(\delta^{-1} \log T) for convex and strongly convex loss functions, respectively. Our method incorporates the error feedback mechanism into the Follow-the-Regularized-Leader framework to address the coupling between the compression error and the projection error. Furthermore, we employ the online compression strategy to mitigate the accumulated error arising from the bidirectional compression. Our online method has great generality, and can be extended to the offline stochastic setting via online-to-batch conversion. We establish convergence rates of O(\delta^{-1/2}T^{-1/2}) and O(\delta^{-1} T^{-1}) for convex and strongly convex loss functions, respectively, providing the first guarantees for distributed non-smooth optimization with compressed communication and domain constraints.