Adaptive Threshold-Driven Continuous Greedy Method for Scalable Submodular Optimization

arXiv cs.LG / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses submodular maximization under matroid constraints, highlighting that Sequential Greedy (SG) is limited to a 1/2 approximation while Continuous Greedy (CG) can achieve the (1−1/e) optimum but requires very dense decision vectors and heavy communication of feature embeddings.
  • It introduces ATCG (Adaptive Thresholded Continuous Greedy), which adaptively controls computation and communication by gating gradient evaluations and expanding each agent’s active set only when marginal gains are insufficient relative to a per-partition progress ratio.
  • The method aims to bound which feature embeddings are transmitted, reducing communication overhead while maintaining objective quality close to full CG.
  • Theoretical results provide a curvature-aware approximation guarantee using an effective factor \(\tau_{eff}=\max\{\tau,1-c\}\), showing ATCG interpolates between threshold-driven guarantees and the low-curvature regime where it matches CG performance.
  • Experiments on prototype selection (class-balanced) over a CIFAR-10 animal subset demonstrate comparable objective values to full CG with substantially reduced communication costs.

Abstract

Submodular maximization under matroid constraints is a fundamental problem in combinatorial optimization with applications in sensing, data summarization, active learning, and resource allocation. While the Sequential Greedy (SG) algorithm achieves only a \frac{1}{2}-approximation due to irrevocable selections, Continuous Greedy (CG) attains the optimal \bigl(1-\frac{1}{e}\bigr)-approximation via the multilinear relaxation, at the cost of a progressively dense decision vector that forces agents to exchange feature embeddings for nearly every ground-set element. We propose \textit{ATCG} (\underline{A}daptive \underline{T}hresholded \underline{C}ontinuous \underline{G}reedy), which gates gradient evaluations behind a per-partition progress ratio \eta_i, expanding each agent's active set only when current candidates fail to capture sufficient marginal gain, thereby directly bounding which feature embeddings are ever transmitted. Theoretical analysis establishes a curvature-aware approximation guarantee with effective factor \tau_{\mathrm{eff}}=\max\{\tau,1-c\}, interpolating between the threshold-based guarantee and the low-curvature regime where \textit{ATCG} recovers the performance of CG. Experiments on a class-balanced prototype selection problem over a subset of the CIFAR-10 animal dataset show that \textit{ATCG} achieves objective values comparable to those of the full CG method while substantially reducing communication overhead through adaptive active-set expansion.