Optimal Rates for Pure {\varepsilon}-Differentially Private Stochastic Convex Optimization with Heavy Tails

arXiv cs.LG / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates stochastic convex optimization under heavy-tailed gradients while enforcing pure ε-differential privacy, replacing worst-case Lipschitz assumptions with only bounded k-th moment conditions.
  • It resolves an open problem by characterizing the minimax optimal excess-risk rate for pure ε-DP heavy-tailed SCO (up to logarithmic factors), matching the known approximate (ε,δ)-DP rates in this setting.
  • The authors propose a polynomial-time algorithm that achieves the optimal rate with high probability, and—under additional conditions such as polynomially bounded worst-case Lipschitz parameters—runs in polynomial time with probability 1.
  • For several structured loss classes (e.g., hinge/ReLU-type and absolute-value losses over Euclidean balls/ellipsoids/polytopes), the same excess-risk guarantee is obtained with polynomial-time complexity with probability 1 even when the worst-case Lipschitz constant is infinite.
  • The method is built on a new framework for privately optimizing Lipschitz extensions of the empirical loss, supported by a high-probability lower bound on excess risk.

Abstract

We study stochastic convex optimization (SCO) with heavy-tailed gradients under pure epsilon-differential privacy (DP). Instead of assuming a bound on the worst-case Lipschitz parameter of the loss, we assume only a bounded k-th moment. This assumption allows for unbounded, heavy-tailed stochastic gradient distributions, and can yield sharper excess risk bounds. The minimax optimal rate for approximate (epsilon, delta)-DP SCO is known in this setting, but the pure epsilon-DP case has remained open. We characterize the minimax optimal excess-risk rate for pure epsilon-DP heavy-tailed SCO up to logarithmic factors. Our algorithm achieves this rate in polynomial time with high probability. Moreover, it runs in polynomial time with probability 1 when the worst-case Lipschitz parameter is polynomially bounded. For important structured problem classes - including hinge/ReLU-type and absolute-value losses on Euclidean balls, ellipsoids, and polytopes - we achieve the same excess-risk guarantee in polynomial time with probability 1 even when the worst-case Lipschitz parameter is infinite. Our approach is based on a novel framework for privately optimizing Lipschitz extensions of the empirical loss. We complement our excess risk upper bound with a novel high probability lower bound.