Smoothing the Edges: Smooth Optimization for Sparse Regularization using Hadamard Overparametrization

arXiv stat.ML / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a framework for smoothing explicitly regularized optimization problems that target structured sparsity, which are usually non-smooth and may be non-convex.
  • It makes the optimization fully differentiable and compatible with standard gradient descent by using a Hadamard overparametrization of selected parameters plus a change in penalties.
  • The authors prove that the smooth surrogate objective is equivalent to the original sparse regularization objective by matching both global minima and local minima, preventing spurious solutions.
  • Beyond sparse regularization, the theory also yields general results about matching local minima for arbitrary objectives, even when those objectives are not explicitly regularized.
  • The work includes a review of sparsity-inducing parametrizations, theoretical extensions and improvements, and numerical experiments showing effectiveness across sparse learning tasks including sparse neural network training.

Abstract

We present a framework for smooth optimization of explicitly regularized objectives for (structured) sparsity. These non-smooth and possibly non-convex problems typically rely on solvers tailored to specific models and regularizers. In contrast, our method enables fully differentiable and approximation-free optimization and is thus compatible with the ubiquitous gradient descent paradigm in deep learning. The proposed optimization transfer comprises an overparameterization of selected parameters and a change of penalties. In the overparametrized problem, smooth surrogate regularization induces non-smooth, sparse regularization in the base parametrization. We prove that the surrogate objective is equivalent in the sense that it not only has identical global minima but also matching local minima, thereby avoiding the introduction of spurious solutions. Additionally, our theory establishes results of independent interest regarding matching local minima for arbitrary, potentially unregularized, objectives. We comprehensively review sparsity-inducing parametrizations across different fields that are covered by our general theory, extend their scope, and propose improvements in several aspects. Numerical experiments further demonstrate the correctness and effectiveness of our approach on several sparse learning problems ranging from high-dimensional regression to sparse neural network training.