Unveiling Hidden Convexity in Deep Learning: a Sparse Signal Processing Perspective

arXiv cs.LG / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that while deep neural networks (especially those using ReLU) have non-convex loss functions, recent work reveals “hidden convexities” in certain architectures’ loss landscapes.
  • It focuses on convex equivalences of ReLU networks and explains how these can connect deep learning training/understanding to sparse signal processing formulations.
  • The authors aim to provide an accessible, educational overview linking mathematical advances in deep learning with classical signal processing perspectives.
  • The discussion highlights existing results for two-layer ReLU networks and suggests similar convex-structure insights may extend to other deeper or architecturally varied models.

Abstract

Deep neural networks (DNNs), particularly those using Rectified Linear Unit (ReLU) activation functions, have achieved remarkable success across diverse machine learning tasks, including image recognition, audio processing, and language modeling. Despite this success, the non-convex nature of DNN loss functions complicates optimization and limits theoretical understanding. In this paper, we highlight how recently developed convex equivalences of ReLU NNs and their connections to sparse signal processing models can address the challenges of training and understanding NNs. Recent research has uncovered several hidden convexities in the loss landscapes of certain NN architectures, notably two-layer ReLU networks and other deeper or varied architectures. This paper seeks to provide an accessible and educational overview that bridges recent advances in the mathematics of deep learning with traditional signal processing, encouraging broader signal processing applications.