Most ReLU Networks Admit Identifiable Parameters

arXiv cs.LG / 5/6/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies when deep ReLU networks are identifiable, asking under what conditions a realized function determines the network parameters up to only standard symmetries like scaling and permutation.
  • It introduces a new analytical framework using weighted polyhedral complexes to go beyond hidden redundancies and characterize identifiability more precisely.
  • The main theorem states that for architectures where both input and hidden layers have width at least two, there is an open set of parameters with identifiable representations, leading to an exact functional dimension formula.
  • The authors show that even minimal functional representations can still exhibit non-trivial parameter redundancies, and they prove a generic depth hierarchy: for an open set of parameters, the realized function cannot be generically represented by any shallower network.
  • Overall, the results sharpen our understanding of the relationship between network architecture, parameter redundancy, and expressive power across depths for ReLU models.

Abstract

We study the realization map of deep ReLU networks, focusing on when a function determines its parameters up to scaling and permutation. To analyze hidden redundancies beyond these standard symmetries, we introduce a framework based on weighted polyhedral complexes. Our main result shows that for every architecture whose input and hidden layers have width at least two, there exists an open set of identifiable parameters. This implies that the functional dimension of every such architecture is exactly the number of parameters minus the number of hidden neurons. We further show that minimal functional representations can still have non-trivial parameter redundancies. Finally, we establish a generic depth hierarchy, whereby for an open set of parameters the realized function cannot be represented generically by any shallower network.