Functional Similarity Metric for Neural Networks: Overcoming Parametric Ambiguity via Activation Region Analysis

arXiv cs.LG / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses representational ambiguity in ReLU neural networks, where different weight configurations can implement the same functions due to neuron permutation and positive diagonal scaling symmetries.
  • It introduces a stable canonical form and a functional similarity metric that compares networks by analyzing activation-region topology rather than raw parameters.
  • The method removes scaling ambiguity using L2-normalization of weight vectors with layer compensation, then creates discrete activation-region “signatures” by evaluating binary patterns over sampled data.
  • To efficiently compare large binary signatures, it uses Locality-Sensitive Hashing with MinHash to approximate the Jaccard index, and performs cross-network neuron matching via the Hungarian algorithm.
  • The authors show theoretically and experimentally that the metric reduces neuron “flickering” and remains robust under small weight perturbations, supporting model merging, pruning assessment, transfer learning, and Explainable AI.

Abstract

As modern deep learning architectures grow in complexity, representational ambiguity emerges as a critical barrier to their interpretability and reliable merging. For ReLU networks, identical functional mappings can be achieved through entirely different weight configurations due to algebraic symmetries: neuron permutation and positive diagonal scaling. Consequently, traditional parameter-based comparison methods exhibit extreme instability to slight weight perturbations during training. This paper proposes a mathematically grounded approach to constructing a stable canonical representation of neural networks and a robust functional similarity metric. We shift focus from comparing raw weights to analyzing the topology of neuron activation regions. The algorithm first eliminates scaling ambiguity via L2-normalization of weight vectors with subsequent layer compensation. Next, discrete approximations of activation regions are generated as binary functional signatures evaluated over a data sample. To overcome the computational bottleneck of comparing large binary vectors, we adapt Locality-Sensitive Hashing, specifically MinHash, providing a fast and statistically precise approximation of the Jaccard index. The final cross-network neuron matching is formulated as a linear sum assignment problem solved via the Hungarian algorithm. We demonstrate theoretically and experimentally that our metric mitigates the neuron "flickering" effect and exhibits exceptional robustness to minor weight perturbations. This framework provides a solid foundation for model merging, transfer learning, objective assessment during pruning, and Explainable AI paradigms.

Functional Similarity Metric for Neural Networks: Overcoming Parametric Ambiguity via Activation Region Analysis | AI Navigate