Quasi-Equivariant Metanetworks

arXiv cs.LG / 4/28/2026

💬 OpinionModels & Research

Key Points

  • Metanetworks reuse pretrained weights to solve downstream tasks, but because the mapping from parameters to functions is non-injective, distinct weights can correspond to the same input-output behavior, making architectural symmetries easy to miss.
  • The paper argues that reasoning about functional identity is crucial for metanetwork design, leading to equivariant metanetworks that explicitly respect architectural symmetries.
  • Prior work typically enforces strict equivariance, which can be overly rigid and reduce expressivity by producing sparse constraints on models.
  • To overcome this, the authors propose “quasi-equivariance,” a framework that relaxes strict equivariance while still preserving functional identity, improving the symmetry–expressivity balance.
  • The approach is shown to be broadly applicable across feedforward, convolutional, and transformer architectures, with empirical results indicating favorable trade-offs and advancing theory around weight-space learning.

Abstract

Metanetworks are neural architectures designed to operate directly on pretrained weights to perform downstream tasks. However, the parameter space serves only as a proxy for the underlying function class, and the parameter-function mapping is inherently non-injective: distinct parameter configurations may yield identical input-output behaviors. As a result, metanetworks that rely solely on raw parameters risk overlooking the intrinsic symmetries of the architecture. Reasoning about functional identity is therefore essential for effective metanetwork design, motivating the development of equivariant metanetworks, which incorporate equivariance principles to respect architectural symmetries. Existing approaches, however, typically enforce strict equivariance, which imposes rigid constraints and often leads to sparse and less expressive models. To address this limitation, we introduce the novel concept of quasi-equivariance, which allows metanetworks to move beyond the rigidity of strict equivariance while still preserving functional identity. We lay down a principled basis for this framework and demonstrate its broad applicability across diverse neural architectures, including feedforward, convolutional, and transformer networks. Through empirical evaluation, we show that quasi-equivariant metanetworks achieve good trade-offs between symmetry preservation and representational expressivity. These findings advance the theoretical understanding of weight-space learning and provide a principled foundation for the design of more expressive and functionally robust metanetworks.