Universal Hypernetworks for Arbitrary Models

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a Universal Hypernetwork (UHN), designed as a fixed generator that can produce weights for different target architectures using deterministic parameter, architecture, and task descriptors rather than being tied to a specific parameterization.
  • By using this descriptor-based formulation, the authors aim to decouple the hypernetwork’s design from the target model’s architecture, enabling one generator to instantiate heterogeneous models across vision, graph, text, and formula-regression tasks.
  • Experiments claim the same fixed UHN stays competitive with direct training across multiple benchmark types, while also supporting multi-model generalization within an architecture family and multi-task learning across heterogeneous models.
  • The work further reports that UHN can generate models recursively in a stable way, including up to three intermediate generated UHNs before producing the final base model.
  • The authors provide an implementation via GitHub, supporting reproducibility of the described approach and results.

Abstract

Conventional hypernetworks are typically engineered around a specific base-model parameterization, so changing the target architecture often entails redesigning the hypernetwork and retraining it from scratch. We introduce the \emph{Universal Hypernetwork} (UHN), a fixed-architecture generator that predicts weights from deterministic parameter, architecture, and task descriptors. This descriptor-based formulation decouples the generator architecture from target-network parameterization, so one generator can instantiate heterogeneous models across the tested architecture and task families. Our empirical claims are threefold: (1) one fixed UHN remains competitive with direct training across vision, graph, text, and formula-regression benchmarks; (2) the same UHN supports both multi-model generalization within a family and multi-task learning across heterogeneous models; and (3) UHN enables stable recursive generation with up to three intermediate generated UHNs before the final base model. Our code is available at https://github.com/Xuanfeng-Zhou/UHN.