MAGNET: Autonomous Expert Model Generation via Decentralized Autoresearch and BitNet Training

arXiv cs.AI / 3/30/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • MAGNET is proposed as a decentralized framework that can autonomously generate, train, and serve domain-expert language models on commodity hardware using multiple integrated components.
  • The system’s autoresearch pipeline automates end-to-end ML research tasks, including dataset generation, hyperparameter search, evaluation, and error-driven iteration, and is validated via three case studies.
  • MAGNET introduces BitNet b1.58 ternary training intended to enable CPU-native inference (via bitnet.cpp) without requiring GPU hardware, and reports measurable validation-loss improvements through hyperparameter optimization.
  • It combines DiLoCo-based distributed merging to aggregate “domain specialist” models efficiently and uses on-chain contribution tracking on the HOOTi EVM chain to document inputs.
  • Reported results span video safety classification performance gains, improved cryptocurrency directional prediction hit rate, and quantified loss reduction from an automated BitNet hyperparameter sweep.

Abstract

We present MAGNET (Model Autonomously Growing Network), a decentralized system for autonomous generation, training, and serving of domain-expert language models across commodity hardware. MAGNET integrates four components: (1) autoresearch, an autonomous ML research pipeline that automates dataset generation, hyperparameter exploration, evaluation, and error-driven iteration; (2) BitNet b1.58 ternary training, enabling CPU-native inference via bitnet.cpp without GPU hardware; (3) DiLoCo-based distributed merging for communication-efficient aggregation of domain specialists; and (4) on-chain contribution tracking on the HOOTi EVM chain. We validate autoresearch through three case studies: video safety classification (balanced accuracy 0.9287 to 0.9851), cryptocurrency directional prediction (41% to 54.9% hit rate), and BitNet hyperparameter optimization (10-phase sweep, -16.7% validation loss).