Hyperspherical Forward-Forward with Prototypical Representations

arXiv cs.AI / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Hyperspherical Forward-Forward (HFF), a reformulation of the Forward-Forward algorithm to eliminate its class-by-class inference bottleneck.
  • HFF changes each layer’s local objective from a binary goodness-of-fit check into a multi-class classification task in a hyperspherical feature space.
  • It learns class-specific unit-norm prototypes that serve as geometric anchors and implicit negatives, enabling both weight updates and inference in a single forward pass.
  • The authors report that HFF is over 40× faster than the original Forward-Forward approach while also achieving higher accuracy on standard image classification benchmarks.
  • Results include more than 25% top-1 accuracy on ImageNet-1k using greedy local learning, and 65.96% top-1 accuracy with transfer learning, narrowing the performance gap with backpropagation.

Abstract

The Forward-Forward (FF) algorithm presents a compelling, bio-inspired alternative to backpropagation. However, while efficient in training, it has a computationally prohibitive inference process that requires a separate forward pass for every class that is evaluated. In this work, we introduce the Hyperspherical Forward-Forward (HFF), a novel reformulation that resolves this critical bottleneck. Our core innovation is to reframe the local objective of each layer from a binary goodness-of-fit task to a direct multi-class classification problem within a hyperspherical feature space. We achieve this by learning a set of class-specific, unit-norm prototypes that act as geometric anchors and implicit negatives. This architectural innovation preserves the benefits of local training while enabling weight update and inference in a single forward pass, making it >40x faster than the original FF algorithm. Our method is simple to implement, scales effectively to modern convolutional architectures, and achieves superior accuracy on standard image classification benchmarks, closing the gap with backpropagation. Most notably, we are among the first greedy local-learning methods to report over 25% top-1 accuracy on ImageNet-1k, and 65.96% with transfer learning.