AI Navigate

A Synthesizable RTL Implementation of Predictive Coding Networks

arXiv cs.AI / 3/20/2026

💬 OpinionDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper presents a digitally synthesizable RTL architecture that implements discrete-time predictive coding learning dynamics directly in hardware, addressing challenges of online, distributed learning with backpropagation.
  • Each neural core maintains its own activity, prediction error, and synaptic weights and communicates only with adjacent layers via hardwired connections.
  • Supervised learning and inference are enabled using a per-neuron clamping primitive that enforces boundary conditions without changing the local update schedule.
  • The design is deterministic and synthesizable, built around a sequential MAC datapath and a fixed finite-state schedule, avoiding task-specific instruction sequences.
  • The contribution is a complete hardware substrate for predictive coding rather than a new learning rule, enabling hardware realization of predictive-coding dynamics through connectivity, parameters, and boundary conditions.

Abstract

Backpropagation has enabled modern deep learning but is difficult to realize as an online, fully distributed hardware learning system due to global error propagation, phase separation, and heavy reliance on centralized memory. Predictive coding offers an alternative in which inference and learning arise from local prediction-error dynamics between adjacent layers. This paper presents a digital architecture that implements a discrete-time predictive coding update directly in hardware. Each neural core maintains its own activity, prediction error, and synaptic weights, and communicates only with adjacent layers through hardwired connections. Supervised learning and inference are supported via a uniform per-neuron clamping primitive that enforces boundary conditions while leaving the internal update schedule unchanged. The design is a deterministic, synthesizable RTL substrate built around a sequential MAC datapath and a fixed finite-state schedule. Rather than executing a task-specific instruction sequence inside the learning substrate, the system evolves under fixed local update rules, with task structure imposed through connectivity, parameters, and boundary conditions. The contribution of this work is not a new learning rule, but a complete synthesizable digital substrate that executes predictive-coding learning dynamics directly in hardware.