TsetlinWiSARD: On-Chip Training of Weightless Neural Networks using Tsetlin Automata on FPGAs

arXiv cs.LG / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces TsetlinWiSARD, an on-chip training method for weightless neural networks (WNNs) that uses Tsetlin Automata for iterative, probabilistic feedback-driven learning.
  • It addresses a key limitation of prior WiSARD-style WNNs by mitigating overfitting from one-shot memorization-based training and reducing the need for tedious post-training tuning.
  • The authors present an FPGA-based training architecture designed for efficient learning with continuous binary feedback, targeting edge requirements like low latency and improved privacy/security.
  • Reported results show over 1000x faster training versus traditional WiSARD, along with 22% lower FPGA resource usage, 93.3% lower latency, and 64.2% lower power compared with other FPGA ML training accelerators.
  • Overall, the work positions WNN training on FPGAs as a hardware-efficient alternative to multiply-accumulate-heavy deep learning approaches for edge ML scenarios.

Abstract

Increasing demands for adaptability, privacy, and security at the edge have persistently pushed the frontiers for a new generation of machine learning (ML) algorithms with training and inference capabilities on-chip. Weightless Neural Network (WNN) is such an algorithm that is principled on lookup table based simple neuron structures. As a result, it offers architectural benefits, such as low-latency, low-complexity inference, compared to deep neural networks that depend heavily on multiply-accumulate operations. However, traditional WNNs rely on memorization-based one-shot training, which either leads to overfitting and reduced accuracy or requires tedious post-training adjustments, limiting their effectiveness for efficient on chip training. In this work, we propose TsetlinWiSARD, a training approach for WNNs that leverages Tsetlin Automata (TAs) to enable probabilistic, feedback-driven learning. It overcomes the overfitting of WiSARD's one-shot training with iterative optimization, while maintaining simple, continuous binary feedback for efficient on-chip training. Central to our approach is a field programmable gate array (FPGA)-based training architecture that delivers state-of-the-art accuracy while significantly improving hardware efficiency. Our approach provides over 1000x faster training when compared with the traditional WiSARD implementation of WNNs. Further, we demonstrate 22% reduced resource usage, 93.3% lower latency, and 64.2% lower power consumption compared to FPGA-based training accelerators implementing other ML algorithms.