AI Navigate

SRAM-Based Compute-in-Memory Accelerator for Linear-decay Spiking Neural Networks

arXiv cs.AI / 3/16/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The authors propose an SRAM-based compute-in-memory (CIM) accelerator for Spiking Neural Networks (SNNs) that co-optimizes algorithm and hardware using Linear Decay Leaky Integrate-and-Fire neurons.
  • They replace the conventional exponential membrane decay with a linear decay, converting multiplications into simple additions with only about 1% accuracy loss.
  • An in-memory parallel update scheme performs in-place decay inside the SRAM array, removing the need for global sequential membrane-potential updates.
  • On benchmark SNN workloads, the method achieves 1.1x to 16.7x reductions in SOP energy and 15.9x to 69x improvements in overall energy efficiency, with negligible accuracy loss.

Abstract

Spiking Neural Networks (SNNs) have emerged as a biologically inspired alternative to conventional deep networks, offering event-driven and energy-efficient computation. However, their throughput remains constrained by the serial update of neuron membrane states. While many hardware accelerators and Compute-in-Memory (CIM) architectures efficiently parallelize the synaptic operation (W x I) achieving O(1) complexity for matrix-vector multiplication, the subsequent state update step still requires O(N) time to refresh all neuron membrane potentials. This mismatch makes state update the dominant latency and energy bottleneck in SNN inference. To address this challenge, we propose an SRAM-based CIM for SNN with Linear Decay Leaky Integrate-and-Fire (LD-LIF) Neuron that co-optimizes algorithm and hardware. At the algorithmic level, we replace the conventional exponential membrane decay with a linear decay approximation, converting costly multiplications into simple additions while accuracy drops only around 1%. At the architectural level, we introduce an in-memory parallel update scheme that performs in-place decay directly within the SRAM array, eliminating the need for global sequential updates. Evaluated on benchmark SNN workloads, the proposed method achieves a 1.1 x to 16.7 x reduction of SOP energy consumption, while providing 15.9 x to 69 x more energy efficiency, with negligible accuracy loss relative to original decay models. This work highlights that beyond accelerating the (W x I) computation, optimizing state-update dynamics within CIM architectures is essential for scalable, low-power, and real-time neuromorphic processing.