Silicon Aware Neural Networks

arXiv cs.CV / 4/22/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper introduces a method to translate Differentiable Logic Gate Networks (DLGNs) into a one-to-one gate-level implementation using a digital CMOS standard cell library.
  • It proposes a custom loss function that encourages the model to minimize circuit area per neuron (and thus can indirectly reduce power) based on the target library’s standard cell characteristics.
  • The authors demonstrate end-to-end silicon design flow by laying out a DLGN as a custom hard macro in SkyWater 130nm and running post-layout power analysis.
  • In simulation, the resulting macro achieves MNIST classification at 97% accuracy, reaching 41.8 million inferences per second while consuming 83.88 mW.
  • Overall, the work suggests DLGNs can be co-optimized for performance and hardware efficiency when targeting real digital standard-cell processes.

Abstract

Recent work in the machine learning literature has demonstrated that deep learning can train neural networks made of discrete logic gate functions to perform simple image classification tasks at very high speeds on CPU, GPU and FPGA platforms. By virtue of being formed by discrete logic gates, these Differentiable Logic Gate Networks (DLGNs) lend themselves naturally to implementation in custom silicon - in this work we present a method to map DLGNs in a one-to-one fashion to a digital CMOS standard cell library by converting the trained model to a gate-level netlist. We also propose a novel loss function whereby the DLGN can optimize the area, and indirectly power consumption, of the resulting circuit by minimizing the expected area per neuron based on the area of the standard cells in the target standard cell library. Finally, we also show for the first time an implementation of a DLGN as a silicon circuit in simulation, performing layout of a DLGN in the SkyWater 130nm process as a custom hard macro using a Cadence standard cell library and performing post-layout power analysis. We find that our custom macro can perform classification on MNIST with 97% accuracy 41.8 million times a second at a power consumption of 83.88 mW.