Physical Foundation Models: Fixed hardware implementations of large-scale neural networks

arXiv cs.LG / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that the foundation-model era (large, general-purpose neural networks adapted to many tasks) creates a new opportunity for hardware engineers to build fixed, task-agnostic hardware implementations released on about a 1-year cadence aligned with model updates.
  • Instead of conventional digital inference with read-only weight memory, it proposes “Physical Foundation Models” (PFMs), where the neural network is implemented directly in physical hardware design and leverages the hardware’s natural dynamics.
  • The authors claim PFMs could deliver orders-of-magnitude improvements in energy efficiency, speed, and parameter density, potentially reducing datacenter energy costs and enabling edge AI for much larger models.
  • They provide back-of-the-envelope scaling calculations using an optical example (a 3D nanostructured glass medium) and discuss feasibility prospects across nanoelectronics and other physical platforms.
  • The work concludes by outlining major research challenges that must be solved to make trillion-parameter PFMs—and potentially even larger (10^15–10^18 parameters)—a practical reality.

Abstract

Foundation models are deep neural networks (such as GPT-5, Gemini~3, and Opus~4) trained on large datasets that can perform diverse downstream tasks -- text and code generation, question answering, summarization, image classification, and so on. The philosophy of foundation models is to put effort into a single, large ({\sim}10^{12}-parameter) general-purpose model that can be adapted to many downstream tasks with no or minimal additional training. We argue that the rise of foundation models presents an opportunity for hardware engineers: in contrast to when different models were used for different tasks, it now makes sense to build special-purpose, fixed hardware implementations of neural networks, manufactured and released at the roughly 1-year cadence of major new foundation-model versions. Beyond conventional digital-electronic inference hardware with read-only weight memory, we advocate a more radical re-thinking: hardware in which the neural network is realized directly at the level of the physical design and operates via the hardware's natural physical dynamics -- \textit{Physical Foundation Models} (PFMs). PFMs could enable orders-of-magnitude advantages in energy efficiency, speed, and parameter density. For {\sim}10^{12}-parameter models, this would both reduce the high energy burden of AI in datacenters and enable AI in edge devices that today are power-constrained to far smaller models. PFMs could also enable inference hardware for models much larger than current ones: 10^{15}- or even 10^{18}-parameter PFMs seem plausible by some measures. We present back-of-the-envelope calculations illustrating PFM scaling using an optical example -- a 3D nanostructured glass medium -- and discuss prospects in nanoelectronics and other physical platforms. We conclude with the major research challenges that must be resolved for trillion-parameter PFMs and beyond to become reality.