A Unified Memory Perspective for Probabilistic Trustworthy AI

arXiv cs.LG / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that trustworthy AI workloads that combine probabilistic sampling with deterministic data access increasingly become constrained by memory performance rather than compute arithmetic units.
  • It proposes a unified perspective where deterministic access can be seen as a special (limiting) case of stochastic sampling, allowing both workload modes to be analyzed within one framework.
  • The authors show that higher stochastic sampling demand can reduce effective data-access efficiency and potentially push systems into “entropy-limited” operation.
  • They introduce memory-centric evaluation criteria—such as unified operation, distribution programmability, efficiency, robustness to hardware non-idealities, and parallel compatibility—to assess and compare architectures.
  • Using these criteria, the paper critiques conventional architectures and surveys probabilistic compute-in-memory approaches, suggesting design pathways for scalable trustworthy AI hardware.

Abstract

Trustworthy artificial intelligence increasingly relies on probabilistic computation to achieve robustness, interpretability, security and privacy. In practical systems, such workloads interleave deterministic data access with repeated stochastic sampling across models, data paths and system functions, shifting performance bottlenecks from arithmetic units to memory systems that must deliver both data and randomness. Here we present a unified data-access perspective in which deterministic access is treated as a limiting case of stochastic sampling, enabling both modes to be analyzed within a common framework. This view reveals that increasing stochastic demand reduces effective data-access efficiency and can drive systems into entropy-limited operation. Based on this insight, we define memory-level evaluation criteria, including unified operation, distribution programmability, efficiency, robustness to hardware non-idealities and parallel compatibility. Using these criteria, we analyze limitations of conventional architectures and examine emerging probabilistic compute-in-memory approaches that integrate sampling with memory access, outlining pathways toward scalable hardware for trustworthy AI.
広告