Active Imitation Learning for Thermal- and Kernel-Aware LFM Inference on 3D S-NUCA Many-Cores

arXiv cs.LG / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the challenge of running large foundation model (LFM) inference efficiently on emerging 3D-stacked Static Non-Uniform Cache Architecture (3D S-NUCA) CPUs, which have better bandwidth/locality but face thermal and cache-latency issues.
  • It proposes AILFM, an Active Imitation Learning scheduling framework that learns near-optimal thermal-aware thread migration and V/f scaling policies using oracle demonstrations.
  • AILFM explicitly models both core-level performance heterogeneity and kernel-specific behaviors across diverse LFM kernels to keep operation within thermal safety constraints.
  • The authors report extensive experimental results showing AILFM outperforms existing state-of-the-art baselines and generalizes across a range of LFM inference workloads.

Abstract

Large Foundation Model (LFM) inference is both memory- and compute-intensive, traditionally relying on GPUs. However, the limited availability and high cost have motivated the adoption of high-performance general-purpose CPUs, especially emerging 3D-stacked Static Non-Uniform Cache Architecture (3D S-NUCA) systems. These architectures offer enhanced bandwidth and locality but suffer from severe thermal challenges and uneven cache latencies due to 3D Networks-on-Chip (NoC). Optimal management of thread migration and V/f scaling is non-trivial due to LFM kernel diversity and system heterogeneity. Existing thermal management approaches often rely on oversimplified analytical models and lack adaptability. We propose AILFM, an Active Imitation Learning (AIL)-based scheduling framework that learns near-optimal thermal-aware scheduling policies from Oracle demonstrations with minimal run-time overhead. AILFM accounts for both core-level performance heterogeneity and kernel-specific behavior in LFMs to maintain thermal safety while maximizing performance. Extensive experiments show that AILFM outperforms state-of-the-art baselines and generalizes well across diverse LFM workloads.