Beyond Fixed Formulas: Data-Driven Linear Predictor for Efficient Diffusion Models
arXiv cs.LG / 4/30/2026
📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper addresses the high sampling cost of Diffusion Transformers (DiTs) and argues that hand-crafted feature-caching formulas break down when aggressive skipping is used.
- It introduces L2P (Learnable Linear Predictor), a training-free-acceleration caching framework that replaces fixed coefficients with learnable, per-timestep weights to reconstruct current features from past trajectories.
- L2P can be trained rapidly (about 20 seconds on a single GPU) and is designed to work efficiently for DiT inference.
- Experiments show substantial performance gains, including 4.55× FLOPs reduction and 4.15× latency speedup on FLUX.1-dev, and strong quality retention up to 7.18× acceleration on Qwen-Image compared with prior baselines.
- The authors provide code publicly and conclude that learning linear predictors is an effective strategy for efficient diffusion model sampling.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat USA
AI Business
Vector DB and ANN vs PHE conflict, is there a practical workaround? [D]
Reddit r/MachineLearning

Agent Amnesia and the Case of Henry Molaison
Dev.to
Azure Weekly: Microsoft and OpenAI Restructure Partnership as GPT-5.5 Lands in Foundry
Dev.to
Proven Patterns for OpenAI Codex in 2026: Prompts, Validation, and Gateway Governance
Dev.to