Diffusion Mental Averages
arXiv cs.CV / 4/1/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces Diffusion Mental Averages (DMA), aiming to generate a single “sharp and realistic” prototype of a concept directly from a diffusion model rather than averaging images outside the model.
- It argues that prior data-centric averaging on diffusion samples yields blur, and proposes instead averaging in the model’s evolving semantic space by aligning multiple denoising trajectories so they converge from coarse to fine semantics.
- DMA is framed as an optimization problem over multiple noise latents, producing a consistent visual summary and a way to probe how concepts are represented and biased in the diffusion process.
- For multimodal concepts (e.g., many dog breeds), the method clusters samples in semantically rich embedding spaces like CLIP and then uses Textual Inversion or LoRA to connect CLIP clusters to diffusion space.
- The authors claim it is the first approach to deliver consistent, realistic averages for both concrete and abstract concepts using this within-model averaging and trajectory-alignment strategy.
Related Articles

Black Hat Asia
AI Business

Knowledge Governance For The Agentic Economy.
Dev.to

AI server farms heat up the neighborhood for miles around, paper finds
The Register

Paperclip: Công Cụ Miễn Phí Biến AI Thành Đội Phát Triển Phần Mềm
Dev.to
Does the Claude “leak” actually change anything in practice?
Reddit r/LocalLLaMA