Neighbor-Aware Localized Concept Erasure in Text-to-Image Diffusion Models
arXiv cs.CV / 3/30/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses “concept erasure” in text-to-image diffusion models, where an unwanted concept is removed while preserving overall generation quality.
- It argues that existing localized erasure can unintentionally weaken semantically related neighboring concepts, hurting fidelity on fine-grained categories.
- The proposed Neighbor-Aware Localized Concept Erasure (NLCE) is a training-free, three-stage method that suppresses target concept embeddings, uses attention to find residual activation regions, and then applies spatially gated hard erasure only where needed.
- Experiments on Oxford Flowers and Stanford Dogs show NLCE removes the target concept more effectively while better preserving closely related neighboring categories.
- Additional tests indicate robustness and generalization across broader erasure scenarios, including celebrity identity, explicit content, and artistic style.
Related Articles

Mr. Chatterbox is a (weak) Victorian-era ethically trained model you can run on your own computer
Simon Willison's Blog
Beyond the Chatbot: Engineering Multi-Agent Ecosystems in 2026
Dev.to

I missed the "fun" part in software development
Dev.to

The Billion Dollar Tax on AI Agents
Dev.to

Hermes Agent: A Self-Improving AI Agent That Runs Anywhere
Dev.to