Closed-Form Concept Erasure via Double Projections

arXiv cs.LG / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a training-free, closed-form linear transformation method for “concept erasure,” aiming to remove unwanted concepts from pretrained generative model representations without iterative optimization.
  • The approach uses two deterministic projection steps: first estimating a proxy projection of the target concept, then applying a constrained transformation in the left null space of known concept directions to avoid disturbing unrelated concepts.
  • Experiments across Stable Diffusion variants and a flow-matching model (FLUX) show the method matches or outperforms state-of-the-art techniques while preserving non-target concepts more faithfully.
  • Because the method runs in only a few seconds and is designed as a drop-in tool, it offers a lightweight pathway for safer, more controlled model editing.
  • The work positions concept erasure as a geometrically interpretable procedure, providing clearer theoretical grounding compared with optimization-based techniques.

Abstract

While modern generative models such as diffusion-based architectures have enabled impressive creative capabilities, they also raise important safety and ethical risks. These concerns have led to growing interest in concept erasure, the process of removing unwanted concepts from model representations. Existing approaches often achieve strong erasure performance but rely on iterative optimization and may inadvertently distort unrelated concepts. In this work, we present a simple yet principled alternative: a linear transformation framework that achieves concept erasure analytically, without any training. Our method adapts a pretrained model through two sequential, closed-form steps: first, computing a proxy projection of the target concept, and second, applying a constrained transformation within the left null space of known concept directions. This design yields a deterministic and geometrically interpretable procedure for safe, efficient, and theory-grounded concept removal. Across a wide range of experiments, including object and style erasure on multiple Stable Diffusion variants and the flow-matching model (FLUX), our approach matches or surpasses the performance of state-of-the-art methods while preserving non-target concepts more faithfully. Requiring only a few seconds to apply, it offers a lightweight and drop-in tool for controlled model editing, advancing the goal of safer and more responsible generative models.