SHIFT: Steering Hidden Intermediates in Flow Transformers

arXiv cs.CV / 4/13/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SHIFT, a lightweight inference-time framework for DiT (Diffusion Transformer) models that removes unwanted visual concepts by manipulating intermediate activations.
  • SHIFT learns steering vectors and applies them dynamically across selected layers and timesteps to suppress specific concepts while retaining prompt-relevant content and image quality.
  • The method is presented as retraining-free (no time-consuming retraining), aiming to control generations effectively across diverse prompts and targets.
  • Beyond suppression, SHIFT can steer outputs into a desired style domain or bias images toward adding/changing target objects, suggesting broader controllability.
  • The approach is inspired by activation steering techniques used in large language models, transferring the idea to diffusion/DiT generation workflows.

Abstract

Diffusion models have become leading approaches for high-fidelity image generation. Recent DiT-based diffusion models, in particular, achieve strong prompt adherence while producing high-quality samples. We propose SHIFT, a simple but effective and lightweight framework for concept removal in DiT diffusion models via targeted manipulation of intermediate activations at inference time, inspired by activation steering in large language models. SHIFT learns steering vectors that are dynamically applied to selected layers and timesteps to suppress unwanted visual concepts while preserving the prompt's remaining content and overall image quality. Beyond suppression, the same mechanism can shift generations into a desired \emph{style domain} or bias samples toward adding or changing target objects. We demonstrate that SHIFT provides effective and flexible control over DiT generation across diverse prompts and targets without time-consuming retraining.