AI Navigate

VeloEdit: Training-Free Consistent and Continuous Instruction-Based Image Editing via Velocity Field Decomposition

arXiv cs.CV / 3/17/2026

📰 NewsModels & Research

Key Points

  • VeloEdit presents a training-free approach to instruction-based image editing that maintains consistency in non-edited regions by partitioning velocity fields into source-preserving and editing components.
  • It automatically identifies editing regions by measuring the discrepancy between the velocity fields responsible for preserving the source content and those driving the desired edits, enabling targeted control over where changes occur.
  • The method enforces preservation-region consistency by substituting the editing velocity with the source-restoring velocity, while enabling continuous modulation of edit strength in target regions via velocity interpolation.
  • Experiments on Flux.1 Kontext and Qwen-Image-Edit demonstrate improved visual consistency and editing continuity with negligible additional computational cost, and the code is released on GitHub.

Abstract

Instruction-based image editing aims to modify source content according to textual instructions. However, existing methods built upon flow matching often struggle to maintain consistency in non-edited regions due to denoising-induced reconstruction errors that cause drift in preserved content. Moreover, they typically lack fine-grained control over edit strength. To address these limitations, we propose VeloEdit, a training-free method that enables highly consistent and continuously controllable editing. VeloEdit dynamically identifies editing regions by quantifying the discrepancy between the velocity fields responsible for preserving source content and those driving the desired edits. Based on this partition, we enforce consistency in preservation regions by substituting the editing velocity with the source-restoring velocity, while enabling continuous modulation of edit intensity in target regions via velocity interpolation. Unlike prior works that rely on complex attention manipulation or auxiliary trainable modules, VeloEdit operates directly on the velocity fields. Extensive experiments on Flux.1 Kontext and Qwen-Image-Edit demonstrate that VeloEdit improves visual consistency and editing continuity with negligible additional computational cost. Code is available at https://github.com/xmulzq/VeloEdit.