OmniDiT: Extending Diffusion Transformer to Omni-VTON Framework

arXiv cs.CV / 3/23/2026

📰 NewsModels & Research

Key Points

  • OmniDiT is a diffusion-transformer based framework that unifies virtual try-on (VTON) and try-off (VTOFF) tasks into a single model.
  • The authors introduce the Omni-TryOn dataset with over 380k garment-model-try-on image pairs and detailed text prompts, built through a self-evolving data curation pipeline.
  • They propose architectural innovations, including token concatenation, adaptive position encoding, and Shifted Window Attention to achieve linear complexity in the diffusion model, along with multiple timestep prediction and an alignment loss to boost fidelity.
  • Experiments show state-of-the-art performance for model-free VTON and VTOFF, with performance comparable to current SOTA methods in model-based VTON.

Abstract

Despite the rapid advancement of Virtual Try-On (VTON) and Try-Off (VTOFF) technologies, existing VTON methods face challenges with fine-grained detail preservation, generalization to complex scenes, complicated pipeline, and efficient inference. To tackle these problems, we propose OmniDiT, an omni Virtual Try-On framework based on the Diffusion Transformer, which combines try-on and try-off tasks into one unified model. Specifically, we first establish a self-evolving data curation pipeline to continuously produce data, and construct a large VTON dataset Omni-TryOn, which contains over 380k diverse and high-quality garment-model-tryon image pairs and detailed text prompts. Then, we employ the token concatenation and design an adaptive position encoding to effectively incorporate multiple reference conditions. To relieve the bottleneck of long sequence computation, we are the first to introduce Shifted Window Attention into the diffusion model, thus achieving a linear complexity. To remedy the performance degradation caused by local window attention, we utilize multiple timestep prediction and an alignment loss to improve generation fidelity. Experiments reveal that, under various complex scenes, our method achieves the best performance in both the model-free VTON and VTOFF tasks and a performance comparable to current SOTA methods in the model-based VTON task.