Dress-ED: Instruction-Guided Editing for Virtual Try-On and Try-Off
arXiv cs.CV / 3/25/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The paper introduces Dress-ED, a new large-scale benchmark that unifies Virtual Try-On (VTON), Virtual Try-Off (VTOFF), and text-guided garment editing under one dataset framework.
- Each Dress-ED sample provides an in-shop garment image, a person image wearing the garment, edited results, and an accompanying natural-language instruction describing the desired modification.
- Dress-ED is built with a fully automated multimodal pipeline using MLLM-based garment understanding, diffusion-based editing, and LLM-guided verification, and contains 146k+ verified quadruplets across 3 garment categories and 7 edit types.
- The work also proposes a unified multimodal diffusion framework that jointly conditions on linguistic instructions and visual garment cues, aiming to serve as a baseline for instruction-driven VTON/VTOFF.
- The authors state that the dataset and code will be publicly available, enabling researchers to develop and evaluate controllable, interactive fashion editing systems.
Related Articles

The Complete Guide to Model Context Protocol (MCP): Building AI-Native Applications in 2026
Dev.to

AI Agent Skill Security Report — 2026-03-25
Dev.to

Origin raises $30M Series A+ to improve global benefits efficiency
Tech.eu
[R] Adversarial Machine Learning
Reddit r/MachineLearning

Sora's app and API are dead but OpenAI hints the video model lives on inside ChatGPT
THE DECODER