FIRE-CIR: Fine-grained Reasoning for Composed Fashion Image Retrieval

arXiv cs.CV / 4/13/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • FIRE-CIR tackles composed image retrieval by adding fine-grained compositional reasoning to decide which visual attributes to preserve versus modify from a text instruction, improving both accuracy and interpretability in fashion-specific settings.
  • Instead of relying only on embedding similarity, the model generates attribute-focused visual questions from the modification text and checks visual evidence across the reference and candidate images.
  • The approach is trained using a newly constructed large-scale fashion visual question answering dataset with questions that require single- or dual-image analysis.
  • On the Fashion IQ benchmark, FIRE-CIR achieves higher retrieval accuracy than state-of-the-art methods and provides attribute-level, explainable justifications for why particular candidates are re-ranked or filtered.

Abstract

Composed image retrieval (CIR) aims to retrieve a target image that depicts a reference image modified by a textual description. While recent vision-language models (VLMs) achieve promising CIR performance by embedding images and text into a shared space for retrieval, they often fail to reason about what to preserve and what to change. This limitation hinders interpretability and yields suboptimal results, particularly in fine-grained domains like fashion. In this paper, we introduce FIRE-CIR, a model that brings compositional reasoning and interpretability to fashion CIR. Instead of relying solely on embedding similarity, FIRE-CIR performs question-driven visual reasoning: it automatically generates attribute-focused visual questions derived from the modification text, and verifies the corresponding visual evidence in both reference and candidate images. To train such a reasoning system, we automatically construct a large-scale fashion-specific visual question answering dataset, containing questions requiring either single- or dual-image analysis. During retrieval, our model leverages this explicit reasoning to re-rank candidate results, filtering out images inconsistent with the intended modifications. Experimental results on the Fashion IQ benchmark show that FIRE-CIR outperforms state-of-the-art methods in retrieval accuracy. It also provides interpretable, attribute-level insights into retrieval decisions.