Direct Segmentation without Logits Optimization for Training-Free Open-Vocabulary Semantic Segmentation

arXiv cs.CV / 4/10/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper targets open-vocabulary semantic segmentation by removing the usual pixel-level vision-language alignment step that relies on cosine-similarity “logits” and iterative optimization.
  • It proposes a training-free method that derives an analytic solution for the semantic segmentation map instead of optimizing logits with time-consuming training or model-specific attention modulation.
  • The core hypothesis is that the distribution discrepancy between visual and linguistic features encodes semantics, showing intra-category consistency across image patches and inter-category inconsistency.
  • By directly using the analytic solution of this distribution discrepancy, the approach avoids iterative training and still achieves state-of-the-art results across eight benchmark datasets.

Abstract

Open-vocabulary semantic segmentation (OVSS) aims to segment arbitrary category regions in images using open-vocabulary prompts, necessitating that existing methods possess pixel-level vision-language alignment capability. Typically, this capability involves computing the cosine similarity, \ie, logits, between visual and linguistic features, and minimizing the distribution discrepancy between the logits and the ground truth (GT) to generate optimal logits that are subsequently used to construct segmentation maps, yet it depends on time-consuming iterative training or model-specific attention modulation. In this work, we propose a more direct approach that eschews the logits-optimization process by directly deriving an analytic solution for the segmentation map. We posit a key hypothesis: the distribution discrepancy encodes semantic information; specifically, this discrepancy exhibits consistency across patches belonging to the same category but inconsistency across different categories. Based on this hypothesis, we directly utilize the analytic solution of this distribution discrepancy as the semantic maps. In other words, we reformulate the optimization of the distribution discrepancy as deriving its analytic solution, thereby eliminating time-consuming iterative training, freeing us from model-specific attention modulation, and achieving state-of-the-art performance on eight benchmark datasets.