dinov3.seg: Open-Vocabulary Semantic Segmentation with DINOv3

arXiv cs.AI / 3/23/2026

💬 OpinionModels & Research

Key Points

  • The paper introduces dinov3.seg, a dedicated framework for Open-Vocabulary Semantic Segmentation (OVSS) built on the DINOv3 backbone to handle open-set text-defined categories.
  • It jointly aligns text embeddings with both the global CLS token and local patch-level visual features, enabling strong semantic discrimination alongside fine-grained spatial locality.
  • The approach performs early refinement of visual representations prior to image-text interaction and late refinement of the resulting image-text correlation features to improve dense predictions in cluttered scenes.
  • A high-resolution local-global inference strategy based on sliding-window aggregation preserves spatial detail while maintaining global context, and experiments on five OVSS benchmarks show consistent gains over state-of-the-art methods.

Abstract

Open-Vocabulary Semantic Segmentation (OVSS) assigns pixel-level labels from an open set of text-defined categories, demanding reliable generalization to unseen classes at inference. Although modern vision-language models (VLMs) support strong open-vocabulary recognition, their representations learned through global contrastive objectives remain suboptimal for dense prediction, prompting many OVSS methods to depend on limited adaptation or refinement of image-text similarity maps. This, in turn, restricts spatial precision and robustness in complex, cluttered scenes. We introduce dinov3.seg, extending dinov3.txt into a dedicated framework for OVSS. Our contributions are four-fold. First, we design a task-specific architecture tailored to this backbone, systematically adapting established design principles from prior open-vocabulary segmentation work. Second, we jointly leverage text embeddings aligned with both the global [CLS] token and local patch-level visual features from ViT-based encoder, effectively combining semantic discrimination with fine-grained spatial locality. Third, unlike prior approaches that rely primarily on post hoc similarity refinement, we perform early refinement of visual representations prior to image-text interaction, followed by late refinement of the resulting image-text correlation features, enabling more accurate and robust dense predictions in cluttered scenes. Finally, we propose a high-resolution local-global inference strategy based on sliding-window aggregation, which preserves spatial detail while maintaining global context. We conduct extensive experiments on five widely adopted OVSS benchmarks to evaluate our approach. The results demonstrate its effectiveness and robustness, consistently outperforming current state-of-the-art methods.