AI Navigate

Rethinking MLLM Itself as a Segmenter with a Single Segmentation Token

arXiv cs.CV / 3/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The SELF1E paper investigates decoder-free segmentation for MLLMs by using a single segmentation embedding, aiming to remove the need for external mask decoders.
  • It addresses resolution loss by keeping image features at original resolution and refilling them with residuals from LLM-processed compressed features to improve precision.
  • It introduces pixel-unshuffle operations and a dual-path attention mask (image-to-image and image-to-segmentation) to enrich feature interaction between pixels and the segmentation token.
  • Experiments show SELF1E achieves competitive results with decoder-based methods across multiple segmentation tasks, demonstrating the feasibility of decoder-free segmentation in MLLMs. Project page: https://github.com/ANDYZAQ/SELF1E.

Abstract

Recent segmentation methods leveraging Multi-modal Large Language Models (MLLMs) have shown reliable object-level segmentation and enhanced spatial perception. However, almost all previous methods predominantly rely on specialist mask decoders to interpret masks from generated segmentation-related embeddings and visual features, or incorporate multiple additional tokens to assist. This paper aims to investigate whether and how we can unlock segmentation from MLLM itSELF with 1 segmentation Embedding (SELF1E) while achieving competitive results, which eliminates the need for external decoders. To this end, our approach targets the fundamental limitation of resolution reduction in pixel-shuffled image features from MLLMs. First, we retain image features at their original uncompressed resolution, and refill them with residual features extracted from MLLM-processed compressed features, thereby improving feature precision. Subsequently, we integrate pixel-unshuffle operations on image features with and without LLM processing, respectively, to unleash the details of compressed features and amplify the residual features under uncompressed resolution, which further enhances the resolution of refilled features. Moreover, we redesign the attention mask with dual perception pathways, i.e., image-to-image and image-to-segmentation, enabling rich feature interaction between pixels and the segmentation token. Comprehensive experiments across multiple segmentation tasks validate that SELF1E achieves performance competitive with specialist mask decoder-based methods, demonstrating the feasibility of decoder-free segmentation in MLLMs. Project page: https://github.com/ANDYZAQ/SELF1E.