ID-Selection: Importance-Diversity Based Visual Token Selection for Efficient LVLM Inference

arXiv cs.CV / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ID-Selection, an importance-and-diversity based visual token selection strategy to accelerate LVLM inference while avoiding the typical redundancy vs. information-loss trade-off.
  • ID-Selection assigns each visual token an importance score and then iteratively selects high-scoring tokens while suppressing scores of similar tokens to enforce diversity.
  • Experiments across five LVLM backbones and sixteen benchmarks show ID-Selection improves both efficiency and accuracy, with the biggest gains under extreme visual token pruning ratios.
  • For instance, on LLaVA-1.5-7B it prunes 97.2% of visual tokens, keeps 16 tokens, cuts inference FLOPs by over 97%, and retains 91.8% of the original performance without additional training.

Abstract

Recent advances have explored visual token pruning to accelerate the inference of large vision-language models (LVLMs). However, existing methods often struggle to balance token importance and diversity: importance-based methods tend to retain redundant tokens, whereas diversity-based methods may overlook informative ones. This trade-off becomes especially problematic under high reduction ratios, where preserving only a small subset of visual tokens is critical. To address this issue, we propose ID-Selection, a simple yet effective token selection strategy for efficient LVLM inference. The key idea is to couple importance estimation with diversity-aware iterative selection: each token is first assigned an importance score, after which high-scoring tokens are selected one by one while the scores of similar tokens are progressively suppressed. In this way, ID-Selection preserves informative tokens while reducing redundancy in a unified selection process. Extensive experiments across 5 LVLM backbones and 16 main benchmarks demonstrate that ID-Selection consistently achieves superior performance and efficiency, especially under extreme pruning ratios. For example, on LLaVA-1.5-7B, ID-Selection prunes 97.2% of visual tokens, retaining only 16 tokens, while reducing inference FLOPs by over 97% and preserving 91.8% of the original performance, all without additional training.