Understanding Pruning Regimes in Vision-Language Models Through Domain-Aware Layer Selection

arXiv cs.CV / 3/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies structured decoder layer pruning in vision-language models by using domain-aware activation similarity to determine which layers least change representations for math versus non-math inputs.
  • It introduces math-aware, non-math-aware, and mixed layer-ranking criteria based on how layer transformations differ across targeted domains.
  • Experiments on two state-of-the-art VLMs across math and general multimodal benchmarks reveal a consistent three-regime behavior: high sensitivity at low pruning budgets, convergence at moderate budgets, and continuity/spacing effects dominating at high budgets.
  • The proposed domain-aware ranking is reported to be most effective for stability in the ranking-sensitive regime and to match or outperform structure-aware baselines when pruning is more aggressive, yielding an interpretable approach to reducing depth.

Abstract

Transformer-based vision-language models (VLMs) contain substantial depth redundancy, yet the effect of removing specific decoder layers remains poorly understood, especially for domains that require tight coupling between perception and multi-step reasoning. We study structured decoder layer pruning through the lens of domain-aware activation similarity, measuring how strongly each layer transforms representations for math versus non-math inputs. This yields simple math-aware, non-math-aware, and mixed ranking criteria that identify layers whose input-output activations change least within a target domain. Across two state-of-the-art VLMs and a broad suite of math and general multimodal benchmarks, we uncover a consistent three-regime structure: at low pruning budgets, performance is highly sensitive to which layers are removed; at moderate budgets, methods converge as structural damage accumulates; and at high budgets, structural continuity dominates, favoring spacing-aware strategies. Our domain-aware rankings achieve the strongest stability in the ranking-sensitive regime, while matching or exceeding structure-aware baselines at larger budgets. These results provide a clearer picture of how depth contributes to domain-specific behavior in VLMs and offer a practical, interpretable approach to reducing model depth without sacrificing essential mathematical or general vision-language capabilities.