Efficient Inference for Large Vision-Language Models: Bottlenecks, Techniques, and Prospects

arXiv cs.CL / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper explains that Large Vision-Language Models suffer from an inference efficiency barrier called “visual token dominance,” driven by a mix of high-resolution encoding cost, quadratic attention scaling, and memory bandwidth limits.
  • It proposes an end-to-end efficiency taxonomy across the LVLM inference lifecycle—encoding, prefilling, and decoding—showing how upstream design choices create downstream bottlenecks.
  • It analyzes three key bottleneck themes: compute-bound visual encoding, intensive prefilling for massive long contexts, and a “visual memory wall” in bandwidth-bound decoding.
  • The work reframes optimization as managing information density, long-context attention efficiently, and memory limits, focusing on the trade-off between visual fidelity and system efficiency.
  • It concludes with four future frontiers (hybrid compression, modality-aware decoding, progressive state for streaming, and stage-disaggregated serving via hardware–algorithm co-design) and releases a maintained “living” software literature snapshot.

Abstract

Large Vision-Language Models (LVLMs) enable sophisticated reasoning over images and videos, yet their inference is hindered by a systemic efficiency barrier known as visual token dominance. This overhead is driven by a multi-regime interplay between high-resolution feature extraction, quadratic attention scaling, and memory bandwidth constraints. We present a systematic taxonomy of efficiency techniques structured around the inference lifecycle, consisting of encoding, prefilling, and decoding. Unlike prior reviews focused on isolated optimizations, we analyze the end-to-end pipeline to reveal how upstream decisions dictate downstream bottlenecks, covering compute-bound visual encoding, the intensive prefilling of massive contexts, and the ''visual memory wall'' in bandwidth-bound decoding. By decoupling the efficiency landscape into the axes of shaping information density, managing long-context attention, and overcoming memory limits, this work provides a structured analysis of how isolated optimizations compose to navigate the trade-off between visual fidelity and system efficiency. The survey concludes by outlining four future frontiers supported by pilot empirical insights, including hybrid compression based on functional unit sensitivity, modality-aware decoding with relaxed verification, progressive state management for streaming continuity, and stage-disaggregated serving through hardware-algorithm co-design. The submitted software contains a snapshot of our literature repository, which is designed to be maintained as a living resource for the community.