Unlocking Few-Shot Capabilities in LVLMs via Prompt Conditioning and Head Selection

arXiv cs.CV / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates why large vision-language models (LVLMs) do well on many zero-shot generative tasks yet perform poorly on image classification compared with CLIP-based approaches despite using CLIP-pretrained vision encoders.
  • It argues that CLIP’s class-name matching bias differs from joint visual-text reasoning and shows LVLMs can improve class separability at inference via prompt conditioning.
  • The authors propose Head Ensemble Classifiers (HEC), a training-free method that selects and ensembles the most discriminative attention heads from both vision and text components.
  • HEC is inspired by Gaussian Discriminant Analysis and is designed to close the performance gap between CLIP-based and LVLM-based classification methods.
  • Experiments reportedly achieve state-of-the-art results for zero-shot and few-shot image classification across 12 datasets.

Abstract

Current Large Vision Language Models (LVLMs) excel at many zero-shot tasks like image captioning, visual question answering and OCR. However, these same models suffer from poor performance at image classification tasks, underperforming against CLIP-based methods. Notably, this gap is surprising because many LVLMs use CLIP-pretrained vision encoders. Yet LVLMs are not inherently limited by CLIP's architecture with independent vision and text encoders. In CLIP, this separation biases classification toward class-name matching rather than joint visual-text reasoning. In this paper we show that, despite their poor raw performance, LVLMs can improve visual feature class separability at inference using prompt conditioning, and LVLMs' internal representations, especially attention heads, can outperform the model itself at zero-shot and few-shot classification. We introduce Head Ensemble Classifiers (HEC) to bridge the performance gap between CLIP-based and LVLM-based classification methods. Inspired by Gaussian Discriminant Analysis, HEC ranks the most discriminative vision and text heads and combines them into a training-free classifier. We show that HEC achieves state-of-the-art performance in few-shot and zero-shot classification across 12 datasets.