Beyond Text-Dominance: Understanding Modality Preference of Omni-modal Large Language Models

arXiv cs.AI / 4/21/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies a previously underexplored issue in native omni-modal LLMs: how the models develop a “modality preference” when text and vision are both available.
  • It introduces a conflict-based benchmark and a modality selection rate metric to systematically quantify modality preference across ten representative OLLMs.
  • The results show a shift from traditional VLM “text-dominance” to a strong visual preference in most omni-modal LLMs.
  • Layer-wise probing indicates the preference is not fixed from the start, but gradually emerges in the mid-to-late layers.
  • The authors use these internal signals to diagnose cross-modal hallucinations and report competitive results on three downstream multimodal benchmarks without task-specific training data.

Abstract

Native Omni-modal Large Language Models (OLLMs) have shifted from pipeline architectures to unified representation spaces. However, this native integration gives rise to a critical yet underexplored phenomenon: modality preference. To bridge this gap, we first systematically quantify modality preference of OLLMs using a newly-curated conflict-based benchmark and the modality selection rate metric. Our evaluation of ten representative OLLMs reveals a notable paradigm shift: unlike the ``text-dominance'' of traditional VLMs, most OLLMs exhibit a pronounced visual preference. To further understand the underlying mechanism, we conduct layer-wise probing and demonstrate that such modality preference is not static but emerges progressively in the mid-to-late layers. Building upon these insights, we leverage these internal signals to diagnose cross-modal hallucinations, achieving competitive performance across three downstream multi-modal benchmarks without task-specific data. Our work provides both a mechanistic understanding and a practical tool for building more trustworthy OLLMs. Our code and related resources are publicly available at: https://github.com/icip-cas/OmniPreference