When Choices Become Priors: Contrastive Decoding for Scientific Figure Multiple-Choice QA

arXiv cs.AI / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Scientific figure multiple-choice QA can fail because the answer-choice text functions as a prior, biasing multimodal models toward scientifically plausible options even when the figure indicates otherwise.
  • The paper introduces SCICON, a training-free contrastive decoding method that scores each candidate by subtracting its text-only score from its image-conditioned score to discount choice-induced priors.
  • SCICON differs from earlier contrastive decoding methods by focusing specifically on priors embedded in the candidate text rather than contrasting inputs or perturbing instructions.
  • Experiments across three scientific figure QA benchmarks and three model backbones show consistent accuracy improvements versus standard decoding baselines.
  • The findings suggest that explicitly decoding against choice-induced priors is a straightforward way to improve figure-grounded reasoning in scientific MCQA.

Abstract

Scientific figure multiple-choice question answering (MCQA) requires models to reason over diverse visual evidence, ranging from charts and multipanel figures to microscopy and biomedical images. However, this setting suffers from a distinctive bias: answer choices themselves can act as priors, steering multimodal models toward scientifically plausible options even when the figure supports a different answer. We investigate this failure mode through a simple question: what if decoding explicitly discounts what the model would prefer from text alone, so as to favor figure-grounded evidence? To this end, we propose SCICON, a training-free decoding method that scores each candidate by subtracting a text-only option score from its image-conditioned counterpart. Unlike prior contrastive decoding approaches that mitigate hallucinations by contrasting original inputs with distorted images or perturbed instructions, SCICON directly targets the choice-induced prior encoded in candidate text. Across three scientific figure QA benchmarks and three model backbones, SCICON consistently improves accuracy over standard decoding baselines. These results show that decoding against choice-induced priors is an effective and simple way to improve figure-grounded reasoning in scientific MCQA.