ReXInTheWild: A Unified Benchmark for Medical Photograph Understanding

arXiv cs.CV / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • ReXInTheWild introduces a benchmark of 955 clinician-verified multiple-choice questions across 484 photographs spanning seven clinical topics to test vision-language models on medical content in ordinary images.
  • Leading multimodal LLMs show varied performance (Gemini-3 78%, Claude Opus 4.5 72%, GPT-5 68%), while the medical specialist model MedGemma trails at 37%, highlighting gaps between generalist and domain-specific medical models.
  • An error analysis identifies four categories of mistakes from low-level geometric errors to high-level reasoning failures, suggesting targeted mitigation strategies.
  • The dataset is publicly available on HuggingFace, enabling researchers to benchmark and advance clinically grounded multimodal AI for medical image understanding.
  • Overall, the work emphasizes clinically grounded evaluation at the intersection of natural image understanding and medical reasoning to drive future model development.

Abstract

Everyday photographs taken with ordinary cameras are already widely used in telemedicine and other online health conversations, yet no comprehensive benchmark evaluates whether vision-language models can interpret their medical content. Analyzing these images requires both fine-grained natural image understanding and domain-specific medical reasoning, a combination that challenges both general-purpose and specialized models. We introduce ReXInTheWild, a benchmark of 955 clinician-verified multiple-choice questions spanning seven clinical topics across 484 photographs sourced from the biomedical literature. When evaluated on ReXInTheWild, leading multimodal large language models show substantial performance variation: Gemini-3 achieves 78% accuracy, followed by Claude Opus 4.5 (72%) and GPT-5 (68%), while the medical specialist model MedGemma reaches only 37%. A systematic error analysis also reveals four categories of common errors, ranging from low-level geometric errors to high-level reasoning failures and requiring different mitigation strategies. ReXInTheWild provides a challenging, clinically grounded benchmark at the intersection of natural image understanding and medical reasoning. The dataset is available on HuggingFace.