Appear2Meaning: A Cross-Cultural Benchmark for Structured Cultural Metadata Inference from Images

arXiv cs.CV / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Appear2Meaning, a cross-cultural benchmark aimed at inferring structured cultural metadata (such as creator, origin, and period) from images rather than producing only free-form captions.
  • It evaluates vision-language models using an LLM-as-Judge approach that scores semantic alignment with reference annotations.
  • Performance is assessed with exact-match, partial-match, and attribute-level accuracy, revealing that models often rely on fragmented visual signals.
  • Results show significant variation across cultural regions and metadata types, with predictions that are inconsistent and only weakly grounded.

Abstract

Recent advances in vision-language models (VLMs) have improved image captioning for cultural heritage. However, inferring structured cultural metadata (e.g., creator, origin, period) from visual input remains underexplored. We introduce a multi-category, cross-cultural benchmark for this task and evaluate VLMs using an LLM-as-Judge framework that measures semantic alignment with reference annotations. To assess cultural reasoning, we report exact-match, partial-match, and attribute-level accuracy across cultural regions. Results show that models capture fragmented signals and exhibit substantial performance variation across cultures and metadata types, leading to inconsistent and weakly grounded predictions. These findings highlight the limitations of current VLMs in structured cultural metadata inference beyond visual perception.