Many Dialects, Many Languages, One Cultural Lens: Evaluating Multilingual VLMs for Bengali Culture Understanding Across Historically Linked Languages and Regional Dialects

arXiv cs.CL / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces BanglaVerse, a culturally grounded multilingual vision-language benchmark to evaluate how well models understand Bengali culture across languages and regional dialects.
  • The benchmark is built from 1,152 manually curated images across nine visual domains and expanded to four related languages and five Bangla dialects, producing about 32.3K evaluation artifacts.
  • Results indicate that evaluating only standard Bangla can overestimate model ability, with performance dropping most notably under dialectal variation (especially for caption generation).
  • While historically linked languages like Hindi and Urdu preserve some cultural meaning, models are still weaker for structured reasoning compared with dialectally robust understanding.
  • The study finds that the dominant limitation is missing cultural knowledge in knowledge-intensive categories, not lack of visual grounding, positioning BanglaVerse as a more realistic test bed for culturally nuanced multimodal evaluation.

Abstract

Bangla culture is richly expressed through region, dialect, history, food, politics, media, and everyday visual life, yet it remains underrepresented in multimodal evaluation. To address this gap, we introduce BanglaVerse, a culturally grounded benchmark for evaluating multilingual vision-language models (VLMs) on Bengali culture across historically linked languages and regional dialects. Built from 1,152 manually curated images across nine domains, the benchmark supports visual question answering and captioning, and is expanded into four languages and five Bangla dialects, yielding ~32.3K artifacts. Our experiments show that evaluating only standard Bangla overestimates true model capability: performance drops under dialectal variation, especially for caption generation, while historically linked languages such as Hindi and Urdu retain some cultural meaning but remain weaker for structured reasoning. Across domains, the main bottleneck is missing cultural knowledge rather than visual grounding alone, with knowledge-intensive categories. These findings position BanglaVerse as a more realistic test bed for measuring culturally grounded multimodal understanding under linguistic variation.