When Meaning Isn't Literal: Exploring Idiomatic Meaning Across Languages and Modalities
arXiv cs.CL / 4/14/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that current language models often struggle with idiomatic and culturally grounded meaning because they over-rely on surface-level lexical/semantic cues rather than metaphorical intent.
- It presents “Mediom,” a multilingual, multimodal corpus covering 3,533 Hindi, Bengali, and Thai idioms, with gold-standard explanations, cross-lingual translations, and aligned text–image representations to enable figurative disambiguation evaluation.
- The authors benchmark both large language models (textual reasoning) and vision-language models on Mediom, finding systematic failures in metaphor and idiom comprehension.
- To address these gaps, they propose “HIDE,” a hinting-based idiom explanation framework that uses error-feedback retrieval and diagnostic cues for iterative reasoning improvement.
- Overall, Mediom and HIDE are positioned as a rigorous test bed and methodology for building next-generation AI systems capable of culturally grounded, multimodal idiom understanding.
Related Articles

Black Hat Asia
AI Business

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial