Progressing beyond Art Masterpieces or Touristic Clich\'es: how to assess your LLMs for cultural alignment?

arXiv cs.CL / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights that while cultural misalignment in LLMs has gained attention, there has been limited research on creating datasets specifically for cultural assessment.
  • It reviews existing dataset approaches, pinpoints their key limitations, and proposes concrete design guidelines for annotators.
  • Using these guidelines, the authors construct a new dataset intended to measure cultural alignment more reliably.
  • The study runs contrastive experiments showing that the resulting test sets have stronger discriminative power, separating culture-specialized models from non-specialized ones under comparable conditions.

Abstract

Although the cultural (mis)alignment of Large Language Models (LLMs) has attracted increasing attention -- often framed in terms of cultural bias -- until recently there has been limited work on the design and development of datasets for cultural assessment. Here, we review existing approaches to such datasets and identify their main limitations. To address these issues, we propose design guidelines for annotators and report on the construction of a dataset built according to these principles. We further present a series of contrastive experiments conducted with this dataset. The results demonstrate that our design yields test sets with greater discriminative power, effectively distinguishing between models specialized for a given culture and those that are not, ceteris paribus.