ALBA: A European Portuguese Benchmark for Evaluating Language and Linguistic Dimensions in Generative LLMs

arXiv cs.CL / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ALBA, a linguistically grounded benchmark specifically designed to evaluate European Portuguese (pt-PT) capabilities in generative LLMs, addressing the gap left by pt-BR–centric data and benchmarks.
  • ALBA covers eight linguistic dimensions—such as syntax, morphology, lexicology, discourse analysis, culture-bound semantics, word plays, and phonetics/phonology—to assess proficiency in varied language-related tasks.
  • The benchmark is manually constructed by language experts and evaluated using an LLM-as-a-judge setup to enable scalable assessment of pt-PT generated language.
  • Experiments across multiple LLMs show that performance varies by linguistic dimension, emphasizing the need for variety- and linguistics-sensitive benchmarking for under-represented languages like pt-PT.

Abstract

As Large Language Models (LLMs) expand across multilingual domains, evaluating their performance in under-represented languages becomes increasingly important. European Portuguese (pt-PT) is particularly affected, as existing training data and benchmarks are mainly in Brazilian Portuguese (pt-BR). To address this, we introduce ALBA, a linguistically grounded benchmark designed from the ground up to assess LLM proficiency in linguistic-related tasks in pt-PT across eight linguistic dimensions, including Language Variety, Culture-bound Semantics, Discourse Analysis, Word Plays, Syntax, Morphology, Lexicology, and Phonetics and Phonology. ALBA is manually constructed by language experts and paired with an LLM-as-a-judge framework for scalable evaluation of pt-PT generated language. Experiments on a diverse set of models reveal performance variability across linguistic dimensions, highlighting the need for comprehensive, variety-sensitive benchmarks that support further development of tools in pt-PT.