AI Navigate

Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck

arXiv cs.CL / 3/12/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies translationese bias in multilingual LLM evaluation, where LLMs prefer machine-translated text over human-authored references, especially in low-resource languages.
  • It attributes this bias to spurious correlations with English latent manifold alignment and cross-lingual predictability.
  • It proposes DIBJudge, a robust fine-tuning framework that learns a minimally sufficient, judgment-critical representation via variational information compression, with a dedicated bias branch to isolate spurious factors.
  • It introduces a cross-covariance penalty to explicitly suppress statistical dependence between robust and bias representations to promote effective disentanglement.
  • Experimental results on multilingual reward modeling benchmarks and a translationese bias suite show DIBJudge outperforms baselines and substantially mitigates translationese bias.

Abstract

Large language models (LLMs) have become a standard for multilingual evaluation, yet they exhibit a severe systematic translationese bias. In this paper, translationese bias is characterized as LLMs systematically favoring machine-translated text over human-authored references, particularly in low-resource languages. We attribute this bias to spurious correlations with (i) latent manifold alignment with English and (ii) cross-lingual predictability. To mitigate this bias, we propose DIBJudge, a robust fine-tuning framework that learns a minimally sufficient, judgment-critical representation via variational information compression, while explicitly isolating spurious factors into the dedicated bias branch. Furthermore, we incorporate a cross-covariance penalty that explicitly suppresses statistical dependence between robust and bias representations, thereby encouraging effective disentanglement. Extensive evaluations on multilingual reward modeling benchmarks and a dedicated translationese bias evaluation suite demonstrate that the proposed DIBJudge consistently outperforms strong baselines and substantially mitigates translationese bias.