To Lie or Not to Lie? Investigating The Biased Spread of Global Lies by LLMs

arXiv cs.CL / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates how large language models generate and propagate misinformation differently depending on the target country and language when prompted to lie.
  • It introduces GlobalLies, a multilingual dataset with 440 misinformation prompt templates and 6,867 entities across 8 languages and 195 countries, enabling systematic study of cross-lingual, cross-region bias.
  • Findings show that misinformation generation is higher for many lower-resource languages and for countries with lower Human Development Index (HDI), indicating geographically patterned bias.
  • Human and large-scale “LLM-as-a-judge” evaluations across hundreds of thousands of outputs support the conclusion that these disparities are measurable and systematic.
  • The authors assess mitigations and find uneven protection, including cross-lingual gaps in input safety classifiers and inconsistent performance of retrieval-augmented fact-checking due to unequal information availability across regions, and they release the dataset to support future defenses.

Abstract

Misinformation is on the rise, and the strong writing capabilities of LLMs lower the barrier for malicious actors to produce and disseminate false information. We study how LLMs behave when prompted to spread misinformation across languages and target countries, and introduce GlobalLies, a multilingual parallel dataset of 440 misinformation generation prompt templates and 6,867 entities, spanning 8 languages and 195 countries. Using both human annotations and large-scale LLM-as-a-judge evaluations across hundreds of thousands of generations from state-of-the-art models, we show that misinformation generation varies systematically based on the country being discussed. Propagation of lies by LLMs is substantially higher in many lower-resource languages and for countries with a lower Human Development Index (HDI). We find that existing mitigation strategies provide uneven protection: input safety classifiers exhibit cross-lingual gaps, and retrieval-augmented fact-checking remains inconsistent across regions due to unequal information availability. We release GlobalLies for research purposes, aiming to support the development of mitigation strategies to reduce the spread of global misinformation: https://github.com/zohaib-khan5040/globallies