No Universal Courtesy: A Cross-Linguistic, Multi-Model Study of Politeness Effects on LLMs Using the PLUM Corpus

arXiv cs.CL / 4/20/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study tests how LLMs respond to user prompts that vary in politeness and impoliteness using the PLUM corpus across three languages (English, Hindi, Spanish) and five models (Gemini-Pro, GPT-4o Mini, Claude 3.7 Sonnet, DeepSeek-Chat, Llama 3).
  • Across 22,500 prompt–response pairs, the researchers evaluate politeness on five interaction histories (raw, polite, impolite) and five politeness levels with an eight-factor framework covering quality and safety dimensions such as coherence, clarity, depth, responsiveness, context retention, toxicity, conciseness, and readability.
  • Polite prompts can improve average response quality by up to about 11%, while impolite tones degrade it, but these effects are not universal and vary substantially by language and model.
  • The findings suggest language-specific best practices for tone: English favors courteous or direct wording, Hindi prefers deferential and indirect phrasing, and Spanish performs better with more assertive tone.
  • The paper releases PLUM, a publicly available multilingual dataset of human-validated prompts, to support reproducibility and future hypothesis testing on politeness theory for LLM behavior.

Abstract

This paper explores the response of Large Language Models (LLMs) to user prompts with different degrees of politeness and impoliteness. The Politeness Theory by Brown and Levinson and the Impoliteness Framework by Culpeper form the basis of experiments conducted across three languages (English, Hindi, Spanish), five models (Gemini-Pro, GPT-4o Mini, Claude 3.7 Sonnet, DeepSeek-Chat, and Llama 3), and three interaction histories between users (raw, polite, and impolite). Our sample consists of 22,500 pairs of prompts and responses of various types, evaluated across five levels of politeness using an eight-factor assessment framework: coherence, clarity, depth, responsiveness, context retention, toxicity, conciseness, and readability. The findings show that model performance is highly influenced by tone, dialogue history, and language. While polite prompts enhance the average response quality by up to ~11% and impolite tones worsen it, these effects are neither consistent nor universal across languages and models. English is best served by courteous or direct tones, Hindi by deferential and indirect tones, and Spanish by assertive tones. Among the models, Llama is the most tone-sensitive (11.5% range), whereas GPT is more robust to adversarial tone. These results indicate that politeness is a quantifiable computational variable that affects LLM behaviour, though its impact is language- and model-dependent rather than universal. To support reproducibility and future work, we additionally release PLUM (Politeness Levels in Utterances, Multilingual), a publicly available corpus of 1,500 human-validated prompts across three languages and five politeness categories, and provide a formal supplementary analysis of six falsifiable hypotheses derived from politeness theory, empirically assessed against the dataset.