LSR: Linguistic Safety Robustness Benchmark for Low-Resource West African Languages
arXiv cs.AI / 3/23/2026
📰 NewsModels & Research
Key Points
- LSR introduces the first cross-lingual safety benchmark for West African languages (Yoruba, Hausa, Igbo, Igala) to measure how model refusal behavior degrades when harmful intent is stated in a local language.
- It uses a dual-probe evaluation protocol that submits matched English and target-language prompts to the same model to quantify cross-language refusal degradation.
- It proposes Refusal Centroid Drift (RCD), a metric that quantifies how much of a model's English refusal behavior is lost in a target language.
- The authors evaluate Gemini 2.5 Flash across 14 culturally grounded attack probes in four harm categories, finding English refusals around 90% but West African languages drop to 35-55%, with Igala most affected (RCD = 0.55).
- The benchmark is implemented in Inspect AI and released as a PR-ready contribution to the UK AISI inspect_evals repository, with a live reference implementation and dataset publicly available.
Related Articles

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to
How to settle on a coding LLM ? What parameters to watch out for ?
Reddit r/LocalLLaMA

Andrej Karpathy's autonomous AI research agent ran 700 experiments in 2 days and gave a glimpse of where AI is heading
Reddit r/artificial

So cursor admits that Kimi K2.5 is the best open source model
Reddit r/LocalLLaMA