ReasonXL: Shifting LLM Reasoning Language Without Sacrificing Performance

arXiv cs.CL / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights a persistent gap in multilingual LLMs: even for non-English tasks, models often generate reasoning traces in English, creating a mismatch for non-English use cases.
  • It introduces ReasonXL, a large cross-domain parallel corpus of reasoning traces across five European languages (EN/DE/FR/IT/ES) with 2M+ aligned samples per language, including prompts, reasoning traces, and final outputs.
  • Using ReasonXL, the authors show language-specific reasoning can be achieved via a two-stage pipeline—SFT followed by RL with verifiable rewards (RLVR)—while maintaining or improving baseline performance with minimal general-knowledge degradation.
  • Representational analysis finds early network layers form an “activation bottleneck” that causally governs language identity, while later layers absorb most adaptation-driven changes.
  • RLVR can produce greater behavioral divergence from the base model than SFT even with smaller parameter updates, indicating a more efficient way to reroute representations toward target-language reasoning.

Abstract

Despite advances in multilingual capabilities, most large language models (LLMs) remain English-centric in their training and, crucially, in their production of reasoning traces. Even when tasked with non-English problems, these models predominantly reason in English, creating a fundamental mismatch for non-English usage scenarios. We address this disparity directly with three contributions. (i) We introduce ReasonXL, the first large-scale parallel corpus of cross-domain reasoning traces spanning five European languages (English, German, French, Italian, and Spanish), with over two million aligned samples per language, each comprising prompts, reasoning traces, and final outputs, enabling direct supervision of language-specific reasoning. (ii) Using ReasonXL, we demonstrate that LLMs can be adapted to reason entirely in a desired target language, using a simple two-stage pipeline of supervised fine-tuning (SFT) followed by reinforcement learning with verifiable rewards (RLVR). The resulting models match or exceed baseline performance, with minimal loss in general knowledge and broadly preserved cross-lingual transfer. (iii) We conduct an extensive representational analysis of the adaptation and find a clear functional division across model depth: early layers contain an activation bottleneck that causally determines language identity, while upper layers concentrate the weight and activation changes driven by adaptation. We further find that RLVR achieves greater behavioral divergence from the base model with smaller parameter updates than SFT, suggesting a more efficient representational rerouting despite much smaller weight updates.