TAPO: Translation Augmented Policy Optimization for Multilingual Mathematical Reasoning

arXiv cs.CL / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the gap between strong English math reasoning in LLMs and weaker multilingual performance, attributing the disparity primarily to language understanding shortcomings.
  • It proposes Translation-Augmented Policy Optimization (TAPO), a reinforcement learning framework built on GRPO that uses English as a pivot with an explicit understand-then-reason alignment strategy.
  • TAPO introduces a step-level relative advantage mechanism to decouple understanding from reasoning, enabling translation-quality reward signals without causing optimization conflicts.
  • Experiments show TAPO improves multilingual mathematical reasoning and translation performance, works across multiple model types, and generalizes to unseen languages and out-of-domain tasks.

Abstract

Large Language Models (LLMs) have demonstrated remarkable proficiency in English mathematical reasoning, yet a significant performance disparity persists in multilingual contexts, largely attributed to deficiencies in language understanding. To bridge this gap, we introduce Translation-Augmented Policy Optimization (TAPO), a novel reinforcement learning framework built upon GRPO. TAPO enforces an explicit alignment strategy where the model leverages English as a pivot and follows an understand-then-reason paradigm. Crucially, we employ a step-level relative advantage mechanism that decouples understanding from reasoning, allowing the integration of translation quality rewards without introducing optimization conflicts. Extensive experiments reveal that TAPO effectively synergizes language understanding with reasoning capabilities and is compatible with various models. It outperforms baseline methods in both multilingual mathematical reasoning and translation tasks, while generalizing well to unseen languages and out-of-domain tasks.