IRIS: Interleaved Reinforcement with Incremental Staged Curriculum for Cross-Lingual Mathematical Reasoning

arXiv cs.CL / 4/28/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces IRIS, a two-axis curriculum framework for improving cross-lingual mathematical reasoning by combining progressively harder supervised fine-tuning with reverse-curriculum reinforcement learning to lessen dependence on step-by-step prompts.
  • IRIS uses a composite reward that covers answer correctness, step-level alignment, continuity of reasoning, and numeric incentives, optimized with Group Relative Policy Optimization (GRPO).
  • The authors release CL-Math, a dataset of 29k multilingual math problems with step-level annotations in English, Hindi, and Marathi, intended to support multilingual reasoning research.
  • Experiments across benchmarks and curated multilingual test sets show consistent performance gains, particularly for low-resource and bilingual settings, with smaller improvements for high-resource languages.

Abstract

Curriculum learning helps language models tackle complex reasoning by gradually increasing task difficulty. However, it often fails to generate consistent step-by-step reasoning, especially in multilingual and low-resource settings where cross-lingual transfer from English to Indian languages remains limited. We propose IRIS: Interleaved Reinforcement with Incremental Staged Curriculum, a two-axis framework that combines Supervised Fine-Tuning on progressively harder problems (vertical axis) with Reverse Curriculum Reinforcement Learning to reduce reliance on step-by-step guidance (horizontal axis). We design a composite reward combining correctness, step-wise alignment, continuity, and numeric incentives, optimized via Group Relative Policy Optimization (GRPO). We release CL-Math, a dataset of 29k problems with step-level annotations in English, Hindi, and Marathi. Across standard benchmarks and curated multilingual test sets, IRIS consistently improves performance, with strong results on math reasoning tasks and substantial gains in low-resource and bilingual settings, alongside modest improvements in high-resource languages.