Right at My Level: A Unified Multilingual Framework for Proficiency-Aware Text Simplification

arXiv cs.CL / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Re-RIGHT, a unified reinforcement-learning framework for proficiency-aware, adaptive multilingual text simplification without needing parallel corpus supervision.
  • It finds that prompt-based lexical simplification using target proficiency levels (CEFR/JLPT/TOPIK/HSK) performs poorly for easier proficiency bands and for non-English languages even with strong LLMs.
  • To overcome this, the authors collect 43K vocabulary-level data points across English, Japanese, Korean, and Chinese and train a compact 4B policy model.
  • Re-RIGHT uses three reward modules—vocabulary coverage, semantic preservation, and coherence—to better target readability levels while keeping meaning and fluency.
  • Experiments show improved lexical coverage at target proficiency levels compared with stronger LLM baselines while maintaining semantic and fluency quality.

Abstract

Text simplification supports second language (L2) learning by providing comprehensible input, consistent with the Input Hypothesis. However, constructing personalized parallel corpora is costly, while existing large language model (LLM)-based readability control methods rely on pre-labeled sentence corpora and primarily target English. We propose Re-RIGHT, a unified reinforcement learning framework for adaptive multilingual text simplification without parallel corpus supervision. We first show that prompting-based lexical simplification at target proficiency levels (CEFR, JLPT, TOPIK, and HSK) performs poorly at easier levels and for non-English languages, even with state-of-the-art LLMs such as GPT-5.2 and Gemini 2.5. To address this, we collect 43K vocabulary-level data across four languages (English, Japanese, Korean, and Chinese) and train a compact 4B policy model using Re-RIGHT, which integrates three reward modules: vocabulary coverage, semantic preservation, and coherence. Compared to the stronger LLM baselines, Re-RIGHT achieves higher lexical coverage at target proficiency levels while maintaining original meaning and fluency.