Exploring the Capability Boundaries of LLMs in Mastering of Chinese Chouxiang Language

arXiv cs.CL / 4/20/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Mouse, a specialized benchmark to evaluate how well LLMs can handle NLP tasks in Chouxiang Language, a subcultural language on the Chinese internet.
  • Experiments across six tasks find that current SOTA LLMs show clear weaknesses on several tasks, while they perform relatively well when contextual semantic understanding is required.
  • The study investigates why performance is generally low on Chouxiang Language, including an examination of whether LLM-as-a-judge for translation matches human judgments and values.
  • It analyzes the key factors that drive Chouxiang translation quality and encourages further NLP research focused on multicultural integration and evolving online language dynamics.
  • The authors make their code and data publicly available to support follow-up research.

Abstract

While large language models (LLMs) have achieved remarkable success in general language tasks, their performance on Chouxiang Language, a representative subcultural language in the Chinese internet context, remains largely unexplored. In this paper, we introduce Mouse, a specialized benchmark designed to evaluate the capabilities of LLMs on NLP tasks involving Chouxiang Language across six tasks. Experimental results show that, current state-of-the-art (SOTA) LLMs exhibit clear limitations on multiple tasks, while performing well on tasks that involve contextual semantic understanding. In addition, we further discuss the reasons behind the generally low performance of SOTA LLMs on Chouxiang Language, examine whether the LLM-as-a-judge approach adopted for translation tasks aligns with human judgments and values, and analyze the key factors that influence Chouxiang translation. Our study aims to promote further research in the NLP community on multicultural integration and the dynamics of evolving internet languages. Our code and data are publicly available.