Is Mathematical Problem-Solving Expertise in Large Language Models Associated with Assessment Performance?

arXiv cs.AI / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates whether an LLM’s math problem-solving skill translates into better step-level assessment of learners’ reasoning, using PROCESSBENCH with GSM8K and MATH subsets.
  • Two math tutor-agent setups based on GPT-4 and GPT-5 are tested on the same problems: one setting solves the problems, and the other predicts the earliest erroneous step in a provided solution.
  • Results show a consistent within-model pattern: the same model achieves substantially higher assessment accuracy on items it solves correctly than on items it solves incorrectly, with statistically significant associations across models and datasets.
  • Assessment is still harder than direct problem solving, particularly when the input solutions already contain errors, indicating that diagnosis requires more than raw solving ability.
  • The findings imply that AI-supported adaptive instructional systems for formative assessment should incorporate additional capabilities for step tracking, monitoring, and accurate error localization.

Abstract

Large Language Models (LLMs) are increasingly used in math education not only as problem solvers but also as assessors of learners' reasoning. However, it remains unclear whether stronger math problem-solving ability is associated with stronger step-level assessment performance. This study examines that relationship using the GSM8K and MATH subsets of PROCESSBENCH, a human-annotated benchmark for identifying the earliest erroneous step in mathematical reasoning. We evaluate two LLM-based math tutor agent settings, instantiated with GPT-4 and GPT-5, in two independent tasks on the same math problems: solving the original problem and assessing a benchmark-provided solution by predicting the earliest erroneous step. Results show a consistent within-model pattern: assessment accuracy is substantially higher on math problem items the same model solved correctly than on items it solved incorrectly, with statistically significant associations across both models and datasets. At the same time, assessment remains more difficult than direct problem solving, especially on error-present solutions. These findings suggest that math problem-solving expertise supports stronger assessment performance, but reliable step-level diagnosis also requires additional capabilities such as step tracking, monitoring, and precise error localization. The results have implications for the design and evaluation of AI-supported Adaptive Instructional Systems (AISs) for formative assessment in math education.