StressEval: Failure-Driven Dynamic Benchmarking for Knowledge-Intensive Reasoning in Large Language Models

arXiv cs.CL / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • Static LLM benchmarks are increasingly unreliable for knowledge-intensive reasoning due to contamination and overfitting, while some dynamic benchmarks trade off answerability and controllability.
  • The paper introduces StressEval, a failure-driven framework that converts observed model failures into dynamic test instances by identifying the failed reasoning step, synthesizing targeted new problems, and filtering for grounded, unambiguous cases.
  • StressEval uses a difficulty “card” to capture root causes and difficulty factors, then performs dual-perspective data synthesis aimed at both knowledge gaps and reasoning breakdowns.
  • Using multiple knowledge-intensive reasoning datasets, the authors create Dynamic OneEval and show that it causes substantially larger performance drops across several leading LLMs while preserving explicit difficulty factors for more actionable iteration.

Abstract

Static benchmarks for LLMs are increasingly compromised by contamination and overfitting especially on knowledge intensive reasoning tasks While recent dynamic benchmarks can alleviate staleness they often increase difficulty at the expense of answerability and controllability In this paper we propose StressEval a failure driven data synthesis framework that turns observed model failures into dynamic challenging and controllable test instances StressEval consists of three stages first it constructs a semi structured difficulty card that identifies the failed reasoning step and its root cause second it applies a dual perspective instance synthesis method that targets both knowledge gaps and reasoning breakdowns while preserving the underlying difficulty factors and third it applies a gating mechanism to retain only grounded unambiguous instances Seeding from multiple knowledge intensive reasoning datasets we employ StressEval to build Dynamic OneEval a focused suite of challenging dynamic benchmark Across several state of the art LLMs Dynamic OneEval yields substantially larger performance drops than the original benchmarks while retaining explicit difficulty factors enabling more actionable iteration