TopBench: A Benchmark for Implicit Prediction and Reasoning over Tabular Question Answering

arXiv cs.CL / 5/1/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces TopBench, a new benchmark for evaluating how LLMs handle implicit prediction and reasoning in tabular question answering beyond simple lookup or aggregation.
  • TopBench contains 779 samples across four sub-tasks, including single-point prediction, decision making, treatment effect analysis, and complex filtering with outputs that must include reasoning text and structured tables.
  • The study finds that current models frequently fail at intent recognition, often defaulting to straightforward retrieval rather than performing the required predictive inference.
  • It concludes that correct latent-intent disambiguation is a key prerequisite for achieving better predictive behavior, and that improving prediction precision will likely require more sophisticated modeling or reasoning.
  • Models are evaluated in both text-based and agentic workflows to compare performance under different interaction patterns.

Abstract

Large Language Models (LLMs) have advanced Table Question Answering, where most queries can be answered by extracting information or simple aggregation. However, a common class of real-world queries is implicitly predictive, requiring the inference of unobserved answers from historical patterns rather than mere retrieval. These queries introduce two challenges: recognizing latent intent and reliable predictive reasoning over massive tables. To assess LLMs in such Tabular questiOn answering with implicit Prediction tasks, we introduce TopBench, a benchmark consisting of 779 samples across four sub-tasks, ranging from single-point prediction to decision making, treatment effect analysis, and complex filtering, requiring models to generate outputs spanning reasoning text and structured tables. We evaluate diverse models under both text-based and agentic workflows. Experiments reveal that current models often struggle with intent recognition, defaulting to just lookups. Deeper analysis identifies that accurate intent disambiguation serves as the prerequisite for leading these predictive behaviors. Furthermore, elevating the upper bound of prediction precision requires the integration of more sophisticated modeling or reasoning capabilities.