ODUTQA-MDC: A Task for Open-Domain Underspecified Tabular QA with Multi-turn Dialogue-based Clarification

arXiv cs.CL / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ODUTQA-MDC, a new benchmark task aimed at open-domain tabular question answering when queries contain underspecified or uncertain expressions.
  • It provides a large-scale dataset (209 tables, 25,105 QA pairs) plus a fine-grained labeling scheme for more detailed evaluation than prior benchmarks.
  • The benchmark includes a dynamic clarification interface that simulates interactive user feedback, enabling assessment of multi-turn clarification behavior.
  • The authors propose MAIC-TQA, a multi-agent framework designed to detect ambiguities, carry out dialogue-based clarifications, and improve final answer quality.
  • Experiments reportedly validate both the benchmark and MAIC-TQA, positioning them as resources for advancing conversational, underspecification-aware Tabular QA research.

Abstract

The advancement of large language models (LLMs) has enhanced tabular question answering (Tabular QA), yet they struggle with open-domain queries exhibiting underspecified or uncertain expressions. To address this, we introduce the ODUTQA-MDC task and the first comprehensive benchmark to tackle it. This benchmark includes: (1) a large-scale ODUTQA dataset with 209 tables and 25,105 QA pairs; (2) a fine-grained labeling scheme for detailed evaluation; and (3) a dynamic clarification interface that simulates user feedback for interactive assessment. We also propose MAIC-TQA, a multi-agent framework that excels at detecting ambiguities, clarifying them through dialogue, and refining answers. Experiments validate our benchmark and framework, establishing them as a key resource for advancing conversational, underspecification-aware Tabular QA research.