FollowTable: A Benchmark for Instruction-Following Table Retrieval

arXiv cs.CL / 5/4/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that traditional table retrieval is overly based on topical semantic similarity, and proposes a new instruction-driven task called Instruction-Following Table Retrieval (IFTR) for LLM-based agentic systems.
  • IFTR is defined as requiring models to satisfy both topical relevance and fine-grained instruction constraints, including content-scope rules (inclusion/exclusion) and schema-grounded requirements (column semantics and representation granularity).
  • The authors introduce FollowTable, a large-scale benchmark for IFTR built using a taxonomy-driven annotation pipeline to enable systematic evaluation.
  • They also propose an Instruction Responsiveness Score metric to measure whether retrieval rankings adapt to user instructions compared with a topic-only baseline.
  • Experimental results show existing retrieval models often fail at fine-grained instruction following for tables, displaying biases toward surface-level semantics and difficulty with schema-aware constraints.

Abstract

Table Retrieval (TR) has traditionally been formulated as an ad-hoc retrieval problem, where relevance is primarily determined by topical semantic similarity. With the growing adoption of LLM-based agentic systems, access to structured data is increasingly instruction-driven, where relevance is conditional on explicit content and schema constraints rather than topical similarity alone. We therefore formalize Instruction-Following Table Retrieval (IFTR), a new task that requires models to jointly satisfy topical relevance and fine-grained instruction constraints. We identify two core challenges in IFTR: (i) sensitivity to content scope, such as inclusion and exclusion constraints, and (ii) awareness of schema-grounded requirements, including column semantics and representation granularity--capabilities largely absent in existing retrievers. To support systematic evaluation, we introduce FollowTable, the first large-scale benchmark for IFTR, constructed via a taxonomy-driven annotation pipeline. We further propose a new metric, termed the Instruction Responsiveness Score, to evaluate whether retrieval rankings consistently adapt to user instructions relative to a topic-only baseline. Our results indicate that existing retrieval models struggle to follow fine-grained instructions over tabular data. In particular, they exhibit systematic biases toward surface-level semantic cues and remain limited in handling schema-grounded constraints, highlighting substantial room for future improvements.