TDATR: Improving End-to-End Table Recognition via Table Detail-Aware Learning and Cell-Level Visual Alignment

arXiv cs.CV / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces TDATR, an end-to-end table recognition approach that improves integration between table structure and cell/content understanding versus traditional modular pipelines.
  • TDATR uses a “perceive-then-fuse” design: it first performs table detail-aware learning via multiple structure- and content-focused tasks framed under a language modeling paradigm to boost robustness across varied document types.
  • It then generates structured HTML outputs by fusing learned implicit table details, aiming to make training more efficient and effective in data-constrained settings.
  • A structure-guided cell localization module is added to locate cells and strengthen vision-language alignment, improving both interpretability and accuracy.
  • The method reports state-of-the-art or highly competitive results on seven benchmarks without dataset-specific fine-tuning, suggesting strong generalization.

Abstract

Tables are pervasive in diverse documents, making table recognition (TR) a fundamental task in document analysis. Existing modular TR pipelines separately model table structure and content, leading to suboptimal integration and complex workflows. End-to-end approaches rely heavily on large-scale TR data and struggle in data-constrained scenarios. To address these issues, we propose TDATR (Table Detail-Aware Table Recognition) improves end-to-end TR through table detail-aware learning and cell-level visual alignment. TDATR adopts a ``perceive-then-fuse'' strategy. The model first performs table detail-aware learning to jointly perceive table structure and content through multiple structure understanding and content recognition tasks designed under a language modeling paradigm. These tasks can naturally leverage document data from diverse scenarios to enhance model robustness. The model then integrates implicit table details to generate structured HTML outputs, enabling more efficient TR modeling when trained with limited data. Furthermore, we design a structure-guided cell localization module integrated into the end-to-end TR framework, which efficiently locates cell and strengthens vision-language alignment. It enhances the interpretability and accuracy of TR. We achieve state-of-the-art or highly competitive performance on seven benchmarks without dataset-specific fine-tuning.