ReSS: Learning Reasoning Models for Tabular Data Prediction via Symbolic Scaffold

arXiv cs.AI / 4/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes ReSS, a framework that bridges symbolic decision-tree logic and neural LLM reasoning for tabular prediction in high-stakes domains like healthcare and finance.
  • ReSS extracts instance-level decision paths as symbolic scaffolds and uses them to condition an LLM to produce grounded natural-language explanations that strictly follow the decision logic.
  • The generated scaffold-grounded dataset is then used to fine-tune a pretrained LLM into a specialized tabular reasoning model, with scaffold-invariant data augmentation to improve generalization and explainability.
  • The authors introduce quantitative faithfulness evaluation metrics (e.g., hallucination rate, explanation necessity, and explanation sufficiency) to verify that generated reasoning reflects the underlying logic.
  • Experiments on medical and financial benchmarks report up to 10% improvement over traditional decision trees and standard fine-tuning, alongside more faithful and consistent reasoning.

Abstract

Tabular data remains prevalent in high-stakes domains such as healthcare and finance, where predictive models are expected to provide both high accuracy and faithful, human-understandable reasoning. While symbolic models offer verifiable logic, they lack semantic expressiveness. Meanwhile, general-purpose LLMs often require specialized fine-tuning to master domain-specific tabular reasoning. To address the dual challenges of scalable data curation and reasoning consistency, we propose ReSS, a systematic framework that bridges symbolic and neural reasoning models. ReSS leverages a decision-tree model to extract instance-level decision paths as symbolic scaffolds. These scaffolds, alongside input features and labels, guide an LLM to generate grounded natural-language reasoning that strictly adheres to the underlying decision logic. The resulting high-quality dataset is used to fine-tune a pretrained LLM into a specialized tabular reasoning model, further enhanced by a scaffold-invariant data augmentation strategy to improve generalization and explainability. To rigorously assess faithfulness, we introduce quantitative metrics including hallucination rate, explanation necessity, and explanation sufficiency. Experimental results on medical and financial benchmarks demonstrate that ReSS-trained models improve traditional decision trees and standard fine-tuning approaches up to 10\% while producing faithful and consistent reasoning