Self-Improving Tabular Language Models via Iterative Group Alignment

arXiv cs.LG / 4/22/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies two key drawbacks of current tabular data generation approaches: static fine-tuning prevents models from learning from their own outputs, and autoregressive training can miss important global statistical properties.
  • It proposes TabGRAA (Tabular Group-Relative Advantage Alignment), a self-improving framework that iteratively uses an automated quality signal to split generated samples into high- and low-quality groups.
  • TabGRAA then optimizes a group-relative advantage objective that strengthens realistic patterns while penalizing artifacts, with the quality-signal mechanism chosen modularly rather than fixed.
  • The approach recalculates quality signals using newly generated synthetic samples each iteration, fine-tuning the language model only on these self-generated signals to reduce data-leakage risk beyond the initial supervised fine-tuning.
  • Experiments indicate TabGRAA improves fidelity, utility, and privacy, and performs competitively with or better than diffusion-based tabular synthesizers while moving beyond static replication toward dynamic self-improving generation.

Abstract

While language models have been adapted for tabular data generation, two fundamental limitations remain: (1) static fine-tuning produces models that cannot learn from their own generated samples and adapt to self-correct, and (2) autoregressive objectives preserve local token coherence but neglect global statistical properties, degrading tabular quality. Reinforcement learning offers a potential solution but requires designing reward functions that balance competing objectives -- impractical for tabular data. To fill the gap, we introduce TabGRAA (Tabular Group-Relative Advantage Alignment), the first self-improving framework for tabular data generation via automated feedback. At each iteration, TabGRAA uses an \emph{automated quality signal} -- such as a two-sample distinguishability classifier or a distance-based reward -- to partition newly generated samples into high- and low-quality groups, then optimizes a group-relative advantage objective that reinforces realistic patterns while penalizing artifacts. The specific signal is a modular choice rather than a fixed component of the framework. This establishes a virtuous feedback cycle, where the quality signal is re-computed against newly \emph{generated synthetic} samples at each round; the language model is only fine-tuned on these self-generated signals, so no additional real record is exposed during alignment, mitigating data-leakage risk beyond the initial supervised fine-tuning. Experiments show TabGRAA outperforms existing methods in fidelity, utility, and privacy, while matching or exceeding diffusion-based synthesizers, advancing tabular synthesis from static statistical replication to dynamic, self-improving generation.

Self-Improving Tabular Language Models via Iterative Group Alignment | AI Navigate