Selecting Feature Interactions for Generalized Additive Models by Distilling Foundation Models

arXiv cs.LG / 4/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces TabDistill, a method to identify meaningful (including higher-order or context-dependent) feature interactions for tabular modeling by using tabular foundation models plus post-hoc distillation.
  • TabDistill works by first fitting a tabular foundation model to the dataset and then applying interaction attribution to extract salient feature interactions from that model.
  • The extracted interactions are evaluated by inserting them as terms in generalized additive models (GAMs), aiming to improve both predictive performance and interpretability.
  • Experiments across tasks show that interactions discovered by TabDistill produce consistent gains in the downstream GAMs’ predictive accuracy compared with prior interaction-selection approaches.
  • Overall, the work positions tabular foundation models as data-driven guides that bridge high-capacity representation learning with interpretable additive modeling frameworks.

Abstract

Identifying meaningful feature interactions is a central challenge in building accurate and interpretable models for tabular data. Generalized additive models (GAMs) have shown great success at modeling tabular data, but often rely on heuristic procedures to select interactions, potentially missing higher-order or context-dependent effects. To meet this challenge, we propose TabDistill, a method that leverages tabular foundation models and post-hoc distillation methods. Our key intuition is that tabular foundation models implicitly learn rich, adaptive feature dependencies through large-scale representation learning. Given a dataset, TabDistill first fits a tabular foundation model to the dataset, and then applies a post-hoc interaction attribution method to extract salient feature interactions from it. We evaluate these interactions by then using them as terms in a GAM. Across tasks, we find that interactions identified by TabDistill lead to consistent improvements in downstream GAMs' predictive performance. Our results suggest that tabular foundation models can serve as effective, data-driven guides for interaction discovery, bridging high-capacity models and interpretable additive frameworks.