KumoRFM-2: Scaling Foundation Models for Relational Learning

arXiv cs.LG / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • KumoRFM-2 is a new pre-trained foundation model designed for relational data, enabling in-context learning and fine-tuning across diverse predictive tasks.
  • Unlike tabular foundation models, it natively processes connected tables together without manual flattening or target-variable generation, while preserving temporal consistency.
  • The model is pre-trained on a mix of synthetic and real-world data across four representation axes (row/column within tables and foreign-key/cross-sample across databases).
  • A key architectural/training change is injecting task information earlier, which improves selection of task-relevant columns and robustness to noisy inputs.
  • Experiments on 41 benchmarks show up to 8% gains over supervised and other foundation approaches, including strong performance under cold-start and noisy conditions, and it scales to billion-scale relational datasets.

Abstract

We introduce KumoRFM-2, the next iteration of a pre-trained foundation model for relational data. KumoRFM-2 supports in-context learning as well as fine-tuning and is applicable to a wide range of predictive tasks. In contrast to tabular foundation models, KumoRFM-2 natively operates on relational data, processing one or more connected tables simultaneously without manual table flattening or target variable generation, all while preserving temporal consistency. KumoRFM-2 leverages a large corpus of synthetic and real-world data to pre-train across four axes: the row and column dimensions at the individual table level, and the foreign key and cross-sample dimensions at the database level. In contrast to its predecessor, KumoRFM-2 injects task information as early as possible, enabling sharper selection of task-relevant columns and improved robustness to noisy data. Through extensive experiments on 41 challenging benchmarks and analysis around expressivity and sensitivity, we demonstrate that KumoRFM-2 outperforms supervised and foundational approaches by up to 8%, while maintaining strong performance under extreme settings of cold start and noisy data. To our knowledge, this is the first time a few-shot foundation model has been shown to surpass supervised approaches on common benchmark tasks, with performance further improving upon fine-tuning. Finally, while KumoRFM-1 was limited to small-scale in-memory datasets, KumoRFM-2 scales to billion-scale relational datasets.