Multitask-Informed Prior for In-Context Learning on Tabular Data: Application to Steel Property Prediction

arXiv cs.LG / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a multitask learning framework that adapts TabPFN, a transformer-based foundation model for in-context learning on tabular data, by injecting multitask awareness into its prior via novel fine-tuning strategies.
  • It proposes two complementary adaptation methods—target averaging to maintain TabPFN’s single-target interface and task-specific adapters to provide task-wise supervision—so the model can better capture correlations across steel mechanical properties.
  • Experiments on an industrial Thin Slab Direct Rolling (TSDR) steel dataset show the multitask-adapted approach outperforms classical ML and several recent tabular learning methods on multiple metrics.
  • The authors report improvements in both predictive accuracy and computational efficiency compared with task-specific fine-tuning, positioning the method as more scalable for automated industrial quality control and process optimization.

Abstract

Accurate prediction of mechanical properties of steel during hot rolling processes, such as Thin Slab Direct Rolling (TSDR), remains challenging due to complex interactions among chemical compositions, processing parameters, and resultant microstructures. Traditional empirical and experimental methodologies, while effective, are often resource-intensive and lack adaptability to varied production conditions. Moreover, most existing approaches do not explicitly leverage the strong correlations among key mechanical properties, missing an opportunity to improve predictive accuracy through multitask learning. To address this, we present a multitask learning framework that injects multitask awareness into the prior of TabPFN--a transformer-based foundation model for in-context learning on tabular data--through novel fine-tuning strategies. Originally designed for single-target regression or classification, we augment TabPFN's prior with two complementary approaches: (i) target averaging, which provides a unified scalar signal compatible with TabPFN's single-target architecture, and (ii) task-specific adapters, which introduce task-specific supervision during fine-tuning. These strategies jointly guide the model toward a multitask-informed prior that captures cross-property relationships among key mechanical metrics. Extensive experiments on an industrial TSDR dataset demonstrate that our multitask adaptations outperform classical machine learning methods and recent state-of-the-art tabular learning models across multiple evaluation metrics. Notably, our approach enhances both predictive accuracy and computational efficiency compared to task-specific fine-tuning, demonstrating that multitask-aware prior adaptation enables foundation models for tabular data to deliver scalable, rapid, and reliable deployment for automated industrial quality control and process optimization in TSDR.