AI Navigate

Embedding World Knowledge into Tabular Models: Towards Best Practices for Embedding Pipeline Design

arXiv cs.LG / 3/19/2026

💬 OpinionModels & Research

Key Points

  • The paper systematically benchmarks 256 embedding-based pipeline configurations for tabular prediction, covering 8 preprocessing strategies, 16 embedding models, and 2 downstream models.
  • It finds that the benefit of incorporating LLM-derived world knowledge depends strongly on the specific pipeline design, with concatenating embeddings generally outperforming replacing original columns.
  • Larger embedding models tend to yield better performance, while public leaderboard rankings and model popularity are poor indicators of actual performance.
  • Gradient boosting decision trees emerge as strong downstream models in these embedding pipelines.
  • The study offers practical guidance for researchers and practitioners on designing more effective embedding pipelines for tabular prediction tasks.

Abstract

Embeddings are a powerful way to enrich data-driven machine learning models with the world knowledge of large language models (LLMs). Yet, there is limited evidence on how to design effective LLM-based embedding pipelines for tabular prediction. In this work, we systematically benchmark 256 pipeline configurations, covering 8 preprocessing strategies, 16 embedding models, and 2 downstream models. Our results show that it strongly depends on the specific pipeline design whether incorporating the prior knowledge of LLMs improves the predictive performance. In general, concatenating embeddings tends to outperform replacing the original columns with embeddings. Larger embedding models tend to yield better results, while public leaderboard rankings and model popularity are poor performance indicators. Finally, gradient boosting decision trees tend to be strong downstream models. Our findings provide researchers and practitioners with guidance for building more effective embedding pipelines for tabular prediction tasks.