Vecchia-Inducing-Points Full-Scale Approximations for Gaussian Processes

arXiv stat.ML / 3/30/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper introduces Vecchia-inducing-points full-scale (VIF) Gaussian-process approximations that merge global inducing-point ideas with local Vecchia approximations to improve scalability on large datasets.
  • It uses an efficient correlation-based neighbor-finding strategy for the residual process, implemented via a modified cover tree algorithm, to better handle different input dimensionalities and covariance smoothness regimes.
  • For non-Gaussian likelihoods, the authors develop iterative training and prediction methods with new preconditioners and theoretical convergence guarantees, aiming to drastically reduce computation versus Cholesky-based approaches under a Laplace approximation.
  • Extensive experiments on simulated and real data indicate VIF is more accurate, numerically stable, and computationally efficient than state-of-the-art alternatives.
  • The approach is released via the open-source C++ GPBoost library with Python and R interfaces for practical adoption.

Abstract

Gaussian processes are flexible, probabilistic, non-parametric models widely used in machine learning and statistics. However, their scalability to large data sets is limited by computational constraints. To overcome these challenges, we propose Vecchia-inducing-points full-scale (VIF) approximations combining the strengths of global inducing points and local Vecchia approximations. Vecchia approximations excel in settings with low-dimensional inputs and moderately smooth covariance functions, while inducing point methods are better suited to high-dimensional inputs and smoother covariance functions. Our VIF approach bridges these two regimes by using an efficient correlation-based neighbor-finding strategy for the Vecchia approximation of the residual process, implemented via a modified cover tree algorithm. We further extend our framework to non-Gaussian likelihoods by introducing iterative methods that substantially reduce computational costs for training and prediction by several orders of magnitudes compared to Cholesky-based computations when using a Laplace approximation. In particular, we propose and compare novel preconditioners and provide theoretical convergence results. Extensive numerical experiments on simulated and real-world data sets show that VIF approximations are both computationally efficient as well as more accurate and numerically stable than state-of-the-art alternatives. All methods are implemented in the open source C++ library GPBoost with high-level Python and R interfaces.