BITS for GAPS: Bayesian Information-Theoretic Sampling for hierarchical GAussian Process Surrogates
arXiv stat.ML / 3/24/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes BITS for GAPS, a framework for information-theoretic experimental design of hierarchical Gaussian-process surrogate models that explicitly accounts for hyperparameter uncertainty.
- Rather than using fixed or point-estimated hyperparameters in acquisition functions, the method uses Bayesian hierarchical modeling to propagate uncertainty from both the latent GP function and its hyperparameters into the sampling criterion.
- The authors derive theoretical support, including a closed-form approximation and a lower bound on posterior differential entropy, to characterize the information gain used for adaptive sampling.
- In a vapor-liquid equilibrium hybrid modeling case study, the approach improves expected information gain and predictive accuracy by preferentially sampling regions with high uncertainty in the Wilson activity model.
- The results demonstrate how partial physical knowledge can be encoded through hierarchical priors in a GP surrogate and then used to inform downstream distillation design.
Related Articles

Interactive Web Visualization of GPT-2
Reddit r/artificial
Stop Treating AI Interview Fraud Like a Proctoring Problem
Dev.to
[R] Causal self-attention as a probabilistic model over embeddings
Reddit r/MachineLearning
The 5 software development trends that actually matter in 2026 (and what they mean for your startup)
Dev.to
InVideo AI Review: Fast Finished
Dev.to