Speeding Up Mixed-Integer Programming Solvers with Sparse Learning for Branching

arXiv cs.LG / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes using sparse learning to build interpretable, lightweight ML models that approximate strong branching (SB) scores in mixed-integer programming branch-and-bound solvers.
  • The resulting models use less than 4% of the parameters of a state-of-the-art graph neural network (GNN) while maintaining competitive accuracy in predicting branching decisions.
  • Reported speedups show that the CPU-only models outperform SCIP’s default built-in branching rules and are faster than a GPU-accelerated GNN approach.
  • The authors emphasize practical deployment advantages: the models are simple to train, effective even with small training datasets, and suitable for low-resource settings without heavy GPU parallelization.
  • Extensive experiments across multiple problem classes are used to demonstrate the efficiency and robustness of the sparse learning approach for branching.

Abstract

Machine learning is increasingly used to improve decisions within branch-and-bound algorithms for mixed-integer programming. Many existing approaches rely on deep learning, which often requires very large training datasets and substantial computational resources for both training and deployment, typically with GPU parallelization. In this work, we take a different path by developing interpretable models that are simple but effective. We focus on approximating strong branching (SB) scores, a highly effective yet computationally expensive branching rule. Using sparse learning methods, we build models with fewer than 4% of the parameters of a state-of-the-art graph neural network (GNN) while achieving competitive accuracy. Relative to SCIP's built-in branching rules and the GNN-based model, our CPU-only models are faster than the default solver and the GPU-accelerated GNN. The models are simple to train and deploy, and they remain effective with small training sets, which makes them practical in low-resource settings. Extensive experiments across diverse problem classes demonstrate the efficiency of this approach.