Unrealized Expectations: Comparing AI Methods vs Classical Algorithms for Maximum Independent Set

arXiv stat.ML / 4/28/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper benchmarks GPU-based AI-inspired methods (including generative-model and reinforcement-learning approaches) against classical CPU solvers for the Maximum Independent Set (MIS) problem.
  • Across even in-distribution random graphs, the state-of-the-art classical solver KaMIS on a single CPU consistently outperforms leading AI-inspired approaches, which often also fail to beat a simple degree-based greedy heuristic.
  • Adding post-processing such as local search does not close the gap; AI-inspired methods still underperform compared with KaMIS.
  • The authors introduce a “serialization” analysis showing that some non-backtracking AI-inspired methods (e.g., LTFT based on GFlowNets) end up reasoning in ways effectively similar to the simplest degree-greedy approach, explaining their poor performance.
  • The results argue for a rethink of current AI-for-CO approaches, emphasizing more rigorous benchmarking and principled integration of classical heuristics, while also finding that KaMIS performs strongly on sparse random graphs and that a shattering-threshold conjecture may not hold for real-world problem sizes (e.g., around 10^6 nodes).

Abstract

AI methods, such as generative models and reinforcement learning, have recently been applied to combinatorial optimization (CO) problems, especially NP-hard ones. This paper compares such GPU-based methods with classical CPU-based methods on the Maximum Independent Set (MIS) problem. Strikingly, even on in-distribution random graphs, leading AI-inspired methods are consistently outperformed by the state-of-the-art classical solver KaMIS running on a single CPU, and some AI-inspired methods frequently fail to surpass even the simplest degree-based greedy heuristic. Even with post-processing techniques like local search, AI-inspired methods still perform worse than CPU-based solvers. To better understand the source of these failures, we introduce a novel analysis, serialization, which reveals that non-backtracking AI-inspired methods, e.g. LTFT (which is based on GFlowNets), end up reasoning similarly to the simplest degree-based greedy, and thus worse than KaMIS. More generally, our findings suggest a need for a rethinking of current approaches in AI for CO, advocating for more rigorous benchmarking and the principled integration of classical heuristics. Additionally, we also find that CPU-based algorithm KaMIS have strong performance on sparse random graphs, which appears to show that the shattering threshold conjecture for large independent sets proposed by Coja-Oghlan & Efthymiou (2015) does not apply for real-life sizes (such as 10^6 nodes).