Machine Learning for Two-Stage Graph Sparsification for the Travelling Salesman Problem

arXiv cs.LG / 4/23/2026

📰 NewsModels & Research

Key Points

  • The paper proposes a two-stage graph sparsification method for the Travelling Salesman Problem (TSP) that balances edge reduction with high recall of the optimal tour’s edges.
  • In Stage 1, it forms the candidate set by taking the union of α-Nearest and POPMUSIC to maximize recall, then uses Stage 2 to train a single model to further reduce candidate-graph density.
  • Experiments across multiple TSPLIB distance types, spatial distributions, and TSP sizes (50–500) show substantial density reduction while maintaining strong coverage.
  • The approach generalizes across distance types and distributions and outperforms recent neural sparsification methods that are largely limited to Euclidean distances.
  • The benefit grows for larger problem instances where single-stage heuristics typically degrade in performance and reliability.

Abstract

High-performance TSP solvers like LKH search within a sparsified candidate graph rather than over all possible edges. Graph sparsification is non-trivial: keep too many edges and the solver wastes time; cut too many and it loses edges that belong to the optimal tour. The two leading heuristic methods, \alpha-Nearest and POPMUSIC, produce high-quality candidate graphs, but no single heuristic is both sparse and reliable across all instance sizes and distributions. Machine learning methods can potentially learn better sparsification models. However, existing approaches operate on the complete graph, which is expensive and mostly restricted to Euclidean distances. To address this issue, we propose a two-stage graph sparsification approach: Stage~1 takes the union of \alpha-Nearest and POPMUSIC to maximise recall; Stage~2 trains a single model to reduce density. We conducted experiments across four TSPLIB distance types, five spatial distributions, and problem sizes from 50 to 500. The two-stage approach substantially reduces candidate-graph density while retaining high coverage, generalises across distance types and distributions, outperforms recent neural sparsification methods that are restricted to Euclidean distances, and becomes increasingly valuable at larger scales where single-stage heuristics degrade.