Infeasibility Aware Large Language Models for Combinatorial Optimization

arXiv cs.LG / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that most LLM approaches to NP-hard combinatorial optimization focus on finding feasible solutions but lack explicit mechanisms to detect and certify infeasibility.
  • It proposes an infeasibility-aware framework that builds certifiable training datasets, uses supervised fine-tuning, and applies LLM-assisted downstream search.
  • For the minor-embedding problem, the authors introduce a new mathematical programming formulation that supports provable zero-phase infeasibility screening, enabling training instances to be labeled as feasible (with certificates) or certifiably infeasible.
  • Using this exact pipeline-generated data, they fine-tune an 8B-parameter LLM to perform both solution generation and infeasibility detection.
  • In experiments, the fine-tuned model improves accuracy by up to 30% over GPT-5.2 and LLM-guided warm starts yield up to 2× speedups in downstream local search even when LLM outputs are imperfect.

Abstract

Large language models (LLMs) are increasingly explored for NP-hard combinatorial optimization problems, but most existing methods emphasize feasible-instance solution generation and do not explicitly address infeasibility detection. We propose an infeasibility-aware framework that combines certifiable dataset construction, supervised fine-tuning, and LLM-assisted downstream search. For the minor-embedding problem, we introduce a new mathematical programming formulation together with provable zero-phase infeasibility screening, which enables scalable construction of training instances labeled either as feasible with structured certificates or as certifiably infeasible. Using training data generated through this exact optimization pipeline, we show that an 8B-parameter LLM can be fine-tuned to jointly perform solution generation and infeasibility detection. We further utilize LLM outputs as warm starts for downstream local search, providing a practical way to accelerate optimization even when the LLM outputs are imperfect. Experiments show that our fine-tuned model improves overall accuracy by up to 30\% over GPT-5.2; meanwhile LLM-guided warm starts provide up to 2\times speedup compared with starting from scratch in downstream local search.