Query-Efficient Quantum Approximate Optimization via Graph-Conditioned Trust Regions

arXiv cs.LG / 4/29/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a graph-conditioned trust-region method to reduce the dominant query (objective evaluation) cost in low-depth QAOA implementations rather than focusing on circuit depth.
  • A graph neural network outputs a Gaussian distribution over QAOA angles, using the mean to initialize local optimization and the covariance to define an ellipsoidal trust region that limits the search.
  • The model’s predicted uncertainty sets an instance-dependent evaluation budget, effectively creating a learned search policy instead of providing only an initial parameter guess.
  • Under stated assumptions (smoothness, curvature, calibration, and noise), the authors derive theoretical guarantees including bounds on objective degradation, lower bounds on gradient variance, ordering preservation under depolarizing noise, and finite-sample coverage.
  • Experiments on MaxCut with QAOA depth p=2 across several random graph families show a large reduction in mean circuit evaluations (from 343/85 down to about 45±7) while keeping approximation quality within ~3 percentage points of strong heuristics.

Abstract

In low-depth implementations of the Quantum Approximate Optimization Algorithm (QAOA), the dominant cost is often the number of objective evaluations rather than circuit depth. We introduce a graph-conditioned trust-region method for reducing this query cost. A graph neural network predicts a Gaussian distribution N(mu, Sigma) over QAOA angles. The mean initializes a local optimizer, the covariance defines an ellipsoidal trust region that constrains the search, and the predicted uncertainty determines an instance-dependent evaluation budget. Thus the learned distribution defines a search policy rather than only an initial parameter estimate. Under explicit assumptions on local smoothness, curvature, calibration, and noise, we derive bounds on objective degradation within the trust region, lower bounds on gradient variance, preservation of expected objective ordering under depolarizing noise, and finite-sample coverage guarantees. We evaluate the method for MaxCut at depth p = 2 on Erdos-Renyi, 3-regular, Barabasi-Albert, and Watts-Strogatz graphs with n = 8-16 vertices. Relative to random restarts and the strongest learned point-prediction baseline, the method reduces the mean number of circuit evaluations from 343 and 85 to 45 +/- 7, while maintaining sampled approximation ratios within 3 percentage points of concentration-based heuristics. The method does not improve absolute approximation ratios; its advantage is reduced query cost at comparable solution quality. The predictive uncertainty is calibrated in the experiments, with ECE = 0.052 and Spearman correlation rho = 0.770, and the learned trust regions transfer to graph sizes not used during training. The results identify a low-depth, query-dominated regime in which graph-conditioned trust regions reduce the query cost of QAOA without modifying the ansatz.