The Cost of Relaxation: Evaluating the Error in Convex Neural Network Verification

arXiv cs.AI / 4/22/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes how convex relaxations used in neural network verification can lose soundness by allowing outputs that the original network cannot produce.
  • It models the relaxation set as a lattice, where the fully relaxed model (all neurons linearized) is the top element and the original network is the bottom element.
  • The authors derive analytical bounds on the worst-case difference (measured via the ℓ∞-distance) between original-network outputs and fully relaxed outputs.
  • They find the output divergence grows exponentially with network depth and linearly with input radius, while the misclassification probability shows a step-like dependence on input radius.
  • Experiments on MNIST, Fashion MNIST, and random networks support the theoretical predictions.

Abstract

Many neural network (NN) verification systems represent the network's input-output relation as a constraint program. Sound and complete, representations involve integer constraints, for simulating the activations. Recent works convexly relax the integer constraints, improving performance, at the cost of soundness. Convex relaxations consider outputs that are unreachable by the original network. We study the worst case divergence between the original network and its convex relaxations; both qualitatively and quantitatively. The relaxations' space forms a lattice, where the top element corresponds to a full relaxation, with every neuron linearized. The bottom element corresponds to the original network. We provide analytical upper and lower bounds for the \ell_\infty-distance between the fully relaxed and original outputs. This distance grows exponentially, w.r.t. the network's depth, and linearly w.r.t. the input's radius. The misclassification probability exhibits a step-like behavior, w.r.t. input radius. Our results are supported by experiments on MNIST, Fashion MNIST and random networks.