Gap Safe Screening Rules for Fast Training of Robust Support Vector Machines under Feature Noise

arXiv cs.LG / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes safe sample screening rules that speed up training of robust support vector machines (R-SVMs) under feature noise without changing the optimal solution.
  • It identifies training points whose uncertainty sets are guaranteed to lie entirely on one side of the margin hyperplane, allowing those samples to be safely removed and shrinking the optimization problem.
  • Because R-SVMs have nonstandard structure, the screening rules are derived using Lagrangian duality rather than the Fenchel-Rockafellar duality used in many prior screening methods.
  • The authors start from an ideal screening rule and then derive a practical GAP-based rule adapted to the robust setting.
  • Experiments show the method significantly reduces training time while preserving classification accuracy, demonstrating effectiveness in robust supervised learning.

Abstract

Robust Support Vector Machines (R-SVMs) address feature noise by adopting a worst-case robust formulation that explicitly incorporates uncertainty sets into training. While this robustness improves reliability, it also leads to increased computational cost. In this work, we develop safe sample screening rules for R-SVMs that reduce the training complexity without affecting the optimal solution. To the best of our knowledge, this is the first study to apply safe screening techniques to worst-case robust models in supervised machine learning. Our approach safely identifies training samples whose uncertainty sets are guaranteed to lie entirely on either side of the margin hyperplane, thereby reducing the problem size and accelerating optimization. Owing to the nonstandard structure of R-SVMs, the proposed screening rules are derived from the Lagrangian duality rather than the Fenchel-Rockafellar duality commonly used in recent methods. Based on this analysis, we first establish an ideal screening rule, and then derive a practical rule by adapting GAP-based safe regions to the robust setting. Experiments demonstrate that the proposed method significantly reduces training time while preserving classification accuracy.