GF-Score: Certified Class-Conditional Robustness Evaluation with Fairness Guarantees
arXiv cs.LG / 4/15/2026
📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper introduces GF-Score, a framework that produces certified, class-conditional robustness profiles instead of a single aggregate score to expose how robustness varies across classes.
- It defines four fairness/welfare-economics-grounded metrics (RDI, NRGC, WCR, and FP-GREAT) to quantify disparity and worst-case class performance under certified robustness guarantees.
- GF-Score also removes reliance on adversarial attacks via a self-calibration procedure that tunes a temperature parameter using only clean accuracy correlations.
- Experiments on 22 RobustBench models for CIFAR-10 and ImageNet show that the decomposition matches the original method exactly and highlight consistent vulnerability patterns, such as “cat” being weakest in most CIFAR-10 models.
- The authors provide an attack-free auditing pipeline for diagnosing where certified robustness fails to protect classes evenly and release code on GitHub.




