Procedural Fairness via Group Counterfactual Explanation
arXiv cs.AI / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- GCIG is an in-processing regularization framework that enforces explanation invariance across protected groups by using Group Counterfactual Integrated Gradients.
- For each input, GCIG computes explanations relative to multiple Group Conditional baselines and penalizes cross-group variation in these attributions during training, formalizing procedural fairness as Group Counterfactual explanation stability.
- Empirical comparisons against six state-of-the-art methods show GCIG substantially reduces cross-group explanation disparity while preserving predictive performance and favorable accuracy–fairness trade-offs.
- The authors argue that aligning model reasoning across groups offers a principled, practical path to advancing fairness beyond outcome parity, complementing existing prediction-focused fairness objectives.




