Procedural Fairness via Group Counterfactual Explanation
arXiv cs.AI / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- GCIG is an in-processing regularization framework that enforces explanation invariance across protected groups by using Group Counterfactual Integrated Gradients.
- For each input, GCIG computes explanations relative to multiple Group Conditional baselines and penalizes cross-group variation in these attributions during training, formalizing procedural fairness as Group Counterfactual explanation stability.
- Empirical comparisons against six state-of-the-art methods show GCIG substantially reduces cross-group explanation disparity while preserving predictive performance and favorable accuracy–fairness trade-offs.
- The authors argue that aligning model reasoning across groups offers a principled, practical path to advancing fairness beyond outcome parity, complementing existing prediction-focused fairness objectives.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
How I Built an AI SDR Agent That Finds Leads and Writes Personalized Cold Emails
Dev.to
Complete Guide: How To Make Money With Ai
Dev.to
I Analyzed My Portfolio with AI and Scored 53/100 — Here's How I Fixed It to 85+
Dev.to
The Demethylation
Dev.to