Fairness of Classifiers in the Presence of Constraints between Features
arXiv cs.AI / 5/4/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines how standard fairness notions for classifiers—specifically independence from protected attributes like gender—can fail when constraints exist between features, masking underlying dependencies.
- It proposes a new fairness criterion based on the existence of a “fair explanation,” defined as a prime-implicant reason for the decision that contains no protected features, accounting for feature constraints.
- The authors find that ignoring constraints can drastically change whether a decision is considered fair under this explanation-based definition, even when protected and non-protected features are not directly constrained.
- They analyze three definitions of classifier fairness (all decisions have fair explanations, at least one decision has a fair explanation, or outcomes are invariant under changes to protected features) and study the computational complexity of testing these properties.
Related Articles
AnnouncementsBuilding a new enterprise AI services company with Blackstone, Hellman & Friedman, and Goldman Sachs
Anthropic News

Dara Khosrowshahi on replacing Uber drivers — and himself — with AI
The Verge

CLMA Frame Test
Dev.to

You Are Right — You Don't Need CLAUDE.md
Dev.to

Governance and Liability in AI Agents: What I Built Trying to Answer Those Questions
Dev.to