Fairness under uncertainty in sequential decisions

arXiv cs.LG / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies fairness in online, sequential decision-making, emphasizing that uncertainty and missing counterfactuals can disproportionately harm under-represented groups.
  • It proposes a taxonomy of uncertainty in sequential decisions—covering model uncertainty, feedback uncertainty, and prediction uncertainty—along with a shared vocabulary for evaluating fairness when uncertainty is uneven across groups.
  • The authors formalize parts of model and feedback uncertainty using counterfactual logic and reinforcement learning, showing how ignoring unobserved space can compound exclusion and reduce access.
  • Through algorithmic examples and simulations, the paper demonstrates that uncertainty-aware exploration can reduce variance and improve fairness metrics while maintaining institutional objectives like expected utility.
  • The framework is positioned as a tool for practitioners to diagnose, audit, and govern fairness risks in socio-technical decision systems where data is selectively observed.

Abstract

Fair machine learning (ML) methods help identify and mitigate the risk that algorithms encode or automate social injustices. Algorithmic approaches alone cannot resolve structural inequalities, but they can support socio-technical decision systems by surfacing discriminatory biases, clarifying trade-offs, and enabling governance. Although fairness is well studied in supervised learning, many real ML applications are online and sequential, with prior decisions informing future ones. Each decision is taken under uncertainty due to unobserved counterfactuals and finite samples, with dire consequences for under-represented groups, systematically under-observed due to historical exclusion and selective feedback. A bank cannot know whether a denied loan would have been repaid, and may have less data on marginalized populations. This paper introduces a taxonomy of uncertainty in sequential decision-making -- model, feedback, and prediction uncertainty -- providing shared vocabulary for assessing systems where uncertainty is unevenly distributed across groups. We formalize model and feedback uncertainty via counterfactual logic and reinforcement learning, and illustrate harms to decision makers (unrealized gains/losses) and subjects (compounding exclusion, reduced access) of policies that ignore the unobserved space. Algorithmic examples show it is possible to reduce outcome variance for disadvantaged groups while preserving institutional objectives (e.g. expected utility). Experiments on data simulated with varying bias show how unequal uncertainty and selective feedback produce disparities, and how uncertainty-aware exploration alters fairness metrics. The framework equips practitioners to diagnose, audit, and govern fairness risks. Where uncertainty drives unfairness rather than incidental noise, accounting for it is essential to fair and effective decision-making.