Unbiased Model Prediction Without Using Protected Attribute Information

arXiv cs.CV / 4/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses persistent bias in deep learning, noting that many existing fairness methods require protected attribute data that is often unavailable in real-world settings.
  • It introduces the Non-Protected Attribute-based Debiasing (NPAD) algorithm, which performs bias mitigation using only auxiliary information from non-protected attributes.
  • The authors propose two fairness-oriented objectives—Debiasing via Attribute Cluster Loss (DACL) and Filter Redundancy Loss (FRL)—to train models toward reduced subgroup disparities.
  • Experiments on LFWA and CelebA for facial attribute prediction report significant bias reductions across gender and age subgroups.

Abstract

The problem of bias persists in the deep learning community as models continue to provide disparate performance across different demographic subgroups. Therefore, several algorithms have been proposed to improve the fairness of deep models. However, a majority of these algorithms utilize the protected attribute information for bias mitigation, which severely limits their application in real-world scenarios. To address this concern, we have proposed a novel algorithm, termed as \textbf{Non-Protected Attribute-based Debiasing (NPAD)} algorithm for bias mitigation, that does not require the protected attribute information. The proposed NPAD algorithm utilizes the auxiliary information provided by the non-protected attributes to optimize the model for bias mitigation. Further, two different loss functions, \textbf{Debiasing via Attribute Cluster Loss (DACL)} and \textbf{Filter Redundancy Loss (FRL)} have been proposed to optimize the model for fairness goals. Multiple experiments are performed on the LFWA and CelebA datasets for facial attribute prediction, and a significant reduction in bias across different gender and age subgroups is observed.