Auditing Demographic Bias in Facial Landmark Detection for Fair Human-Robot Interaction

arXiv cs.CV / 4/9/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a systematic audit of demographic bias in facial landmark detection, focusing on age, gender, and race impacts relevant to fair human-robot interaction (HRI).
  • It introduces a controlled statistical methodology to separate demographic effects from confounding visual factors such as head pose and image resolution.
  • Using a representative baseline model, the study finds that demographic attributes initially appear less influential than confounders, with pose and resolution dominating performance differences.
  • After controlling for confounders, gender and race-related performance disparities largely disappear, but a statistically significant age effect remains, with higher bias for older individuals.
  • The authors conclude that fairness risks can originate in low-level vision components like landmark detection and propagate through the HRI perception pipeline, potentially harming vulnerable groups, underscoring the need for auditing and correction.

Abstract

Fairness in human-robot interaction critically depends on the reliability of the perceptual models that enable robots to interpret human behavior. While demographic biases have been widely studied in high-level facial analysis tasks, their presence in facial landmark detection remains unexplored. In this paper, we conduct a systematic audit of demographic bias in this task, analyzing the age, gender and race biases. To this end we introduce a controlled statistical methodology to disentangle demographic effects from confounding visual factors. Evaluations of a standard representative model demonstrate that confounding visual factors, particularly head pose and image resolution, heavily outweigh the impact of demographic attributes. Notably, after accounting for these confounders, we show that performance disparities across gender and race vanish. However, we identify a statistically significant age-related effect, with higher biases observed for older individuals. This shows that fairness issues can emerge even in low-level vision components and can propagate through the HRI pipeline, disproportionately affecting vulnerable populations. We argue that auditing and correcting such biases is a necessary step toward trustworthy and equitable robot perception systems.