Operationalizing Fairness in Text-to-Image Models: A Survey of Bias, Fairness Audits and Mitigation Strategies

arXiv cs.CV / 4/21/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Text-to-Image (T2I) generation models are widely used but are often criticized for producing outputs that reflect societal stereotypes.
  • The paper highlights conceptual ambiguity in the field, noting that terms such as “bias” and “fairness” are inconsistently defined and operationalized.
  • It presents a systematic survey that organizes T2I fairness research into a taxonomy of bias types and fairness notions, and evaluates the mismatch between “target fairness” ideals and “threshold fairness” decision rules.
  • The survey covers mitigation strategies spanning prompt engineering and modifications to the diffusion process.
  • It proposes a framework to operationalize fairness through target-based, rigorous testing rather than relying only on descriptive evaluation metrics, aiming to improve accountability in generative AI development.

Abstract

Text-to-Image (T2I) generation models have been widely adopted across various industries, yet are criticized for frequently exhibiting societal stereotypes. While a growing body of research has emerged to evaluate and mitigate these biases, the field at present contends with conceptual ambiguity, for example terms like "bias" and "fairness" are not always clearly distinguished and often lack clear operational definitions. This paper provides a comprehensive systematic review of T2I fairness literature, organizing existing work into a taxonomy of bias types and fairness notions. We critically assess the gap between "target fairness" (normative ideals in T2I outputs) and "threshold fairness" (normative standards with actionable decision rules). Furthermore, we survey the landscape of mitigation strategies, ranging from prompt engineering to diffusion process manipulation. We conclude by proposing a new framework for operationalizing fairness that moves beyond descriptive metrics towards rigorous, target-based testing, offering an approach for more accountable generative AI development.