AI Navigate

A Closed-Form Solution for Debiasing Vision-Language Models with Utility Guarantees Across Modalities and Tasks

arXiv cs.CV / 3/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a training-free debiasing method for Vision-Language Models that yields a closed-form solution in cross-modal space with Pareto-optimal fairness and bounded utility loss.
  • It requires no annotated data and can jointly debias both visual and textual modalities across downstream tasks without retraining.
  • The method achieves fairness across group and intersectional metrics on tasks such as zero-shot image classification, text-to-image retrieval, and text-to-image generation while preserving task performance.
  • Extensive experiments show the approach outperforms existing debiasing methods across diverse datasets and fairness metrics.

Abstract

While Vision-Language Models (VLMs) have achieved remarkable performance across diverse downstream tasks, recent studies have shown that they can inherit social biases from the training data and further propagate them into downstream applications. To address this issue, various debiasing approaches have been proposed, yet most of them aim to improve fairness without having a theoretical guarantee that the utility of the model is preserved. In this paper, we introduce a debiasing method that yields a \textbf{closed-form} solution in the cross-modal space, achieving Pareto-optimal fairness with \textbf{bounded utility losses}. Our method is \textbf{training-free}, requires \textbf{no annotated data}, and can jointly debias both visual and textual modalities across downstream tasks. Extensive experiments show that our method outperforms existing methods in debiasing VLMs across diverse fairness metrics and datasets for both group and \textbf{intersectional} fairness in downstream tasks such as zero-shot image classification, text-to-image retrieval, and text-to-image generation while preserving task performance.