Locating Demographic Bias at the Attention-Head Level in CLIP's Vision Encoder
arXiv cs.CV / 3/13/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- Introduces a mechanistic fairness audit that locates demographic bias at the level of individual attention heads in vision transformers, using projected residual-stream decomposition, zero-shot Concept Activation Vectors, and bias-augmented TextSpan analysis.
- Applies the pipeline to the CLIP ViT-L-14 encoder across 42 profession classes in the FACET benchmark, auditing gender and age bias.
- For gender, identifies four terminal-layer heads whose ablation reduces global bias with a small accuracy gain, and a layer-matched random control confirms the effect is head-specific.
- Finds that a single final-layer head accounts for most of the bias reduction in the most stereotyped classes, shifting predictions toward correct occupations.
- For age, the same approach yields weaker, more diffuse effects, suggesting age bias is encoded more diffusely than gender in this model, and indicating head-level localisation may vary by attribute.
Related Articles

Attacks On Data Centers, Qwen3.5 In All Sizes, DeepSeek’s Huawei Play, Apple’s Multimodal Tokenizer
The Batch

Your AI generated code is "almost right", and that is actually WORSE than it being "wrong".
Dev.to

Lessons from Academic Plagiarism Tools for SaaS Product Development
Dev.to

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

KI in der amtlichen Recherche beim DPMA: Was Patentanwälte bei Neuanmeldungen jetzt beachten sollten (Stand: März 2026)
Dev.to