From Measurement to Mitigation: Quantifying and Reducing Identity Leakage in Image Representation Encoders with Linear Subspace Removal

arXiv cs.CV / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies identity leakage risks when using frozen visual image encoders (e.g., CLIP, DINOv2/v3, SSCD) on face-containing data, arguing that current practice lacks measured, deployable mitigations.
  • It introduces an attacker-aware benchmark including open-set verification at low false-accept rates, a calibrated diffusion-based template inversion check, and face-context attribution via equal-area perturbations.
  • The authors propose a one-shot linear “identity sanitization projection” (ISP) that removes an estimated identity subspace while preserving the remaining feature space to maintain downstream task utility.
  • Experiments on CelebA-20 and VGGFace2 show that leakage varies by encoder (CLIP higher than DINOv2/v3 and SSCD), that performance is robust to template inversion, and that ISP reduces linear access to near-chance while retaining high non-biometric utility.
  • The approach is reported to transfer across datasets with minor degradation, presenting what the authors call the first attacker-calibrated facial privacy audit for non-face-recognition (FR) encoders.

Abstract

Frozen visual embeddings (e.g., CLIP, DINOv2/v3, SSCD) power retrieval and integrity systems, yet their use on face-containing data is constrained by unmeasured identity leakage and a lack of deployable mitigations. We take an attacker-aware view and contribute: (i) a benchmark of visual embeddings that reports open-set verification at low false-accept rates, a calibrated diffusion-based template inversion check, and face-context attribution with equal-area perturbations; and (ii) propose a one-shot linear projector that removes an estimated identity subspace while preserving the complementary space needed for utility, which for brevity we denote as the identity sanitization projection ISP. Across CelebA-20 and VGGFace2, we show that these encoders are robust under open-set linear probes, with CLIP exhibiting relatively higher leakage than DINOv2/v3 and SSCD, robust to template inversion, and are context-dominant. In addition, we show that ISP drives linear access to near-chance while retaining high non-biometric utility, and transfers across datasets with minor degradation. Our results establish the first attacker-calibrated facial privacy audit of non-FR encoders and demonstrate that linear subspace removal achieves strong privacy guarantees while preserving utility for visual search and retrieval.

From Measurement to Mitigation: Quantifying and Reducing Identity Leakage in Image Representation Encoders with Linear Subspace Removal | AI Navigate