From Measurement to Mitigation: Quantifying and Reducing Identity Leakage in Image Representation Encoders with Linear Subspace Removal
arXiv cs.CV / 4/8/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies identity leakage risks when using frozen visual image encoders (e.g., CLIP, DINOv2/v3, SSCD) on face-containing data, arguing that current practice lacks measured, deployable mitigations.
- It introduces an attacker-aware benchmark including open-set verification at low false-accept rates, a calibrated diffusion-based template inversion check, and face-context attribution via equal-area perturbations.
- The authors propose a one-shot linear “identity sanitization projection” (ISP) that removes an estimated identity subspace while preserving the remaining feature space to maintain downstream task utility.
- Experiments on CelebA-20 and VGGFace2 show that leakage varies by encoder (CLIP higher than DINOv2/v3 and SSCD), that performance is robust to template inversion, and that ISP reduces linear access to near-chance while retaining high non-biometric utility.
- The approach is reported to transfer across datasets with minor degradation, presenting what the authors call the first attacker-calibrated facial privacy audit for non-face-recognition (FR) encoders.
Related Articles

Black Hat Asia
AI Business

Meta's latest model is as open as Zuckerberg's private school
The Register

AI fuels global trade growth as China-US flows shift, McKinsey finds
SCMP Tech

Why multi-agent AI security is broken (and the identity patterns that actually work)
Dev.to
BANKING77-77: New best of 94.61% on the official test set (+0.13pp) over our previous tests 94.48%.
Reddit r/artificial