SALMUBench: A Benchmark for Sensitive Association-Level Multimodal Unlearning

arXiv cs.CV / 3/30/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SALMUBench, a new benchmark focused on evaluating sensitive association-level “unlearning” for contrastively trained multimodal encoders.
  • It uses a synthetic dataset of 60K persona–attribute associations and compares a “Compromised” model polluted with that data against a “Clean” model, both retrained from scratch on the same retain base to isolate unlearning effects.
  • The authors propose a structured evaluation protocol with specific holdout sets (e.g., holdout identity and holdout association) to measure both deletion efficacy and collateral damage.
  • Results indicate that utility-efficient deletion may be achievable, but existing unlearning methods show distinct failure modes—either under-forgetting or over-generalizing and erasing too much.
  • SALMUBench is released with dataset, models, evaluation scripts, and leaderboards to support further research on comprehensive unlearning evaluation.

Abstract

As multimodal models like CLIP become integral to downstream systems, the need to remove sensitive information is critical. However, machine unlearning for contrastively-trained encoders remains underexplored, and existing evaluations fail to diagnose fine-grained, association-level forgetting. We introduce SALMUBench (Sensitive Association-Level Multimodal Unlearning), a benchmark built upon a synthetic dataset of 60K persona-attribute associations and two foundational models: a Compromised model polluted with this data, and a Clean model without it. To isolate unlearning effects, both are trained from scratch on the same 400M-pair retain base, with the Compromised model additionally trained on the sensitive set. We propose a novel evaluation protocol with structured holdout sets (holdout identity, holdout association) to precisely measure unlearning efficacy and collateral damage. Our benchmark reveals that while utility-efficient deletion is feasible, current methods exhibit distinct failure modes: they either fail to forget effectively or over-generalize by erasing more than intended. SALMUBench sets a new standard for comprehensive unlearning evaluation, and we publicly release our dataset, models, evaluation scripts, and leaderboards to foster future research.