AI Navigate

Exact Federated Continual Unlearning for Ridge Heads on Frozen Foundation Models

arXiv cs.LG / 3/16/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper studies exact unlearning for a frozen foundation model with a ridge-head in a federated setting, addressing the right-to-be-forgotten with exactness rather than approximate retraining.
  • It shows the optimal unlearning solution depends only on two additive sufficient statistics, enabling a communication protocol that handles an arbitrary stream of add and delete requests via fixed-size messages.
  • The server maintains a head that, in exact arithmetic, is pointwise identical to centralized retraining after every request, with guarantees of retrain-equivalence and invariance to order and data partitioning.
  • The authors present two server-side variants and a Bayesian certificate of zero KL divergence, plus empirical results on four benchmarks confirming the guarantees and near-identical performance to centralized ridge retraining (within 1e-9 relative Frobenius error).

Abstract

Foundation models are commonly deployed as frozen feature extractors with a small trainable head to adapt to private, user-generated data in federated settings. The ``right to be forgotten'' requires removing the influence of specific samples or users from the trained model on demand. Existing federated unlearning methods target general deep models and rely on approximate reconstruction or selective retraining, making exactness costly or elusive. We study this problem in a practically relevant but under-explored regime: a frozen foundation model with a ridge-regression head. The exact optimum depends on the data only through two additive sufficient statistics, which we turn into a communication protocol supporting an arbitrary stream of \emph{add} and \emph{delete} requests via fixed-size messages. The server maintains a head that is, in exact arithmetic, \emph{pointwise identical} to centralized retraining after every request. We provide deterministic retrain-equivalence guarantees, order and partition invariance, two server-side variants, and a Bayesian certificate of zero KL divergence. Experiments on four benchmarks confirm the guarantees: both variants match centralized ridge retraining to within 10^{-9} relative Frobenius error and complete each request at orders-of-