AI Navigate

Parameter-Efficient Token Embedding Editing for Clinical Class-Level Unlearning

arXiv cs.AI / 3/23/2026

📰 NewsModels & Research

Key Points

  • STEU is a parameter-efficient method for clinical language models that performs class-level unlearning by updating only PMI-selected token embeddings and a small classifier head, while keeping all encoder layers frozen.
  • It achieves near-complete forgetting on MIMIC-IV (forget F1 = 0.0004) and maintains retained task performance (avg F1 = 0.4766) with only about 0.19% of parameters modified.
  • The approach was evaluated across MIMIC-IV, MIMIC-III, and eICU using BioClinicalBERT, BERT-base, and DistilBERT, showing consistent forgetting with minimal utility loss.
  • This work suggests targeted behavioral unlearning can be achieved via sparse embedding edits without modifying deeper encoder representations, offering privacy-preserving model maintenance.

Abstract

Machine unlearning is increasingly important for clinical language models, where privacy regulations and institutional policies may require removing sensitive information from deployed systems without retraining from scratch. In practice, deletion requests must balance effective forgetting of targeted information with preservation of model utility and minimal parameter modification. We introduce Sparse Token Embedding Unlearning (STEU), a parameter-efficient method for behavioral class-level unlearning that updates only PMI-selected token embeddings together with a small classifier head while keeping all encoder layers frozen. Across experiments on MIMIC-IV, MIMIC-III, and eICU using BioClinicalBERT, BERT-base, and DistilBERT, STEU consistently suppresses the target class while largely preserving retained task performance. In the primary MIMIC-IV setting, STEU achieves near-complete forgetting (forget F1 = 0.0004) while maintaining competitive retained utility (retain avg F1 = 0.4766) after modifying only 0.19\% of model parameters. These results suggest that targeted behavioral unlearning can be achieved through sparse embedding edits without modifying deeper encoder representations.