Unlearning What Matters: Token-Level Attribution for Precise Language Model Unlearning

arXiv cs.CL / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces TokenUnlearn, a token-level framework for machine unlearning in large language models that avoids the limitations of existing sequence-level methods.
  • It uses knowledge-aware masking and entropy-aware signals to compute token importance scores, enabling more precise targeting of the subset of tokens that actually encode the knowledge to be removed.
  • Two strategies are proposed: hard selection (apply unlearning only to high-importance tokens) and soft weighting (scale gradient contributions by importance).
  • The authors provide theoretical evidence that token-level selection improves the gradient signal-to-noise ratio and reduces suboptimal forgetting.
  • Experiments on TOFU and WMDP across three model architectures show consistent gains in forgetting effectiveness while preserving overall utility compared with sequence-level baselines.

Abstract

Machine unlearning has emerged as a critical capability for addressing privacy, safety, and regulatory concerns in large language models (LLMs). Existing methods operate at the sequence level, applying uniform updates across all tokens despite only a subset encoding the knowledge targeted for removal. This introduces gradient noise, degrades utility, and leads to suboptimal forgetting. We propose TokenUnlearn, a token-level attribution framework that identifies and selectively targets critical tokens. Our approach combines knowledge-aware signals via masking, and entropy-aware signals to yield importance scores for precise token selection. We develop two complementary strategies: hard selection, applying unlearning only to high-importance tokens, and soft weighting, modulating gradient contributions based on importance scores. Both extend existing methods to token-level variants. Theoretical analysis shows token-level selection improves gradient signal-to-noise ratio. Experiments on TOFU and WMDP benchmarks across three model architectures demonstrate consistent improvements over sequence-level baselines in both forgetting effectiveness and utility preservation.