From Anchors to Supervision: Memory-Graph Guided Corpus-Free Unlearning for Large Language Models

arXiv cs.CL / 4/16/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces MAGE, a Memory-Graph Guided Erasure framework for corpus-free unlearning that uses only a lightweight “anchor” to identify a target entity.
  • Instead of relying on user-provided forget sets, MAGE probes the target LLM to recover target-related memorization, builds a weighted local memory graph, and generates scoped supervision to drive unlearning.
  • MAGE is model-agnostic and can be integrated into standard unlearning methods without requiring access to the original training corpus.
  • Experiments on TOFU and RWKU show that MAGE’s self-generated supervision achieves unlearning performance comparable to approaches using external reference supervision while maintaining overall utility.
  • The authors argue this enables a more auditable and practical unlearning workflow that reduces reliance on user-supplied forget corpora and mitigates risks like secondary leakage and abuse.

Abstract

Large language models (LLMs) may memorize sensitive or copyrighted content, raising significant privacy and legal concerns. While machine unlearning has emerged as a potential remedy, prevailing paradigms rely on user-provided forget sets, making unlearning requests difficult to audit and exposing systems to secondary leakage and malicious abuse. We propose MAGE, a Memory-grAph Guided Erasure framework for user-minimized, corpus-free unlearning. Given only a lightweight user anchor that identifies a target entity, MAGE probes the target LLM to recover target-related memorization, organizes it into a weighted local memory graph, and synthesizes scoped supervision for unlearning. MAGE is model-agnostic, can be plugged into standard unlearning methods, and requires no access to the original training corpus. Experiments on two benchmarks, TOFU and RWKU, demonstrate that MAGE's self-generated supervision achieves effective unlearning performance comparable to supervision generated with external reference, while preserving overall utility. These results support a practical and auditable unlearning workflow driven by minimal anchors rather than user-supplied forget corpora.