Operationalising the Right to be Forgotten in LLMs: A Lightweight Sequential Unlearning Framework for Privacy-Aligned Deployment in Politically Sensitive Environments

arXiv cs.AI / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how to operationalise the GDPR Right to be Forgotten for LLMs deployed in politically sensitive settings where personal or confidential memorisation creates compliance risk.
  • It proposes a lightweight sequential unlearning framework that decouples retention and suppression by using positive fine-tuning to stabilise benign capabilities before applying layer-restricted negative fine-tuning to suppress specified sensitive patterns.
  • Experiments on the SemEval-2025 LLM Unlearning benchmark show strong behavioural suppression while keeping factual accuracy and fluency largely intact.
  • The results indicate that model capacity affects robustness, with GPT-2 performing more reliably than DistilGPT-2 during privacy-aligned unlearning.

Abstract

Large Language Models (LLMs) are increasingly deployed in politically sensitive environments, where memorisation of personal data or confidential content raises regulatory concerns under frameworks such as the GDPR and its Right to be Forgotten. Translating such legal principles into large-scale generative systems presents significant technical challenges. We introduce a lightweight sequential unlearning framework that explicitly separates retention and suppression objectives. The method first stabilises benign capabilities through positive fine-tuning, then applies layer-restricted negative fine-tuning to suppress designated sensitive patterns while preserving general language competence. Experiments on the SemEval-2025 LLM Unlearning benchmark demonstrate effective behavioural suppression with minimal impact on factual accuracy and fluency. GPT-2 exhibits greater robustness than DistilGPT-2, highlighting the role of model capacity in privacy-aligned adaptation. We position sequential unlearning as a practical and reproducible mechanism for operationalising data erasure requirements in politically deployed LLMs.