Less is More: Geometric Unlearning for LLMs with Minimal Data Disclosure
arXiv cs.CL / 5/5/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses the need for post-hoc unlearning in deployed LLMs to remove specific sensitive content while preserving overall usefulness.
- It proposes Geometric Unlearning (GU), which performs alignment using prompt-time planning states rather than requiring access to the original training corpus.
- GU distills a compact low-rank “geometry” of safe behavior from a small set of safe reference prompts and uses lightweight anchor-in-context synthetic prompts to localize alignment.
- A teacher-distillation regularizer on synthetic non-target anchors is used to reduce collateral drift and protect non-target knowledge.
- Experiments on privacy-focused benchmarks (ToFU and UnlearnPII) show strong suppression of targeted content with minimal degradation on non-target performance, using limited synthetic data.
Related Articles

Singapore's Fraud Frontier: Why AI Scam Detection Demands Regulatory Precision
Dev.to

How AI is Changing the Way We Code in 2026: The Shift from Syntax to Strategy
Dev.to

13 CLAUDE.md Rules That Make AI Write Modern PHP (Not PHP 5 Resurrected)
Dev.to

MCP annotations are a UX layer, not a security layer
Dev.to
From OOM to 262K Context: Running Qwen3-Coder 30B Locally on 8GB VRAM
Dev.to