CURE:Circuit-Aware Unlearning for LLM-based Recommendation

arXiv cs.AI / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses privacy-driven “unlearning” for LLM-based recommender systems, arguing that current methods mix forgetting and retaining objectives in ways that create gradient conflicts and unstable training.
  • It proposes CURE, a circuit-aware unlearning framework that identifies causally responsible computation subgraphs (“circuits”) for recommendation behavior and isolates which modules affect forget vs. retain.
  • CURE dissects model components into forget-specific, retain-specific, and task-shared groups, applying different update rules to each to reduce gradient conflicts.
  • Experiments on real-world datasets indicate CURE produces more effective unlearning than prior baseline approaches, while aiming to better preserve overall recommendation utility.
  • The work also improves the transparency of unlearning by moving away from largely black-box update procedures toward a module/circuit-level explanation of what gets changed.

Abstract

Recent advances in large language models (LLMs) have opened new opportunities for recommender systems by enabling rich semantic understanding and reasoning about user interests and item attributes. However, as privacy regulations tighten, incorporating user data into LLM-based recommendation (LLMRec) introduces significant privacy risks, making unlearning algorithms increasingly crucial for practical deployment. Despite growing interest in LLMRec unlearning, most existing approaches formulate unlearning as a weighted combination of forgetting and retaining objectives while updating model parameters in a uniform manner. Such formulations inevitably induce gradient conflicts between the two objectives, leading to unstable optimization and resulting in either ineffective unlearning or severe degradation of model utility. Moreover, the unlearning procedure remains largely black-box, undermining its transparency and trustworthiness. To tackle these challenges, we propose CURE, a circuit-aware unlearning framework that disentangles model components into functionally distinct subsets and selectively updates them. Here, a circuit refers to a computational subgraph that is causally responsible for task-specific behaviors. Specifically, we extract the core circuits underlying item recommendation and analyze how individual modules within these circuits contribute to the forget and retain objectives. Based on this analysis, these modules are categorized into forget-specific, retain-specific, and task-shared groups, each subject to function-specific update rules to mitigate gradient conflicts during unlearning. Experiments on real-world datasets show that our approach achieves more effective unlearning than existing baselines.