The Anatomy of an Edit: Mechanism-Guided Activation Steering for Knowledge Editing

arXiv cs.CL / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how knowledge editing (KE) actually takes effect inside LLMs by applying neuron-level knowledge attribution (NLKA) with post-edit contrasts between successful and failed edits.
  • It finds a consistent mechanism across KE methods: mid-to-late attention helps promote the new target, while attention and feed-forward network (FFN) components jointly suppress the original fact.
  • Based on these findings, the authors introduce MEGA (MEchanism-Guided Activation steering), which performs attention-residual interventions in attribution-aligned regions without changing model weights.
  • Experiments on CounterFact and Popular show that MEGA delivers strong editing performance across KE metrics on GPT-2 XL and LLaMA 2 7B, and the work frames post-edit attribution as an engineering signal rather than only analysis.

Abstract

Large language models (LLMs) are increasingly used as knowledge bases, but keeping them up to date requires targeted knowledge editing (KE). However, it remains unclear how edits are implemented inside the model once applied. In this work, we take a mechanistic view of KE using neuron-level knowledge attribution (NLKA). Unlike prior work that focuses on pre-edit causal tracing and localization, we use post-edit attribution -- contrasting successful and failed edits -- to isolate the computations that shift when an edit succeeds. Across representative KE methods, we find a consistent pattern: mid-to-late attention predominantly promotes the new target, while attention and FFN modules cooperate to suppress the original fact. Motivated by these findings, we propose MEGA, a MEchanism-Guided Activation steering method that performs attention-residual interventions in attribution-aligned regions without modifying model weights. On CounterFact and Popular, MEGA achieves strong editing performance across KE metrics on GPT2-XL and LLaMA2-7B. Overall, our results elevate post-edit attribution from analysis to engineering signal: by pinpointing where and how edits take hold, it powers MEGA to deliver reliable, architecture-agnostic knowledge edits.