Distributed Multi-Layer Editing for Rule-Level Knowledge in Large Language Models

arXiv cs.CL / 4/10/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how to edit rule-level knowledge in large language models, arguing that unlike fact-level edits, rule edits must stay consistent across multiple interdependent representations.
  • It extends the RuleEdit benchmark from 80 to 200 manually verified rules in mathematics and physics and uses fine-grained causal tracing to show that rule knowledge is organized across transformer layers in a form-specific way.
  • The authors find formulas and natural-language descriptions tend to concentrate in earlier layers, while concrete instances align more with middle layers, implying that rule edits cannot be reliably done with a single-layer or contiguous-block intervention.
  • They propose Distributed Multi-Layer Editing (DMLE), applying coordinated updates to multiple layers (shared early-layer updates for formulas/descriptions and separate middle-layer updates for instances) to improve rule-level editing.
  • Experiments across GPT-J-6B, Qwen2.5-7B, Qwen2-7B, and LLaMA-3-8B show DMLE maintains competitiveness on standard metrics while delivering much stronger rule-level performance, including sizable gains in instance portability and rule understanding.

Abstract

Large language models store not only isolated facts but also rules that support reasoning across symbolic expressions, natural language explanations, and concrete instances. Yet most model editing methods are built for fact-level knowledge, assuming that a target edit can be achieved through a localized intervention. This assumption does not hold for rule-level knowledge, where a single rule must remain consistent across multiple interdependent forms. We investigate this problem through a mechanistic study of rule-level knowledge editing. To support this study, we extend the RuleEdit benchmark from 80 to 200 manually verified rules spanning mathematics and physics. Fine-grained causal tracing reveals a form-specific organization of rule knowledge in transformer layers: formulas and descriptions are concentrated in earlier layers, while instances are more associated with middle layers. These results suggest that rule knowledge is not uniformly localized, and therefore cannot be reliably edited by a single-layer or contiguous-block intervention. Based on this insight, we propose Distributed Multi-Layer Editing (DMLE), which applies a shared early-layer update to formulas and descriptions and a separate middle-layer update to instances. While remaining competitive on standard editing metrics, DMLE achieves substantially stronger rule-level editing performance. On average, it improves instance portability and rule understanding by 13.91 and 50.19 percentage points, respectively, over the strongest baseline across GPT-J-6B, Qwen2.5-7B, Qwen2-7B, and LLaMA-3-8B. The code is available at https://github.com/Pepper66/DMLE.