SWE-Edit: Rethinking Code Editing for Efficient SWE-Agent

arXiv cs.CL / 4/30/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies a context-coupling problem in current LLM-based code editing workflows, where code inspection, planning, and edit execution are mixed in a single context window.
  • It introduces SWE-Edit, a two-subagent approach that separates viewing (task-relevant code extraction) from editing (executing changes from high-level plans) to keep reasoning and context-heavy operations cleanly separated.
  • The authors study editing-model design and show that the common find-and-replace interface is error-prone, leading them to train Qwen3-8B with GRPO to dynamically choose editing modes.
  • Experiments on SWE-bench Verified show a 2.1% improvement in resolved rate alongside a 17.9% reduction in inference cost, and the work also proposes a benchmark to better predict downstream agent performance.
  • The authors release the SWE-Edit code publicly, supporting adoption and further evaluation by the community.

Abstract

Large language model agents have achieved remarkable progress on software engineering tasks, yet current approaches suffer from a fundamental context coupling problem: the standard code editing interface conflates code inspection, modification planning, and edit execution within a single context window, forcing agents to interleave exploratory viewing with strictly formatted edit generation. This causes irrelevant information to accumulate and degrades agent performance. To address this, we propose SWE-Edit, which decomposes code editing into two specialized subagents: a Viewer that extracts task-relevant code on demand, and an Editor that executes modifications from high-level plans--allowing the main agent to focus on reasoning while delegating context-intensive operations to clean context windows. We further investigate what makes an effective editing model: observing that the prevalent find-and-replace format is error-prone, we train Qwen3-8B with GRPO to adaptively select editing modes, yielding improved editing efficiency over single-format baselines. On SWE-bench Verified, SWE-Edit improves resolved rate by 2.1% while reducing inference cost by 17.9%. We additionally propose a code editing benchmark that reliably predicts downstream agentic performance, providing practical guidance for editing model selection. Our code is publicly available at https://github.com/microsoft/SWE-Edit.