One Word at a Time: Incremental Completion Decomposition Breaks LLM Safety

arXiv cs.CL / 4/30/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Incremental Completion Decomposition (ICD), a jailbreak approach that extracts a malicious response from an LLM by requesting a sequence of one-word continuations before the full answer.
  • ICD includes multiple variants—such as manually selecting the next word, having the model generate it, and pre-filling the final response step—while aiming to improve reliability of the attack.
  • Across several model families, the authors report higher Attack Success Rate (ASR) on benchmarks like AdvBench, JailbreakBench, and StrongREJECT compared with prior methods.
  • The work provides both theoretical reasoning and mechanistic evidence, suggesting that successful ICD trajectories suppress refusal-related representations and move internal activations away from safety-aligned states.

Abstract

Large Language Models (LLMs) are trained to refuse harmful requests, yet they remain vulnerable to jailbreak attacks that exploit weaknesses in conversational safety mechanisms. We introduce Incremental Completion Decomposition (ICD), a trajectory-based jailbreak strategy that elicits a sequence of single-word continuations related to a malicious request before eliciting the full response. In addition, we propose variants of ICD by manually picking or model-generating the one-word continuation, as well as prefilling when eliciting the full model response in the final step. We systematically evaluate these variants across a broad set of model families, demonstrating superior Attack Success Rate (ASR) on AdvBench, JailbreakBench, and StrongREJECT compared to existing methods. In addition, we provide a theoretical account of why ICD is effective and present mechanistic evidence that successful attack trajectories systematically suppress refusal-related representations and shift activations away from safety-aligned states.