The Depth Ceiling: On the Limits of Large Language Models in Discovering Latent Planning

arXiv cs.LG / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates whether large language models can discover multi-step latent planning strategies and execute them in a single forward pass without intermediate-step supervision.
  • Experiments on controlled graph path-finding tasks show a clear depth limit that is not solved by scaling: tiny transformers learn up to three latent steps, fine-tuned GPT-4o and Qwen3-32B reach five, and GPT-5.4 reaches seven via few-shot prompting.
  • While models can learn latent planning depth up to five during training, the learned strategy can generalize to execute up to eight latent steps at test time.
  • The findings suggest a dissociation between discovering a latent planning strategy from final-answer supervision and successfully executing it at greater latent depths once discovered, implying constraints relevant to chain-of-thought monitoring assumptions.
  • The authors argue that if similar limitations generalize, multi-step coordinated latent planning may require explicit instruction or externalization, supporting the usefulness (and limitations) of CoT monitoring.

Abstract

The viability of chain-of-thought (CoT) monitoring hinges on models being unable to reason effectively in their latent representations. Yet little is known about the limits of such latent reasoning in LLMs. We test these limits by studying whether models can discover multi-step planning strategies without supervision on intermediate steps and execute them latently, within a single forward pass. Using graph path-finding tasks that precisely control the number of required latent planning steps, we uncover a striking limitation unresolved by massive scaling: tiny transformers trained from scratch discover strategies requiring up to three latent steps, fine-tuned GPT-4o and Qwen3-32B reach five, and GPT-5.4 attains seven under few-shot prompting. Although the maximum latent planning depth models can learn during training is five, the discovered strategy generalizes up to eight latent steps at test-time. This reveals a dissociation between the ability to discover a latent strategy under final-answer supervision alone and the ability to execute it once discovered. If similar limits hold more broadly, strategies requiring multiple coordinated latent planning steps may need to be explicitly taught or externalized, lending credence to CoT monitoring.