Stability and Generalization in Looped Transformers

arXiv cs.LG / 4/17/2026

📰 NewsModels & Research

Key Points

  • The paper proposes a fixed-point analysis framework for looped transformers to determine which architectural choices enable generalization to harder test-time inputs.
  • The authors analyze stability along three axes—reachability, input-dependence, and geometry—and provide theoretical conditions under which fixed-point iteration yields meaningful predictions.
  • They prove that looped networks without recall have countably many fixed points and cannot achieve strong input-dependence in any spectral regime, limiting their extrapolation ability.
  • By combining recall with outer normalization, the study identifies a regime where fixed points are reachable, locally smooth with respect to inputs, and support stable backpropagation.
  • Experiments on chess, sudoku, and prefix-sums show downstream performance aligns with the framework’s predictions, and a new “internal recall” variant can outperform standard recall placement when outer normalization is used.

Abstract

Looped transformers promise test-time compute scaling by spending more iterations on harder problems, but it remains unclear which architectural choices let them extrapolate to harder problems at test time rather than memorize training-specific solutions. We introduce a fixed-point based framework for analyzing looped architectures along three axes of stability -- reachability, input-dependence, and geometry -- and use it to characterize when fixed-point iteration yields meaningful predictions. Theoretically, we prove that looped networks without recall have countable fixed points and cannot achieve strong input-dependence at any spectral regime, while recall combined with outer normalization reliably produces a regime in which fixed points are simultaneously reachable, locally smooth in the input, and supported by stable backpropagation. Empirically, we train single-layer looped transformers on chess, sudoku, and prefix-sums and find that downstream performance tracks the framework's predictions across tasks and architectural configurations. We additionally introduce internal recall, a novel recall placement variant, and show that it becomes competitive with -- and on sudoku, substantially better than -- standard recall placement once outer normalization is applied.