A Theory of LLM Information Susceptibility

arXiv cs.LG / 2026/3/26

💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper proposes a “theory of LLM information susceptibility” to explain fundamental limits on how much an LLM can improve the performance of strategies in agentic optimization settings.
  • It argues that with sufficiently large computational resources, using a fixed LLM intervention may not increase a strategy set’s performance susceptibility to available budget.
  • The authors extend the framework to multi-variable, multi-channel budget architectures and identify conditions under which co-scaling across channels can exceed the susceptibility bound.
  • Empirical validation across diverse domains and multiple model scales finds that nested/co-scaling architectures can create additional response channels not available to fixed configurations.
  • The work suggests that statistical-physics-style tools can predict constraints for designing AI systems, and that nested architectures may be structurally necessary for open-ended agentic self-improvement if the hypothesis generalizes.

Abstract

Large language models (LLMs) are increasingly deployed as optimization modules in agentic systems, yet the fundamental limits of such LLM-mediated improvement remain poorly understood. Here we propose a theory of LLM information susceptibility, centred on the hypothesis that when computational resources are sufficiently large, the intervention of a fixed LLM does not increase the performance susceptibility of a strategy set with respect to budget. We develop a multi-variable utility-function framework that generalizes this hypothesis to architectures with multiple co-varying budget channels, and discuss the conditions under which co-scaling can exceed the susceptibility bound. We validate the theory empirically across structurally diverse domains and model scales spanning an order of magnitude, and show that nested, co-scaling architectures open response channels unavailable to fixed configurations. These results clarify when LLM intervention helps and when it does not, demonstrating that tools from statistical physics can provide predictive constraints for the design of AI systems. If the susceptibility hypothesis holds generally, the theory suggests that nested architectures may be a necessary structural condition for open-ended agentic self-improvement.