Steered LLM Activations are Non-Surjective

arXiv cs.AI / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies activation steering as a surjectivity question: whether every activation state produced by steering can be reached by some discrete text prompt through the model’s standard forward pass.
  • Under practical assumptions, it proves that steering drives the residual stream off the manifold of activation states reachable from prompts, meaning most steered internal behaviors have no prompt pre-image.
  • The authors report empirical evidence across three widely used LLMs that supports the theoretical non-surjectivity result.
  • The findings formally separate “white-box” steerability from “black-box” prompt-based realizability, suggesting steering success should not be taken as evidence of interpretability or vulnerability via prompts.
  • The work recommends evaluation protocols that explicitly decouple white-box interventions (steering) from black-box prompting when assessing interpretability and safety risks.

Abstract

Activation steering is a popular white-box control technique that modifies model activations to elicit an abstract change in output behavior. It has also become a standard tool in interpretability (e.g., probing truthfulness, or translating activations into human-readable explanations and safety research (e.g., studying jailbreakability). However, it is unclear whether steered activation states are realizable by any textual prompt. In this work, we cast this question as a surjectivity problem: for a fixed model, does every steered activation admit a pre-image under the model's natural forward pass? Under practical assumptions, we prove that activation steering pushes the residual stream off the manifold of states reachable from discrete prompts. Almost surely, no prompt can reproduce the same internal behavior induced by steering. We also illustrate this finding empirically across three widely used LLMs. Our results establish a formal separation between white-box steerability and black-box prompting. We therefore caution against interpreting the ease and success of activation steering as evidence of prompt-based interpretability or vulnerability, and argue for evaluation protocols that explicitly decouple white-box and black-box interventions.