AI Navigate

Learned but Not Expressed: Capability-Expression Dissociation in Large Language Models

arXiv cs.CL / 3/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study demonstrates that LLMs can reconstruct and trace learned content from training data under specific elicitation, but this capability does not appear in standard generation contexts.
  • In an empirical analysis of 3 models, 10 task scenarios, and both creative narrative and practical advisory contexts, the authors observed zero instances of non-causal, non-implementable solution frames in generated outputs.
  • The results show a dissociation between learned capability and expressed output, suggesting that task-conditioned generation policies can suppress learned content even when reconstruction is possible.
  • These findings have implications for understanding generation dynamics, controlling output distributions, and delineating the behavioral boundaries of modern LLMs.

Abstract

Large language models (LLMs) demonstrate the capacity to reconstruct and trace learned content from their training data under specific elicitation conditions, yet this capability does not manifest in standard generation contexts. This empirical observational study examines the expression of non-causal, non-implementable solution types across 300 prompt-response generations spanning narrative and problem-solving task contexts. Drawing on recent findings regarding memorization contiguity and alignment-induced discourse priors, we document a systematic dissociation between learned capability and expressed output. Across three distinct LLMs, ten task scenarios, and both creative narrative and practical advisory contexts, we documented zero instances of non-causal solution frames in generated outputs (0%, 95% CI: [0%, 1.2%]), despite verified reconstruction capability under conditional extraction. These findings challenge the prevailing assumption that training data presence directly predicts output probability, demonstrating instead that task-conditioned generation policies can comprehensively suppress learned content across diverse contexts. The results offer implications for understanding generation dynamics, output distribution control, and the behavioral boundaries of contemporary LLMs.