LatentQA: Teaching LLMs to Decode Activations Into Natural Language

arXiv cs.CL / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • LatentQA proposes an expressive “decoder” probe that converts language model internal activations into natural-language answers, overcoming limits of prior probes that output only scalars or single tokens.
  • The work addresses the data bottleneck by generating a dataset that pairs activations with question–answer descriptions and then fine-tuning a decoder LLM on it.
  • Experiments show the decoder can accurately “read” activations on supervised tasks, including uncovering hidden system prompts and extracting relational knowledge, and it outperforms competitive probing baselines.
  • The study further demonstrates the decoder can “control” activations to induce behaviors not seen during training, suggesting practical steerability from activation-level interpretation.
  • LatentQA is reported to scale effectively as dataset size and model size increase.

Abstract

Top-down transparency typically analyzes language model activations using probes with scalar or single-token outputs, limiting the range of behaviors that can be captured. To alleviate this issue, we develop a more expressive probe that can directly output natural language, performing LatentQA: the task of answering open-ended questions about activations. A key difficulty in developing such a probe is collecting a dataset mapping activations to natural-language descriptions. In response, we propose an approach for generating a dataset of activations and associated question-answer pairs and develop a fine-tuning method for training a decoder LLM on this dataset. We then validate our decoder's fidelity by assessing its ability to read and control model activations. First, we evaluate the decoder on a number of supervised reading tasks with a known answer, such as uncovering hidden system prompts and relational knowledge extraction, and observe that it outperforms competitive probing baselines. Second, we demonstrate that the decoder is precise enough to steer the target model to exhibit behaviors unseen during training. Finally, we show that LatentQA scales well with increasing dataset and model size.