Why isn’t LLM reasoning done in vector space instead of natural language?

Reddit r/LocalLLaMA / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The article asks why LLM reasoning is typically presented in natural language (e.g., step-by-step chain-of-thought) even though the models internally manipulate high-dimensional vectors.
  • It considers a hypothetical design where an LLM would “reason in vectors” and only translate the final conclusions into language at the end.
  • The piece evaluates potential upsides of vector-space reasoning, such as being faster, more compressed, and better aligned with intuition-like tasks.
  • It also raises concerns that vector-based reasoning could be too opaque, difficult to verify, and less reliable for domains requiring precise symbolic behavior like math, programming, or legal logic.
  • The overall goal is to spark discussion among researchers and engineers about what would practically change in model design, interpretability, and trustworthiness if reasoning were made explicit in latent space.

Why don’t LLMs use explicit vector-based reasoning instead of language-based chain-of-thought? What would happen if they did?

Most LLM reasoning we see is expressed through language: step-by-step text, explanations, chain-of-thought style outputs, etc. But internally, models already operate on high-dimensional vectors.

So my question is:

Why don’t we have models that reason more explicitly in latent/vector space instead of producing intermediate reasoning in natural language?

Would vector-based reasoning be faster, more compressed, and better for intuition-like tasks? Or would it make reasoning too opaque, hard to verify, and unreliable for math/programming/legal logic?

In other words:

Could an LLM “think” in vectors and only translate the final reasoning into language at the end?

Curious how researchers/engineers think about this.

submitted by /u/ZeusZCC
[link] [comments]