Knowledge Vector of Logical Reasoning in Large Language Models

arXiv cs.CL / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines how large language models internally represent different logical reasoning types—deductive, inductive, and abductive—and studies how these representations relate to each other.
  • It finds that each reasoning type can be represented as a distinct, reasoning-specific knowledge vector in a linear embedding space, with relatively weak dependence between the vectors.
  • Motivated by cognitive science and evidence that reasoning chains from one type can help another, the authors propose a refinement method to make these vectors complementary rather than isolated.
  • The proposed complementary subspace-constrained refinement framework uses a complementary loss (to share helpful auxiliary knowledge) and a subspace constraint loss (to preserve unique characteristics), leading to consistent performance gains in steering experiments.
  • A mechanism-interpretability analysis further identifies which features of reasoning are shared versus unique across the different logical reasoning vectors in LLMs.

Abstract

Logical reasoning serve as a central capability in LLMs and includes three main forms: deductive, inductive, and abductive reasoning. In this work, we study the knowledge representations of these reasoning types in LLMs and analyze the correlations among them. Our analysis shows that each form of logical reasoning can be captured as a reasoning-specific knowledge vector in a linear representation space, yet these vectors are largely independent of each other. Motivated by cognitive science theory that these subforms of logical reasoning interact closely in the human brain, as well as our observation that the reasoning process for one type can benefit from the reasoning chain produced by another, we further propose to refine the knowledge representations of each reasoning type in LLMs to encourage complementarity between them. To this end, we design a complementary subspace-constrained refinement framework, which introduces a complementary loss that enables each reasoning vector to leverage auxiliary knowledge from the others, and a subspace constraint loss that prevents erasure of their unique characteristics. Through steering experiments along reasoning vectors, we find that refined vectors incorporating complementary knowledge yield consistent performance gains. We also conduct a mechanism-interpretability analysis of each reasoning vector, revealing insights into the shared and specific features of different reasoning in LLMs.