Grounding vs. Compositionality: On the Non-Complementarity of Reasoning in Neuro-Symbolic Systems

arXiv cs.AI / 4/30/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that compositional generalization is still a key weakness of modern neural networks, especially for out-of-distribution reasoning tasks.
  • It directly tests a common neuro-symbolic assumption that compositional reasoning should emerge automatically from symbol grounding, separating the effects of grounding and reasoning.
  • The authors introduce iLTN (Iterative Logic Tensor Network), a fully differentiable architecture for multi-step deduction, along with a taxonomy that probes generalization over novel entities, unseen relations, and composed rules.
  • Results show that training only on a grounding objective does not yield generalization, whereas jointly training perceptual grounding plus explicit multi-step reasoning enables high zero-shot performance across all tested tasks.
  • The study concludes that symbol grounding alone is necessary but insufficient, and that reasoning must be explicitly learned rather than expected to emerge as a byproduct.

Abstract

Compositional generalization remains a foundational weakness of modern neural networks, limiting their robustness and applicability in domains requiring out-of-distribution reasoning. A central, yet unverified, assumption in neuro-symbolic AI is that compositional reasoning will emerge as a byproduct of successful symbol grounding. This work presents the first systematic empirical analysis to challenge this assumption by disentangling the contributions of grounding and reasoning. To operationalize this investigation, we introduce the Iterative Logic Tensor Network (iLTN), a fully differentiable architecture designed for multi-step deduction. Using a formal taxonomy of generalization -- probing for novel entities, unseen relations, and complex rule compositions -- we demonstrate that a model trained solely on a grounding objective fails to generalize. In contrast, our full iLTN, trained jointly on perceptual grounding and multi-step reasoning, achieves high zero-shot accuracy across all tasks. Our findings provide conclusive evidence that symbol grounding, while necessary, is insufficient for generalization, establishing that reasoning is not an emergent property but a distinct capability that requires an explicit learning objective.