Automatic Ontology Construction Using LLMs as an External Layer of Memory, Verification, and Planning for Hybrid Intelligent Systems

arXiv cs.AI / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a hybrid intelligent-system architecture that augments LLMs with an external ontological memory layer built as an RDF/OWL knowledge graph for persistent, verifiable reasoning.
  • It introduces an automated pipeline that constructs and continuously updates the ontology from heterogeneous sources (documents, APIs, and dialogue logs) via entity recognition, relation extraction, normalization, triple generation, and constraint-based validation using SHACL/OWL.
  • During inference, the system combines vector-based retrieval with graph-based reasoning and external tool interaction, integrating symbolic and neural approaches.
  • Experiments on planning tasks (including the Tower of Hanoi benchmark) show that ontology augmentation improves multi-step reasoning performance versus baseline LLM systems.
  • The ontology layer also supports formal validation and verification, turning generation into a generation–verification–correction loop aimed at reliability and explainability for agent and enterprise/robotics use cases.

Abstract

This paper presents a hybrid architecture for intelligent systems in which large language models (LLMs) are extended with an external ontological memory layer. Instead of relying solely on parametric knowledge and vector-based retrieval (RAG), the proposed approach constructs and maintains a structured knowledge graph using RDF/OWL representations, enabling persistent, verifiable, and semantically grounded reasoning. The core contribution is an automated pipeline for ontology construction from heterogeneous data sources, including documents, APIs, and dialogue logs. The system performs entity recognition, relation extraction, normalization, and triple generation, followed by validation using SHACL and OWL constraints, and continuous graph updates. During inference, LLMs operate over a combined context that integrates vector-based retrieval with graph-based reasoning and external tool interaction. Experimental observations on planning tasks, including the Tower of Hanoi benchmark, indicate that ontology augmentation improves performance in multi-step reasoning scenarios compared to baseline LLM systems. In addition, the ontology layer enables formal validation of generated outputs, transforming the system into a generation-verification-correction pipeline. The proposed architecture addresses key limitations of current LLM-based systems, including lack of long-term memory, weak structural understanding, and limited reasoning capabilities. It provides a foundation for building agent-based systems, robotics applications, and enterprise AI solutions that require persistent knowledge, explainability, and reliable decision-making.