Why Cognitive Robotics Matters: Lessons from OntoAgent and LLM Deployment in HARMONIC for Safety-Critical Robot Teaming

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that deploying embodied, embodied AI agents in physical environments requires long-horizon planning that is reliable, deterministic, and transparent, which motivates “cognitive robotics.”
  • It introduces HARMONIC, a cognitive-robotic architecture that integrates OntoAgent’s content-centric, metacognitive self-monitoring and consequence-based action selection with a modular reactive tactical layer.
  • HARMONIC is used as an evaluation platform to test whether multiple LLMs can replicate OntoAgent-like cognitive capabilities under identical robotic conditions, including native and knowledge-equalized settings.
  • The study finds that LLMs often fail to accurately assess their own knowledge state before acting, leading to failures in domain-grounded diagnostics and subsequent action selection.
  • The authors conclude that these shortcomings are largely architectural (not simply missing knowledge) and recommend keeping deterministic cognitive architectures as primary authority in safety-critical robotic reasoning.

Abstract

Deploying embodied AI agents in the physical world demands cognitive capabilities for long-horizon planning that execute reliably, deterministically, and transparently. We present HARMONIC, a cognitive-robotic architecture that pairs OntoAgent, a content-centric cognitive architecture providing metacognitive self-monitoring, domain-grounded diagnosis, and consequence-based action selection over ontologically structured knowledge, with a modular reactive tactical layer. HARMONIC's modular design enables a functional evaluation of whether LLMs can replicate OntoAgent's cognitive capabilities, evaluated within the same robotic system under identical conditions. Six LLMs spanning frontier and efficient tiers replace OntoAgent in a collaborative maintenance scenario under native and knowledge-equalized conditions. Results reveal that LLMs do not consistently assess their own knowledge state before acting, causing downstream failures in diagnostic reasoning and action selection. These deficits persist even with equivalent procedural knowledge, indicating the issues are architectural rather than knowledge-based. These findings support the design of physically embodied systems in which cognitive architectures retain primary authority for reasoning, owing to their deterministic and transparent characteristics.