GRAIL: Autonomous Concept Grounding for Neuro-Symbolic Reinforcement Learning

arXiv cs.AI / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces GRAIL, a framework for autonomous concept grounding in neuro-symbolic reinforcement learning by learning relational concepts (e.g., “left of”, “close by”) directly from environment interaction.
  • Instead of relying on manually defined concepts, GRAIL uses large language models as weak supervision to generate generic relational concept representations and then refines them to fit environment-specific semantics.
  • The approach is designed to mitigate two key challenges in underdetermined settings: sparse reward signals and misalignment between intended and actually learned concept meanings.
  • Experiments on Atari games (Kangaroo, Seaquest, and Skiing) show that GRAIL can match or outperform agents using hand-crafted concepts in simplified settings, while in the full environment it highlights trade-offs between maximizing rewards and completing higher-level goals.

Abstract

Neuro-symbolic Reinforcement Learning (NeSy-RL) combines symbolic reasoning with gradient-based optimization to achieve interpretable and generalizable policies. Relational concepts, such as "left of" or "close by", serve as foundational building blocks that structure how agents perceive and act. However, conventional approaches require human experts to manually define these concepts, limiting adaptability since concept semantics vary across environments. We propose GRAIL (Grounding Relational Agents through Interactive Learning), a framework that autonomously grounds relational concepts through environmental interaction. GRAIL leverages large language models (LLMs) to provide generic concept representations as weak supervision, then refines them to capture environment-specific semantics. This approach addresses both sparse reward signals and concept misalignment prevalent in underdetermined environments. Experiments on the Atari games Kangaroo, Seaquest, and Skiing demonstrate that GRAIL matches or outperforms agents with manually crafted concepts in simplified settings, and reveals informative trade-offs between reward maximization and high-level goal completion in the full environment.