Memory-Augmented Vision-Language Agents for Persistent and Semantically Consistent Object Captioning

arXiv cs.CV / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a memory-augmented vision-language agent that aims to produce persistent, semantically consistent object captions across viewpoints for embodied agents.
  • It unifies data association, object captioning, and exploration policy in a single autoregressive framework using object-level episodic memory serialized into tokens.
  • Training is self-supervised via a disagreement-based exploration policy and a pseudo-captioning approach that enforces consistency across multi-view caption histories.
  • Experiments in photorealistic 3D environments show gains of up to +11.86% in captioning scores and +7.39% in caption self-similarity versus baseline models, with a compact scene representation for scalability.
  • The authors provide code, model weights, and data publicly through their GitHub repository.

Abstract

Vision-Language Models (VLMs) often yield inconsistent descriptions of the same object across viewpoints, hindering the ability of embodied agents to construct consistent semantic representations over time. Previous methods resolved inconsistencies using offline multi-view aggregation or multi-stage pipelines that decouple exploration, data association, and caption learning, with limited capacity to reason over previously observed objects. In this paper, we introduce a unified, memory-augmented Vision-Language agent that simultaneously handles data association, object captioning, and exploration policy within a single autoregressive framework. The model processes the current RGB observation, a top-down explored map, and an object-level episodic memory serialized into object-level tokens, ensuring persistent object identity and semantic consistency across extended sequences. To train the model in a self-supervised manner, we collect a dataset in photorealistic 3D environments using a disagreement-based policy and a pseudo-captioning model that enforces consistency across multi-view caption histories. Extensive evaluation on a manually annotated object-level test set, demonstrate improvements of up to +11.86% in standard captioning scores and +7.39% in caption self-similarity over baseline models, while enabling scalable performance through a compact scene representation. Code, model weights, and data are available at https://github.com/hsp-iit/epos-vlm