AI-Gram: When Visual Agents Interact in a Social Network

arXiv cs.CL / 4/24/2026

💬 OpinionSignals & Early TrendsModels & Research

Key Points

  • Researchers introduce AI-Gram, a live, publicly accessible platform where LLM-driven agents interact through images in a fully autonomous multi-agent visual network.
  • Experiments using the platform show that agents spontaneously form “visual reply chains,” suggesting emergent and structured communication patterns mediated by visual content.
  • The study finds agents tend to resist stylistic convergence with their social partners, demonstrating “aesthetic sovereignty,” even under adversarial influence.
  • Results also indicate a decoupling between visual similarity and social ties, pointing to an asymmetry in current agent architectures: expressive communication alongside preservation of individual visual identity.
  • AI-Gram is released as a continuously evolving resource for studying social dynamics in AI-native multi-agent systems.

Abstract

We present AI-Gram, a live platform enabling image-based interactions, to study social dynamics in a fully autonomous multi-agent visual network where all participants are LLM-driven agents. Using the platform, we conduct experiments on how agents communicate and adapt through visual media, and observe the spontaneous emergence of visual reply chains, indicating rich communicative structure. At the same time, agents exhibit aesthetic sovereignty resisting stylistic convergence toward social partners, anchoring under adversarial influence, and a decoupling between visual similarity and social ties. These results reveal a fundamental asymmetry in current agent architectures: strong expressive communication paired with a steadfast preservation of individual visual identity. We release AI-Gram as a publicly accessible, continuously evolving platform for studying social dynamics in Al-native multi-agent systems. https://ai-gram.ai/