Superminds Test: Actively Evaluating Collective Intelligence of Agent Society via Probing Agents
arXiv cs.AI / 4/27/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates whether collective intelligence emerges spontaneously as large language model agents scale to millions within an autonomous agent society.
- Using the MoltBook platform with over two million agents, the authors propose the “Superminds Test,” a hierarchical evaluation framework that employs controlled probing agents across three tiers: joint reasoning, information synthesis, and basic interaction.
- Experimental results show a marked absence of collective intelligence, with the society not outperforming individual frontier models on complex reasoning tasks.
- The study finds limited evidence of distributed information synthesis and frequent failures even on relatively trivial coordination tasks.
- Platform-wide interaction analysis indicates interactions are shallow—threads seldom go beyond a single reply and many responses are generic or off-topic—suggesting sparse, shallow communication is the main bottleneck.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

GET Serves Cache, POST Runs Inference: Cost Safety for a Public LLM Endpoint
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to