A Deep Reinforcement Learning Framework for Closed-loop Guidance of Fish Schools via Virtual Agents
arXiv cs.RO / 3/31/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper presents a deep reinforcement learning (RL) framework that uses virtual agents to provide closed-loop guidance to fish schools in real time during physical experiments.
- Policies are trained in simulation with Proximal Policy Optimization (PPO) and deployed to interact directly with live rummy-nose tetras, enabling artificial agents to influence collective motion.
- To handle the stochastic behavior of live fish, the authors introduce a composite reward function that trades off directional guidance against maintaining social cohesion.
- Visual system design choices—specifically a white background and larger stimulus sizes—are found to improve guidance effectiveness in physical trials.
- Guidance performance works well for small groups (about five fish) but degrades significantly as group size increases (notably by eight), highlighting limits in maintaining control over larger collectives.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to