Insect-inspired modular architectures as inductive biases for reinforcement learning

arXiv cs.LG / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an insect-inspired reinforcement learning (RL) policy architecture that decomposes control into multiple interacting modules rather than using a single centralized latent state.
  • The modular design includes sensory encoding, heading representation, sparse associative memory, recurrent command generation, local motor control, and a learned arbitration mechanism that selectively allocates motor authority across modules.
  • On a 2D navigation task requiring simultaneous food seeking, obstacle avoidance, and predator escape, the modular policy outperforms tested centralized baselines (including a gated recurrent unit and a multilayer perceptron) in final mean performance.
  • The modular approach also shows the lowest final value loss and more stable PPO optimization statistics, with very low module-assignment entropy, suggesting the arbitration becomes highly selective.
  • Overall, the findings indicate that distributed control structures can provide a useful inductive bias for RL settings where multiple behavior objectives dynamically compete.

Abstract

Most reinforcement-learning (RL) controllers used in continuous control are architecturally centralized: observations are compressed into a single latent state from which both value estimates and actions are produced. Biological control systems are often organized differently. Insects, in particular, coordinate navigation, heading stabilization, memory, and context-dependent action selection through distributed circuits rather than a single monolithic controller. Motivated by this contrast, we study an RL policy architecture that decomposes control into interacting modules for sensory encoding, heading representation, sparse associative memory, recurrent command generation, and local motor control, with a learned arbitration mechanism that allocates motor authority across modules. The model is evaluated on a two-dimensional navigation task that require simultaneous food seeking, obstacle avoidance, and predator escape. In a six-seed predator-navigation experiment trained with Proximal Policy Optimization (PPO) for 75 updates, the modular policy achieves the strongest final mean performance among the tested controllers, with final episodic return -2798.8\pm964.4 versus -3778.0\pm628.1 for a centralized gated recurrent unit (GRU) and -4727.5\pm772.5 for a centralized multilayer perceptron (MLP). The modular policy also attains the lowest final value loss and stable PPO optimization statistics while driving module-assignment entropy to 0.0457\pm0.0244, indicating highly selective control allocation. These results suggest that distributed control can serve as a useful inductive bias for RL problems involving dynamically competing behavioral objectives.