Multi-action Tangled Program Graphs for Multi-task Reinforcement Learning with Continuous Control
arXiv cs.AI / 4/29/2026
📰 NewsModels & Research
Key Points
- The paper proposes Multi-Action Tangled Program Graph (MATPG), a genetic-programming approach that aggregates MAPLE-style agents and uses a control-flow mechanism to activate multiple behaviors for continuous-control tasks.
- While MATPG was previously tested mainly on single-task reinforcement learning, the authors introduce a new multi-task benchmark using MuJoCo HalfCheetah with five randomly placed obstacles, each requiring distinct behaviors.
- Experiments on the new continuous multi-task setting show that MATPG performs strongly, and the authors report superiority when MATPG is combined with lexicase selection.
- The study also evaluates interpretability, finding that the evolved program graph’s decision flow is fully understandable, supporting explainable policy structure.
- Overall, the work positions MATPG as an effective GP-based solution for continuous Multi-Task Reinforcement Learning and provides a new evaluation scenario to test such methods.
Related Articles
v0.22.1
Ollama Releases

The best of Cloud Next '26: Gemini Enterprise Agent Platform. The perfect combination of Intelligence and Automation to generate VALUE.
Dev.to

Open source memory layer so any AI agent can do what Claude.ai and ChatGPT do
Dev.to

Sources: Anthropic could raise a new $50B round at a valuation of $900B
TechCrunch

Satya Nadella says he’s ready to ‘exploit’ the new OpenAI deal
TechCrunch