Discovering Reinforcement Learning Interfaces with Large Language Models

arXiv cs.LG / 5/6/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper tackles the challenge of automatically discovering full reinforcement learning (RL) task interfaces—both observation mappings and reward functions—starting from raw simulator state.
  • It proposes LIMEN, an LLM-guided evolutionary framework that generates candidate interfaces as executable programs and improves them iteratively using feedback from policy training.
  • Experiments on discrete gridworld tasks and continuous control (including locomotion and manipulation) show that jointly evolving observations and rewards can succeed with only trajectory-level success metrics.
  • The study finds that optimizing only the observation mapping or only the reward function fails in at least one domain, highlighting the importance of co-design.
  • The authors argue that this automatic interface construction can significantly reduce manual engineering effort for new RL tasks.

Abstract

Reinforcement learning systems rely on environment interfaces that specify observations and reward functions, yet constructing these interfaces for new tasks often requires substantial manual effort. While recent work has automated reward design using large language models (LLMs), these approaches assume fixed observations and do not address the broader challenge of synthesizing complete task interfaces. We study RL task interface discovery from raw simulator state, where both observation mappings and reward functions must be generated. We propose LIMEN (Code available at https://github.com/Lossfunk/LIMEN), a LLM guided evolutionary framework that produces candidate interfaces as executable programs and iteratively refines them using policy training feedback. Across novel discrete gridworld tasks and continuous control domains spanning locomotion and manipulation, joint evolution of observations and rewards discovers effective interfaces given only a trajectory-level success metric, while optimizing either component alone fails on at least one domain. These results demonstrate that automatic construction of RL interfaces from raw state can substantially reduce manual engineering and that observation and reward components often benefit from co-design, as single-component optimization fails catastrophically on at least one domain in our evaluation suite.