Enhancing sample efficiency in reinforcement-learning-based flow control: replacing the critic with an adaptive reduced-order model
arXiv cs.LG / 4/8/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes an adaptive reduced-order-model (ROM) reinforcement-learning framework for active flow control to address model-free DRL’s low sample efficiency by estimating gradient information needed for controller optimization without a conventional critic.
- The ROM blends a physics-informed linear dynamical system with a data-driven neural ordinary differential equation (NODE) to capture flow nonlinearity, with linear parameters identified via operator inference and the NODE trained using gradient-based optimization.
- During controller–environment interactions, the ROM is continuously updated using newly collected data, and the controller is optimized through differentiable simulation of the learned ROM.
- Experiments on two benchmark flow control problems—Blasius boundary layer flow and flow past a square cylinder—show fewer exploration samples and improved performance, including drag reduction with significantly less data than typical DRL methods.
- The authors argue the method tackles a key bottleneck in model-free DRL control and provides a foundation for more sample-efficient DRL-based active flow controllers.
Related Articles

Black Hat Asia
AI Business
[N] Just found out that Milla Jovovich is a dev, invested in AI, and just open sourced a project
Reddit r/MachineLearning

ALTK‑Evolve: On‑the‑Job Learning for AI Agents
Hugging Face Blog

Context Windows Are Getting Absurd — And That's a Good Thing
Dev.to

Every AI Agent Registry in 2026, Compared
Dev.to