Bi-Level Reinforcement Learning Control for an Underactuated Blimp via Center-of-Mass Reconfiguration
arXiv cs.RO / 5/5/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies goal-directed tracking control for underactuated blimps by using center-of-mass (CoM) reconfiguration with a compact hardware setup of two thrusters and a movable internal slider.
- It argues that this design improves energy efficiency and payload capacity but creates strong nonlinear coupling and significant underactuation between CoM dynamics and vehicle motion.
- To manage these difficulties, the authors propose a bi-level reinforcement learning framework that separates CoM planning (outer policy) from thrust control for reference tracking (inner policy).
- A two-stage learning strategy and convergence analysis are introduced to stabilize the bi-level RL training process.
- Extensive simulations and real-world experiments on a 27-goal evaluation set show improved tracking accuracy and robustness over fixed-CoM baselines and PID controllers, with reliable sim-to-real transfer.
Related Articles

Singapore's Fraud Frontier: Why AI Scam Detection Demands Regulatory Precision
Dev.to

How AI is Changing the Way We Code in 2026: The Shift from Syntax to Strategy
Dev.to

13 CLAUDE.md Rules That Make AI Write Modern PHP (Not PHP 5 Resurrected)
Dev.to

MCP annotations are a UX layer, not a security layer
Dev.to
From OOM to 262K Context: Running Qwen3-Coder 30B Locally on 8GB VRAM
Dev.to