Soft MPCritic: Amortized Model Predictive Value Iteration

arXiv cs.LG / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces “soft MPCritic,” an RL-MPC framework that learns in a soft value space while using sample-based planning for both online control and generating value targets.
  • It implements MPC via MPPI (model predictive path integral control) and learns a terminal Q-function using fitted value iteration to better align the learned value function with the planner and effectively extend planning horizons.
  • The approach adds an amortized warm-start method that reuses previously planned open-loop action sequences from online observations to compute batched MPPI-based value targets more efficiently.
  • Soft MPCritic uses scenario-based planning with an ensemble of learned dynamics models for next-step prediction accuracy, enabling robust learning from short-horizon planning.
  • The authors argue the combination of value-learning, planner-aligned targets, and amortization makes the framework computationally practical and scalable for both simple and complex control tasks.

Abstract

Reinforcement learning (RL) and model predictive control (MPC) offer complementary strengths, yet combining them at scale remains computationally challenging. We propose soft MPCritic, an RL-MPC framework that learns in (soft) value space while using sample-based planning for both online control and value target generation. soft MPCritic instantiates MPC through model predictive path integral control (MPPI) and trains a terminal Q-function with fitted value iteration, aligning the learned value function with the planner and implicitly extending the effective planning horizon. We introduce an amortized warm-start strategy that recycles planned open-loop action sequences from online observations when computing batched MPPI-based value targets. This makes soft MPCritic computationally practical, while preserving solution quality. soft MPCritic plans in a scenario-based fashion with an ensemble of dynamic models trained for next-step prediction accuracy. Together, these ingredients enable soft MPCritic to learn effectively through robust, short-horizon planning on classic and complex control tasks. These results establish soft MPCritic as a practical and scalable blueprint for synthesizing MPC policies in settings where policy extraction and direct, long-horizon planning may fail.