DeepStock: Reinforcement Learning with Policy Regularizations for Inventory Management

arXiv cs.LG / 3/23/2026

💬 OpinionIndustry & Market MovesModels & Research

Key Points

  • The authors show that policy regularizations grounded in classical inventory concepts like Base Stock can significantly accelerate hyperparameter tuning for DRL and improve final performance.
  • Policy regularizations reduce sensitivity to training hyperparameters, making DRL-based inventory policies more robust in practice.
  • The work reports a 100% deployment of DRL with policy regularizations on Alibaba's Tmall, demonstrating real-world viability at scale.
  • Additional synthetic experiments indicate that policy regularizations influence which DRL method is considered best for inventory management, reshaping practical recommendations.

Abstract

Deep Reinforcement Learning (DRL) provides a general-purpose methodology for training inventory policies that can leverage big data and compute. However, off-the-shelf implementations of DRL have seen mixed success, often plagued by high sensitivity to the hyperparameters used during training. In this paper, we show that by imposing policy regularizations, grounded in classical inventory concepts such as "Base Stock", we can significantly accelerate hyperparameter tuning and improve the final performance of several DRL methods. We report details from a 100% deployment of DRL with policy regularizations on Alibaba's e-commerce platform, Tmall. We also include extensive synthetic experiments, which show that policy regularizations reshape the narrative on what is the best DRL method for inventory management.