Near-Optimal Sample Complexities of Divergence-based S-rectangular Distributionally Robust Reinforcement Learning

arXiv stat.ML / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies divergence-based S-rectangular distributionally robust reinforcement learning (DR-RL), focusing on the empirical value iteration algorithm under adversaries that better reflect real-world distribution shifts.
  • It derives near-optimal sample complexity bounds of \(\widetilde{O}(|\mathcal{S}||\mathcal{A}|(1-\gamma)^{-4}\varepsilon^{-2})\), quantifying how many samples are needed to reach target accuracy \(\varepsilon\).
  • The authors claim these results are the first to achieve (at once) near-optimal dependence on the state size \(|\mathcal{S}|\), action size \(|\mathcal{A}|\), and accuracy \(\varepsilon\) for divergence-based S-rectangular models.
  • Numerical experiments on robust inventory control and a worst-case theoretical example are used to confirm the predicted fast learning behavior.
  • The work highlights a trade-off in DR-RL modeling: S-rectangular adversaries are potentially more expressive than SA-rectangular ones, enabling stronger robust randomized policies even while maintaining tractable analysis.

Abstract

Distributionally robust reinforcement learning (DR-RL) has recently gained significant attention as a principled approach that addresses discrepancies between training and testing environments. To balance robustness, conservatism, and computational traceability, the literature has introduced DR-RL models with SA-rectangular and S-rectangular adversaries. While most existing statistical analyses focus on SA-rectangular models, owing to their algorithmic simplicity and the optimality of deterministic policies, S-rectangular models more accurately capture distributional discrepancies in many real-world applications and often yield more effective robust randomized policies. In this paper, we study the empirical value iteration algorithm for divergence-based S-rectangular DR-RL and establish near-optimal sample complexity bounds of \widetilde{O}(|\mathcal{S}||\mathcal{A}|(1-\gamma)^{-4}\varepsilon^{-2}), where \varepsilon is the target accuracy, |\mathcal{S}| and |\mathcal{A}| denote the cardinalities of the state and action spaces, and \gamma is the discount factor. To the best of our knowledge, these are the first sample complexity results for divergence-based S-rectangular models that achieve optimal dependence on |\mathcal{S}|, |\mathcal{A}|, and \varepsilon simultaneously. We further validate this theoretical dependence through numerical experiments on a robust inventory control problem and a theoretical worst-case example, demonstrating the fast learning performance of our proposed algorithm.