Optimizing Resource-Constrained Non-Pharmaceutical Interventions for Multi-Cluster Outbreak Control Using Hierarchical Reinforcement Learning

arXiv cs.LG / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses allocating scarce non-pharmaceutical interventions (NPIs) across multiple asynchronous outbreak clusters using a hierarchical reinforcement learning framework.
  • It formulates the problem as a constrained restless multi-armed bandit and introduces a global controller that learns a continuous action cost multiplier to shape global resource demand, alongside local policies that estimate the marginal value of allocating resources within each cluster.
  • The framework is evaluated in a realistic agent-based SARS-CoV-2 simulator, demonstrating improvements over RMAB-inspired and heuristic baselines by 20-30% across various system scales and testing budgets.
  • The approach scales to as many as 40 concurrently active clusters and enables faster decision-making than the RMAB-inspired method.

Abstract

Non-pharmaceutical interventions (NPIs), such as diagnostic testing and quarantine, are crucial for controlling infectious disease outbreaks but are often constrained by limited resources, particularly in early outbreak stages. In real-world public health settings, resources must be allocated across multiple outbreak clusters that emerge asynchronously, vary in size and risk, and compete for a shared resource budget. Here, a cluster corresponds to a group of close contacts generated by a single infected index case. Thus, decisions must be made under uncertainty and heterogeneous demands, while respecting operational constraints. We formulate this problem as a constrained restless multi-armed bandit and propose a hierarchical reinforcement learning framework. A global controller learns a continuous action cost multiplier that adjusts global resource demand, while a generalized local policy estimates the marginal value of allocating resources to individuals within each cluster. We evaluate the proposed framework in a realistic agent-based simulator of SARS-CoV-2 with dynamically arriving clusters. Across a wide range of system scales and testing budgets, our method consistently outperforms RMAB-inspired and heuristic baselines, improving outbreak control effectiveness by 20%-30%. Experiments on up to 40 concurrently active clusters further demonstrate that the hierarchical framework is highly scalable and enables faster decision-making than the RMAB-inspired method.