When Can You Poison Rewards? A Tight Characterization of Reward Poisoning in Linear MDPs

arXiv cs.LG / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes “reward poisoning” in reinforcement learning, where an adversary alters rewards within a limited budget to steer an agent toward attacker-chosen behaviors.
  • It provides the first tight necessity-and-sufficiency characterization of when reward poisoning is attackable in linear MDPs, separating vulnerable instances from intrinsically robust ones.
  • The authors establish a “bright line” indicating which RL settings cannot be targeted effectively without prohibitive attack costs, even if the agent uses standard (non-robust) RL algorithms.
  • Beyond linear MDPs, the work argues that approximating deep RL environments as linear MDPs can make the framework general enough to both distinguish vulnerability and efficiently attack susceptible environments in practice.

Abstract

We study reward poisoning attacks in reinforcement learning (RL), where an adversary manipulates rewards within constrained budgets to force the target RL agent to adopt a policy that aligns with the attacker's objectives. Prior works on reward poisoning mainly focused on sufficient conditions to design a successful attacker, while only a few studies discussed the infeasibility of targeted attacks. This paper provides the first precise necessity and sufficiency characterization of the attackability of a linear MDP under reward poisoning attacks. Our characterization draws a bright line between the vulnerable RL instances, and the intrinsically robust ones which cannot be attacked without large costs even running vanilla non-robust RL algorithms. Our theory extends beyond linear MDPs -- by approximating deep RL environments as linear MDPs, we show that our theoretical framework effectively distinguishes the attackability and efficiently attacks the vulnerable ones, demonstrating both the theoretical and practical significance of our characterization.