Pseudo-Quantized Actor-Critic Algorithm for Robustness to Noisy Temporal Difference Error

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses instability in reinforcement learning caused by noisy temporal-difference (TD) errors that arise from bootstrapping, which can destabilize value and policy learning.
  • It revisits the control-as-inference perspective and proposes a robust learning rule using a sigmoid-based distribution model of optimality, where large TD errors likely due to noise lead to gradient vanishing and are implicitly excluded from updates.
  • The method analyzes how forward vs. reverse KL divergences differently affect gradient-vanishing behavior, and uses this insight to design a learning update that remains stable under noisy TD signals.
  • It further decomposes optimality into multiple levels to “pseudo-quantize” TD errors for additional noise reduction, and derives an approximate Jensen–Shannon divergence-based alternative that combines favorable properties.
  • Experiments on RL benchmarks show stable learning even in settings where common heuristics (e.g., target networks, ensembles) or noisy rewards are not sufficient.

Abstract

In reinforcement learning (RL), temporal difference (TD) errors are widely adopted for optimizing value and policy functions. However, since the TD error is defined by a bootstrap method, its computation tends to be noisy and destabilize learning. Heuristics to improve the accuracy of TD errors, such as target networks and ensemble models, have been introduced so far. While these are essential approaches for the current deep RL algorithms, they cause side effects like increased computational cost and reduced learning efficiency. Therefore, this paper revisits the TD learning algorithm based on control as inference, deriving a novel algorithm capable of robust learning against noisy TD errors. First, the distribution model of optimality, a binary random variable, is represented by a sigmoid function. Alongside forward and reverse Kullback-Leibler divergences, this new model derives a robust learning rule: when the sigmoid function saturates with a large TD error probably due to noise, the gradient vanishes, implicitly excluding it from learning. Furthermore, the two divergences exhibit distinct gradient-vanishing characteristics. Building on these analyses, the optimality is decomposed into multiple levels to achieve pseudo-quantization of TD errors, aiming for further noise reduction. Additionally, a Jensen-Shannon divergence-based approach is approximately derived to inherit the characteristics of both divergences. These benefits are verified through RL benchmarks, demonstrating stable learning even when heuristics are insufficient or rewards contain noise.