Malliavin Calculus for Counterfactual Gradient Estimation in Adaptive Inverse Reinforcement Learning

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies adaptive inverse reinforcement learning (IRL), where an IRL method reconstructs a forward learner’s loss function by passively observing gradient information during reinforcement learning.
  • It introduces a passive Langevin-based algorithm whose training requires counterfactual gradients (gradients conditioned on probability-zero events), making naive Monte Carlo estimation inefficient.
  • To address this, the authors use Malliavin calculus to rewrite counterfactual conditional expectations as ratios of unconditioned expectations augmented with Malliavin quantities, enabling efficient estimation with standard convergence rates.
  • The work derives the needed Malliavin derivatives and expresses them via adjoint Skorohod integrals for a general Langevin formulation, culminating in a concrete counterfactual gradient estimation algorithm.
  • Overall, the contribution is a mathematically grounded estimation framework that targets the core bottleneck of counterfactual gradient estimation in adaptive IRL.

Abstract

Inverse reinforcement learning (IRL) recovers the loss function of a forward learner from its observed responses adaptive IRL aims to reconstruct the loss function of a forward learner by passively observing its gradients as it performs reinforcement learning (RL). This paper proposes a novel passive Langevin-based algorithm that achieves adaptive IRL. The key difficulty in adaptive IRL is that the required gradients in the passive algorithm are counterfactual, that is, they are conditioned on events of probability zero under the forward learner's trajectory. Therefore, naive Monte Carlo estimators are prohibitively inefficient, and kernel smoothing, though common, suffers from slow convergence. We overcome this by employing Malliavin calculus to efficiently estimate the required counterfactual gradients. We reformulate the counterfactual conditioning as a ratio of unconditioned expectations involving Malliavin quantities, thus recovering standard estimation rates. We derive the necessary Malliavin derivatives and their adjoint Skorohod integral formulations for a general Langevin structure, and provide a concrete algorithmic approach which exploits these for counterfactual gradient estimation.