Adversary-Free Counterfactual Prediction via Information-Regularized Representations

arXiv stat.ML / 4/28/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles counterfactual prediction under assignment (treatment) bias by proposing an information-theoretic method that removes dependence between the learned representation and treatment without using adversarial training.
  • It starts from a theoretical bound relating the counterfactual–factual risk gap to mutual information, and trains a stochastic representation Z that is outcome-predictive while minimizing I(Z; T).
  • The authors derive a variational training objective that tractably upper-bounds the information term and integrates it with a supervised decoder, aiming for stability and reduced tuning complexity.
  • The approach is extended to dynamic (sequential decision) settings by applying the information penalty at each decision time using sequential representations.
  • Experiments on controlled simulations and a real clinical dataset show favorable results versus state-of-the-art balancing, reweighting, and adversarial baselines across likelihood, counterfactual error, and policy evaluation metrics.

Abstract

We study counterfactual prediction under assignment bias and propose a mathematically grounded, information-theoretic approach that removes treatment-covariate dependence without adversarial training. Starting from a bound that links the counterfactual-factual risk gap to mutual information, we learn a stochastic representation Z that is predictive of outcomes while minimizing I(Z; T). We derive a tractable variational objective that upper-bounds the information term and couples it with a supervised decoder, yielding a stable, provably motivated training criterion. The framework extends naturally to dynamic settings by applying the information penalty to sequential representations at each decision time. We evaluate the method on controlled numerical simulations and a real-world clinical dataset, comparing against recent state-of-the-art balancing, reweighting, and adversarial baselines. Across metrics of likelihood, counterfactual error, and policy evaluation, our approach performs favorably while avoiding the training instabilities and tuning burden of adversarial schemes.