Orthogonal Learner for Estimating Heterogeneous Long-Term Treatment Effects

arXiv cs.LG / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • 本稿は、短期RCTデータと長期の観測データを組み合わせて個別化された長期治療効果(HLTE)を推定する際、特にサブポピュレーションでの「オーバーラップ不足」により推定が不安定になる問題を扱っています。
  • その解決として、LT-O-Learners(Long-Term Orthogonal Learners)を提案し、オーバーラップの低いサンプルを抑える「カスタムのオーバーラップ重み」で学習目的を再ターゲティングします。
  • 提案手法の損失は重み付きのオラクル損失と等価で、Neyman直交性を満たすため、ヌイサンス推定の誤差に対して頑健であることを理論的に示しています。
  • 一般の誤差境界や、準オラクル率(quasi-oracle rate)を達成する条件も提示し、さらにモデル非依存(model-agnostic)に任意の機械学習モデルへ実装可能であると述べています。
  • 合成・準実データのベンチマークで低オーバーラップ設定でも理論特性、とりわけ頑健性が確認されたと報告しています。

Abstract

Estimation of heterogeneous long-term treatment effects (HLTEs) is widely used for personalized decision-making in marketing, economics, and medicine, where short-term randomized experiments are often combined with long-term observational data. However, HLTE estimation is challenging due to limited overlap in treatment or in observing long-term outcomes for certain subpopulations, which can lead to unstable HLTE estimates with large finite-sample variance. To address this challenge, we introduce the LT-O-learners (Long-Term Orthogonal Learners), a set of novel orthogonal learners for HLTE estimation. The learners are designed for the canonical HLTE setting that combines a short-term randomized dataset \mathcal{D}_1 with a long-term historical dataset \mathcal{D}_2. The key idea of our LT-O-Learners is to retarget the learning objective by introducing custom overlap weights that downweight samples with low overlap in treatment or in long-term observation. We show that the retargeted loss is equivalent to the weighted oracle loss and satisfies Neyman-orthogonality, which means our learners are robust to errors in the nuisance estimation. We further provide a general error bound for the LT-O-Learners and give the conditions under which quasi-oracle rate can be achieved. Finally, our LT-O-learners are model-agnostic and can thus be instantiated with arbitrary machine learning models. We conduct empirical evaluations on synthetic and semi-synthetic benchmarks to confirm the theoretical properties of our LT-O-Learners, especially the robustness in low-overlap settings. To the best of our knowledge, ours are the first orthogonal learners for HLTE estimation that are robust to low overlap that is common in long-term outcomes.