Causally-Guided Diffusion for Stable Feature Selection
arXiv cs.LG / 2026/3/24
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper argues that conventional feature selection methods often overfit to one data distribution and may pick spurious features that break under distribution shifts.
- It introduces CGDFS, which frames stable feature selection as approximate posterior inference over feature subsets with objectives that jointly favor low prediction error and low cross-environment variance.
- CGDFS uses a diffusion model as a learned prior over continuous selection masks, capturing structural dependencies among features while enabling scalable search over a large subset space.
- The method applies guided, annealed Langevin sampling that combines the diffusion prior with a stability-aware likelihood inspired by causal invariance, avoiding hard discrete optimization.
- Experiments on real datasets with distribution shifts show CGDFS selects more stable and transferable features and improves out-of-distribution performance versus several sparsity-, tree-, and stability-based baselines.

