AI Navigate

SCE-LITE-HQ: Smooth visual counterfactual explanations with generative foundation models

arXiv cs.LG / 3/19/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • SCE-LITE-HQ introduces a scalable framework for generating counterfactual explanations by operating in the latent space of pretrained generative foundation models and without task-specific retraining.
  • The method uses smoothed gradients to improve optimization stability and employs mask-based diversification to produce realistic and structurally diverse counterfactuals.
  • It is evaluated on natural and medical datasets, showing that the generated counterfactuals are valid, realistic, and diverse, and it competes with or outperforms existing baselines while avoiding heavy model training.
  • The approach aims to improve interpretability for high-resolution visual data, enabling broader scalability of CFEs in practical applications.

Abstract

Modern neural networks achieve strong performance but remain difficult to interpret in high-dimensional visual domains. Counterfactual explanations (CFEs) provide a principled approach to interpreting black-box predictions by identifying minimal input changes that alter model outputs. However, existing CFE methods often rely on dataset-specific generative models and incur substantial computational cost, limiting their scalability to high-resolution data. We propose SCE-LITE-HQ, a scalable framework for counterfactual generation that leverages pretrained generative foundation models without task-specific retraining. The method operates in the latent space of the generator, incorporates smoothed gradients to improve optimization stability, and applies mask-based diversification to promote realistic and structurally diverse counterfactuals. We evaluate SCE-LITE-HQ on natural and medical datasets using a desiderata-driven evaluation protocol. Results show that SCE-LITE-HQ produces valid, realistic, and diverse counterfactuals competitive with or outperforming existing baselines, while avoiding the overhead of training dedicated generative models.