AutoAWG: Adverse Weather Generation with Adaptive Multi-Controls for Automotive Videos

arXiv cs.CV / 4/22/2026

📰 NewsModels & Research

Key Points

  • AutoAWG is a controllable adverse-weather video generation framework designed to improve perception robustness for autonomous driving by addressing the lack of real adverse-weather video data.
  • The method uses semantics-guided adaptive fusion of multiple controls to achieve a balance between strong weather stylization and the preservation of safety-critical targets.
  • It introduces vanishing point–anchored temporal synthesis to generate training sequences from static images, reducing dependence on fully synthetic data.
  • Masked training is used to improve stability for long-horizon video generation.
  • On the nuScenes validation set, AutoAWG significantly improves over prior methods, with large reductions in FID/FVD and additional gains when using first-frame conditioning.

Abstract

Perception robustness under adverse weather remains a critical challenge for autonomous driving, with the core bottleneck being the scarcity of real-world video data in adverse weather. Existing weather generation approaches struggle to balance visual quality and annotation reusability. We present AutoAWG, a controllable Adverse Weather video Generation framework for Autonomous driving. Our method employs a semantics-guided adaptive fusion of multiple controls to balance strong weather stylization with high-fidelity preservation of safety-critical targets; leverages a vanishing point-anchored temporal synthesis strategy to construct training sequences from static images, thereby reducing reliance on synthetic data; and adopts masked training to enhance long-horizon generation stability. On the nuScenes validation set, AutoAWG significantly outperforms prior state-of-the-art methods: without first-frame conditioning, FID and FVD are relatively reduced by 50.0% and 16.1%; with first-frame conditioning, they are further reduced by 8.7% and 7.2%, respectively. Extensive qualitative and quantitative results demonstrate advantages in style fidelity, temporal consistency, and semantic--structural integrity, underscoring the practical value of AutoAWG for improving downstream perception in autonomous driving. Our code is available at: https://github.com/higherhu/AutoAWG