Evaluation of Randomization through Style Transfer for Enhanced Domain Generalization

arXiv cs.CV / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study tackles the Sim2Real generalization problem in computer vision, focusing on unresolved contradictions in style-transfer-based domain generalization regarding style-pool diversity, texture complexity, and style source choice.
  • Experiments indicate that using a larger style pool improves performance more than repeatedly augmenting with only a few styles, while texture complexity matters little once the style pool is sufficiently diverse.
  • The paper finds that diverse artistic styles outperform domain-aligned (more “target-like”) alternatives for scene understanding under GTAV-to-real-world benchmarks.
  • Based on these findings, the authors introduce StyleMixDG, a lightweight, model-agnostic augmentation recipe that improves domain generalization without architectural changes or extra losses.
  • On the GTAV → {BDD100k, Cityscapes, Mapillary Vistas} benchmark, StyleMixDG delivers consistent gains over strong baselines, and the code is planned for release on GitHub.

Abstract

Deep learning models for computer vision often suffer from poor generalization when deployed in real-world settings, especially when trained on synthetic data due to the well-known Sim2Real gap. Despite the growing popularity of style transfer as a data augmentation strategy for domain generalization, the literature contains unresolved contradictions regarding three key design axes: the diversity of the style pool, the role of texture complexity, and the choice of style source. We present a systematic empirical study that isolates and evaluates each of these factors for driving scene understanding, resolving inconsistencies in prior work. Our findings show that (i) expanding the style pool yields larger gains than repeated augmentation with few styles, (ii) texture complexity has no significant effect when the pool is sufficiently large, and (iii) diverse artistic styles outperform domain-aligned alternatives. Guided by these insights, we derive StyleMixDG (Style-Mixing for Domain Generalization), a lightweight, model-agnostic augmentation recipe that requires no architectural modifications or additional losses. Evaluated on the GTAV \rightarrow {BDD100k, Cityscapes, Mapillary Vistas} benchmark, StyleMixDG demonstrates consistent improvements over strong baselines, confirming that the empirically identified design principles translate into practical gains. The code will be released on GitHub.