AI Navigate

A Systematic Benchmark of GAN Architectures for MRI-to-CT Synthesis

arXiv cs.CV / 3/17/2026

📰 NewsTools & Practical UsageModels & Research

Key Points

  • The paper benchmarks ten GAN architectures for MRI-to-CT translation on the SynthRAD2025 dataset across abdomen, thorax, and head‑and‑neck, using a unified validation protocol with identical preprocessing and optimization settings.
  • It assesses voxel-wise accuracy, structural fidelity, perceptual quality, and distribution-level realism, in addition to computational complexity, to provide a comprehensive performance comparison.
  • Supervised paired models consistently outperform unpaired approaches, highlighting the importance of voxel-wise supervision in this task.
  • Pix2Pix achieves the most balanced performance across districts with a favorable quality-to-complexity trade-off, while multi-district training improves structural robustness and intra-district training enhances voxel-wise fidelity.
  • The study offers quantitative guidance for model selection in MRI-only radiotherapy workflows and provides a reproducible framework with public code and results.

Abstract

The translation from Magnetic resonance imaging (MRI) to Computed tomography (CT) has been proposed as an effective solution to facilitate MRI-only clinical workflows while limiting exposure to ionizing radiation. Although numerous Generative Adversarial Network (GAN) architectures have been proposed for MRI-to-CT translation, systematic and fair comparisons across heterogeneous models remain limited. We present a comprehensive benchmark of ten GAN architectures evaluated on the SynthRAD2025 dataset across three anatomical districts (abdomen, thorax, head-and-neck). All models were trained under a unified validation protocol with identical preprocessing and optimization settings. Performance was assessed using complementary metrics capturing voxel-wise accuracy, structural fidelity, perceptual quality, and distribution-level realism, alongside an analysis of computational complexity. Supervised Paired models consistently outperformed Unpaired approaches, confirming the importance of voxel-wise supervision. Pix2Pix achieved the most balanced performance across districts while maintaining a favorable quality-to-complexity trade-off. Multi-district training improved structural robustness, whereas intra-district training maximized voxel-wise fidelity. This benchmark provides quantitative and computational guidance for model selection in MRI-only radiotherapy workflows and establishes a reproducible framework for future comparative studies. To ensure the reproducibility of our experiments we make our code public, together with the overall results, at the following link:https://github.com/arco-group/MRI_TO_CT.git