Benchmarking CNN-based Models against Transformer-based Models for Abdominal Multi-Organ Segmentation on the RATIC Dataset
arXiv cs.CV / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study benchmarks three hybrid transformer-based models (UNETR, SwinUNETR, UNETR++) against a CNN baseline (SegResNet) for volumetric multi-organ segmentation on the RATIC dataset, comprising 206 CT scans from 23 institutions across five abdominal organs.
- Under identical preprocessing and training conditions, the CNN-based SegResNet achieves the highest overall Dice Similarity Coefficient, outperforming all transformer-based models on all organs.
- Among transformer approaches, UNETR++ is the most competitive, while UNETR demonstrates faster convergence with fewer training iterations.
- The findings imply that for small- to medium-sized heterogeneous datasets, well-optimized CNN architectures can remain highly competitive and may surpass hybrid transformer designs.
Related Articles
The massive shift toward edge computing and local processing
Dev.to
Self-Refining Agents in Spec-Driven Development
Dev.to
Week 3: Why I'm Learning 'Boring' ML Before Building with LLMs
Dev.to
The Three-Agent Protocol Is Transferable. The Discipline Isn't.
Dev.to

has anyone tried this? Flash-MoE: Running a 397B Parameter Model on a Laptop
Reddit r/LocalLLaMA