NEMESIS: Noise-suppressed Efficient MAE with Enhanced Superpatch Integration Strategy
arXiv cs.CV / 4/3/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- NEMESIS is a masked autoencoder (MAE) framework for self-supervised learning on 3D CT volumes that uses local 128×128×128 “superpatches” to reduce memory demands while maintaining anatomical detail.
- The method improves pretext learning with a noise-enhanced reconstruction task and uses Masked Anatomical Transformer Blocks (MATB) that apply dual masking via parallel plane-wise and axis-wise token removal.
- It adds NEMESIS Tokens (NT) for cross-scale context aggregation to better capture anisotropic CT structure that conventional masking fails to represent well.
- On the BTCV multi-organ benchmark, NEMESIS achieves 0.9633 mean AUROC with a frozen backbone plus linear classifier, outperforming fully fine-tuned SuPreM and VoCo.
- In a low-label setting with only 10% annotations, it still reaches 0.9075 AUROC and significantly reduces compute (31.0 GFLOPs) versus a full-volume baseline (985.8 GFLOPs).
Related Articles

Black Hat Asia
AI Business

90000 Tech Workers Got Fired This Year and Everyone Is Blaming AI but Thats Not the Whole Story
Dev.to

Microsoft’s $10 Billion Japan Bet Shows the Next AI Battleground Is National Infrastructure
Dev.to

TII Releases Falcon Perception: A 0.6B-Parameter Early-Fusion Transformer for Open-Vocabulary Grounding and Segmentation from Natural Language Prompts
MarkTechPost

Portable eye scanner powered by AI expands access to low-cost community screening
Reddit r/artificial