Addressing Data Scarcity in 3D Trauma Detection through Self-Supervised and Semi-Supervised Learning with Vertex Relative Position Encoding
arXiv cs.CV / 3/16/2026
📰 NewsModels & Research
Key Points
- The paper proposes a label-efficient approach combining self-supervised pre-training (masked image modeling) with semi-supervised detection for 3D medical imaging, addressing scarce annotated data.
- It pre-trains a 3D U-Net encoder on 1,206 CT volumes without annotations, learning robust anatomical representations via patch-based MIM.
- Downstream tasks include 3D injury detection using VDETR with Vertex Relative Position Encoding and multi-label injury classification, achieving 56.57% val mAP@0.50 and 45.30% test mAP@0.50 with 144 labeled samples, a 115% improvement over supervised training.
- For classification, with 2,244 labeled samples, the model attains 94.07% test accuracy across seven injury categories using a frozen encoder, showing transferability of self-supervised features.
Related Articles
[D] Matryoshka Representation Learning
Reddit r/MachineLearning
Two new Qwen3.5 “Neo” fine‑tunes focused on fast, efficient reasoning
Reddit r/LocalLLaMA

HKIC, Gobi Partners and HKU team up for fund backing university research start-ups
SCMP Tech
Yann LeCun’s New LeWorldModel (LeWM) Research Targets JEPA Collapse in Pixel-Based Predictive World Modeling
MarkTechPost
Streaming experts
Simon Willison's Blog