Addressing Data Scarcity in 3D Trauma Detection through Self-Supervised and Semi-Supervised Learning with Vertex Relative Position Encoding
arXiv cs.CV / 3/16/2026
📰 NewsModels & Research
Key Points
- The paper proposes a label-efficient approach combining self-supervised pre-training (masked image modeling) with semi-supervised detection for 3D medical imaging, addressing scarce annotated data.
- It pre-trains a 3D U-Net encoder on 1,206 CT volumes without annotations, learning robust anatomical representations via patch-based MIM.
- Downstream tasks include 3D injury detection using VDETR with Vertex Relative Position Encoding and multi-label injury classification, achieving 56.57% val mAP@0.50 and 45.30% test mAP@0.50 with 144 labeled samples, a 115% improvement over supervised training.
- For classification, with 2,244 labeled samples, the model attains 94.07% test accuracy across seven injury categories using a frozen encoder, showing transferability of self-supervised features.
Related Articles

Jeff Bezos reportedly wants $100 billion to buy and transform old manufacturing firms with AI
TechCrunch
[R] Weekly digest: arXiv AI security papers translated for practitioners -- Cascade (cross-stack CVE+Rowhammer attacks on compound AI), LAMLAD (dual-LLM adversarial ML, 97% evasion), OpenClaw (4 vuln classes in agent frameworks)
Reddit r/MachineLearning
My Experience with Qwen 3.5 35B
Reddit r/LocalLLaMA

Cursor’s new coding model Composer 2 is here: It beats Claude Opus 4.6 but still trails GPT-5.4
VentureBeat
Qwen 3.5 122B completely falls apart at ~ 100K context
Reddit r/LocalLLaMA