Patient4D: Temporally Consistent Patient Body Mesh Recovery from Monocular Operating Room Video
arXiv cs.CV / 3/19/2026
📰 NewsModels & Research
Key Points
- Patient4D is a stationarity-constrained reconstruction pipeline that recovers dense 3D body meshes from monocular operating-room video by exploiting a stationarity prior to handle occlusion and changing viewpoints.
- The method combines image-level foundation models for perception with lightweight geometric components, notably Pose Locking and Rigid Fallback, to enforce temporal consistency while remaining compatible with off-the-shelf HMR models.
- Evaluation on 4,680 synthetic surgical sequences and three public HMR benchmarks shows mean IoU of 0.75 under drape occlusion and a reduction of failure frames from 30.5% to 1.3% compared with the best baseline.
- The results indicate that leveraging stationarity priors can substantially improve monocular 3D reconstruction in clinical AR scenarios and may broaden the applicability of HMR in surgical settings.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
[P] Vibecoded on a home PC: building a ~2700 Elo browser-playable neural chess engine with a Karpathy-inspired AI-assisted research loop
Reddit r/MachineLearning
Meet DuckLLM 1.0 My First Model!
Reddit r/LocalLLaMA
Since FastFlowLM added support for Linux, I decided to benchmark all the models they support, here are some results
Reddit r/LocalLLaMA
What measure do I use to compare nested models and non nested models in high dimensional survival analysis [D]
Reddit r/MachineLearning