MoViD: View-Invariant 3D Human Pose Estimation via Motion-View Disentanglement
arXiv cs.CV / 4/7/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- MoViD is a new framework for viewpoint-invariant 3D human pose estimation that separates viewpoint information from motion features to improve generalization to unseen camera angles.
- It uses a dedicated view estimator (based on key joint relationships) plus an orthogonal projection module to disentangle view and motion representations, strengthened by physics-grounded contrastive alignment across views.
- For efficiency in real-time edge deployment, MoViD uses a frame-by-frame inference pipeline with a view-aware strategy that adaptively activates flip refinement depending on the estimated viewpoint.
- Experiments on nine public datasets and newly collected multiview UAV and gait datasets report over 24.2% lower pose error versus state-of-the-art methods, robustness under severe occlusions with 60% less training data, and real-time performance at 15 FPS on NVIDIA edge devices.
Related Articles

Black Hat Asia
AI Business
v0.20.5
Ollama Releases

Inside Anthropic's Project Glasswing: The AI Model That Found Zero-Days in Every Major OS
Dev.to
Gemma 4 26B fabricated an entire code audit. I have the forensic evidence from the database.
Reddit r/LocalLLaMA

SoloEngine: Low-Code Agentic AI Development Platform with Native Support for Multi-Agent Collaboration, MCP, and Skill System
Dev.to