VisionNVS: Self-Supervised Inpainting for Novel View Synthesis under the Virtual-Shift Paradigm
arXiv cs.CV / 3/19/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- VisionNVS presents a camera-only framework for novel view synthesis in autonomous driving by reframing the task as self-supervised inpainting under a Virtual-Shift paradigm.
- The Virtual-Shift strategy uses monocular depth proxies to simulate occlusion patterns and map them to the original view, enabling pixel-perfect supervision from raw images and reducing domain gaps.
- The Pseudo-3D Seam Synthesis method aggregates data from adjacent cameras during training to model real-world photometric discrepancies and calibration errors for improved spatial consistency.
- Experiments demonstrate that VisionNVS achieves superior geometric fidelity and visual quality compared with LiDAR-dependent baselines, supporting scalable driving simulation.




