Reliev3R: Relieving Feed-forward Reconstruction from Multi-View Geometric Annotations

arXiv cs.CV / 4/2/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Reliev3R, a weakly-supervised training framework for feed-forward reconstruction models that reduces dependence on expensive multi-view geometric annotations like 3D point maps and camera poses.
  • Instead of relying on costly structure-from-motion preprocessing, it leverages monocular relative depths and sparse image correspondences derived from zero-shot predictions from pretrained models to obtain 3D knowledge.
  • Reliev3R proposes an ambiguity-aware relative depth loss and a trigonometry-based reprojection loss to enforce multi-view geometric consistency during training.
  • Experiments claim that training from scratch with less data allows Reliev3R to reach performance comparable to fully-supervised FFRMs, aiming to make 3D reconstruction supervision more scalable and lower-cost.

Abstract

With recent advances, Feed-forward Reconstruction Models (FFRMs) have demonstrated great potential in reconstruction quality and adaptiveness to multiple downstream tasks. However, the excessive reliance on multi-view geometric annotations, e.g. 3D point maps and camera poses, makes the fully-supervised training scheme of FFRMs difficult to scale up. In this paper, we propose Reliev3R, a weakly-supervised paradigm for training FFRMs from scratch without cost-prohibitive multi-view geometric annotations. Relieving the reliance on geometric sensory data and compute-exhaustive structure-from-motion preprocessing, our method draws 3D knowledge directly from monocular relative depths and image sparse correspondences given by zero-shot predictions of pretrained models. At the core of Reliev3R, we design an ambiguity-aware relative depth loss and a trigonometry-based reprojection loss to facilitate supervision for multi-view geometric consistency. Training from scratch with the less data, Reliev3R catches up with its fully-supervised sibling models, taking a step towards low-cost 3D reconstruction supervisions and scalable FFRMs.