Which Reconstruction Model Should a Robot Use? Routing Image-to-3D Models for Cost-Aware Robotic Manipulation

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how robots should choose among multiple Image-to-3D reconstruction methods with different cost–quality tradeoffs for tasks requiring either fine detail or coarse, collision-safe geometry.
  • It proposes SCOUT, a routing framework that separates reconstruction scoring into (1) viewpoint-dependent model performance modeled by a learned probability distribution and (2) an overall image difficulty estimate via a scalar partition function.
  • SCOUT is designed so that view-invariant reconstruction pipelines can be added, removed, or reconfigured without retraining because the learned component only operates over viewpoint-dependent models.
  • The framework supports arbitrary, multi-dimensional cost constraints at inference time, making it suitable for real robotic systems where compute, latency, and quality requirements vary.
  • Experiments on several 3D reconstruction datasets and robotic grasping/dexterous manipulation show consistent improvements over routing baselines, and the authors release code and additional results.

Abstract

Robotic manipulation tasks require 3D mesh reconstructions of varying quality: dexterous manipulation demands fine-grained surface detail, while collision-free planning tolerates coarser representations. Multiple reconstruction methods offer different cost-quality tradeoffs, from Image-to-3D models - whose output quality depends heavily on the input viewpoint - to view-invariant methods such as structured light scanning. Querying all models is computationally prohibitive, motivating per-input model selection. We propose SCOUT, a novel routing framework that decouples reconstruction scores into two components: (1) the relative performance of viewpoint-dependent models, captured by a learned probability distribution, and (2) the overall image difficulty, captured by a scalar partition function estimate. As the learned network operates only over the viewpoint-dependent models, view-invariant pipelines can be added, removed, or reconfigured without retraining. SCOUT also supports arbitrary cost constraints at inference time, accommodating the multi-dimensional cost constraints common in robotics. We evaluate on the Google Scanned Objects, BigBIRD, and YCB datasets under multiple mesh quality metrics, demonstrating consistent improvements over routing baselines adapted from the LLM literature across various cost constraints. We further validate the framework through robotic grasping and dexterous manipulation experiments. We release the code and additional results on our website.