Depth-Guided Privacy-Preserving Visual Localization Using 3D Sphere Clouds

arXiv cs.CV / 5/4/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses privacy risks in visual localization that uses sparse private 3D point maps, where deep models can potentially reconstruct high-fidelity scene details.
  • It critiques the common defense of lifting map points to randomly oriented 3D lines because such lines can be recovered via density-based attacks that infer geometry from line neighborhood statistics.
  • To counter this, the authors propose a new privacy-preserving representation called “sphere cloud,” built by lifting points to 3D lines that pass through the map centroid, forming a sphere-like pattern intended to mislead density-based recovery.
  • The authors also identify drawbacks of sphere cloud—namely vulnerability to a new kind of attack and an unresolved translation-scale issue—and mitigate them by using absolute depth maps from on-device time-of-flight (ToF) sensors in a localization framework.
  • Experiments on public RGB-D datasets show sphere cloud maintains competitive privacy protection and runtime performance without significantly harming pose accuracy relative to other depth-guided methods.

Abstract

The emergence of deep neural networks capable of revealing high-fidelity scene details from sparse 3D point clouds has raised significant privacy concerns in visual localization involving private maps. Lifting map points to randomly oriented 3D lines is a well-known approach for obstructing undesired recovery of the scene images, but these lines are vulnerable to a density-based attack that can recover the point cloud geometry by observing the neighborhood statistics of lines. With the aim of nullifying this attack, we present a new privacy-preserving scene representation called \emph{sphere cloud}, which is constructed by lifting all points to 3D lines crossing the centroid of the map, resembling points on the unit sphere. Since lines are most dense at the map centroid, the sphere cloud mislead the density-based attack algorithm to incorrectly yield points at the centroid, effectively neutralizing the attack. Nevertheless, this advantage comes at the cost of i) a new type of attack that may directly recover images from this cloud representation and ii) unresolved translation scale for camera pose estimation. To address these issues, we introduce a simple yet effective cloud construction strategy to thwart new attack and propose an efficient localization framework to guide the translation scale by utilizing absolute depth maps acquired from on-device time-of-flight (ToF) sensors. Experimental results on public RGB-D datasets demonstrate sphere cloud achieves competitive privacy-preserving ability and localization runtime while not excessively compensating the pose estimation accuracy compared to other depth-guided localization methods.