Differentially Private Manifold Denoising

arXiv cs.LG / 4/2/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a differentially private manifold denoising framework that uses sensitive reference datasets to correct noisy, non-private query points while preserving formal , -differential privacy guarantees for the reference data.
  • It uses an iterative pipeline that privately estimates local means and tangent geometry, then projects and corrects query points toward the manifold using modular, budgeted update steps.
  • The approach includes rigorous privacy accounting across multiple iterations and queries and provides a scheduler to allocate the DP privacy budget.
  • The authors derive high-probability non-asymptotic utility bounds under standard manifold regularity, sampling density, and noise assumptions, showing convergence rates depending on sample size, noise, bandwidth, and privacy budget.
  • Experiments and case studies indicate the method can recover geometric signal under moderate privacy budgets, illustrating practical utility-privacy trade-offs for downstream embedding, clustering, and visualization.

Abstract

We introduce a differentially private manifold denoising framework that allows users to exploit sensitive reference datasets to correct noisy, non-private query points without compromising privacy. The method follows an iterative procedure that (i) privately estimates local means and tangent geometry using the reference data under calibrated sensitivity, (ii) projects query points along the privately estimated subspace toward the local mean via corrective steps at each iteration, and (iii) performs rigorous privacy accounting across iterations and queries using (\varepsilon,\delta)-differential privacy (DP). Conceptually, this framework brings differential privacy to manifold methods, retaining sufficient geometric signal for downstream tasks such as embedding, clustering, and visualization, while providing formal DP guarantees for the reference data. Practically, the procedure is modular and scalable, separating DP-protected local geometry (means and tangents) from budgeted query-point updates, with a simple scheduler allocating privacy budget across iterations and queries. Under standard assumptions on manifold regularity, sampling density, and measurement noise, we establish high-probability utility guarantees showing that corrected queries converge toward the manifold at a non-asymptotic rate governed by sample size, noise level, bandwidth, and the privacy budget. Simulations and case studies demonstrate accurate signal recovery under moderate privacy budgets, illustrating clear utility-privacy trade-offs and providing a deployable DP component for manifold-based workflows in regulated environments without reengineering privacy systems.