Integrating Feature Correlation in Differential Privacy with Applications in DP-ERM

arXiv stat.ML / 5/6/2026

💬 OpinionModels & Research

Key Points

  • The paper addresses a limitation of standard differential privacy, which applies the same privacy protection to all features even though real datasets often contain both sensitive and insensitive features.
  • It introduces a relaxed privacy definition, CorrDP, that treats selected “insensitive” features as less strictly protected while still accounting for their correlation with sensitive features using total variation distance.
  • The authors develop CorrDP-compatible algorithms for differentially private empirical risk minimization (DP-ERM), using distance-dependent noise in gradients to improve theoretical utility.
  • If the feature correlation distance is not known, the framework estimates it from the dataset and still delivers comparable privacy-utility guarantees.
  • Experiments on synthetic and real-world data show that CorrDP-based DP-ERM can outperform standard DP methods when insensitive features exist.

Abstract

Standard differential privacy imposes uniform privacy constraints across all features, overlooking the inherent distinction between sensitive and insensitive features in practice. In this paper, we introduce a relaxed definition of differential privacy that accounts for such heterogeneity, allowing certain features to be treated as insensitive even when correlated with sensitive ones. We propose a correlation-aware framework, \textsf{CorrDP}, which relaxes privacy for insensitive features while accounting for their correlations with sensitive features, with the correlations quantified using total variation distance. We design algorithms for differentially private empirical risk minimization (DP-ERM) under the \textsf{CorrDP} framework, incorporating distance-dependent noise into gradients for improved theoretical utility guarantees. When the correlation distance is unknown, we estimate it from the dataset and show that it achieves a comparable privacy-utility guarantee. We perform experiments on synthetic and real-world datasets and show that \textsf{CorrDP}-based DP-ERM algorithms consistently outperform the standard DP framework in the presence of insensitive features.