Amplified Patch-Level Differential Privacy for Free via Random Cropping

arXiv cs.LG / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how random cropping can probabilistically remove spatially localized sensitive content (e.g., faces or license plates) from vision model inputs, adding a new privacy-relevant randomness source during DP-SGD training.
  • It introduces a patch-level neighboring relation for images and derives tight differential privacy bounds for DP-SGD when combined with random cropping.
  • The analysis quantifies patch inclusion probabilities and explains how this interacts with minibatch sampling, effectively reducing the privacy accounting’s sampling rate.
  • Experiments across multiple segmentation architectures and datasets show improved privacy-utility trade-offs from patch-level privacy amplification without changing model architectures or the training procedure.
  • The authors argue that incorporating domain structure into privacy accounting—by leveraging existing stochastic training components—can strengthen privacy guarantees at no added computational or implementation cost.

Abstract

Random cropping is one of the most common data augmentation techniques in computer vision, yet the role of its inherent randomness in training differentially private machine learning models has thus far gone unexplored. We observe that when sensitive content in an image is spatially localized, such as a face or license plate, random cropping can probabilistically exclude that content from the model's input. This introduces a third source of stochasticity in differentially private training with stochastic gradient descent, in addition to gradient noise and minibatch sampling. This additional randomness amplifies differential privacy without requiring changes to model architecture or training procedure. We formalize this effect by introducing a patch-level neighboring relation for vision data and deriving tight privacy bounds for differentially private stochastic gradient descent (DP-SGD) when combined with random cropping. Our analysis quantifies the patch inclusion probability and shows how it composes with minibatch sampling to yield a lower effective sampling rate. Empirically, we validate that patch-level amplification improves the privacy-utility trade-off across multiple segmentation architectures and datasets. Our results demonstrate that aligning privacy accounting with domain structure and additional existing sources of randomness can yield stronger guarantees at no additional cost.