Refined Differentially Private Linear Regression via Extension of a Free Lunch Result

arXiv cs.LG / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the growing need for privacy-preserving linear regression under the add-remove differential privacy (DP) model by focusing on privately estimating dataset-size–dependent quantities for regression.
  • Building on earlier “free lunch” results, it extends the technique using refined, multidimensional simplex transformations for variables and functions bounded in [0,1].
  • The authors show that these transformations improve the private estimation of sufficient statistics required for private simple linear regression via ordinary least squares.
  • They provide analytical and numerical evidence that the proposed approach outperforms prior methods, indicating better accuracy under DP constraints.
  • The transformation framework is presented as broadly adaptable, including for differentially private polynomial regression.

Abstract

As data-privacy regulations tighten and statistical models are increasingly deployed on sensitive human-sourced data, privacy-preserving linear regression has become a critical necessity. For the add-remove DP model, Kulesza et al. (2024) and Fitzsimons et al. (2024) have independently shown that the size of the dataset -- an important statistic for linear regression -- can be privately estimated for "free", via a simplex transformation of bounded variables and private sum queries on the transformed variables. In this work, we extend this free lunch result via carefully crafted multidimensional simplex transformations to variables and functions that are bounded in the interval [0,1]. We show that these transformations can be applied to refine the estimates of sufficient statistics needed for private simple linear regression based on ordinary least squares. We provide both analytical and numerical results to demonstrate the superiority of our approach. Our proposed transformations have general applicability and can be readily adapted for differentially private polynomial regression.