Uncovering Memorization in Timeseries Imputation models: LBRM Membership Inference and its link to attribute Leakage

arXiv cs.LG / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies privacy risks in deep learning time series imputation models, showing that they can be attacked via black-box inference despite prior focus on memorization in generative models.
  • It introduces a two-stage framework that first performs a membership inference attack using a reference model to improve detection accuracy, including against models that resist overfitting-based attacks.
  • It also presents what it claims is the first attribute inference attack for time series imputation, predicting sensitive characteristics of training data.
  • Experiments across attention-based and autoencoder architectures (trained from scratch and fine-tuned with access to initial weights) show the membership attack retrieves a significant portion of training data, outperforming a naive baseline at tpr@top25%.
  • The authors find the membership attack can predict whether attribute inference will be effective, achieving higher precision (90% vs 78% in general), linking memorization behavior to attribute leakage.

Abstract

Deep learning models for time series imputation are now essential in fields such as healthcare, the Internet of Things (IoT), and finance. However, their deployment raises critical privacy concerns. Beyond the well-known issue of unintended memorization, which has been extensively studied in generative models, we demonstrate that time series models are vulnerable to inference attacks in a black-box setting. In this work, we introduce a two-stage attack framework comprising: (1) a novel membership inference attack based on a reference model that improves detection accuracy, even for models robust to overfitting-based attacks, and (2) the first attribute inference attack that predicts sensitive characteristics of the training data for timeseries imputation model. We evaluate these attacks on attention-based and autoencoder architectures in two scenarios: models that are trained from scratch, and fine-tuned models where the adversary has access to the initial weights. Our experimental results demonstrate that the proposed membership attack retrieves a significant portion of the training data with a tpr@top25% score significantly higher than a naive attack baseline. We show that our membership attack also provides a good insight of whether attribute inference will work (with a precision of 90% instead of 78% in the genral case).