Preserving Temporal Dynamics in Time Series Generation

arXiv cs.LG / 5/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The article addresses a key limitation of existing GAN-based time-series generation methods: they often match only marginal distributions and fail to preserve natural temporal dynamics in multivariate data.
  • It proposes a model-agnostic, MCMC-based framework that mitigates distribution shift and temporal drift by enforcing consistency with empirical transition statistics between neighboring time points.
  • The authors provide a theoretical analysis showing how deviations accumulate in sequential conditional generation and argue that the proposed MCMC correction can counteract these discrepancies.
  • Experiments on Lorenz, Licor, ETTh, and ILI datasets (using several GAN variants and related models) show consistent improvements across multiple fidelity, statistical, and predictive metrics, including autocorrelation alignment and predictive scores.
  • Overall, the work suggests that high-quality synthetic time series require explicitly preserving transition laws rather than relying solely on adversarial distribution matching.

Abstract

Time-series data augmentation plays a crucial role in regression-oriented forecasting tasks, where limited data restricts the performance of deep learning models. While Generative Adversarial Networks (GANs) have shown promise in synthetic time-series generation, existing approaches primarily focus on matching marginal data distributions and often overlook the temporal dynamics that naturally exist in the original multivariate time series. When generating multivariate time series, this mismatch leads to distribution shift and temporal drift, thereby degrading the fidelity of the synthetic sequences. In this work, we propose a model-agnostic Markov Chain Monte Carlo (MCMC)-based framework to mitigate distribution shift and preserve temporal dynamics in synthetic time series. We provide a theoretical analysis of how conditional generative models accumulate deviations under sequential generation and demonstrate that the MCMC algorithm can correct these discrepancies by enforcing consistency with empirical transition statistics between neighboring time points. Extensive experiments on the Lorenz, Licor, ETTh, and ILI datasets using RCGAN, GCWGAN, TimeGAN, SigCWGAN, and AECGAN demonstrate that the proposed MCMC framework consistently improves autocorrelation alignment, skewness error, kurtosis error, R^2, discriminative score, and predictive score. These results suggest that synthetic time series consistent with the original data require explicit preservation of transition laws rather than solely relying on adversarial distribution matching, thereby offering a principled direction for improving generative modeling of time-series data.