Synthetic Data in Education: Empirical Insights from Traditional Resampling and Deep Generative Models

arXiv cs.LG / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper benchmarks traditional resampling methods (SMOTE, Bootstrap, Random Oversampling) against deep generative models (Autoencoder, Variational Autoencoder, Copula-GAN) for synthetic student performance data.
  • It evaluates both distributional fidelity and downstream ML utility using metrics like KS distance, Jensen-Shannon divergence, and Train-on-Synthetic/Test-on-Real (TSTR).
  • Privacy preservation is assessed via “Distance to Closest Record” (DCR), revealing that resampling can provide near-perfect utility (TSTR ≈ 0.997) while failing privacy (DCR ≈ 0.00).
  • Deep generative models, especially VAEs, deliver strong privacy protection (DCR ≈ 1.00) but incur a meaningful drop in utility; VAEs achieve the best trade-off with 83.3% predictive performance while maintaining complete privacy protection.
  • The study proposes a practical decision framework: use traditional resampling for internal development under controlled privacy, and use VAEs for external data sharing when privacy is critical.

Abstract

Synthetic data generation offers promise for addressing data scarcity and privacy concerns in educational technology, yet practitioners lack empirical guidance for selecting between traditional resampling techniques and modern deep learning approaches. This study presents the first systematic benchmark comparing these paradigms using a 10,000-record student performance dataset. We evaluate three resampling methods (SMOTE, Bootstrap, Random Oversampling) against three deep learning models (Autoencoder, Variational Autoencoder, Copula-GAN) across multiple dimensions: distributional fidelity (Kolmogorov-Smirnov distance, Jensen-Shannon divergence), machine learning utility such as Train-on-Synthetic-Test-on-Real scores (TSTR), and privacy preservation (Distance to Closest Record). Our findings reveal a fundamental trade-off: resampling methods achieve near-perfect utility (TSTR: 0.997) but completely fail privacy protection (DCR ~ 0.00), while deep learning models provide strong privacy guarantees (DCR ~ 1.00) at significant utility cost. Variational Autoencoders emerge as the optimal compromise, maintaining 83.3% predictive performance while ensuring complete privacy protection. We also provide actionable recommendations: use traditional resampling for internal development where privacy is controlled, and VAEs for external data sharing where privacy is paramount. This work establishes a foundational benchmark and practical decision framework for synthetic data generation in learning analytics.