ParaRNN: An Interpretable and Parallelizable Recurrent Neural Network for Time-Dependent Data

arXiv stat.ML / 5/5/2026

📰 NewsModels & Research

Key Points

  • The paper introduces ParaRNN, a new recurrent neural network architecture made of multiple small recurrent units to address traditional RNNs’ poor interpretability and slow training.
  • ParaRNN provides an additive representation that separates recurrent dynamics into interpretable components, enabling analysis via “recurrence features.”
  • The authors show how this interpretability supports applications such as nonparametric regression for time-dependent data.
  • They establish approximation capacity and non-asymptotic prediction error bounds for ParaRNN in the nonparametric regression setting.
  • Experiments on three sequential modeling tasks indicate ParaRNN matches vanilla RNN performance while improving interpretability and training efficiency.

Abstract

The proliferation of large-scale and structurally complex data has spurred the integration of machine learning methods into statistical modeling. Recurrent neural networks (RNNs), a foundational class of models for time-dependent data, can be viewed as nonlinear extensions of classical autoregressive moving average models. Despite their flexibility and empirical success in machine learning, RNNs often suffer from limited interpretability and slow training, which hinders their use in statistics. This paper proposes the Parallelized RNN (ParaRNN), a novel model composed of multiple small recurrent units. ParaRNN admits an additive representation that decouples recurrent dynamics into interpretable components, whose behavior can be characterized through recurrence features. This interpretability enables its applications in nonparametric regression for time-dependent data, while the design also allows efficient parallelization. The approximation capacity and non-asymptotic prediction error bounds in a nonparametric regression setting are established for ParaRNN. Empirical results on three sequential modeling tasks further demonstrate that ParaRNN achieves performance comparable to vanilla RNNs while offering improved interpretability and efficiency.