On the Role of Depth in the Expressivity of RNNs

arXiv cs.LG / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes how increasing depth affects the expressive power of recurrent neural networks (RNNs), showing that depth boosts memory capacity efficiently relative to parameter count.
  • It argues that deeper RNNs enhance expressivity not only by enabling more complex transformations of current inputs but also by improving retention of past information.
  • The study extends the theory to 2RNNs, where multiplicative interactions between inputs and hidden states yield polynomial transformations whose maximum degree increases with depth.
  • It further demonstrates that, in general, multiplicative interactions in 2RNNs cannot be effectively replaced by simply adding layerwise nonlinearities.
  • The authors support the theoretical claims with experiments on both synthetic setups and real-world tasks.

Abstract

The benefits of depth in feedforward neural networks are well known: composing multiple layers of linear transformations with nonlinear activations enables complex computations. While similar effects are expected in recurrent neural networks (RNNs), it remains unclear how depth interacts with recurrence to shape expressive power. Here, we formally show that depth increases RNNs' memory capacity efficiently with respect to the number of parameters, thus enhancing expressivity both by enabling more complex input transformations and improving the retention of past information. We broaden our analysis to 2RNNs, a generalization of RNNs with multiplicative interactions between inputs and hidden states. Unlike RNNs, which remain linear without nonlinear activations, 2RNNs perform polynomial transformations whose maximal degree grows with depth. We further show that multiplicative interactions cannot, in general, be replaced by layerwise nonlinearities. Finally, we validate these insights empirically on synthetic and real-world tasks.

On the Role of Depth in the Expressivity of RNNs | AI Navigate