Teacher Forcing as Generalized Bayes: Optimization Geometry Mismatch in Switching Surrogates for Chaotic Dynamics

arXiv cs.LG / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies Identity Teacher Forcing (ITF) for training deterministic recurrent surrogates for chaotic dynamical systems, showing it can work well for dynamical system reconstruction (DSR) with RNNs.
  • It argues that ITF’s intervention-based prediction loss (viewed as a generalized Bayes update) can mismatch the free-running model’s marginal likelihood “geometry,” leading to different objective curvatures.
  • Using a probabilistic switching augmentation of almost-linear RNNs (AL-RNNs), the authors compare ITF vs marginal-likelihood curvature and use Louis’ identity to estimate ambiguity-aware observed information.
  • In their switching experiments (including Lorenz-63), conditioning on a single forced regime increases curvature, while marginal likelihood curvature is reduced via a missing-information correction when multiple switching explanations are plausible.
  • They find that while windowed evidence fine-tuning can improve held-out evidence, it may worsen dynamical quantities of interest (QoIs) compared with models pretrained under ITF.

Abstract

Identity teacher forcing (ITF) enables stable training of deterministic recurrent surrogates for chaotic dynamical systems and has been highly effective for dynamical systems reconstruction (DSR) with recurrent neural networks (RNNs), including interpretable almost-linear RNNs (AL-RNNs). However, as an intervention-based prediction loss (and thus a generalized Bayes update), teacher forcing need not match the free-running model's marginal likelihood geometry. We compare the objective-induced curvatures of ITF and marginal likelihood in a probabilistic switching augmentation of AL-RNNs, estimating ambiguity-aware observed information via Louis' identity. In the switching setting studied here, conditioning on a single forced regime path (as ITF does) inflates curvature, while marginal likelihood curvature is reduced by a missing-information correction when multiple switching explanations remain plausible. In Lorenz-63 experiments, windowed evidence fine-tuning improves held-out evidence but can degrade dynamical quantities of interest (QoIs) relative to ITF-pretrained models.