Tracking High-order Evolutions via Cascading Low-rank Fitting

arXiv cs.LG / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies higher-order diffusion models that learn not only first-order velocity fields but also derivatives like acceleration and jerk to expand the family of generative dynamics.
  • It addresses a key scaling bottleneck: naive higher-order models require separate networks per derivative order, increasing parameters linearly with order.
  • The proposed method, cascading low-rank fitting, reuses a shared base function while adding sequential low-rank components to approximate successive derivatives more efficiently.
  • The authors provide theory on how matrix ranks evolve across derivative orders, proving monotonic non-increase under a decomposability assumption and showing rank growth can occur without it via the General Leibniz Rule.
  • They also show that, under conditions, derivative-rank sequences can be shaped to realize arbitrary permutations and include a simple algorithm for efficient computation.

Abstract

Diffusion models have become the de facto standard for modern visual generation, including well-established frameworks such as latent diffusion and flow matching. Recently, modeling high-order dynamics has emerged as a promising frontier in generative modeling. Rather than only learning the first-order velocity field that transports random noise to a target data distribution, these approaches simultaneously learn higher-order derivatives, such as acceleration and jerk, yielding a diverse family of higher-order diffusion variants. To represent higher-order derivatives, naive approaches instantiate separate neural networks for each order, which scales the parameter space linearly with the derivative order. To overcome this computational bottleneck, we introduce cascading low-rank fitting, an ordinary differential equation inspired method that approximates successive derivatives by applying a shared base function augmented with sequentially accumulated low-rank components. Theoretically, we analyze the rank dynamics of these successive matrix differences. We prove that if the initial difference is linearly decomposable, the generic ranks of high-order derivatives are guaranteed to be monotonically non-increasing. Conversely, we demonstrate that without this structural assumption, the General Leibniz Rule allows ranks to strictly increase. Furthermore, we establish that under specific conditions, the sequence of derivative ranks can be designed to form any arbitrary permutation. Finally, we present a straightforward algorithm to efficiently compute the proposed cascading low-rank fitting.