SLE-FNO: Single-Layer Extensions for Task-Agnostic Continual Learning in Fourier Neural Operators

arXiv cs.LG / 2026/3/24

💬 オピニオンIdeas & Deep AnalysisModels & Research

要点

  • The paper proposes SLE-FNO, an architecture-based continual learning method that extends Fourier Neural Operators (FNO) using Single-Layer Extensions (SLE) to adapt to distribution shifts without reusing past training data.
  • SLE-FNO is evaluated on an image-to-image regression fluid dynamics problem: mapping transient concentration fields to time-averaged wall shear stress (TAWSS) in pulsatile aneurysmal blood flow.
  • Using 230 CFD simulations split into four sequential, out-of-distribution task configurations, the authors compare SLE-FNO against established continual learning baselines including EWC, LwF, replay methods, OGD, GEM, PiggyBack, and LoRA.
  • Results indicate replay-based methods and architecture-based approaches (PiggyBack, LoRA, and SLE-FNO) yield the best retention, with SLE-FNO achieving the strongest overall trade-off between plasticity and stability.
  • The study reports that SLE-FNO delivers strong accuracy with zero forgetting and minimal added parameters, positioning it as a promising way to update baseline models when extrapolation is required.

Abstract

Scientific machine learning is increasingly used to build surrogate models, yet most models are trained under a restrictive assumption in which future data follow the same distribution as the training set. In practice, new experimental conditions or simulation regimes may differ significantly, requiring extrapolation and model updates without re-access to prior data. This creates a need for continual learning (CL) frameworks that can adapt to distribution shifts while preventing catastrophic forgetting. Such challenges are pronounced in fluid dynamics, where changes in geometry, boundary conditions, or flow regimes induce non-trivial changes to the solution. Here, we introduce a new architecture-based approach (SLE-FNO) combining a Single-Layer Extension (SLE) with the Fourier Neural Operator (FNO) to support efficient CL. SLE-FNO was compared with a range of established CL methods, including Elastic Weight Consolidation (EWC), Learning without Forgetting (LwF), replay-based approaches, Orthogonal Gradient Descent (OGD), Gradient Episodic Memory (GEM), PiggyBack, and Low-Rank Approximation (LoRA), within an image-to-image regression setting. The models were trained to map transient concentration fields to time-averaged wall shear stress (TAWSS) in pulsatile aneurysmal blood flow. Tasks were derived from 230 computational fluid dynamics simulations grouped into four sequential and out-of-distribution configurations. Results show that replay-based methods and architecture-based approaches (PiggyBack, LoRA, and SLE-FNO) achieve the best retention, with SLE-FNO providing the strongest overall balance between plasticity and stability, achieving accuracy with zero forgetting and minimal additional parameters. Our findings highlight key differences between CL algorithms and introduce SLE-FNO as a promising strategy for adapting baseline models when extrapolation is required.