DecompKAN: Decomposed Patch-KAN for Long-Term Time Series Forecasting

arXiv cs.LG / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • DecompKAN is a new, lightweight time-series forecasting model that avoids attention and combines trend–residual decomposition, channel-wise patching, learned instance normalization, and B-spline Kolmogorov-Arnold Network (KAN) edge functions.
  • The model’s KAN edges learn explicit, inspectable 1D scalar functions over learned patch-embedding coordinates, enabling visualization of latent nonlinear transformations.
  • On multiple standard benchmarks, DecompKAN achieves best or tied-best MSE in a large fraction of dataset–horizon and cross-dataset comparisons, with particular gains on datasets exhibiting smooth temporal dynamics and on physiological series like PPG-DaLiA.
  • Ablation results suggest the main performance driver is the overall architectural pipeline (decomposition, patching, normalization), while the KAN formulation primarily improves interpretability by making learned transformations inspectable across domains.
  • Domain visualizations indicate qualitatively different nonlinearities learned in different scientific/physiological settings, supporting the transparency motivation of the work.

Abstract

Accurate time series forecasting in scientific domains such as climate modeling, physiological monitoring, and energy systems benefits from both competitive predictions and model transparency. This work proposes DecompKAN, a lightweight attention-free architecture that combines trend-residual decomposition, channel-wise patching, learned instance normalization, and B-spline Kolmogorov-Arnold Network (KAN) edge functions. Each KAN edge learns an explicit, inspectable 1D scalar function over learned patch-embedding coordinates that can be directly visualized. On standard benchmarks, DecompKAN achieves best or tied-best MSE on 15 of 32 dataset-horizon combinations among selected published baselines, and achieves best or tied-best MSE on 20 of 36 comparisons under a controlled same-recipe evaluation across 9 datasets including the physiological PPG-DaLiA benchmark. The architecture shows particular strength on datasets with smooth temporal dynamics (Solar -17%, ECL -10% vs. iTransformer, Weather) and physiological time series. Visualization of learned edge functions reveals qualitatively different latent nonlinearities across domains. Ablation analysis shows that the architectural pipeline (decomposition, patching, normalization) drives performance more than the choice of nonlinear layer, while the KAN formulation enables inspection of learned latent transformations.