Extraction of linearized models from pre-trained networks via knowledge distillation

arXiv cs.LG / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a framework that uses Koopman operator theory combined with knowledge distillation to extract a linearized classification model from an existing pre-trained neural network.
  • It targets scenarios where only linear operations are feasible or desirable after a simple nonlinear preprocessing step, motivated by advances in photonic integrated circuits and optical hardware.
  • Experiments on MNIST and Fashion-MNIST show the resulting linearized model achieves better classification accuracy than a conventional least-squares-based Koopman approximation.
  • The authors also report improved numerical stability relative to the baseline approach, suggesting more reliable training/inference for the linearized formulation.

Abstract

Recent developments in hardware, such as photonic integrated circuits and optical devices, are driving demand for research on constructing machine learning architectures tailored for linear operations. Hence, it is valuable to explore methods for constructing learning machines with only linear operations after simple nonlinear preprocessing. In this study, we propose a framework to extract a linearized model from a pre-trained neural network for classification tasks by integrating Koopman operator theory with knowledge distillation. Numerical demonstrations on the MNIST and the Fashion-MNIST datasets reveal that the proposed model consistently outperforms the conventional least-squares-based Koopman approximation in both classification accuracy and numerical stability.

Extraction of linearized models from pre-trained networks via knowledge distillation | AI Navigate