Natural Gradient Descent for Online Continual Learning
arXiv cs.LG / 2026/3/24
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper targets Online Continual Learning (OCL) for image classification, where models must learn from a data stream without assuming i.i.d. data and must avoid catastrophic forgetting.
- It proposes a training approach using Natural Gradient Descent with an approximation of the Fisher Information Matrix via Kronecker-Factored Approximate Curvature (KFAC) to improve convergence in the online setting.
- The method yields substantial performance gains across multiple existing OCL methods, indicating the optimizer/curvature component is broadly beneficial.
- Experiments on Split CIFAR-100, CORE50, and Split miniImageNet show the improvements are especially pronounced when the proposed optimizer is combined with other OCL “tricks.”

