CI-CBM: Class-Incremental Concept Bottleneck Model for Interpretable Continual Learning

arXiv cs.LG / 4/17/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • Catastrophic forgetting is a major problem in continual learning, and it is particularly severe for class-incremental learning (CIL) where models must learn new classes without losing old knowledge.
  • The paper proposes CI-CBM (Class-Incremental Concept Bottleneck Model) to preserve interpretability while combating forgetting, using concept regularization and pseudo-concept generation.
  • Across evaluations on seven datasets, CI-CBM matches black-box model performance and improves over prior interpretable methods in CIL by an average of 36% accuracy.
  • The method produces both input-level interpretable decisions and global, human-understandable decision rules, and it works in pretrained and from-scratch training settings.
  • The authors make the code publicly available on GitHub for replication and further experimentation.

Abstract

Catastrophic forgetting remains a fundamental challenge in continual learning, in which models often forget previous knowledge when fine-tuned on a new task. This issue is especially pronounced in class incremental learning (CIL), which is the most challenging setting in continual learning. Existing methods to address catastrophic forgetting often sacrifice either model interpretability or accuracy. To address this challenge, we introduce ClassIncremental Concept Bottleneck Model (CI-CBM), which leverage effective techniques, including concept regularization and pseudo-concept generation to maintain interpretable decision processes throughout incremental learning phases. Through extensive evaluation on seven datasets, CI-CBM achieves comparable performance to black-box models and outperforms previous interpretable approaches in CIL, with an average 36% accuracy gain. CICBM provides interpretable decisions on individual inputs and understandable global decision rules, as shown in our experiments, thereby demonstrating that human understandable concepts can be maintained during incremental learning without compromising model performance. Our approach is effective in both pretrained and non-pretrained scenarios; in the latter, the backbone is trained from scratch during the first learning phase. Code is publicly available at github.com/importAmir/CI-CBM.