Compensating Visual Insufficiency with Stratified Language Guidance for Long-Tail Class Incremental Learning

arXiv cs.AI / 2026/3/24

💬 オピニオンIdeas & Deep AnalysisModels & Research

要点

  • The paper addresses long-tail class incremental learning (LT CIL), where scarce tail-class samples both slow learning and worsen catastrophic forgetting under shifting, imbalanced data streams.
  • It proposes using language knowledge from large language models (LLMs) by analyzing the LT CIL data distribution to build a stratified language tree that organizes semantics from coarse to fine granularity.
  • It introduces stratified adaptive language guidance that uses learnable weights to merge multi-scale semantic representations, enabling supervision to dynamically adjust for tail classes despite imbalance.
  • It also presents stratified alignment language guidance that constrains optimization using the structural stability of the language tree to improve semantic visual alignment and reduce catastrophic forgetting.
  • Experiments across multiple benchmarks reportedly achieve state-of-the-art performance, indicating the approach is effective for LT CIL.

Abstract

Long-tail class incremental learning (LT CIL) remains highly challenging because the scarcity of samples in tail classes not only hampers their learning but also exacerbates catastrophic forgetting under continuously evolving and imbalanced data distributions. To tackle these issues, we exploit the informativeness and scalability of language knowledge. Specifically, we analyze the LT CIL data distribution to guide large language models (LLMs) in generating a stratified language tree that hierarchically organizes semantic information from coarse to fine grained granularity. Building upon this structure, we introduce stratified adaptive language guidance, which leverages learnable weights to merge multi-scale semantic representations, thereby enabling dynamic supervisory adjustment for tail classes and alleviating the impact of data imbalance. Furthermore, we introduce stratified alignment language guidance, which exploits the structural stability of the language tree to constrain optimization and reinforce semantic visual alignment, thereby alleviating catastrophic forgetting. Extensive experiments on multiple benchmarks demonstrate that our method achieves state of the art performance.