One Adapter for All: Towards Unified Representation in Step-Imbalanced Class-Incremental Learning
arXiv cs.CV / 3/12/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses step-imbalanced class-incremental learning (CIL), where task sizes vary and large tasks dominate learning, causing unstable updates from smaller tasks.
- It proposes One-A, a unified, imbalance-aware adapter that incrementally merges task updates into a single adapter to maintain constant inference cost.
- Key components include asymmetric subspace alignment to preserve dominant subspaces learned from large tasks, information-adaptive weighting to balance base and new adapters, and directional gating to selectively fuse updates along singular directions for stability and plasticity.
- Across benchmarks with step-imbalanced streams, One-A achieves competitive accuracy while keeping low inference overhead, demonstrating that a single asymmetrically fused adapter can adapt to dynamic task sizes and streamline deployment.




