JumpLoRA: Sparse Adapters for Continual Learning in Large Language Models

arXiv cs.LG / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces JumpLoRA, a framework for continual learning in large language models that adaptively induces sparsity in LoRA blocks to reduce catastrophic forgetting.
  • It uses JumpReLU gating to enable dynamic parameter isolation, helping prevent task-to-task interference during sequential task learning.
  • The authors position JumpLoRA as modular and compatible with existing LoRA-based continual learning methods rather than requiring a complete redesign.
  • Experiments show significant performance gains for IncLoRA and that JumpLoRA outperforms the current state-of-the-art continual learning method ELLA.

Abstract

Adapter-based methods have become a cost-effective approach to continual learning (CL) for Large Language Models (LLMs), by sequentially learning a low-rank update matrix for each task. To mitigate catastrophic forgetting, state-of-the-art approaches impose constraints on new adapters with respect to the previous ones, by targeting either subspace or coordinate-wise interference. In this paper, we propose JumpLoRA, a novel framework to adaptively induce sparsity in the Low-Rank Adaptation (LoRA) blocks through the use of JumpReLU gating. The method achieves dynamic parameter isolation, which helps prevent task interference. We demonstrate that our method is highly modular and compatible with LoRA-based CL approaches. Specifically, it significantly boosts the performance of IncLoRA and outperforms the leading state-of-the-art CL method, ELLA.