Low-Rank Adaptation Reduces Catastrophic Forgetting in Sequential Transformer Encoder Fine-Tuning: Controlled Empirical Evidence and Frozen-Backbone Representation Probes
arXiv cs.LG / 2026/3/31
📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- The paper presents a controlled empirical study of Low-Rank Adaptation (LoRA) for sequential fine-tuning of pretrained transformer encoders, focusing on whether it reduces catastrophic forgetting versus full fine-tuning.
- In five reruns on a BERT-base sequence (RTE→MRPC→CoLA→SST-2), full fine-tuning shows about 19.9%±4.8% average forgetting, while standard LoRA (r=8 on query/value modules) reduces forgetting to about 0.6%±1.4% with statistically significant improvement.
- Task-level analyses and secondary experiments on RoBERTa-base confirm that LoRA’s reduced forgetting is not just an aggregate artifact, outperforming the strongest Elastic Weight Consolidation (EWC) baseline (≈15.5%±1.4% forgetting).
- A six-task extension demonstrates that low average forgetting can mask substantial task-level heterogeneity, highlighting the need for more granular evaluation in continual learning settings.
- Freezing and representation-probe ablations indicate a mechanistic account: forgetting drops notably once frozen parameters exceed ~95%, and probes suggest backbone freezing preserves a more stable shared feature scaffold, with full fine-tuning diverging most clearly at the final transformer layer.
