Label-Free Cross-Task LoRA Merging with Null-Space Compression

arXiv cs.AI / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a new label-free LoRA model merging technique called Null-Space Compression (NSC) that merges independently fine-tuned checkpoints without requiring joint multi-task training.
  • NSC derives merge weights from the geometry of LoRA adapters, using the observation that null-space compression in LoRA’s down-projection factor (A in ΔW = BA) correlates with downstream performance.
  • Unlike prior entropy-based surrogate approaches that struggle with regression and are costly for large language models, NSC is output-agnostic and designed to generalize across classification, regression, and sequence generation.
  • Experiments report state-of-the-art results across twenty heterogeneous vision tasks, with more balanced gains than previous methods that tended to overfit subsets of tasks.
  • The method also outperforms baselines on six NLI benchmarks and on vision-language evaluations for VQA and image captioning, suggesting scalability beyond single-modality tasks.

Abstract

Model merging combines independently fine-tuned checkpoints without joint multi-task training. In the era of foundation-model, fine-tuning with Low-Rank Adaptation (LoRA) is prevalent, making LoRA merging a promising target. Existing approaches can work in homogeneous settings where all target tasks are classification but often fail when tasks span classification and regression. Approaches using entropy-based surrogates do not apply to regression and are costly for large language models due to long token sequences. We introduce Null-Space Compression (NSC) Merging, a label-free, output-agnostic method that sets merge weights from adapter geometry. Our key observation is that during LoRA finetuning the down-projection factor A in \Delta W = BA compresses its null space, and the compression correlates with performance. NSC uses this as an optimization signal for merging that can generalize across classification, regression, and sequence generation. NSC achieves state-of-the-art performance across twenty heterogeneous vision tasks with balanced gains where prior methods overfit subsets of tasks. It also outperforms baselines on six NLI benchmarks and on vision-language evaluations for VQA and image captioning, demonstrating scalability and effectiveness.