SOLAR: Communication-Efficient Model Adaptation via Subspace-Oriented Latent Adapter Reparametrization

arXiv cs.CL / 4/10/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SOLAR, a post-training compression framework for parameter-efficient fine-tuning (PEFT) adapters that targets reduced communication and storage costs.
  • SOLAR reparameterizes PEFT updates as linear combinations of basis vectors derived from the foundation model’s singular vectors, using controlled random perturbations to keep representations compact.
  • By leveraging subspace similarity between the foundation model and task-specific updates, SOLAR decouples adapter size from the original PEFT structure while maintaining expressiveness.
  • The approach is model-agnostic and compatible with existing PEFT methods such as LoRA and AdaLoRA, and the authors provide a theoretical reconstruction-error bound.
  • Experiments across language and vision tasks (including LLaMA, GPT, and ViT) show SOLAR preserves task performance while significantly reducing adapter representation sizes for deployment in distributed systems and edge devices.

Abstract

Parameter-efficient fine-tuning (PEFT) methods, such as LoRA, enable scalable adaptation of foundation models by injecting low-rank adapters. However, their communication and storage costs remain a major bottleneck in resource-constrained settings. We propose SOLAR (Subspace-Oriented Latent Adapter Reparameterization), a post-training compression framework that substantially reduces the communication cost (i.e., the number of parameters to transmit or store) of PEFT adapters. SOLAR expresses each PEFT update as a linear combination of basis vectors formed from the foundation model's singular vectors with controlled random perturbations. By exploiting the subspace similarity (the alignment of principal directions) between the foundation model and task-specific fine-tuned updates, SOLAR decouples the adapter size from PEFT structure and ensures compact yet expressive representations. It is model-agnostic and compatible with existing PEFT methods, including LoRA, AdaLoRA, and other adapter modules. We theoretically establish a bound on the reconstruction error. Experiments on language and vision tasks using LLaMA, GPT, and ViT models demonstrate that SOLAR preserves task performance while significantly reducing model representation sizes, offering an effective and communication-efficient solution for deployment in distributed systems and edge devices.