CoLA: Cross-Modal Low-rank Adaptation for Multimodal Downstream Tasks

arXiv cs.CL / 4/7/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces CoLA (Cross-Modal Low-rank Adaptation), a parameter-efficient fine-tuning framework that extends LoRA to better capture interactions in multimodal dual-stream architectures.
  • CoLA adds a dedicated inter-modal adaptation pathway in parallel with the usual intra-modal LoRA, aiming to improve cross-modal learning without interference with modality-specific adaptation.
  • Experiments on vision-language benchmarks (RefCOCO, RefCOCO+, RefCOCOg) and audio-visual benchmarks (AVE, AVS) show consistent improvements over standard LoRA, with reported relative gains of about 3% and 2%.
  • The authors claim CoLA enables a “first” multi-task PEFT approach for visual grounding, addressing a gap in efficient adaptation for multimodal downstream tasks.
  • The method maintains parameter efficiency while improving multimodal task performance, making it a practical research direction for adapting large foundation models to multimodal applications.

Abstract

Foundation models have revolutionized AI, but adapting them efficiently for multimodal tasks, particularly in dual-stream architectures composed of unimodal encoders, such as DINO and BERT, remains a significant challenge. Parameter-Efficient Fine-Tuning (PEFT) methods like Low-Rank Adaptation (LoRA) enable lightweight adaptation, yet they operate in isolation within each modality, limiting their ability in capturing cross-modal interactions. In this paper, we take a step in bridging this gap with Cross-Modal Low-Rank Adaptation (CoLA), a novel PEFT framework that extends LoRA by introducing a dedicated inter-modal adaptation pathway alongside the standard intra-modal one. This dual-path design enables CoLA to adapt unimodal foundation models to multimodal tasks effectively, without interference between modality-specific and cross-modal learning. We evaluate CoLA across a range of vision-language (RefCOCO, RefCOCO+, RefCOCOg) and audio-visual (AVE, AVS) benchmarks, where it consistently outperforms LORA, achieving a relative gain of around 3\% and 2\%, respectively, while maintaining parameter efficiency. Notably, CoLA enables the first multi-task PEFT framework for visual grounding, bridging a key gap in efficient multimodal adaptation.