FAAR: Efficient Frequency-Aware Multi-Task Fine-Tuning via Automatic Rank Selection

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The FAAR paper proposes an efficient parameter-efficient fine-tuning approach for multi-task learning that avoids the high cost of full fine-tuning as model sizes and the number of tasks grow.
  • Instead of using a fixed low-rank setting, FAAR uses Performance-Driven Rank Shrinking (PDRS) to automatically allocate an optimal rank per adapter location and per task.
  • To better capture inter-task relationships and spatial information, FAAR introduces a Task-Spectral Pyramidal Decoder (TS-PD) that leverages the image frequency spectrum for input-specific context in spatial bias learning.
  • Experiments on dense visual task benchmarks show FAAR improves accuracy and efficiency over prior PEFT methods for MTL, including reducing parameters by up to 9× versus traditional MTL fine-tuning.
  • The authors provide code, enabling others to reproduce and adopt the FAAR method for efficient multi-task adaptation workflows.

Abstract

Adapting models pre-trained on large-scale datasets is a proven way to reach strong performance quickly for down-stream tasks. However, the growth of state-of-the-art mod-els makes traditional full fine-tuning unsuitable and difficult, especially for multi-task learning (MTL) where cost scales with the number of tasks. As a result, recent studies investigate parameter-efficient fine-tuning (PEFT) using low-rank adaptation to significantly reduce the number of trainable parameters. However, these existing methods use a single, fixed rank, which may not be optimal for differ-ent tasks or positions in the MTL architecture. Moreover, these methods fail to learn spatial information that cap-tures inter-task relationships and helps to improve diverse task predictions. This paper introduces Frequency-Aware and Automatic Rank (FAAR) for efficient MTL fine-tuning. Our method introduces Performance-Driven Rank Shrink-ing (PDRS) to allocate the optimal rank per adapter location and per task. Moreover, by analyzing the image frequency spectrum, FAAR proposes a Task-Spectral Pyramidal Decoder (TS-PD) that injects input-specific context into spatial bias learning to better reflect cross-task relationships. Experiments performed on dense visual task benchmarks show the superiority of our method in terms of both accuracy and efficiency compared to other PEFT methods in MTL. FAAR reduces the number of parameters by up to 9 times compared to traditional MTL fine-tuning whilst improving overall performance. Our code is available.