AI Navigate

IGU-LoRA: Adaptive Rank Allocation via Integrated Gradients and Uncertainty-Aware Scoring

arXiv cs.LG / 3/17/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • IGU-LoRA targets the limitation of uniform rank allocation in LoRA by computing within-layer Integrated Gradients sensitivities and aggregating them into a layer-level score for adaptive rank allocation.
  • It adds an uncertainty-aware mechanism that uses exponential moving averages and deviation tracking to suppress noisy updates and better calibrate the number of ranks assigned per layer.
  • The authors provide a theoretical bound on the error of the parameter-space Integrated Gradients under a pathwise Hessian-Lipschitz condition to guide the quadrature budget.
  • Empirical results show IGU-LoRA consistently outperforms strong PEFT baselines at matched parameter budgets across diverse tasks and architectures, boosting downstream accuracy and robustness.
  • Ablation experiments confirm the importance of pathwise within-layer sensitivities and the uncertainty-aware rank selection, and the code is publicly available at GitHub.

Abstract

As large language models (LLMs) scale to billions of parameters, full-parameter fine-tuning becomes compute- and memory-prohibitive. Parameter-efficient fine-tuning (PEFT) mitigates this issue by updating only a small set of task-specific parameters while keeping the base model frozen. Among PEFT approaches, low-rank adaptation (LoRA) is widely adopted; however, it enforces a uniform rank across layers despite substantial variation in layer importance, motivating {layerwise} rank allocation. Recent adaptive-rank variants (e.g., AdaLoRA) allocate ranks based on importance scores, yet typically rely on instantaneous gradients that capture only local sensitivity, overlooking non-local, pathwise effects within the same layer, which yields unstable and biased scores. To address this limitation, we introduce IGU-LoRA, an adaptive-rank LoRA that (i) computes within-layer Integrated Gradients (IG) sensitivities and aggregates them into a layer-level score for rank allocation, and (ii) applies an uncertainty-aware scheme using exponential moving averages with deviation tracking to suppress noisy updates and calibrate rank selection. Theoretically, we prove an upper bound on the composite trapezoidal rule approximation error for parameter-space IG under a pathwise Hessian-Lipschitz condition, which informs the quadrature budget. Across diverse tasks and architectures, IGU-LoRA consistently outperforms strong PEFT baselines at matched parameter budgets, improving downstream accuracy and robustness. Ablations confirm the contributions of pathwise within-layer sensitivity estimates and uncertainty-aware selection to effective rank allocation. Our code is publicly available at https://github.com/withyou12/igulora.git