AI Navigate

NeuroLoRA: Context-Aware Neuromodulation for Parameter-Efficient Multi-Task Adaptation

arXiv cs.LG / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • NeuroLoRA introduces a mixture-of-experts based LoRA with a lightweight neuromodulation gate that contextually rescales the projection space before expert selection, preserving the efficiency of frozen random projections.
  • It adds a Contrastive Orthogonality Loss to explicitly separate expert subspaces, improving task decoupling and continual learning.
  • The method achieves consistent improvements over FlyLoRA and other baselines on MMLU, GSM8K, and ScienceQA across single-task, multi-task merging, and sequential continual learning scenarios.
  • Inspired by biological neuromodulation, NeuroLoRA enables context-aware adaptation without increasing parameter overhead.

Abstract

Parameter-Efficient Fine-Tuning (PEFT) techniques, particularly Low-Rank Adaptation (LoRA), have become essential for adapting Large Language Models (LLMs) to downstream tasks. While the recent FlyLoRA framework successfully leverages bio-inspired sparse random projections to mitigate parameter interference, it relies on a static, magnitude-based routing mechanism that is agnostic to input context. In this paper, we propose NeuroLoRA, a novel Mixture-of-Experts (MoE) based LoRA framework inspired by biological neuromodulation -- the dynamic regulation of neuronal excitability based on context. NeuroLoRA retains the computational efficiency of frozen random projections while introducing a lightweight, learnable neuromodulation gate that contextually rescales the projection space prior to expert selection. We further propose a Contrastive Orthogonality Loss to explicitly enforce separation between expert subspaces, enhancing both task decoupling and continual learning capacity. Extensive experiments on MMLU, GSM8K, and ScienceQA demonstrate that NeuroLoRA consistently outperforms FlyLoRA and other strong baselines across single-task adaptation, multi-task model merging, and sequential continual learning scenarios, while maintaining comparable parameter efficiency.