AI Navigate

FedTreeLoRA: Reconciling Statistical and Functional Heterogeneity in Federated LoRA Fine-Tuning

arXiv cs.LG / 3/17/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • FedTreeLoRA proposes a tree-structured aggregation for layer-wise alignment in federated LoRA fine-tuning to address both statistical heterogeneity across clients and functional heterogeneity across LLM layers.
  • The framework dynamically builds an aggregation hierarchy that shares broad consensus on shallow trunks while allowing deeper branches to specialize for individual layers, aligning parameter sharing with client similarity.
  • Experimental results on NLU and NLG benchmarks show FedTreeLoRA outperforms existing personalized FL methods by better reconciling generalization and personalization.
  • The work reframes horizontal and vertical heterogeneity as orthogonal yet coupled dimensions in FL, offering a path toward more efficient, privacy-preserving fine-tuning of large language models.

Abstract

Federated Learning (FL) with Low-Rank Adaptation (LoRA) has become a standard for privacy-preserving LLM fine-tuning. However, existing personalized methods predominantly operated under a restrictive Flat-Model Assumption: they addressed client-side \textit{statistical heterogeneity} but treated the model as a monolithic block, ignoring the \textit{functional heterogeneity} across LLM layers. We argue that these two statistical (horizontal) and functional (vertical) dimensions, are \textit{orthogonal in source yet coupled in interaction}, implying that the optimal depth of parameter sharing is functionally dependent on client similarity. To address this, we propose \textbf{FedTreeLoRA}, a framework employing tree-structured aggregation for fine-grained, layer-wise alignment. By dynamically constructing an aggregation hierarchy, FedTreeLoRA allows clients to share broad consensus on shallow `trunks' while progressively specializing on deep `branches'. Experiments on NLU and NLG benchmarks demonstrate that FedTreeLoRA significantly outperforms state-of-the-art methods by effectively reconciling generalization and personalization.