FedProxy: Federated Fine-Tuning of LLMs via Proxy SLMs and Heterogeneity-Aware Fusion
arXiv cs.LG / 4/22/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- FedProxy addresses the “federated fine-tuning trilemma” by targeting three issues at once: LLM IP protection, client privacy, and performance degradation on heterogeneous data.
- The work finds that prior IP-preserving approaches such as Offsite-Tuning (OT) rely on weak adapters and therefore hit a performance bottleneck that trails centralized training.
- FedProxy improves fidelity by replacing lightweight adapters with a single, unified Proxy Small Language Model (SLM) compressed from the proprietary LLM to act as a surrogate for collaborative fine-tuning.
- The proposed three-stage design combines server-guided compression, an interference-mitigating aggregation method for heterogeneity, and a training-free “plug-in” fusion step to merge the learned improvements back into the full LLM.
- Experiments indicate FedProxy substantially outperforms OT and can approach centralized fine-tuning performance, while also setting a new benchmark for secure, high-performance federated LLM adaptation.


