FedHarmony: Harmonizing Heterogeneous Label Correlations in Federated Multi-Label Learning

arXiv cs.LG / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper studies federated multi-label learning where clients have heterogeneous label distributions and highlights a new issue called “label correlation drift.”
  • It proposes FedHarmony, which uses “consensus correlation” as a global teacher to correct biased correlation estimates on individual clients.
  • FedHarmony’s aggregation weighs each client using both its data size and the quality of its learned label correlations, aiming to improve robustness under heterogeneity.
  • The authors introduce an accelerated optimization method and prove that it achieves faster convergence without degrading accuracy.
  • Experiments on real federated multi-label datasets indicate FedHarmony consistently outperforms existing state-of-the-art approaches.

Abstract

Federated Multi-Label Learning is a distributed paradigm where multiple clients possess heterogeneous multi-label data and perform collaborative learning under privacy constraints without sharing raw data. However, modeling label correlations under heterogeneous distributions remains challenging. Due to client-specific label spaces and varying co-occurrence patterns, correlations learned by individual clients inevitably deviate from the global structure, a phenomenon we term label correlation drift. To address this, we propose FedHarmony, a framework that harmonizes heterogeneous label correlations across clients. It introduces consensus correlation, capturing agreement among other clients and serving as a global teacher to correct biased local estimates. During aggregation, FedHarmony evaluates each client by both data size and correlation quality, assigning weights accordingly. Moreover, we develop an accelerated optimization algorithm for FedHarmony and theoretically establish faster convergence without sacrificing accuracy. Experiments on real-world federated multi-label datasets show that FedHarmony consistently outperforms state-of-the-art methods.