On the Spectral Structure and Objective Equivalence of Orthogonal Multilabel Fisher Discriminants

arXiv stat.ML / 5/6/2026

💬 OpinionModels & Research

Key Points

  • The paper delivers a unified theoretical framework for Linear Discriminant Analysis extended to multi-label settings, using multilabel scatter matrices and Stiefel orthogonality constraints.
  • It shows that the effective discriminant dimensionality in multilabel problems can exceed the classic single-label upper bound of C−1, and it provides a variance partition plus conditions under which four Fisher objectives become equivalent.
  • The authors prove a two-sided bound linking projected distances to Hamming distances between labels, offering a way to interpret embedding geometry in label space.
  • On the statistical side, they derive near-minimax optimal finite-sample rates for subspace estimation under sub-Gaussian noise, including both an upper bound and a matching minimax lower bound.
  • Numerical experiments on synthetic data validate the algebraic identities and the multilabel-specific quantities that control the theoretical guarantees, while real-dataset application evaluation is deferred to future work.

Abstract

We provide a unified theoretical analysis of Linear Discriminant Analysis with simultaneous multilabel scatter matrix formulations and Stiefel orthogonality constraints. Our contributions span both algebraic structure and statistical guarantees. On the algebraic side, we characterize the rank of the multilabel between-class scatter matrix, showing that the effective discriminant dimensionality can strictly exceed the classical single-label bound of C-1; we establish a multilabel partition of variance and prove that all four Fisher objectives are equivalent under the W^\top S_t^{ML} W = I_r constraint while characterizing their divergence under the Stiefel constraint; and we prove a two-sided label-distance preservation bound relating projected distances to Hamming distances in label space. On the statistical side, we establish a finite-sample O(k_{\max}\sqrt{d\log d/n}/gap_r) bound on the subspace estimation error under sub-Gaussian noise with a matching \Omega(\sigma^2 d/(n\,gap_r)) minimax lower bound, establishing a near-minimax-optimal rate (matching up to logarithmic and k_{\max} factors) for multilabel discriminant subspace estimation. We further provide high-probability distance concentration, robustness guarantees under label interactions, and a regularization analysis preserving the spectral structure when d \gg n. All results are verified numerically on synthetic data generated from the linear label-effect model, covering both the algebraic identities and the multilabel-specific quantities (k_{\max}, \kappa(S_t^{ML}), \|\Gamma/n\|_2, \Delta_r) that govern the statistical bounds. The numerical experiments are designed as a sanity check for the theorems rather than as an empirical benchmark; evaluation on real multilabel datasets is left to future work targeting application-oriented venues.