Domain-Aware Hierarchical Contrastive Learning for Semi-Supervised Generalization Fault Diagnosis

arXiv cs.LG / 4/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies semi-supervised domain generalization fault diagnosis (SSDGFD) for cases where target operating conditions are unseen and labeled data are scarce.
  • It identifies two key shortcomings in prior work: pseudo-labels for unlabeled domains suffer from systematic cross-domain bias due to ignored domain-specific geometric differences, and unlabeled samples are often filtered with rigid accept-or-discard thresholds that can waste data and add noise.
  • The authors propose a unified framework, domain-aware hierarchical contrastive learning (DAHCL), which adds a domain-aware learning (DAL) module to learn source-domain geometric characteristics and calibrate pseudo-label predictions across heterogeneous sources.
  • DAHCL also introduces a hierarchical contrastive learning (HCL) module that uses dynamic confidence stratification and fuzzy contrastive supervision so uncertain samples can still improve representation learning without relying on unreliable hard labels.
  • Experiments on three benchmark datasets (with engineering noise incorporated into the evaluation) show the approach improves supervision quality and better utilizes unlabeled data compared with existing SSDGFD methods.

Abstract

Fault diagnosis under unseen operating conditions remains highly challenging when labeled data are scarce. Semi-supervised domain generalization fault diagnosis (SSDGFD) provides a practical solution by jointly exploiting labeled and unlabeled source domains. However, existing methods still suffer from two coupled limitations. First, pseudo-labels for unlabeled domains are typically generated primarily from knowledge learned on the labeled source domain, which neglects domain-specific geometric discrepancies and thus induces systematic cross-domain pseudo-label bias. Second, unlabeled samples are commonly handled with a hard accept-or-discard strategy, where rigid thresholding causes imbalanced sample utilization across domains, while hard-label assignment for uncertain samples can easily introduce additional noise. To address these issues, we propose a unified framework termed domain-aware hierarchical contrastive learning (DAHCL) for SSDGFD. Specifically, DAHCL introduces a domain-aware learning (DAL) module to explicitly capture source-domain geometric characteristics and calibrate pseudo-label predictions across heterogeneous source domains, thereby mitigating cross-domain bias in pseudo-label generation. In addition, DAHCL develops a hierarchical contrastive learning (HCL) module that combines dynamic confidence stratification with fuzzy contrastive supervision, enabling uncertain samples to contribute to representation learning without relying on unreliable hard labels. In this way, DAHCL jointly improves the quality of supervision and the utilization of unlabeled samples. Furthermore, to better reflect practical industrial scenarios, we incorporate engineering noise into the SSDGFD evaluation protocol. Extensive experiments on three benchmark datasets demonstrate that...

Continue reading this article on the original site.

Read original →