DuSCN-FusionNet: An Interpretable Dual-Channel Structural Covariance Fusion Framework for ADHD Classification Using Structural MRI

arXiv cs.CV / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DuSCN-FusionNet, an interpretable structural MRI (sMRI) framework for classifying ADHD by modeling inter-regional morphological relationships using dual-channel Structural Covariance Networks (SCNs).
  • It builds two SCN channels from ROI-wise mean intensity and intra-regional variability/heterogeneity descriptors, then encodes them with an SCN-CNN and improves predictions via late-stage fusion with auxiliary variability and global statistical features.
  • On the ADHD-200 dataset (Peking University site), the method reports a mean balanced accuracy of 80.59% and AUC of 0.778 using stratified 10-fold cross-validation and a 5-seed ensemble.
  • The authors adapt Grad-CAM to the SCN domain to generate ROI-level importance scores, aiming to support clinical trust by highlighting structurally relevant brain regions as candidate biomarkers.

Abstract

Attention Deficit Hyperactivity Disorder (ADHD) is a highly prevalent neurodevelopmental condition; however, its neurobiological diagnosis remains challenging due to the lack of reliable imaging-based biomarkers, particularly anatomical markers. Structural MRI (sMRI) provides a non-invasive modality for investigating brain alterations associated with ADHD; nevertheless, most deep learning approaches function as black-box systems, limiting clinical trust and interpretability. In this work, we propose DuSCN-FusionNet, an interpretable sMRI-based framework for ADHD classification that leverages dual-channel Structural Covariance Networks (SCNs) to capture inter-regional morphological relationships. ROI-wise mean intensity and intra-regional variability descriptors are used to construct intensity-based and heterogeneity-based SCNs, which are processed through an SCN-CNN encoder. In parallel, auxiliary ROI-wise variability features and global statistical descriptors are integrated via late-stage fusion to enhance performance. The model is evaluated using stratified 10-fold cross-validation with a 5-seed ensemble strategy, achieving a mean balanced accuracy of 80.59% and an AUC of 0.778 on the Peking University site of the ADHD-200 dataset. DuSCN-FusionNet further achieves precision, recall, and F1-scores of 81.66%, 80.59%, and 80.27%, respectively. Moreover, Grad-CAM is adapted to the SCN domain to derive ROI-level importance scores, enabling the identification of structurally relevant brain regions as potential biomarkers.