Hierarchical Mesh Transformers with Topology-Guided Pretraining for Morphometric Analysis of Brain Structures

arXiv cs.CV / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a Hierarchical Mesh Transformer that can learn from heterogeneous brain meshes (volumetric and surface) using topology-guided hierarchical partitions built from arbitrary-order simplicial complexes.
  • It introduces a feature projection module to integrate variable-length, clinically relevant morphometric descriptors (e.g., cortical thickness, curvature, sulcal depth, myelin content) while decoupling geometric structure from feature dimensionality.
  • The method uses self-supervised pretraining via masked reconstruction of both mesh coordinates and morphometric channels on large unlabeled neuroimaging cohorts to produce a transferable encoder for multiple downstream tasks.
  • Experiments on ADNI (Alzheimer’s disease classification and amyloid burden prediction) and MELD (focal cortical dysplasia detection) report state-of-the-art performance across the tested benchmarks.
  • Overall, the framework aims to enable more generalizable representation learning across different imaging pipelines without requiring topology-specific architectural changes.

Abstract

Representation learning on large-scale unstructured volumetric and surface meshes poses significant challenges in neuroimaging, especially when models must incorporate diverse vertex-level morphometric descriptors, such as cortical thickness, curvature, sulcal depth, and myelin content, which carry subtle disease-related signals. Current approaches either ignore these clinically informative features or support only a single mesh topology, restricting their use across imaging pipelines. We introduce a hierarchical transformer framework designed for heterogeneous mesh analysis that operates on spatially adaptive tree partitions constructed from simplicial complexes of arbitrary order. This design accommodates both volumetric and surface discretizations within a single architecture, enabling efficient multi-scale attention without topology-specific modifications. A feature projection module maps variable-length per-vertex clinical descriptors into the spatial hierarchy, separating geometric structure from feature dimensionality and allowing seamless integration of different neuroimaging feature sets. Self-supervised pretraining via masked reconstruction of both coordinates and morphometric channels on large unlabeled cohorts yields a transferable encoder backbone applicable to diverse downstream tasks and mesh modalities. We validate our approach on Alzheimer's disease classification and amyloid burden prediction using volumetric brain meshes from ADNI, as well as focal cortical dysplasia detection on cortical surface meshes from the MELD dataset, achieving state-of-the-art results across all benchmarks.