EmbedPart: Embedding-Driven Graph Partitioning for Scalable Graph Neural Network Training

arXiv cs.LG / 4/2/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the challenge of scaling Graph Neural Network (GNN) training to massive graphs by partitioning them across machines to minimize inter-machine communication and balance load.
  • It proposes EmbedPart, an embedding-driven partitioning method that clusters dense node embeddings generated during the actual GNN training workload instead of partitioning directly on the original irregular graph structure.
  • The authors report that EmbedPart delivers over 100x speedup versus Metis while maintaining competitive partitioning quality, leading to faster distributed GNN training.
  • EmbedPart is designed to support graph updates and efficient repartitioning, and it can also be used for graph reordering to improve data locality and speed up single-machine training.

Abstract

Graph Neural Networks (GNNs) are widely used for learning on graph-structured data, but scaling GNN training to massive graphs remains challenging. To enable scalable distributed training, graphs are divided into smaller partitions that are distributed across multiple machines such that inter-machine communication is minimized and computational load is balanced. In practice, existing partitioning approaches face a fundamental trade-off between partitioning overhead and partitioning quality. We propose EmbedPart, an embedding-driven partitioning approach that achieves both speed and quality. Instead of operating directly on irregular graph structures, EmbedPart leverages node embeddings produced during the actual GNN training workload and clusters these dense embeddings to derive a partitioning. EmbedPart achieves more than 100x speedup over Metis while maintaining competitive partitioning quality and accelerating distributed GNN training. Moreover, EmbedPart naturally supports graph updates and fast repartitioning, and can be applied to graph reordering to improve data locality and accelerate single-machine GNN training. By shifting partitioning from irregular graph structures to dense embeddings, EmbedPart enables scalable and high-quality graph data optimization.