Exascale Multi-Task Graph Foundation Models for Imbalanced, Multi-Fidelity Atomistic Data
arXiv cs.AI / 4/20/2026
💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research
Key Points
- The paper introduces an exascale materials-discovery workflow using atomistic graph foundation models built on HydraGNN.
- It jointly trains on 16 open first-principles datasets (544M+ structures, 85+ elements) with a multi-task architecture and a scalable ADIOS2/DDStore data pipeline.
- At Frontier, the authors run six large-scale DeepHyper hyperparameter-optimization campaigns (FP64) and then train the best message-passing models on sustained 2,048-node runs to produce a PaiNN-based lead model.
- The lead model supports billion-scale screening by evaluating 1.1B atomistic structures in about 50 seconds, enabling fast and data-scarce fine-tuning across diverse downstream tasks.
- The work analyzes precision/compute tradeoffs across BF16/FP32/FP64 and demonstrates transfer to twelve chemically diverse downstream tasks while validating strong and weak scaling across Frontier, Aurora, and Perlmutter.
Related Articles
Awesome Open-Weight Models: The Practitioner's Guide to Open-Source LLMs (2026 Edition) [P]
Reddit r/MachineLearning

The Mythos vs GPT-5.4-Cyber debate is missing the benchmark
Dev.to

Beyond the Crop: Automating "Ghost Mannequin" Effects with Depth-Aware Inpainting
Dev.to

Everyone is using Claude Code wrong.
Dev.to

A Claude Code hook that warns you before calling a low-trust MCP server
Dev.to