PRIME: Protein Representation via Physics-Informed Multiscale Equivariant Hierarchies

arXiv cs.LG / 5/5/2026

📰 NewsModels & Research

Key Points

  • PRIME is a new protein representation learning framework that models proteins as a nested hierarchy of five physics-grounded structural graphs across multiple spatial resolutions.
  • It connects adjacent structural levels using deterministic, physics-informed assignment operators to enable bidirectional information flow through bottom-up aggregation and top-down refinement.
  • Benchmark experiments show competitive results overall, with particularly large improvements on Fold Classification (Superfamily and Fold splits) versus the strongest geometric GNN baseline.
  • PRIME achieves state-of-the-art performance on Reaction Class prediction, reaching 84.10% accuracy and outperforming baseline methods including ESM.
  • Ablation and cross-attention analyses indicate each structural level provides complementary, non-redundant information and that PRIME adaptively selects the most task-relevant resolutions during inference.
  • The authors have released the source code publicly at the provided GitHub repository.

Abstract

Proteins are inherently multiscale physical systems whose functional properties emerge from coordinated structural organization across multiple spatial resolutions, ranging from atomic interactions to global fold topology. However, existing protein representation learning methods typically operate at a single structural level or treat different sources of structural information as parallel modalities, without explicitly modeling their hierarchical relationships. We introduce PRIME (Protein Representation via Physics-Informed Multiscale Equivariant Hierarchies), a unified framework that models proteins as a nested family of five physically grounded structural graphs spanning surface, atomic, residue, secondary-structure, and protein levels. Adjacent levels are connected through deterministic, physics-informed assignment operators, enabling bidirectional information exchange via bottom-up aggregation and top-down contextual refinement. Experiments on standard protein representation learning benchmarks demonstrate strong and competitive performance across diverse tasks, with particularly notable gains on the Fold Classification benchmark, where PRIME outperforms the strongest geometric GNN baseline by margins of 13.80 and 18.30 points on the harder Superfamily and Fold splits, and achieves a state-of-the-art accuracy of 84.10% on Reaction Class prediction, surpassing all baseline methods, including ESM. Ablation studies confirm that each structural level contributes complementary and non-redundant information, and adaptive cross-attention analysis reveals that PRIME autonomously identifies the most task-relevant structural resolutions at prediction time. Our source code is publicly available at https://github.com/HySonLab/PRIME