AI Navigate

Approximate Subgraph Matching with Neural Graph Representations and Reinforcement Learning

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a Reinforcement Learning based Approximate Subgraph Matching (RL-ASM) method that uses Graph Transformer representations and an RL policy to tackle ASM.
  • It builds on a branch-and-bound framework, performing pairwise node matches rather than relying on heuristics, with an imitation-learning stage followed by PPO fine-tuning.
  • Extensive experiments on synthetic and real-world datasets show that RL-ASM outperforms existing methods in both effectiveness and efficiency.
  • The authors provide open-source code at GitHub to enable replication and further research.

Abstract

Approximate subgraph matching (ASM) is a task that determines the approximate presence of a given query graph in a large target graph. Being an NP-hard problem, ASM is critical in graph analysis with a myriad of applications ranging from database systems and network science to biochemistry and privacy. Existing techniques often employ heuristic search strategies, which cannot fully utilize the graph information, leading to sub-optimal solutions. This paper proposes a Reinforcement Learning based Approximate Subgraph Matching (RL-ASM) algorithm that exploits graph transformers to effectively extract graph representations and RL-based policies for ASM. Our model is built upon the branch-and-bound algorithm that selects one pair of nodes from the two input graphs at a time for potential matches. Instead of using heuristics, we exploit a Graph Transformer architecture to extract feature representations that encode the full graph information. To enhance the training of the RL policy, we use supervised signals to guide our agent in an imitation learning stage. Subsequently, the policy is fine-tuned with the Proximal Policy Optimization (PPO) that optimizes the accumulative long-term rewards over episodes. Extensive experiments on both synthetic and real-world datasets demonstrate that our RL-ASM outperforms existing methods in terms of effectiveness and efficiency. Our source code is available at https://github.com/KaiyangLi1992/RL-ASM.