Graph Fusion Across Languages using Large Language Models

arXiv cs.CL / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a cross-lingual knowledge-graph fusion framework that uses LLM in-context reasoning and multilingual semantic priors to reconcile semantic differences across languages.
  • It converts graph triples into natural-language-like sequences via structural linearization (e.g., [head] [relation] [tail]) so the LLM can align entities/relations when integrating a candidate graph into an evolving fused graph.
  • Experiments on DBP15K show that LLMs can act as a “universal semantic bridge” to resolve cross-lingual discrepancies and support sequential merging of heterogeneous KGs.
  • The approach is positioned as scalable and modular for continuous knowledge synthesis in multi-source, multilingual environments.

Abstract

Combining multiple knowledge graphs (KGs) across linguistic boundaries is a persistent challenge due to semantic heterogeneity and the complexity of graph environments. We propose a framework for cross-lingual graph fusion, leveraging the in-context reasoning and multilingual semantic priors of Large Language Models (LLMs). The framework implements structural linearization by mapping triplets directly into natural language sequences (e.g., [head] [relation] [tail]), enabling the LLM to map relations and reconcile entities between an evolving fused graph (G_{c}^{(t-1)}) and a new candidate graph (G_{t}). Evaluated on the DBP15K dataset, this exploratory study demonstrates that LLMs can serve as a universal semantic bridge to resolve cross-lingual discrepancies. Results show the successful sequential agglomeration of multiple heterogeneous graphs, offering a scalable, modular solution for continuous knowledge synthesis in multi-source, multilingual environments.