Factual and Edit-Sensitive Graph-to-Sequence Generation via Graph-Aware Adaptive Noising

arXiv cs.CL / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DLM4G, a non-autoregressive diffusion-based model for graph-to-sequence (G2S) generation that targets two key weaknesses of fine-tuned autoregressive approaches: factual grounding and edit sensitivity.
  • DLM4G uses graph-to-sequence alignment and an adaptive noising scheme that adjusts noise per token based on denoising error, helping preserve graph structure during generation.
  • The method supports localized updates under graph edits, aiming to improve how generated text changes when the input graph is modified.
  • Across three datasets, DLM4G outperforms other diffusion G2S baselines on both surface-form and embedding-based metrics, and it also surpasses fine-tuned autoregressive baselines up to much larger scales.
  • The authors report improvements over strong PLM and diffusion baselines in factual grounding (FGT@0.5) and edit sensitivity (ESR), and they show generality by extending experiments to molecule captioning beyond purely textual graphs.

Abstract

Fine-tuned autoregressive models for graph-to-sequence generation (G2S) often struggle with factual grounding and edit sensitivity. To tackle these issues, we propose a non-autoregressive diffusion framework that generates text by iterative refinement conditioned on an input graph, named as Diffusion Language Model for Graphs (DLM4G). By aligning graph components (entities/relations) with their corresponding sequence tokens, DLM4G employs an adaptive noising strategy. The proposed strategy uses per-token denoising error as a signal to adaptively modulate noise on entity and relation tokens, improving preservation of graph structure and enabling localized updates under graph edits. Evaluated on three datasets, DLM4G consistently outperforms competitive G2S diffusion baselines trained on identical splits across both surface-form and embedding-based metrics. DLM4G further exceeds fine-tuned autoregressive baselines up to 12x larger (e.g., T5-Large) and is competitive with zero-shot LLM transfer baselines up to 127x larger. Relative to the strongest fine-tuned PLM baseline, DLM4G improves factual grounding (FGT@0.5) by +5.16% and edit sensitivity (ESR) by +7.9%; compared to the best diffusion baseline, it yields gains of +3.75% in FGT@0.5 and +23.6% in ESR. We additionally demonstrate applicability beyond textual graphs through experiments on molecule captioning, indicating the method's generality for scientific G2S generation.