Render-in-the-Loop: Vector Graphics Generation via Visual Self-Feedback

arXiv cs.CV / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that existing text-to-SVG approaches for multimodal LLMs are often “open-loop,” generating SVG code without actually seeing intermediate render states, which limits visuo-spatial reasoning.
  • It proposes “Render-in-the-Loop,” a step-wise SVG generation paradigm that repeatedly renders partial code into a cumulative canvas so the model can condition subsequent tokens on evolving visual context.
  • The authors show that naively adding a visual loop to off-the-shelf models underperforms, so they introduce fine-grained path decomposition and a Visual Self-Feedback (VSF) training strategy to better learn incremental visual-to-code mappings.
  • For inference, they add a Render-and-Verify (RaV) mechanism to filter degenerate or redundant drawing primitives, and the resulting system outperforms strong open-weight baselines on MMSVGBench for both Text-to-SVG and Image-to-SVG.
  • Overall, the work highlights improved data efficiency and generalization by using visual self-feedback and verification rather than treating SVG as purely symbolic code generation.

Abstract

Multimodal Large Language Models (MLLMs) have shown promising capabilities in generating Scalable Vector Graphics (SVG) via direct code synthesis. However, existing paradigms typically adopt an open-loop "blind drawing" approach, where models generate symbolic code sequences without perceiving intermediate visual outcomes. This methodology severely underutilizes the powerful visual priors embedded in MLLMs vision encoders, treating SVG generation as a disjointed textual sequence modeling task rather than an integrated visuo-spatial one. Consequently, models struggle to reason about partial canvas states and implicit occlusion relationships, which are visually explicit but textually ambiguous. To bridge this gap, we propose Render-in-the-Loop, a novel generation paradigm that reformulates SVG synthesis as a step-wise, visual-context-aware process. By rendering intermediate code states into a cumulative canvas, the model explicitly observes the evolving visual context at each step, leveraging on-the-fly feedback to guide subsequent generation. However, we demonstrate that applying this visual loop naively to off-the-shelf models is suboptimal due to their inability to leverage incremental visual-code mappings. To address this, we first utilize fine-grained path decomposition to construct dense multi-step visual trajectories, and then introduce a Visual Self-Feedback (VSF) training strategy to condition the next primitive generation on intermediate visual states. Furthermore, a Render-and-Verify (RaV) inference mechanism is proposed to effectively filter degenerate and redundant primitives. Our framework, instantiated on a multimodal foundation model, outperforms strong open-weight baselines on the standard MMSVGBench. This result highlights the remarkable data efficiency and generalization capability of our Render-in-the-Loop paradigm for both Text-to-SVG and Image-to-SVG tasks.