AI Navigate

Wrong Code, Right Structure: Learning Netlist Representations from Imperfect LLM-Generated RTL

arXiv cs.LG / 3/11/2026

Ideas & Deep AnalysisModels & Research

Key Points

  • The scarcity of labeled netlist datasets due to intellectual property protection limits learning effective netlist representations for realistic circuit designs.
  • Large Language Models (LLMs) can generate RTL code at scale but often produce functionally incorrect designs, hindering their direct use for circuit analysis.
  • The authors observe that despite functional errors, LLM-generated RTL preserves structural patterns useful for understanding netlists.
  • They propose a data augmentation and training framework that leverages imperfect LLM-generated RTL to improve netlist representation learning, enabling an end-to-end pipeline from automated code generation to downstream circuit understanding tasks.
  • Evaluations show that models trained on noisy, synthetic data generalize well to real-world netlists, outperforming those trained on limited high-quality datasets and overcoming the data bottleneck in circuit representation learning.

Computer Science > Machine Learning

arXiv:2603.09161 (cs)
[Submitted on 10 Mar 2026]

Title:Wrong Code, Right Structure: Learning Netlist Representations from Imperfect LLM-Generated RTL

View a PDF of the paper titled Wrong Code, Right Structure: Learning Netlist Representations from Imperfect LLM-Generated RTL, by Siyang Cai and 3 other authors
View PDF HTML (experimental)
Abstract:Learning effective netlist representations is fundamentally constrained by the scarcity of labeled datasets, as real designs are protected by Intellectual Property (IP) and costly to annotate. Existing work therefore focuses on small-scale circuits with clean labels, limiting scalability to realistic designs. Meanwhile, Large Language Models (LLMs) can generate Register-Transfer-Level (RTL) at scale, but their functional incorrectness has hindered their use in circuit analysis. In this work, we make a key observation: even when LLM-Generated RTL is functionally imperfect, the synthesized netlists still preserve structural patterns that are strongly indicative of the intended functionality. Building on this insight, we propose a cost-effective data augmentation and training framework that systematically exploits imperfect LLM-Generated RTL as training data for netlist representation learning, forming an end-to-end pipeline from automated code generation to downstream tasks. We conduct evaluations on circuit functional understanding tasks, including sub-circuit boundary identification and component classification, across benchmarks of increasing scales, extending the task scope from operator-level to IP-level. The evaluations demonstrate that models trained on our noisy synthetic corpus generalize well to real-world netlists, matching or even surpassing methods trained on scarce high-quality data and effectively breaking the data bottleneck in circuit representation learning.
Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Hardware Architecture (cs.AR)
Cite as: arXiv:2603.09161 [cs.LG]
  (or arXiv:2603.09161v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2603.09161
Focus to learn more
arXiv-issued DOI via DataCite

Submission history

From: Siyang Cai [view email]
[v1] Tue, 10 Mar 2026 03:52:41 UTC (1,804 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled Wrong Code, Right Structure: Learning Netlist Representations from Imperfect LLM-Generated RTL, by Siyang Cai and 3 other authors
  • View PDF
  • HTML (experimental)
  • TeX Source
Current browse context:
cs.LG
< prev   |   next >
Change to browse by:

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
IArxiv recommender toggle
IArxiv Recommender (What is IArxiv?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.