POET: Power-Oriented Evolutionary Tuning for LLM-Based RTL PPA Optimization

arXiv cs.AI / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • POET is a framework that applies large language models (LLMs) to RTL code optimization to improve power, performance, and area (PPA).
  • It tackles two key challenges: maintaining functional correctness despite LLM hallucination and systematically prioritizing power reduction within the PPA trade-off space.
  • For correctness, POET introduces a differential-testing-based testbench generation pipeline that uses the original design as a functional oracle and deterministic simulation to create golden references, removing LLM hallucination from verification.
  • For optimization, POET uses an LLM-driven evolutionary mechanism with non-dominated sorting, power-first intra-level ranking, and proportional survivor selection to steer the search toward low-power regions of the Pareto front without manual weight tuning.
  • Evaluated on the RTL-OPT benchmark across 40 RTL designs, POET achieves 100% functional correctness and the best power on all designs, with competitive area and delay improvements.

Abstract

Applying large language models (LLMs) to RTL code optimization for improved power, performance, and area (PPA) faces two key challenges: ensuring functional correctness of optimized designs despite LLM hallucination, and systematically prioritizing power reduction within the multi-objective PPA trade-off space. We propose POET (Power-Oriented Evolutionary Tuning), a framework that addresses both challenges. For functional correctness, POET introduces a differential-testing-based testbench generation pipeline that treats the original design as a functional oracle, using deterministic simulation to produce golden references and eliminating LLM hallucination from the verification process. For PPA optimization, POET employs an LLM-driven evolutionary mechanism with non-dominated sorting, power-first intra-level ranking, and proportional survivor selection to steer the search toward the low-power region of the Pareto front without manual weight tuning. Evaluated on the RTL-OPT benchmark across 40 diverse RTL designs, POET achieves 100% functional correctness, the best power on all 40 designs, and competitive area and delay improvements.