DOne: Decoupling Structure and Rendering for High-Fidelity Design-to-Code Generation
arXiv cs.CV / 4/3/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces DOne, an end-to-end design-to-code framework that decouples structure understanding from element rendering to avoid common layout distortions in vision-language approaches.
- DOne uses a learned layout segmentation module, a hybrid element retriever for UI components with extreme aspect ratios/densities, and a schema-guided generation approach to connect layout representation with code output.
- To better evaluate high-complexity UIs, the authors introduce HiFi2Code, a benchmark with significantly more layout complexity than prior datasets.
- Experiments on HiFi2Code show DOne improves both high-level visual similarity (including over 10% in GPT Score) and fine-grained element alignment versus existing methods.
- Human evaluations report a roughly 3× productivity gain while maintaining higher visual fidelity, indicating practical benefits beyond metric improvements.
Related Articles

Black Hat Asia
AI Business

90000 Tech Workers Got Fired This Year and Everyone Is Blaming AI but Thats Not the Whole Story
Dev.to

Microsoft’s $10 Billion Japan Bet Shows the Next AI Battleground Is National Infrastructure
Dev.to

TII Releases Falcon Perception: A 0.6B-Parameter Early-Fusion Transformer for Open-Vocabulary Grounding and Segmentation from Natural Language Prompts
MarkTechPost

Portable eye scanner powered by AI expands access to low-cost community screening
Reddit r/artificial