Decompose, Structure, and Repair: A Neuro-Symbolic Framework for Autoformalization via Operator Trees

arXiv cs.LG / 4/22/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DSR (Decompose, Structure, and Repair), a neuro-symbolic framework for statement autoformalization that restructures the task into a modular pipeline rather than an end-to-end flat sequence approach.
  • DSR decomposes natural-language mathematical statements into logical components and represents them as structured operator trees to localize mistakes and refine only the problematic sub-trees.
  • It proposes PRIME, a newly created benchmark of 156 undergraduate and graduate theorems from canonical textbooks, annotated in Lean 4 by experts.
  • Experimental results claim DSR achieves new state-of-the-art performance, outperforming existing baselines under the same computational budgets.
  • The authors state that the datasets, model, and code will be released publicly soon.

Abstract

Statement autoformalization acts as a critical bridge between human mathematics and formal mathematics by translating natural language problems into formal language. While prior works have focused on data synthesis and diverse training paradigms to optimize end-to-end Large Language Models (LLMs), they typically treat formal code as flat sequences, neglecting the hierarchical logic inherent in mathematical statements. In this work, we introduce Decompose, Structure, and Repair (DSR), a neuro-symbolic framework that restructures autoformalization into a modular pipeline. DSR decomposes statements into logical components and maps them to structured operator trees, leveraging this topological blueprint to precisely localize and repair errors via sub-tree refinement. Furthermore, we introduce PRIME, a benchmark of 156 undergraduate and graduate-level theorems selected from canonical textbooks and expertly annotated in Lean 4. Experimental results demonstrate that DSR establishes a new state-of-the-art, consistently outperforming baselines under equivalent computational budgets. The datasets, model, and code will be released to the public soon.