Modeling Co-Pilots for Text-to-Model Translation

arXiv cs.AI / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces two related resources—Text2Model (an LLM co-pilot suite with an online leaderboard) and Text2Zinc (a cross-domain dataset plus an interactive editor with an AI assistant) for translating natural language into formal combinatorial models.
  • Text2Model uses multiple LLM strategies at different complexity levels, including zero-shot prompting, chain-of-thought, intermediate representations via knowledge graphs, grammar-based syntax encoding, and agentic multi-step decomposition.
  • Unlike prior work that often targets solver-specific model formats, the approach is solver-agnostic by leveraging MiniZinc’s solver-and-paradigm-agnostic modeling capabilities.
  • The authors emphasize a unified architecture and dataset that integrate both satisfaction and optimization problems rather than treating them as separate translation pipelines.
  • Experimental results suggest the methods are competitive but still not “push-button” for combinatorial modeling, and the release of the co-pilots, leaderboard, dataset, and editor aims to close this performance gap.

Abstract

There is growing interest in leveraging large language models (LLMs) for text-to-model translation and optimization tasks. This paper aims to advance this line of research by introducing \textsc{Text2Model} and \textsc{Text2Zinc}. \textsc{Text2Model} is a suite of co-pilots based on several LLM strategies with varying complexity, along with an online leaderboard. \textsc{Text2Zinc} is a cross-domain dataset for capturing optimization and satisfaction problems specified in natural language, along with an interactive editor with built-in AI assistant. While there is an emerging literature on using LLMs for translating combinatorial problems into formal models, our work is the first attempt to integrate \textit{both} satisfaction and optimization problems within a \textit{unified architecture} and \textit{dataset}. Moreover, our approach is \textit{solver-agnostic} unlike existing work that focuses on translation to a solver-specific model. To achieve this, we leverage \textsc{MiniZinc}'s solver-and-paradigm-agnostic modeling capabilities to formulate combinatorial problems. We conduct comprehensive experiments to compare execution and solution accuracy across several single- and multi-call strategies, including; zero-shot prompting, chain-of-thought reasoning, intermediate representations via knowledge-graphs, grammar-based syntax encoding, and agentic approaches that decompose the model into sequential sub-tasks. Our co-pilot strategies are competitive, and in parts improve, recent research in this domain. Our findings indicate that while LLMs are promising they are not yet a push-button technology for combinatorial modeling. We contribute \textsc{Text2Model} co-pilots and leaderboard, and \textsc{Text2Zinc} and interactive editor to open-source to support closing this performance gap.