Dual-Cluster Memory Agent: Resolving Multi-Paradigm Ambiguity in Optimization Problem Solving

arXiv cs.CL / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • LLMs can struggle with structural ambiguity in optimization tasks when multiple modeling paradigms fit the same problem, leading to conflicting solution generation strategies.
  • The paper introduces the Dual-Cluster Memory Agent (DCM-Agent), which builds training-free memory by clustering historical solutions into modeling and coding groups and distilling them into structured knowledge (Approach, Checklist, Pitfall).
  • Using memory-augmented inference, DCM-Agent dynamically selects and switches reasoning paths, detects and repairs errors, and adapts solution trajectories based on the stored guidance.
  • Experiments on seven optimization benchmarks show an average performance gain of 11%–21% compared with baselines.
  • The authors find a “knowledge inheritance” effect where memory created by larger models can help smaller models perform better, suggesting scalability and efficiency of the framework.

Abstract

Large Language Models (LLMs) often struggle with structural ambiguity in optimization problems, where a single problem admits multiple related but conflicting modeling paradigms, hindering effective solution generation. To address this, we propose Dual-Cluster Memory Agent (DCM-Agent) to enhance performance by leveraging historical solutions in a training-free manner. Central to this is Dual-Cluster Memory Construction. This agent assigns historical solutions to modeling and coding clusters, then distills each cluster's content into three structured types: Approach, Checklist, and Pitfall. This process derives generalizable guidance knowledge. Furthermore, this agent introduces Memory-augmented Inference to dynamically navigate solution paths, detect and repair errors, and adaptively switch reasoning paths with structured knowledge. The experiments across seven optimization benchmarks demonstrate that DCM-Agent achieves an average performance improvement of 11%- 21%. Notably, our analysis reveals a ``knowledge inheritance'' phenomenon: memory constructed by larger models can guide smaller models toward superior performance, highlighting the framework's scalability and efficiency.