AI Navigate

Resource-constrained Amazons chess decision framework integrating large language models and graph attention

arXiv cs.AI / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a lightweight hybrid framework for the Game of the Amazons that combines Graph Attention Autoencoder, Monte Carlo Tree Search, Stochastic Graph Genetic Algorithm, and GPT-4o-mini to operate in resource-constrained environments.
  • It leverages a Graph Attention Autoencoder to inform a multi-step Monte Carlo Tree Search, and uses a Stochastic Graph Genetic Algorithm to optimize evaluation signals while GPT-4o-mini generates synthetic training data.
  • The Graph Attention mechanism acts as a structural filter, denoising the LLM outputs to improve decision quality.
  • Experimental results on a 10x10 Amazons board show 15%-56% improvement in decision accuracy over baselines and competitive win rates at lower node counts, illustrating feasibility of evolving high-performance game AI from foundation models under tight resources.

Abstract

Artificial intelligence has advanced significantly through the development of intelligent game-playing systems, providing rigorous testbeds for decision-making, strategic planning, and adaptive learning. However, resource-constrained environments pose critical challenges, as conventional deep learning methods heavily rely on extensive datasets and computational resources. In this paper, we propose a lightweight hybrid framework for the Game of the Amazons, which explores the paradigm of weak-to-strong generalization by integrating the structural reasoning of graph-based learning with the generative capabilities of large language models. Specifically, we leverage a Graph Attention Autoencoder to inform a multi-step Monte Carlo Tree Search, utilize a Stochastic Graph Genetic Algorithm to optimize evaluation signals, and harness GPT-4o-mini to generate synthetic training data. Unlike traditional approaches that rely on expert demonstrations, our framework learns from noisy and imperfect supervision. We demonstrate that the Graph Attention mechanism effectively functions as a structural filter, denoising the LLM's outputs. Experiments on a 10\times10 Amazons board show that our hybrid approach not only achieves a 15\%--56\% improvement in decision accuracy over baselines but also significantly outperforms its teacher model (GPT-4o-mini), achieving a competitive win rate of 45.0\% at N=30 nodes and a decisive 66.5\% at only N=50 nodes. These results verify the feasibility of evolving specialized, high-performance game AI from general-purpose foundation models under stringent computational constraints.