NeuroGame Transformer: Gibbs-Inspired Attention Driven by Game Theory and Statistical Physics
arXiv cs.AI / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The NeuroGame Transformer introduces a dual perspective on attention by treating tokens as players in a cooperative game and as interacting spins in a Gibbs-based physical system.
- It uses Shapley values for global attribution and Banzhaf indices for local influence, combined via a learnable gate to form an external magnetic field that modulates attention.
- Pairwise interactions are captured by an Ising-like energy with attention weights emerging as marginal probabilities under a Gibbs distribution, computed efficiently via mean-field equations.
- To scale to long sequences, the method employs importance-weighted Monte Carlo estimators with Gibbs-distributed weights and provides theoretical convergence and a fairness-sensitivity trade-off controlled by an interpolation parameter.
- Experimental results on SNLI and MNLI-matched show strong performance, surpassing ALBERT-Base and remaining highly competitive with RoBERTa-Base, with code released on GitHub.
Related Articles
Is AI becoming a bubble, and could it end like the dot-com crash?
Reddit r/artificial

Externalizing State
Dev.to

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to
How to settle on a coding LLM ? What parameters to watch out for ?
Reddit r/LocalLLaMA