AI Navigate

Self-Aware Markov Models for Discrete Reasoning

arXiv cs.LG / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a self-aware Markov model for discrete reasoning by learning a Markov transition kernel that can remask tokens and correct previous mistakes using its own outputs.
  • It removes dependence on a fixed denoising schedule by introducing a trained stopping criterion that adapts the number of function evaluations to problem difficulty.
  • Two lightweight prediction heads are added to enable reuse and fine-tuning of existing pretrained models, facilitating efficient adaptation.
  • Empirical results on Sudoku-Extreme (95% validity) and Countdown-4 (about 10 steps to solve ~96%) show the method outperforms prior flow-based approaches.

Abstract

Standard masked discrete diffusion models face limitations in reasoning tasks due to their inability to correct their own mistakes on the masking path. Since they rely on a fixed number of denoising steps, they are unable to adjust their computation to the complexity of a given problem. To address these limitations, we introduce a method based on learning a Markov transition kernel that is trained on its own outputs. This design enables tokens to be remasked, allowing the model to correct its previous mistakes. Furthermore, we do not need a fixed time schedule but use a trained stopping criterion. This allows for adaptation of the number of function evaluations to the difficulty of the reasoning problem. Our adaptation adds two lightweight prediction heads, enabling reuse and fine-tuning of existing pretrained models. On the Sudoku-Extreme dataset we clearly outperform other flow based methods with a validity of 95%. For the Countdown-4 we only need in average of 10 steps to solve almost 96% of them correctly, while many problems can be solved already in 2 steps.