Do Not Imitate, Reinforce: Iterative Classification via Belief Refinement

arXiv cs.LG / 4/27/2026

📰 NewsModels & Research

Key Points

  • The paper argues that standard supervised classification rigidly imitates fixed ground-truth labels in a single pass, which limits compute usage and can produce overconfident outputs at evaluation time.
  • It introduces Reinforced Iterative Classification (RIC), an RL-based approach where a recurrent agent iteratively refines a class probability distribution across steps.
  • RIC uses a value function to estimate how much further improvement is possible, providing a natural stopping (halting) criterion and enabling an “anytime” classifier.
  • The authors show theoretically that the iterative RL formulation can recover the same optimal predictions as cross-entropy, and experimentally that it matches supervised accuracy while improving calibration and adaptively allocating computation.

Abstract

Standard supervised classification trains models to imitate the exact labels provided by a perfect oracle. This imitation happens in a single pass, restricting the model to a fixed compute budget even when inputs vary in complexity. Moreover, the rigid training objective forces the model to express absolute certainty on its training data, resulting in overconfident predictions during evaluation. We propose Reinforced Iterative Classification (RIC), which replaces the imitative objective with Reinforcement Learning (RL). RIC deploys a recurrent agent that iteratively updates a predictive distribution over classes, receiving reward for stepwise improvement in prediction quality. The value function provides a natural halting criterion by estimating the remaining scope for improvement. We prove that the iterative formulation recovers the same optimal predictions as cross-entropy while yielding an anytime classifier. On image classification benchmarks, RIC matches the accuracy of supervised baselines with improved calibration and learns to allocate computation adaptively across inputs.