Vision Tiny Recursion Model (ViTRM): Parameter-Efficient Image Classification via Recursive State Refinement

arXiv cs.CV / 3/23/2026

📰 NewsModels & Research

Key Points

  • The Vision Tiny Recursion Model (ViTRM) replaces the L-layer ViT encoder with a tiny 3-layer block applied recursively a fixed number of times to perform iterative state refinement.
  • It achieves up to 6x fewer parameters than CNN-based models and up to 84x fewer than ViT, while maintaining competitive accuracy on CIFAR-10 and CIFAR-100.
  • The approach demonstrates that recursive computation can substitute for deep architectural stacks in vision tasks without sacrificing performance.
  • By enabling parameter-efficient vision models, ViTRM could broaden deployment in resource-constrained environments and influence future model design decisions.

Abstract

The success of deep learning in computer vision has been driven by models of increasing scale, from deep Convolutional Neural Networks (CNN) to large Vision Transformers (ViT). While effective, these architectures are parameter-intensive and demand significant computational resources, limiting deployment in resource-constrained environments. Inspired by Tiny Recursive Models (TRM), which show that small recursive networks can solve complex reasoning tasks through iterative state refinement, we introduce the \textbf{Vision Tiny Recursion Model (ViTRM)}: a parameter-efficient architecture that replaces the L-layer ViT encoder with a single tiny k-layer block (k{=}3) applied recursively N times. Despite using up to 6 \times and 84 \times fewer parameters than CNN based models and ViT respectively, ViTRM maintains competitive performance on CIFAR-10 and CIFAR-100. This demonstrates that recursive computation is a viable, parameter-efficient alternative to architectural depth in vision.