Do All Vision Transformers Need Registers? A Cross-Architectural Reassessment

arXiv cs.LG / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses training artifacts in Vision Transformer (ViT) attention maps and how these artifacts affect interpretability.
  • It reproduces prior work that proposes adding empty “register” tokens to store global information beyond the [CLS] token, showing that the approach can improve attention-map clarity.
  • The authors reassess generalizability across several vision transformer families (including DINO, DINOv2, OpenCLIP, and DeiT3) and find that some earlier claims are not universal.
  • They investigate how model size changes the findings, extending the discussion to smaller models.
  • The work also resolves terminology inconsistencies from the original paper and explains how those differences can mislead cross-model comparisons.

Abstract

Training Vision Transformers (ViTs) presents significant challenges, one of which is the emergence of artifacts in attention maps, hindering their interpretability. Darcet et al. (2024) investigated this phenomenon and attributed it to the need of ViTs to store global information beyond the [CLS] token. They proposed a novel solution involving the addition of empty input tokens, named registers, which successfully eliminate artifacts and improve the clarity of attention maps. In this work, we reproduce the findings of Darcet et al. (2024) and evaluate the generalizability of their claims across multiple models, including DINO, DINOv2, OpenCLIP, and DeiT3. While we confirm the validity of several of their key claims, our results reveal that some claims do not extend universally to other models. Additionally, we explore the impact of model size, extending their findings to smaller models. Finally, we untie terminology inconsistencies found in the original paper and explain their impact when generalizing to a wider range of models.