SSL-R1: Self-Supervised Visual Reinforcement Post-Training for Multimodal Large Language Models

arXiv cs.CV / 4/23/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SSL-R1, a self-supervised reinforcement learning (RL) post-training framework that generates verifiable rewards from images for multimodal LLMs (MLLMs).
  • It addresses limitations of existing RL with verifiable rewards (RLVR) that often depend on language-centric priors and costly manual annotations by avoiding human or external-model supervision.
  • SSL-R1 revisits visual self-supervised learning (SSL) and reformulates common SSL tasks into “verifiable visual puzzles” suitable for RL post-training.
  • Experiments report substantial gains for MLLMs on multimodal understanding and reasoning benchmarks, suggesting vision-centric SSL tasks can improve intrinsic visual reasoning.
  • The authors provide the project code and argue the approach offers reusable experience for designing self-supervised, scalable, verifiable rewards for RL.

Abstract

Reinforcement learning (RL) with verifiable rewards (RLVR) has demonstrated the great potential of enhancing the reasoning abilities in multimodal large language models (MLLMs). However, the reliance on language-centric priors and expensive manual annotations prevents MLLMs' intrinsic visual understanding and scalable reward designs. In this work, we introduce SSL-R1, a generic self-supervised RL framework that derives verifiable rewards directly from images. To this end, we revisit self-supervised learning (SSL) in visual domains and reformulate widely-used SSL tasks into a set of verifiable visual puzzles for RL post-training, requiring neither human nor external model supervision. Training MLLMs on these tasks substantially improves their performance on multimodal understanding and reasoning benchmarks, highlighting the potential of leveraging vision-centric self-supervised tasks for MLLM post-training. We think this work will provide useful experience in devising effective self-supervised verifiable rewards to enable RL at scale. Project page: https://github.com/Jiahao000/SSL-R1.