DIRCR: Dual-Inference Rule-Contrastive Reasoning for Solving RAVENs

arXiv cs.AI / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The DIRCR (Dual-Inference Rule-Contrastive Reasoning) model addresses visual reasoning failures where prior approaches overemphasize either global context or local row-wise relations while missing intermediate feature constraints.
  • DIRCR’s Dual-Inference Reasoning Module runs both a local row-wise analogical reasoning path and a global holistic inference path, then fuses them using a gated attention mechanism.
  • A Rule-Contrastive Learning Module uses pseudo-labels to generate positive/negative rule samples and applies contrastive learning to improve feature separability and encourage abstract, transferable rule learning.
  • Experiments on three RAVEN datasets show that DIRCR improves robustness and generalization, and the authors provide code via the linked GitHub repository.

Abstract

Abstract visual reasoning remains challenging as existing methods often prioritize either global context or local row-wise relations, failing to integrate both, and lack intermediate feature constraints, leading to incomplete rule capture and entangled representations. To address these issues, we propose the Dual-Inference Rule-Contrastive Reasoning (DIRCR) model. Its core component, the Dual-Inference Reasoning Module, combines a local path for row-wise analogical reasoning and a global path for holistic inference, integrated via a gated attention mechanism. Additionally, a Rule-Contrastive Learning Module introduces pseudo-labels to construct positive and negative rule samples, applying contrastive learning to enhance feature separability and promote abstract, transferable rule learning. Experimental results on three RAVEN datasets demonstrate that DIRCR significantly enhances reasoning robustness and generalization. Codes are available at https://github.com/csZack-Zhang/DIRCR.