SIMON: Saliency-aware Integrative Multi-view Object-centric Neural Decoding

arXiv cs.CV / 5/4/2026

📰 NewsSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The paper introduces SIMON, a saliency-aware multi-view framework for zero-shot EEG-to-image retrieval that addresses the common center-bias assumption in prior work.
  • SIMON uses foreground segmentation and saliency prediction to choose fixation centers via Saliency-Aware Sampling (SAS), then generates foveated views that highlight informative object regions and reduce background noise.
  • On the THINGS-EEG dataset, SIMON achieves state-of-the-art results for both intra-subject and inter-subject retrieval, with average Top-1 accuracy of 69.7% and 19.6% respectively.
  • The authors report robustness through ablations and analyses across sampling granularity, EEG channel topology, and different visual/brain encoder backbones.
  • The research provides publicly available code and models via the linked GitHub repository.

Abstract

Recent EEG-to-image retrieval methods leverage pretrained vision encoders and foveation-inspired priors, but typically assume a fixed, center-focused view. This center bias conflicts with content-driven human attention, creating a geometric-semantic dissociation between visual features and EEG responses. We propose SIMON, a saliency-aware multi-view framework for zero-shot EEG-to-image retrieval. SIMON combines foreground segmentation and saliency prediction to select fixation centers via Saliency-Aware Sampling (SAS), then generates foveated views that emphasize informative object regions while suppressing background clutter. On THINGS-EEG, SIMON achieves state-of-the-art performance in both intra-subject and inter-subject settings, reaching an average Top-1 accuracy of 69.7% and 19.6%, respectively, consistently outperforming recent competitive baselines. Analyses across sampling granularity, EEG channel topology, and visual/brain encoder backbones further support the robustness of saliency-aware multi-view integration. Our code and models are publicly available at https://github.com/simonlink666/SIMON.