ViBE: Visual-to-M/EEG Brain Encoding via Spatio-Temporal VAE and Distribution-Aligned Projection
arXiv cs.CV / 4/30/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces ViBE, a new brain encoding framework that generates MEG/EEG signals from visual stimuli to support both neuroscience understanding and potential visual prosthesis applications.
- ViBE uses a spatio-temporal convolutional variational autoencoder (TSC-VAE) to reconstruct neural responses by learning the spatio-temporal structure of M/EEG signals.
- To align visual and neural modalities, the method employs Q-Former to map CLIP image embeddings into the TSC-VAE latent space as neural proxy embeddings.
- For cross-modal alignment, ViBE combines point-wise feature matching with MSE loss and distribution-level alignment using sliced Wasserstein distance (SWD).
- Experiments on the THINGS-EEG2 and THINGS-MEG datasets show that the approach can produce high-quality MEG/EEG signals from images.
Related Articles

Building a Local AI Agent (Part 2): Six UX and UI Design Challenges
Dev.to

We Built a DNS-Based Discovery Protocol for AI Agents — Here's How It Works
Dev.to

Your first business opportunity in 3 commands: /register_directory in @biznode_bot, wait for matches, then /my_pulse to view...
Dev.to

Building AI Evaluation Pipelines: Automating LLM Testing from Dataset to CI/CD
Dev.to

Function Calling Harness 2: CoT Compliance from 9.91% to 100%
Dev.to