Independent-Component-Based Encoding Models of Brain Activity During Story Comprehension

arXiv cs.CL / 4/29/2026

📰 NewsModels & Research

Key Points

  • The study introduces an independent-component (IC)-based encoding model for fMRI that separates stimulus-driven neural signals from noise-driven and artifact-related signals.
  • The method decomposes continuous fMRI data from naturalistic story listening into ICs, then trains encoding models to predict IC time series from large language model (LLM) representations of linguistic input.
  • Results show that a subset of ICs demonstrates consistently high predictivity across subjects, with spatial and temporal consistency and involvement of cognitive networks associated with story listening (auditory and language).
  • The authors report that key auditory components correlate strongly with acoustic features, improving interpretability, while components identified as noise or motion artifacts (via ICA-AROMA) yield poor predictive performance.
  • Overall, the approach enables functional-network-level analysis that accounts for cross-individual variability in network locations while producing interpretable, comparable results across subjects.

Abstract

Encoding models provide a powerful framework for linking continuous stimulus features to neural activity; however, traditional voxelwise approaches are limited by measurement noise, inter-subject variability, and redundancy arising from spatially correlated voxels encoding overlapping neural signals. Here, we propose an independent component (IC)-based encoding framework that dissociates stimulus-driven and noise-driven signals in fMRI data. We decompose continuous fMRI data from naturalistic story listening into ICs using one subset of the data, and train encoding models on independent data to predict IC time series from large language model representations of linguistic input. Across subjects, a subset of ICs exhibited consistently high predictivity. These ICs were spatially and temporally consistent across subjects and included cognitive networks known to respond during story listening (auditory and language). Auditory component time series were strongly correlated with acoustic stimulus features, highlighting the interpretability of identified component time series. Components identified as noise or motion-related artifacts by ICA-AROMA showed uniformly poor predictive performance, confirming that highly predicted components reflect genuine stimulus-related neural signals rather than confounds. Overall, IC-based encoding models enable analyses at the level of functional networks, accommodating the variability in network locations across individuals and providing interpretable results that are easy to compare across subjects.