Structural Graph Probing of Vision-Language Models
arXiv cs.CV / 3/31/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper analyzes vision-language models by treating each layer as a within-layer correlation graph built from neuron co-activations, enabling a “neural topology” view of computation.
- It finds that correlation topology contains recoverable behavioral signal and varies systematically across modalities and network depth.
- The study shows cross-modal structure becomes more consolidated with depth around a compact set of recurrent hub neurons.
- Targeted perturbation of these hub neurons leads to substantial changes in model outputs, suggesting the identified components are causally influential.
- The authors position neural topology as a useful interpretability intermediate—more informative than local attribution while being more tractable than full circuit-level recovery—and release code publicly.



