Feature-level Interaction Explanations in Multimodal Transformers
arXiv cs.LG / 3/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- Introduces Feature-level I2MoE (FL-I2MoE), a structured Mixture-of-Experts layer that operates on token/patch sequences from frozen pretrained encoders to separate unique, synergistic, and redundant evidence at the feature level.
- Proposes an expert-wise explanation pipeline combining attribution with top-K% masking to assess faithfulness, and introduces Monte Carlo interaction probes including the Shapley Interaction Index (SII) and a redundancy-gap score to quantify cross-modal interactions.
- Demonstrates on MMIMDb, ENRICO, and MMHS150K that FL-I2MoE yields more interaction-specific and concentrated importance patterns than a dense Transformer with the same encoders.
- Provides causal evidence that removing pairs ranked by SII or redundancy-gap degrades performance more than random masking under the same budget, suggesting the identified interactions are causally relevant.




