Understanding DNNs in Feature Interaction Models: A Dimensional Collapse Perspective

arXiv cs.LG / 4/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines why deep neural networks (DNNs) work—or fail to work—in feature interaction recommendation models, focusing on a “dimensional collapse” view of representation quality.
  • It contrasts common claims that DNNs implicitly learn high-order feature interactions with newer evidence that DNNs struggle to learn even second-order dot products reliably.
  • Through extensive experiments on parallel and stacked DNN variants, the authors evaluate DNN effectiveness across complete models and via detailed component-level ablations.
  • The results indicate that both parallel and stacked DNN architectures can reduce dimensional collapse in embeddings, improving robustness of learned representations.
  • A gradient-based theoretical analysis, corroborated by empirical findings, is used to explain the mechanisms driving dimensional collapse.

Abstract

DNNs have gained widespread adoption in feature interaction recommendation models. However, there has been a longstanding debate on their roles. On one hand, some works claim that DNNs possess the ability to implicitly capture high-order feature interactions. Conversely, recent studies have highlighted the limitations of DNNs in effectively learning dot products, specifically second-order interactions, let alone higher-order interactions. In this paper, we present a novel perspective to understand the effectiveness of DNNs: their impact on the dimensional robustness of the representations. In particular, we conduct extensive experiments involving both parallel DNNs and stacked DNNs. Our evaluation encompasses an overall study of complete DNN on two feature interaction models, alongside a fine-grained ablation analysis of components within DNNs. Experimental results demonstrate that both parallel and stacked DNNs can effectively mitigate the dimensional collapse of embeddings. Furthermore, a gradient-based theoretical analysis, supported by empirical evidence, uncovers the underlying mechanisms of dimensional collapse.