Listening Deepfake Detection: A New Perspective Beyond Speaking-Centric Forgery Analysis

arXiv cs.CV / 4/15/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that most deepfake detection research targets speaking manipulations, but real interactive attacks may alternate between speaking and listening to increase realism and persuasiveness.
  • It introduces a new task, Listening Deepfake Detection (LDD), and presents ListenForge, the first dataset tailored to listening forgeries, built using five Listening Head Generation methods.
  • To detect listening-specific artifacts, the authors propose MANet, a Motion-aware and Audio-guided Network that models subtle motion inconsistencies in listener video while using speaker audio semantics for cross-modal fusion.
  • Experimental results show that existing speaking-centric deepfake detectors generalize poorly to listening scenarios, while MANet performs significantly better on ListenForge.
  • The dataset and code are released to support further multimodal forgery analysis in interactive communication settings.

Abstract

Existing deepfake detection research has primarily focused on scenarios where the manipulated subject is actively speaking, i.e., generating fabricated content by altering the speaker's appearance or voice. However, in realistic interaction settings, attackers often alternate between falsifying speaking and listening states to mislead their targets, thereby enhancing the realism and persuasiveness of the scenario. Although the detection of 'listening deepfakes' remains largely unexplored and is hindered by a scarcity of both datasets and methodologies, the relatively limited quality of synthesized listening reactions presents an excellent breakthrough opportunity for current deepfake detection efforts. In this paper, we present the task of Listening Deepfake Detection (LDD). We introduce ListenForge, the first dataset specifically designed for this task, constructed using five Listening Head Generation (LHG) methods. To address the distinctive characteristics of listening forgeries, we propose MANet, a Motion-aware and Audio-guided Network that captures subtle motion inconsistencies in listener videos while leveraging speaker's audio semantics to guide cross-modal fusion. Extensive experiments demonstrate that existing Speaking Deepfake Detection (SDD) models perform poorly in listening scenarios. In contrast, MANet achieves significantly superior performance on ListenForge. Our work highlights the necessity of rethinking deepfake detection beyond the traditional speaking-centric paradigm and opens new directions for multimodal forgery analysis in interactive communication settings. The dataset and code are available at https://anonymous.4open.science/r/LDD-B4CB.