GDPO-Listener: Expressive Interactive Head Generation via Auto-Regressive Flow Matching and Group reward-Decoupled Policy Optimization

arXiv cs.CV / 3/27/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents GDPO-Listener, a new framework for generating expressive 3D head motion in dyadic virtual human interactions, especially improving “listener” motion realism.
  • It uses an Auto-Regressive Flow Matching architecture to enable stable supervised learning for head-motion generation.
  • To address listener “regression-to-the-mean” and static-face collapse, the method applies Group reward-Decoupled Policy Optimization (GDPO) that separates reward normalization across FLAME parameter groups to encourage high-variance expressive motion.
  • The approach also supports explicit semantic text control, allowing customized responses aligned with provided text.
  • Experiments on Seamless Interaction and DualTalk datasets show improved performance over baselines in long-term kinematic variance, visual expressivity, and semantic controllability.

Abstract

Generating realistic 3D head motion for dyadic interactions is a significant challenge in virtual human synthesis. While recent methods achieve impressive results with speaking heads, they frequently suffer from the `Regression-to-the-Mean' problem in listener motions, collapsing into static faces, and lack the parameter space for complex nonverbal motions. In this paper, we propose GDPO-Listener, a novel framework that achieves highly expressive speaking and listening motion generation. First, we introduce an Auto-Regressive Flow Matching architecture enabling stable supervised learning. Second, to overcome kinematic stillness, we apply the Group reward-Decoupled Policy Optimization (GDPO). By isolating reward normalization across distinct FLAME parameter groups, GDPO explicitly incentivizes high variance expressive generations. Finally, we enable explicit semantic text control for customizable responses. Extensive evaluations across the Seamless Interaction and DualTalk datasets demonstrate superior performance compared to existing baselines on long-term kinematic variance, visual expressivity and semantic controllability.