CoInteract: Physically-Consistent Human-Object Interaction Video Synthesis via Spatially-Structured Co-Generation

arXiv cs.CV / 4/22/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • CoInteract introduces an end-to-end framework to synthesize human–object interaction (HOI) videos using conditioning on a person reference image, a product reference image, text prompts, and speech audio.
  • The paper targets two common failure modes of diffusion-based HOI video generation: unstable fine structures (hands/faces) and physically implausible contacts such as hand–object interpenetration.
  • It proposes a Human-Aware Mixture-of-Experts (MoE) with spatially supervised token routing to route image regions to specialized experts, improving structural fidelity without large parameter increases.
  • It also introduces Spatially-Structured Co-Generation, using a dual-stream training setup with an auxiliary HOI-structure stream that injects interaction-geometry priors while removing the HOI branch at inference for zero extra overhead.
  • Experiments report significant improvements over prior methods in structural stability, logical consistency, and interaction realism.

Abstract

Synthesizing human--object interaction (HOI) videos has broad practical value in e-commerce, digital advertising, and virtual marketing. However, current diffusion models, despite their photorealistic rendering capability, still frequently fail on (i) the structural stability of sensitive regions such as hands and faces and (ii) physically plausible contact (e.g., avoiding hand--object interpenetration). We present CoInteract, an end-to-end framework for HOI video synthesis conditioned on a person reference image, a product reference image, text prompts, and speech audio. CoInteract introduces two complementary designs embedded into a Diffusion Transformer (DiT) backbone. First, we propose a Human-Aware Mixture-of-Experts (MoE) that routes tokens to lightweight, region-specialized experts via spatially supervised routing, improving fine-grained structural fidelity with minimal parameter overhead. Second, we propose Spatially-Structured Co-Generation, a dual-stream training paradigm that jointly models an RGB appearance stream and an auxiliary HOI structure stream to inject interaction geometry priors. During training, the HOI stream attends to RGB tokens and its supervision regularizes shared backbone weights; at inference, the HOI branch is removed for zero-overhead RGB generation. Experimental results demonstrate that CoInteract significantly outperforms existing methods in structural stability, logical consistency, and interaction realism.