MMControl: Unified Multi-Modal Control for Joint Audio-Video Generation

arXiv cs.CV / 4/22/2026

📰 NewsModels & Research

Key Points

  • MMControl is a new framework for unified multi-modal control in joint audio-video generation, addressing the limitation of prior approaches that only supported video-only control.
  • It uses a dual-stream conditional injection mechanism to feed both visual and acoustic constraints (e.g., reference images, reference audio, depth maps, and pose sequences) into a joint audio-video Diffusion Transformer.
  • The method aims to produce identity-consistent video and timbre-consistent audio simultaneously while respecting structural constraints derived from the provided controls.
  • MMControl also adds modality-specific guidance scaling, letting users independently and dynamically adjust how strongly each visual or acoustic condition affects generation during inference.
  • Experiments reportedly show fine-grained, composable control over attributes such as character identity, voice timbre, body pose, and scene layout in synchronized audio-video generation.

Abstract

Recent advances in Diffusion Transformers (DiTs) have enabled high-quality joint audio-video generation, producing videos with synchronized audio within a single model. However, existing controllable generation frameworks are typically restricted to video-only control. This restricts comprehensive controllability and often leads to suboptimal cross-modal alignment. To bridge this gap, we present MMControl, which enables users to perform Multi-Modal Control in joint audio-video generation. MMControl introduces a dual-stream conditional injection mechanism. It incorporates both visual and acoustic control signals, including reference images, reference audio, depth maps, and pose sequences, into a joint generation process. These conditions are injected through bypass branches into a joint audio-video Diffusion Transformer, enabling the model to simultaneously generate identity-consistent video and timbre-consistent audio under structural constraints. Furthermore, we introduce modality-specific guidance scaling, which allows users to independently and dynamically adjust the influence strength of each visual and acoustic condition at inference time. Extensive experiments demonstrate that MMControl achieves fine-grained, composable control over character identity, voice timbre, body pose, and scene layout in joint audio-video generation.