AI Navigate

FrescoDiffusion: 4K Image-to-Video with Prior-Regularized Tiled Diffusion

arXiv cs.CV / 3/19/2026

📰 NewsModels & Research

Key Points

  • FrescoDiffusion introduces a training-free method for coherent large-format image-to-video generation from a single complex image, targeting 4K resolutions.
  • The method augments tiled denoising with a precomputed latent prior by first generating a low-resolution video to obtain a global reference that captures long-range temporal and spatial structure.
  • For 4K generation, per-tile noise predictions are fused with the latent reference at every diffusion timestep using a closed-form least-squares fusion that preserves global coherence while retaining detail.
  • Experiments on the VBench-I2V dataset and a fresco I2V dataset show improved global consistency and fidelity over tiled baselines while remaining computationally efficient.
  • A spatial regularization variable enables region-level control over motion, allowing explicit trade-offs between creativity and consistency.

Abstract

Diffusion-based image-to-video (I2V) models are increasingly effective, yet they struggle to scale to ultra-high-resolution inputs (e.g., 4K). Generating videos at the model's native resolution often loses fine-grained structure, whereas high-resolution tiled denoising preserves local detail but breaks global layout consistency. This failure mode is particularly severe in the fresco animation setting: monumental artworks containing many distinct characters, objects, and semantically different sub-scenes that must remain spatially coherent over time. We introduce FrescoDiffusion, a training-free method for coherent large-format I2V generation from a single complex image. The key idea is to augment tiled denoising with a precomputed latent prior: we first generate a low-resolution video at the underlying model resolution and upsample its latent trajectory to obtain a global reference that captures long-range temporal and spatial structure. For 4K generation, we compute per-tile noise predictions and fuse them with this reference at every diffusion timestep by minimizing a single weighted least-squares objective in model-output space. The objective combines a standard tile-merging criterion with our regularization term, yielding a closed-form fusion update that strengthens global coherence while retaining fine detail. We additionally provide a spatial regularization variable that enables region-level control over where motion is allowed. Experiments on the VBench-I2V dataset and our proposed fresco I2V dataset show improved global consistency and fidelity over tiled baselines, while being computationally efficient. Our regularization enables explicit controllability of the trade-off between creativity and consistency.