LiveMoments: Reselected Key Photo Restoration in Live Photos via Reference-guided Diffusion

arXiv cs.CV / 4/15/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Live Photos store a high-quality key photo plus a short video, but when users reselect an alternative frame as the key photo, visible quality loss can occur because the photo ISP pipeline is better than the video pipeline.
  • The paper proposes LiveMoments, a reference-guided restoration framework that uses the original high-quality key photo to recover the quality of the reselected frame.
  • LiveMoments uses a two-branch neural network, where a reference branch extracts structural/textural cues from the original key photo and a main branch restores the reselected frame guided by those cues.
  • A unified Motion Alignment module provides motion guidance for spatial alignment at both latent and image levels, helping particularly in fast-motion or complex-structure scenes.
  • Experiments on real and synthetic Live Photos show improved perceptual quality and fidelity versus existing methods, with code released on GitHub.

Abstract

Live Photo captures both a high-quality key photo and a short video clip to preserve the precious dynamics around the captured moment. While users may choose alternative frames as the key photo to capture better expressions or timing, these frames often exhibit noticeable quality degradation, as the photo capture ISP pipeline delivers significantly higher image quality than the video pipeline. This quality gap highlights the need for dedicated restoration techniques to enhance the reselected key photo. To this end, we propose LiveMoments, a reference-guided image restoration framework tailored for the reselected key photo in Live Photos. Our method employs a two-branch neural network: a reference branch that extracts structural and textural information from the original high-quality key photo, and a main branch that restores the reselected frame using the guidance provided by the reference branch. Furthermore, we introduce a unified Motion Alignment module that incorporates motion guidance for spatial alignment at both the latent and image levels. Experiments on real and synthetic Live Photos demonstrate that LiveMoments significantly improves perceptual quality and fidelity over existing solutions, especially in scenes with fast motion or complex structures. Our code is available at https://github.com/OpenVeraTeam/LiveMoments.