GenFusion: Feed-forward Human Performance Capture via Progressive Canonical Space Updates

arXiv cs.CV / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces GenFusion, a feed-forward human performance capture method that synthesizes novel views from a single monocular RGB video stream.
  • It addresses missing observations in unseen body regions by maintaining a canonical space that is progressively updated frame-by-frame as the subject moves continuously.
  • The canonical space acts as a time-accumulated “context bank” to provide appearance information when the current frame lacks direct visibility.
  • Rendering is cast as probabilistic regression to better reconcile past (canonical/context) and current (live deformation) observations, yielding sharper results than deterministic regression.
  • Experiments on 4D-Dress (in-domain) and MVHumanNet (out-of-distribution) show improved reconstruction quality and plausible synthesis even where no prior observations exist.

Abstract

We present a feed-forward human performance capture method that renders novel views of a performer from a monocular RGB stream. A key challenge in this setting is the lack of sufficient observations, especially for unseen regions. Assuming the subject moves continuously over time, we take advantage of the fact that more body parts become observable by maintaining a canonical space that is progressively updated with each incoming frame. This canonical space accumulates appearance information over time and serves as a context bank when direct observations are missing in the current live frame. To effectively utilize this context while respecting the deformation of the live state, we formulate the rendering process as probabilistic regression. This resolves conflicts between past and current observations, producing sharper reconstructions than deterministic regression approaches. Furthermore, it enables plausible synthesis even in regions with no prior observations. Experiments on in-domain (4D-Dress) and out-of-distribution (MVHumanNet) datasets demonstrate the effectiveness of our approach.