Two-View Accumulation as the Primary Training Lever for Hybrid-Capture Gaussian Splatting: A Variance-Decomposition View of When Gradient Surgery Helps
arXiv cs.CV / 5/4/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- Hybrid-capture novel view synthesis with 3D Gaussian Splatting (3DGS) under-trains the minority camera-distance regime, losing 1–3 dB PSNR on five benchmarks compared with a standard 30K/one-view-per-step training setup.
- Among several compute-matched training modifications, the key factor that closes the gap is a simple structural change: rendering two views per optimizer step, which improves PSNR while other sophisticated gradient-surgery or pairing strategies do not.
- Experiments show that the specific pairing rule (geometry-defined near/far, random, or active loss-disparity) does not materially affect PSNR beyond randomness across scenes, whereas the two-view accumulation effect consistently matters.
- The paper introduces a variance-decomposition framework arguing that, in bimodal camera-distance regimes, between-regime gradient variance is small relative to within-regime variance in 3DGS, making variance-reduction from two-view accumulation the dominant benefit.
- The findings generalize to Scaffold-GS and Pixel-GS backbones and are presented as a clear characterization of which training-side axes change PSNR (and which do not) for hybrid-capture 3DGS.
Related Articles
A very basic litmus test for LLMs "ok give me a python program that reads my c: and put names and folders in a sorted list from biggest to small"
Reddit r/LocalLLaMA

ALM on Power Platform: ADO + GitHub, the best of both worlds
Dev.to

Experiment: Does repeated usage influence ChatGPT 5.4 outputs in a RAG-like setup?
Dev.to

Find 12 high-volume, low-competition GEO content topics Topify.ai should rank on
Dev.to

When a memorized rule fits your bug too well: a meta-trap of agent workflows
Dev.to