UniSem: Generalizable Semantic 3D Reconstruction from Sparse Unposed Images
arXiv cs.CV / 3/19/2026
📰 NewsModels & Research
Key Points
- UniSem introduces a unified framework for semantic-aware 3D reconstruction from sparse, unposed images, addressing instability and incomplete 3D semantics in prior 3D Gaussian Splatting methods.
- It adds Error-aware Gaussian Dropout (EGD) to suppress redundant Gaussian primitives based on rendering error cues, yielding more stable geometry and improved depth estimation.
- It also proposes Mix-training Curriculum (MTC) to blend 2D segmenter-lifted semantics with emergent 3D semantic priors through object-level prototype alignment, boosting semantic coherence.
- Experiments on ScanNet and Replica show strong depth and open-vocabulary 3D segmentation gains, including a 15.2% reduction in depth error and a 3.7% gain in mAcc with 16 views.
Related Articles
[D] Matryoshka Representation Learning
Reddit r/MachineLearning
Two new Qwen3.5 “Neo” fine‑tunes focused on fast, efficient reasoning
Reddit r/LocalLLaMA

HKIC, Gobi Partners and HKU team up for fund backing university research start-ups
SCMP Tech
Yann LeCun’s New LeWorldModel (LeWM) Research Targets JEPA Collapse in Pixel-Based Predictive World Modeling
MarkTechPost
Streaming experts
Simon Willison's Blog