PLAF: Pixel-wise Language-Aligned Feature Extraction for Efficient 3D Scene Understanding

arXiv cs.CV / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes PLAF, a pixel-wise language-aligned feature extraction framework aimed at improving open-vocabulary 3D scene understanding with both spatial precision and language alignment.
  • It tackles the redundancy problem of propagating dense pixel-level semantics into 3D, which can make large-scale storage and querying inefficient.
  • PLAF performs dense and accurate semantic alignment in 2D while preserving open-vocabulary expressiveness, then extends the representation to support efficient semantic storage and querying across 2D and 3D.
  • The authors report experimental results indicating that PLAF offers an effective and efficient semantic foundation for accurate open-vocabulary 3D scene understanding, with code released on GitHub.

Abstract

Accurate open-vocabulary 3D scene understanding requires semantic representations that are both language-aligned and spatially precise at the pixel level, while remaining scalable when lifted to 3D space. However, existing representations struggle to jointly satisfy these requirements, and densely propagating pixel-wise semantics to 3D often results in substantial redundancy, leading to inefficient storage and querying in large-scale scenes. To address these challenges, we present \emph{PLAF}, a Pixel-wise Language-Aligned Feature extraction framework that enables dense and accurate semantic alignment in 2D without sacrificing open-vocabulary expressiveness. Building upon this representation, we further design an efficient semantic storage and querying scheme that significantly reduces redundancy across both 2D and 3D domains. Experimental results show that \emph{PLAF} provides a strong semantic foundation for accurate and efficient open-vocabulary 3D scene understanding. The codes are publicly available at https://github.com/RockWenJJ/PLAF.