SparseGF: A Height-Aware Sparse Segmentation Framework with Context Compression for Robust Ground Filtering Across Urban to Natural Scenes

arXiv cs.CV / 4/24/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • SparseGF is a new height-aware sparse segmentation framework for robust ground filtering of airborne laser scanning (ALS) point clouds across both urban and natural scenes.
  • The method addresses two key limitations of prior deep-learning ground filtering: losing important context in large-scale processing and misclassifying tall objects due to classification-only optimization.
  • SparseGF combines a convex-mirror-inspired context compression module, a hybrid sparse voxel–point network, and a height-aware loss that enforces elevation priors to reduce random errors on tall structures.
  • Experiments on two large-scale ALS benchmarks show leading performance on complex urban scenes, competitive results on mixed terrains, and moderate (but not catastrophic) accuracy in dense, steep forested areas.
  • The authors position SparseGF as a step toward more truly cross-scene generalization, offering insights for future deep-learning ground filtering research for large-scale environmental monitoring.

Abstract

High-quality digital terrain models derived from airborne laser scanning (ALS) data are essential for a wide range of geospatial analyses, and their generation typically relies on robust ground filtering (GF) to separate point clouds across diverse landscapes into ground and non-ground parts. Although current deep-learning-based GF methods have demonstrated impressive performance, especially in specific challenging terrains, their cross-scene generalization remains limited by two persistent issues: the context-detail dilemma in large-scale processing due to limited computational resources, and the random misclassification of tall objects arising from classification-only optimization. To overcome these limitations, we propose SparseGF, a height-aware sparse segmentation framework enhanced with context compression. It is built upon three key innovations: (1) a convex-mirror-inspired context compression module that condenses expansive contexts into compact representations while preserving central details; (2) a hybrid sparse voxel-point network architecture that effectively interprets compressed representations while mitigating compression-induced geometric distortion; and (3) a height-aware loss function that explicitly enforces topographic elevation priors during training to suppress random misclassification of tall objects. Extensive evaluations on two large-scale ALS benchmark datasets demonstrate that SparseGF delivers robust GF across urban to natural terrains, achieving leading performance in complex urban scenes, competitive results on mixed terrains, and moderate yet non-catastrophic accuracy in densely forested steep areas. This work offers new insights into deep-learning-based GF research and encourages further exploration toward truly cross-scene generalization for large-scale environmental monitoring.