FSKD: Monocular Forest Structure Inference via LiDAR-to-RGBI Knowledge Distillation

arXiv cs.CV / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces FSKD, a LiDAR-to-RGBI knowledge-distillation framework that trains an RGBI-only SegFormer student to infer high-resolution forest structure metrics typically derived from costly airborne LiDAR.
  • Using a multimodal teacher that fuses RGBI imagery with LiDAR-derived planar metrics and vertical profiles via cross-attention, the student achieves state-of-the-art zero-shot canopy height model (CHM) performance on geographically distinct test tiles.
  • On 384 km² of Saxony, Germany data (20 cm GSD), the method reports strong improvements over HRCHM/DAC baselines (29–46% MAE reduction) and outputs CHM, PAI, and foliage height diversity (FHD) jointly, extending beyond many monocular CHM-only approaches.
  • Ablations indicate multimodal fusion boosts accuracy by 10–26% versus RGBI-only training, and that asymmetric distillation plus suitable model capacity are important for best results.
  • The approach tolerates temporal mismatch (e.g., winter LiDAR with summer RGBI), supporting more scalable operational monitoring for digital-twin and national orthophoto-style workflows, though PAI/FHD accuracy remains region-dependent and benefits from local calibration.

Abstract

Very High Resolution (VHR) forest structure data at individual-tree scale is essential for carbon, biodiversity, and ecosystem monitoring. Still, airborne LiDAR remains costly and infrequent despite being the reference for forest structure metrics like Canopy Height Model (CHM), Plant Area Index (PAI), and Foliage Height Diversity (FHD). We propose FSKD: a LiDAR-to-RGB-Infrared (RGBI) knowledge distillation (KD) framework in which a multi-modal teacher fuses RGBI imagery with LiDAR-derived planar metrics and vertical profiles via cross-attention, and an RGBI-only SegFormer student learns to reproduce these outputs. Trained on 384 km^2 of forests in Saxony, Germany (20 cm ground sampling distance (GSD)) and evaluated on eight geographically distinct test tiles, the student achieves state-of-the-art (SOTA) zero-shot CHM performance (MedAE 4.17 m, R^2=0.51, IoU 0.87), outperforming HRCHM/DAC baselines by 29--46% in MAE (5.81 m vs. 8.14--10.84 m) with stronger correlation coefficients (0.713 vs. 0.166--0.652). Ablations show that multi-modal fusion improves performance by 10--26% over RGBI-only training, and that asymmetric distillation with appropriate model capacity is critical. The method jointly predicts CHM, PAI, and FHD, a multi-metric capability not provided by current monocular CHM estimators, although PAI/FHD transfer remains region-dependent and benefits from local calibration. The framework also remains effective under temporal mismatch (winter LiDAR, summer RGBI), removing strict co-acquisition constraints and enabling scalable 20 cm operational monitoring for workflows such as Digital Twin Germany and national Digital Orthophoto programs.