Native-Domain Cross-Attention for Camera-LiDAR Extrinsic Calibration Under Large Initial Perturbations
arXiv cs.CV / 4/1/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses camera–LiDAR extrinsic calibration by improving cross-modal correspondence when the initial extrinsic guess is far from the ground truth.
- It proposes an extrinsic-aware cross-attention framework that matches image patches with LiDAR point groups in their native domains, avoiding 3D distortion from LiDAR-to-depth-map projection.
- The method injects extrinsic parameter hypotheses directly into the attention/correspondence modeling to maintain geometry-consistent fusion across modalities.
- Experiments on KITTI and nuScenes show consistent performance gains over prior state-of-the-art methods in both accuracy and robustness under large perturbations.
- The authors report calibration success rates of 88% on KITTI and 99% on nuScenes under large extrinsic perturbations and have open-sourced the code at the provided GitHub link.
Related Articles

Black Hat Asia
AI Business

Show HN: 1-Bit Bonsai, the First Commercially Viable 1-Bit LLMs
Dev.to

I Built an AI Agent That Can Write Its Own Tools When It Gets Stuck
Dev.to

How to Create AI Videos in 20 Minutes (3 Free Tools, Zero Experience)
Dev.to

Agent Self-Discovery: How AI Agents Find Their Own Wallets
Dev.to