Mamba Learns in Context: Structure-Aware Domain Generalization for Multi-Task Point Cloud Understanding

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes SADG, a structure-aware, Mamba-based in-context learning framework to improve multi-task domain generalization for point cloud understanding where naïve Transformer/Mamba transfer degrades performance.
  • It introduces structure-aware serialization (SAS) using centroid-based topology and geodesic curvature continuity to produce transformation-invariant sequences and reduce structural drift.
  • The method adds hierarchical domain-aware modeling (HDM) to stabilize cross-domain reasoning by consolidating intra-domain structure and fusing inter-domain relations.
  • For test-time adaptation without parameter updates, it proposes a lightweight spectral graph alignment (SGA) that shifts target features toward source prototypes while preserving structural properties.
  • The authors also release MP3DObject, a real-scan object dataset for evaluating multi-task domain generalization, and report consistent state-of-the-art improvements across reconstruction, denoising, and registration.

Abstract

While recent Transformer and Mamba architectures have advanced point cloud representation learning, they are typically developed for single-task or single-domain settings. Directly applying them to multi-task domain generalization (DG) leads to degraded performance. Transformers effectively model global dependencies but suffer from quadratic attention cost and lack explicit structural ordering, whereas Mamba offers linear-time recurrence yet often depends on coordinate-driven serialization, which is sensitive to viewpoint changes and missing regions, causing structural drift and unstable sequential modeling. In this paper, we propose Structure-Aware Domain Generalization (SADG), a Mamba-based In-Context Learning framework that preserves structural hierarchy across domains and tasks. We design structure-aware serialization (SAS) that generates transformation-invariant sequences using centroid-based topology and geodesic curvature continuity. We further devise hierarchical domain-aware modeling (HDM) that stabilizes cross-domain reasoning by consolidating intra-domain structure and fusing inter-domain relations. At test time, we introduce a lightweight spectral graph alignment (SGA) that shifts target features toward source prototypes in the spectral domain without updating model parameters, ensuring structure-preserving test-time feature shifting. In addition, we introduce MP3DObject, a real-scan object dataset for multi-task DG evaluation. Comprehensive experiments demonstrate that the proposed approach improves structural fidelity and consistently outperforms state-of-the-art methods across multiple tasks including reconstruction, denoising, and registration.