Semantic Segmentation of Textured Non-manifold 3D Meshes using Transformers
arXiv cs.CV / 4/3/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a texture-aware transformer for semantic segmentation on textured non-manifold 3D meshes, addressing the difficulty of irregular mesh structure while leveraging texture information from raw face-associated pixels.
- It introduces a hierarchical multi-scale feature aggregation scheme that combines a texture branch (pixel aggregation into a learnable token) with geometric descriptors processed through Two-Stage Transformer Blocks to balance local and global context.
- Experiments on the Semantic Urban Meshes (SUM) benchmark show strong results (81.9% mF1, 94.3% OA), with additional evaluation on a newly curated cultural-heritage roof-tile dataset (49.7% mF1, 72.8% OA).
- The method significantly outperforms existing approaches, indicating that jointly modeling texture plus geometry in transformer architectures can improve per-face semantic/damage-type predictions for complex meshes.
Related Articles

Why I built an AI assistant that doesn't know who you are
Dev.to

DenseNet Paper Walkthrough: All Connected
Towards Data Science

Meta Adaptive Ranking Model: What Instagram Advertisers Gain in 2026 | MKDM
Dev.to

The Facebook insider building content moderation for the AI era
TechCrunch
Qwen3.5 vs Gemma 4: Benchmarks vs real world use?
Reddit r/LocalLLaMA