DouC: Dual-Branch CLIP for Training-Free Open-Vocabulary Segmentation
arXiv cs.CV / 4/29/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- DouC is a training-free, dual-branch CLIP framework for open-vocabulary semantic segmentation that aims to improve both token reliability and spatial coherence.
- It uses OG-CLIP for patch-level reliability through lightweight inference-time token gating, addressing uncertainty in local tokens.
- It uses FADE-CLIP to inject external structural priors via proxy attention guided by frozen vision foundation models, improving structure-aware interactions.
- The two branches are fused at the logit level and can optionally apply instance-aware correction in post-processing to refine predictions.
- Experiments on eight benchmarks and multiple CLIP backbones show DouC consistently outperforms prior training-free methods and scales with backbone capacity without adding learnable parameters or retraining.
Related Articles
LLMs will be a commodity
Reddit r/artificial

Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA

Dex lands $5.3M to grow its AI-driven talent matching platform
Tech.eu

AI Citation Registry: Why Daily Updates Leave No Time for Data Structuring
Dev.to