MedSAD-CLIP: Supervised CLIP with Token-Patch Cross-Attention for Medical Anomaly Detection and Segmentation
arXiv cs.CV / 3/19/2026
📰 NewsModels & Research
Key Points
- MedSAD-CLIP introduces a supervised adaptation of CLIP for medical anomaly detection and segmentation using Token-Patch Cross-Attention to improve lesion localization while preserving CLIP's generalization.
- The approach uses lightweight image adapters and learnable prompt tokens to efficiently tailor the pretrained CLIP encoder to the medical domain with a limited amount of labeled abnormal data.
- A Margin-based image-text Contrastive Loss is proposed to enhance discrimination between normal and abnormal representations at the global feature level.
- Experiments on four datasets (Brain, Retina, Lung, Breast) show superior pixel-level segmentation and image-level classification compared with state-of-the-art methods, with code to be released.
Related Articles
Self-Refining Agents in Spec-Driven Development
Dev.to

has anyone tried this? Flash-MoE: Running a 397B Parameter Model on a Laptop
Reddit r/LocalLLaMA

M2.7 open weights coming in ~2 weeks
Reddit r/LocalLLaMA

MiniMax M2.7 Will Be Open Weights
Reddit r/LocalLLaMA
Best open source coding models for claude code? LB?
Reddit r/LocalLLaMA