Contrastive Semantic Projection: Faithful Neuron Labeling with Contrastive Examples
arXiv cs.LG / 4/27/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- Neuron labeling methods that rely on highly activating examples can produce overly broad or misleading descriptions because they may latch onto incidental dominant visual factors.
- The paper extends contrastive explanation ideas to neuron-level labeling by generating candidate labels with vision-language models (VLMs) using contrastive image sets and then assigning labels via a CLIP-like pipeline.
- It proposes Contrastive Semantic Projection (CSP), integrating contrastive examples into the SemanticLens-style scoring and selection process to improve how labels are chosen.
- Experiments—including a melanoma detection case study—show that contrastive labeling increases both faithfulness to the neuron’s true semantics and semantic granularity compared with state-of-the-art approaches.
- The authors argue that contrastive examples are a simple, underused ingredient that can materially strengthen interpretability and analysis pipelines for deep networks.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to
We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to