Explainable Semantic Textual Similarity via Dissimilar Span Detection
arXiv cs.CL / 3/24/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes Dissimilar Span Detection (DSD) to make Semantic Textual Similarity (STS) more interpretable by locating specific spans that reduce similarity rather than only producing a single overall score.
- It releases a new dataset, the Span Similarity Dataset (SSD), created via a semi-automated pipeline that combines LLM-generated annotations with human verification.
- The authors evaluate multiple baseline approaches for DSD, including unsupervised methods using LIME/SHAP and LLM-based techniques, as well as a supervised model that performs best among the tested baselines.
- Despite improved performance from LLMs and supervised models, overall accuracy remains low, indicating the task’s inherent difficulty and the challenge of reliable negative-span attribution.
- An additional experiment suggests that using DSD signals can improve paraphrase detection performance in related downstream tasks.
Related Articles

Black Hat Asia
AI Business

"The Agent Didn't Decide Wrong. The Instructions Were Conflicting — and Nobody Noticed."
Dev.to
Top 5 LLM Gateway Alternatives After the LiteLLM Supply Chain Attack
Dev.to

Stop Counting Prompts — Start Reflecting on AI Fluency
Dev.to

Reliable Function Calling in Deeply Recursive Union Types: Fixing Qwen Models' Double-Stringify Bug
Dev.to