Towards Verified and Targeted Explanations through Formal Methods
arXiv cs.LG / 4/17/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that current explainable AI (XAI) techniques often provide feature attributions without formal guarantees about how decision boundaries behave under perturbations.
- It highlights that safety-critical misclassifications have different real-world severities, motivating explanations that are targeted toward user-specified critical alternatives.
- The authors introduce ViTaX (Verified and Targeted Explanations), a formal XAI framework that produces targeted semifactual explanations and verifies them using mathematical reachability analysis.
- ViTaX identifies the smallest sensitive subset of features for the transition from an original class y to a user-chosen target class t, then guarantees that perturbing those features within epsilon cannot cause the prediction to flip to t.
- Experiments on MNIST, GTSRB, EMNIST, and TaxiNet show more than 30% improvement in fidelity while using minimal explanation cardinality.


![[2026] OpenTelemetry for LLM Observability — Self-Hosted Setup](/_next/image?url=https%3A%2F%2Fmedia2.dev.to%2Fdynamic%2Fimage%2Fwidth%3D1200%2Cheight%3D627%2Cfit%3Dcover%2Cgravity%3Dauto%2Cformat%3Dauto%2Fhttps%253A%252F%252Fdev-to-uploads.s3.amazonaws.com%252Fuploads%252Farticles%252Flu4b6ttuhur71z5gemm0.png&w=3840&q=75)
