Tiny-ViT: A Compact Vision Transformer for Efficient and Explainable Potato Leaf Disease Classification
arXiv cs.CV / 3/31/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes “Tiny-ViT,” a compact Vision Transformer designed for efficient and explainable potato leaf disease classification in resource-limited settings.
- It targets three classes—Early Blight, Late Blight, and Healthy leaves—using preprocessing steps including resizing, CLAHE, and Gaussian blur to improve image quality.
- Reported performance is extremely high, with test accuracy of 99.85% and mean cross-validation accuracy of 99.82%, outperforming baseline models such as DeiT Small, Swin Tiny, and MobileViT XS.
- The model shows strong reliability and generalization, indicated by a Matthews Correlation Coefficient (MCC) of 0.9990 and very narrow confidence intervals.
- Explainability is enhanced via GRAD-CAM, which highlights diseased regions, and the approach is positioned as suitable for real-time inference due to low computational cost.
Related Articles
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to

Building Real-Time AI Voice Agents with Google Gemini 3.1 Flash Live and VideoSDK
Dev.to

Your Knowledge, Your Model: A Method for Deterministic Knowledge Externalization
Dev.to