Multimodal Deep Learning for Diabetic Foot Ulcer Staging Using Integrated RGB and Thermal Imaging
arXiv cs.CV / 3/31/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The study proposes multimodal deep learning for diabetic foot ulcer (DFU) staging by integrating simultaneously captured RGB and thermal images to improve early diagnosis and monitoring.
- A Raspberry Pi-based portable imaging system was built to collect hospital data, resulting in a labeled dataset of 1,205 expert-annotated samples across six DFU stages.
- Models were trained on three variants (RGB-only, thermal-only, and RGB+Thermal as a 4-channel input) using DenseNet121, EfficientNetV2, InceptionV3, ResNet50, and VGG16.
- Results indicate the combined RGB+Thermal approach outperforms single-modality training, with the best performance coming from VGG16 using RGB+Thermal (accuracy 93.25%, F1 92.53%, MCC 91.03%).
- Grad-CAM visualizations suggest the thermal channel helps localize ulcer-related temperature anomalies while the RGB channel provides complementary structural and texture cues.
Related Articles
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to

Building Real-Time AI Voice Agents with Google Gemini 3.1 Flash Live and VideoSDK
Dev.to

Your Knowledge, Your Model: A Method for Deterministic Knowledge Externalization
Dev.to