Dropout Robustness and Cognitive Profiling of Transformer Models via Stochastic Inference
arXiv cs.LG / 3/19/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper analyzes dropout-induced variability across 19 transformer models using Monte Carlo Dropout with 100 stochastic forward passes per sample to evaluate inference-time robustness.
- It defines dropout robustness as maintaining high accuracy and stable predictions, quantifying stability with the standard deviation of per-run accuracies and a cognitive decomposition into memory and reasoning components.
- In experiments across five dropout configurations, the study performs 95 unique evaluations on 1,000 samples, revealing substantial architectural variation in robustness that is not simply tied to model size.
- Findings show smaller models have highly stable predictions, mid-sized models achieve the best overall accuracy, while larger models excel at memory tasks; importantly, 53% of models suffer severe accuracy degradation under baseline MC Dropout, with task-specific models losing up to 24 percentage points.
- Memory tasks are disproportionately affected by dropout (memory accuracy decreases by 27 percentage points) whereas reasoning loses only 1 percentage point, and 84% of models display memory-biased performance, making this the first comprehensive MC Dropout benchmark for transformers and offering guidance for uncertainty-aware applications.
Related Articles

Attacks On Data Centers, Qwen3.5 In All Sizes, DeepSeek’s Huawei Play, Apple’s Multimodal Tokenizer
The Batch

Your AI generated code is "almost right", and that is actually WORSE than it being "wrong".
Dev.to

Lessons from Academic Plagiarism Tools for SaaS Product Development
Dev.to

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

KI in der amtlichen Recherche beim DPMA: Was Patentanwälte bei Neuanmeldungen jetzt beachten sollten (Stand: März 2026)
Dev.to