UAU-Net: Uncertainty-aware Representation Learning and Evidential Classification for Facial Action Unit Detection
arXiv cs.CV / 4/24/2026
📰 NewsModels & Research
Key Points
- The paper argues that facial action unit (AU) detection is difficult due to AU-specific uncertainty that arises in both feature representation and decision-making, which many existing methods ignore by using deterministic representations.
- It introduces UAU-Net, which models uncertainty explicitly at two stages: CV-AFE, a conditional VAE feature extractor that learns probabilistic representations (means/variances) across multiple spatio-temporal scales and conditions on AU labels to capture inter-AU dependency uncertainty.
- For the decision stage, the authors propose AB-ENN, an asymmetric Beta evidential neural network that represents predictive uncertainty with Beta distributions and reduces overconfident outputs using an asymmetric loss designed for heavily imbalanced AU labels.
- Experiments on BP4D and DISFA demonstrate that UAU-Net delivers strong AU detection performance, and analyses suggest that uncertainty modeling at both stages improves robustness and reliability.
- Overall, the work provides a calibrated, uncertainty-aware alternative to point-estimation classifiers for multi-label AU detection under noisy, subject-varying, and label-imbalanced conditions.
Related Articles

GPT-5.5 is here. So is DeepSeek V4. And honestly, I am tired of version numbers.
Dev.to

I Built an AI Image Workflow with GPT Image 2.0 (+ Fixing Its Biggest Flaw)
Dev.to
Max-and-Omnis/Nemotron-3-Super-64B-A12B-Math-REAP-GGUF
Reddit r/LocalLLaMA

Building a Visual Infrastructure Layer: How We’re Solving the "Visual Trust Gap" for E-com
Dev.to
DeepSeek-V4 Runs on Huawei Ascend Chips at 85% Utilization — Here's What That Means for AI Infrastructure and Pricing
Dev.to