AI Navigate

Multimodal Emotion Regression with Multi-Objective Optimization and VAD-Aware Audio Modeling for the 10th ABAW EMI Track

arXiv cs.AI / 3/17/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The article describes a multimodal emotion regression approach for the EMI estimation track of the 10th ABAW Challenge using the Hume-Vidmimic2 dataset.
  • It finds that, under their pretrained features, direct feature concatenation outperforms more complex fusion strategies, guiding their design choice.
  • The proposed framework combines concatenation-based fusion, a shared six-dimensional regression head, multi-objective optimization (MSE, Pearson, auxiliary supervision), EMA stabilization, and a VAD-inspired latent prior for the acoustic branch.
  • It reports a best mean Pearson Correlation Coefficient of 0.478567 on the official validation set.

Abstract

We participated in the 10th ABAW Challenge, focusing on the Emotional Mimicry Intensity (EMI) Estimation track on the Hume-Vidmimic2 dataset. This task aims to predict six continuous emotion dimensions: Admiration, Amusement, Determination, Empathic Pain, Excitement, and Joy. Through systematic multimodal exploration of pretrained high-level features, we found that, under our pretrained feature setting, direct feature concatenation outperformed the more complex fusion strategies we tested. This empirical finding motivated us to design a systematic approach built upon three core principles: (i) preserving modality-specific attributes through feature-level concatenation; (ii) improving training stability and metric alignment via multi-objective optimization; and (iii) enriching acoustic representations with a VAD-inspired latent prior. Our final framework integrates concatenation-based multimodal fusion, a shared six-dimensional regression head, multi-objective optimization with MSE, Pearson-correlation, and auxiliary branch supervision, EMA for parameter stabilization, and a VAD-inspired latent prior for the acoustic branch. On the official validation set, the proposed scheme achieved our best mean Pearson Correlation Coefficient of 0.478567.