Improving Facial Emotion Recognition through Dataset Merging and Balanced Training Strategies

arXiv cs.CV / 4/23/2026

📰 NewsModels & Research

Key Points

  • The paper proposes a deep learning framework using deep convolutional networks to automatically recognize facial emotions.
  • It improves generalization and robustness by merging three public datasets (CK+, FER+, and KDEF) to expand the training data.
  • Even after merging, minority emotion classes remain underrepresented, so the authors apply online/offline augmentation and random weighted sampling to reduce data imbalance.
  • Experiments report 82% accuracy for recognizing seven basic emotions, indicating the approach effectively addresses class imbalance and improves performance.

Abstract

In this paper, a deep learning framework is proposed for automatic facial emotion based on deep convolutional networks. In order to increase the generalization ability and the robustness of the method, the dataset size is increased by merging three publicly available facial emotion datasets: CK+, FER+ and KDEF. Despite the increase in dataset size, the minority classes still suffer from insufficient number of training samples, leading to data imbalance. The data imbalance problem is minimized by online and offline augmentation techniques and random weighted sampling. Experimental results demonstrate that the proposed method can recognize the seven basic emotions with 82% accuracy. The results demonstrate the effectiveness of the proposed approach in tackling the challenges of data imbalance and improving classification performance in facial emotion recognition.