Multilingual Multi-Label Emotion Classification at Scale with Synthetic Data
arXiv cs.CL / 4/15/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a large-scale synthetic dataset for multilingual, multi-label emotion classification, addressing the lack of non-English and multi-label annotated data in existing corpora.
- It builds 1M+ training samples across 23 languages (50k per language) using culturally adapted generation plus programmatic quality filtering, labeled with 11 emotion categories.
- Six multilingual transformer encoders are trained under identical settings, and the best-performing model is XLM-R-Large (560M), achieving 0.868 F1-micro and 0.987 AUC-micro on the in-domain test set.
- Zero-shot evaluations on human-annotated benchmarks (GoEmotions and SemEval-2018 Task 1 E-c) show the top model matching or outperforming English-specialist baselines on ranking metrics while covering all 23 languages.
- A best base-sized model is released publicly on Hugging Face, enabling others to reuse and benchmark multilingual emotion classifiers.




