Toward Generalized Cross-Lingual Hateful Language Detection with Web-Scale Data and Ensemble LLM Annotations
arXiv cs.CL / 4/14/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper evaluates whether web-scale unlabeled multilingual text plus LLM-generated synthetic labels can improve hateful language detection across four languages (English, German, Spanish, Vietnamese).
- Continued pre-training of BERT on unlabeled web data followed by supervised fine-tuning improves macro-F1 by about 3% on sixteen benchmarks, with larger gains in low-resource settings.
- It compares three LLM ensemble annotation methods (mean averaging, majority voting, and a LightGBM meta-learner), finding that the LightGBM ensemble is consistently best.
- Training smaller models on the synthetic labels yields large improvements (e.g., Llama3.2-1B gains about +11% pooled F1), while larger models see only modest benefit (e.g., Qwen2.5-14B about +0.6%).
- Overall, the authors conclude that combining web-scale unlabeled data with LLM-ensemble annotations is especially valuable for small models and low-resource languages.
Related Articles

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial

One URL for Your AI Agent: HTML, JSON, Markdown, and an A2A Card
Dev.to

One URL for Your AI Agent: HTML, JSON, Markdown, and an A2A Card
Dev.to