M-DaQ: Retrieving Samples with Multilingual Diversity and Quality for Instruction Fine-Tuning Datasets
arXiv cs.CL / 5/1/2026
💬 OpinionTools & Practical UsageModels & Research
Key Points
- The paper introduces M-DaQ, a multilingual diversity-and-quality sampling framework aimed at building higher-quality instruction fine-tuning (IFT) datasets, which are currently scarce.
- M-DaQ combines a fine-tuned quality scoring model with a maximal marginal relevance–inspired selection method to jointly optimize response quality and cross-lingual semantic diversity.
- It also conducts the first systematic study of the Superficial Alignment Hypothesis in multilingual scenarios to understand alignment behavior across languages.
- Experiments covering 18 languages show that models trained on M-DaQ-curated data achieve average win rates above 60% versus strong baselines on Alpaca-Eval and MT-Bench, with supporting human evaluation improvements in cultural relevance and instruction-following.
- The authors release the code publicly to support reproducibility and enable follow-on research.
Related Articles

Black Hat USA
AI Business

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Text-to-image is easy. Chaining LLMs to generate, critique, and iterate on images autonomously is a routing nightmare. AgentSwarms now supports Image generation playground and creative media workflows!
Reddit r/artificial

Announcing the NVIDIA Nemotron 3 Super Build Contest
Dev.to

75% of Sites Blocking AI Bots Still Get Cited. Here Is Why Blocking Does Not Work.
Dev.to