From FusHa to Folk: Exploring Cross-Lingual Transfer in Arabic Language Models
arXiv cs.CL / 4/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The study examines how Arabic language models pretrained mainly on Modern Standard Arabic (MSA) transfer to different Arabic dialects used in speech and online writing.
- Using probing across three NLP tasks and representational similarity analysis, the authors find that cross-dialect transfer is possible but varies significantly between dialects.
- The paper reports that dialect similarity to MSA is partially explained by geographic proximity among dialect regions.
- It also provides evidence of negative interference when models are trained to support all Arabic dialects simultaneously, suggesting added training can reduce effective transfer for some dialects.
- The findings raise concerns about how well “all-dialect” training strategies support cross-lingual transfer in Arabic language models.
Related Articles

Knowledge Governance For The Agentic Economy.
Dev.to

AI server farms heat up the neighborhood for miles around, paper finds
The Register
Does the Claude “leak” actually change anything in practice?
Reddit r/LocalLLaMA

87.4% of My Agent's Decisions Run on a 0.8B Model
Dev.to

AIエージェントをソフトウェアチームに変える無料ツール「Paperclip」
Dev.to