In-context Learning vs. Instruction Tuning: The Case of Small and Multilingual Language Models
arXiv cs.CL / 5/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper contrasts instruction tuning (supervised fine-tuning on curated instruction datasets, sometimes with human-preference alignment) with in-context learning (ICL) as an alternative way to teach instruction following to base LLMs.
- It evaluates whether ICL can reliably produce instruction-following behavior for small and multilingual language models, where instruction tuning is often more resource-intensive.
- The authors find that ICL instruction-following performance degrades in non-English and cross-model-size scenarios.
- They show that applying Direct Preference Optimization (DPO) on base models can partially improve results, but additional approaches are still needed to match the strongest English-centric large models.
- Overall, the work suggests ICL alone is not sufficient for robust multilingual instruction following at smaller scales, highlighting a remaining gap for future methods.
Related Articles

Every handle invocation on BizNode gets a WFID — a universal transaction reference for accountability. Full audit trail,...
Dev.to

I deployed AI agents across AWS, GCP, and Azure without a VPN. Here is how it works.
Dev.to

Panduan Lengkap TestSprite MCP Server — Dokumentasi Getting Started dalam Bahasa Indonesia
Dev.to
AI made learning fun again
Dev.to

MCP, Skills, AI Agents, and New Models: The New Stack for Software Development
Dev.to