AI Navigate

From Language to Action in Arabic: Reliable Structured Tool Calling via Data-Centric Fine-Tuning

arXiv cs.LG / 3/19/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper introduces AISA-AR-FunctionCall, a production-oriented Arabic function-calling framework built on a 270M-parameter FunctionGemma backbone.
  • It leverages systematic dataset auditing, schema repair, tool-aware prompt restructuring, and full-parameter supervised fine-tuning to improve robustness for Arabic across dialects.
  • On held-out tests, fine-tuning reduces parse failures from 87% to below 1%, increases function name accuracy by more than eightfold, and enhances argument alignment across dialects and domains.
  • The study presents a reasoning-augmented LoRA variant with explicit intermediate reasoning before tool invocation and notes that serialization stability and decision-level reasoning are separable challenges, with all datasets and models released under the AISA framework.

Abstract

Function-calling language models are essential for agentic AI systems that translate natural language into executable structured actions, yet existing models exhibit severe structural instability when applied to Arabic. We present AISA-AR-FunctionCall, a production-oriented Arabic function-calling framework built on a 270M-parameter FunctionGemma backbone and trained through systematic dataset auditing, schema repair, tool-aware prompt restructuring, and full-parameter supervised fine-tuning. On a held-out test set, fine-tuning reduces parse failures from 87\% to below 1\%, improves function name accuracy by more than eightfold, and substantially enhances argument alignment across dialects and domains. Error analysis reveals a transition from structural collapse to semantic misalignment, suggesting that serialization stability and decision-level reasoning are separable challenges. We further explore a reasoning-augmented LoRA variant that introduces explicit intermediate reasoning prior to tool invocation. All datasets and models are publicly released under the AISA framework.