AI Navigate

An Agentic System for Schema Aware NL2SQL Generation

arXiv cs.CL / 3/20/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The authors propose a schema-based agentic NL2SQL system that uses small language models as the primary agents and only invokes a large language model when errors are detected, reducing computational overhead.
  • The system achieves substantial cost savings, resolving about 67% of queries with local SLMs and lowering the average cost per query from 0.094 to 0.0085.
  • On the BIRD benchmark, it attains an execution accuracy of 47.78% and a validation efficiency score of 51.05%, demonstrating practical effectiveness with lower resource use.
  • The design targets resource-constrained deployments and aims for near-zero operational costs for locally executed queries, addressing privacy and deployability concerns of LLM-centric approaches.
  • A GitHub repository is provided for implementation and reproducibility.

Abstract

The natural language to SQL (NL2SQL) task plays a pivotal role in democratizing data access by enabling non-expert users to interact with relational databases through intuitive language. While recent frameworks have enhanced translation accuracy via task specialization, their reliance on Large Language Models (LLMs) raises significant concerns regarding computational overhead, data privacy, and real-world deployability in resource-constrained environments. To address these challenges, we propose a schema based agentic system that strategically employs Small Language Models (SLMs) as primary agents, complemented by a selective LLM fallback mechanism. The LLM is invoked only upon detection of errors in SLM-generated output, the proposed system significantly minimizes computational expenditure. Experimental results on the BIRD benchmark demonstrate that our system achieves an execution accuracy of 47.78% and a validation efficiency score of 51.05%, achieving over 90% cost reduction compared to LLM-centric baselines as approximately 67% of queries are resolved using local SLMs. The system achieves an average cost per query of 0.0085 compared to 0.094 for LLM-only systems, achieving near-zero operational costs for locally executed queries. [Github repository: https://github.com/mindslab25/CESMA.]