GhazalBench: Usage-Grounded Evaluation of LLMs on Persian Ghazals
arXiv cs.CL / 3/12/2026
📰 NewsModels & Research
Key Points
- GhazalBench is introduced as a benchmark to evaluate LLMs on Persian ghazals under usage-grounded conditions, focusing on producing faithful paraphrases and accessing canonical verses.
- The evaluation across several proprietary and open-weight multilingual LLMs reveals a consistent dissociation: models generally capture poetic meaning but struggle with exact verse recall in completion-based tasks, while recognition-based tasks reduce this gap.
- An English sonnet benchmark shows markedly higher recall, suggesting the limits are tied to training exposure rather than architectural constraints.
- The authors advocate evaluation frameworks that jointly assess meaning, form, and cue-dependent access to culturally significant texts, and GhazalBench is publicly available at the linked GitHub repository.
Related Articles
Two bots, one confused server: what Nimbus revealed about AI agent identity
Dev.to
PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark forFinance
Dev.to
A Coding Implementation to Build an Uncertainty-Aware LLM System with Confidence Estimation, Self-Evaluation, and Automatic Web Research
MarkTechPost
DNA Memory: Making AI Agents Learn, Forget, and Evolve Like a Human Brain
Dev.to
Tinybox- offline AI device 120B parameters
Hacker News