Can LLM Agents Identify Spoken Dialects like a Linguist?

arXiv cs.CL / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates whether LLMs used as agents can identify spoken dialects (including Swiss German) and compare their performance to established audio-based models like HuBERT.
  • The proposed method uses ASR-generated phonetic transcriptions combined with linguistic resources (e.g., dialect feature maps, vowel history, and rule-based cues) to support dialect classification.
  • Results suggest LLM dialect predictions improve when explicit linguistic information is provided, indicating that grounding and structured linguistic features matter for this task.
  • The authors include both an LLM baseline and a human linguist baseline, concluding that automatically generated transcriptions can help dialect classification while also highlighting room to improve ASR-driven inputs.

Abstract

Due to the scarcity of labeled dialectal speech, audio dialect classification is a challenging task for most languages, including Swiss German. In this work, we explore the ability of large language models (LLMs) as agents in understanding the dialects and whether they can show comparable performance to models such as HuBERT in dialect classification. In addition, we provide an LLM baseline and a human linguist one. Our approach uses phonetic transcriptions produced by ASR systems and combines them with linguistic resources such as dialect feature maps, vowel history, and rules. Our findings indicate that, when linguistic information is provided, the LLM predictions improve. The human baseline shows that automatically generated transcriptions can be beneficial for such classifications, but also presents opportunities for improvement.

Can LLM Agents Identify Spoken Dialects like a Linguist? | AI Navigate