Limited Linguistic Diversity in Embodied AI Datasets

arXiv cs.RO / 4/29/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights that Vision-Language-Action (VLA) models depend heavily on instruction language, but the linguistic properties of commonly used training and evaluation datasets are not well documented.
  • It performs a systematic audit of several widely used VLA corpora, measuring instruction language through lexical variety, duplication/overlap, semantic similarity, and syntactic complexity.
  • The results show that many datasets use repetitive, template-like commands with limited structural variation, leading to a narrow range of instruction forms.
  • The authors frame the work as descriptive documentation of the current “language signal” in VLA data to enable better reporting, more principled dataset selection, and targeted curation or augmentation to broaden linguistic coverage.

Abstract

Language plays a critical role in Vision-Language-Action (VLA) models, yet the linguistic characteristics of the datasets used to train and evaluate these systems remain poorly documented. In this work, we present a systematic dataset audit of several widely used VLA corpora, aiming to characterize what kinds of instructions these datasets actually contain and how much linguistic variety they provide. We quantify instruction language along complementary dimensions--including lexical variety, duplication and overlap, semantic similarity, and syntactic complexity. Our analysis shows that many datasets rely on highly repetitive, template-like commands with limited structural variation, yielding a narrow distribution of instruction forms. We position these findings as descriptive documentation of the language signal available in current VLA training and evaluation data, intended to support more detailed dataset reporting, more principled dataset selection, and targeted curation or augmentation strategies that broaden language coverage.