Multilingual and Domain-Agnostic Tip-of-the-Tongue Query Generation for Simulated Evaluation

arXiv cs.CL / 4/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces multilingual Tip-of-the-Tongue (ToT) retrieval test collections for Chinese, Japanese, Korean, and English to address the field’s previous reliance on English-only benchmarks.
  • It uses an LLM-based query simulation framework to generate synthetic ToT queries and evaluates how prompt language and source-document language influence the fidelity of the simulations.
  • The authors validate simulated queries by measuring system rank correlation against real user queries, showing that language-aware design is crucial for effective ToT simulation.
  • Key findings indicate that non-English source documents are generally important, while English Wikipedia can help when non-English sources lack enough information for query generation.
  • The work releases four large-scale ToT benchmarks (5,000 queries per language across multiple domains) and provides guidance for building realistic ToT datasets for languages beyond English.

Abstract

Tip-of-the-Tongue (ToT) retrieval benchmarks have largely focused on English, limiting their applicability to multilingual information access. In this work, we construct multilingual ToT test collections for Chinese, Japanese, Korean, and English, using an LLM-based query simulation framework. We systematically study how prompt language and source document language affect the fidelity of simulated ToT queries, validating synthetic queries through system rank correlation against real user queries. Our results show that effective ToT simulation requires language-aware design choices: non-English language sources are generally important, while English Wikipedia can be beneficial when non-English sources provide insufficient information for query generation. Based on these findings, we release four ToT test collections with 5,000 queries per language across multiple domains. This work provides the first large-scale multilingual ToT benchmark and offers practical guidance for constructing realistic ToT datasets beyond English.