MUSCAT: MUltilingual, SCientific ConversATion Benchmark

arXiv cs.CL / 4/20/2026

📰 NewsSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The article introduces MUSCAT, a new multilingual speech benchmark aimed at evaluating how well ASR systems handle realistic multilingual conversation scenarios.
  • The benchmark is based on bilingual scientific-paper discussions where multiple speakers talk in different languages, including challenges like mixed-language input, domain-specific vocabulary, and code-switching.
  • It provides a standardized evaluation framework that goes beyond Word Error Rate (WER) to enable fairer comparisons of ASR performance across languages.
  • Initial experiments indicate the dataset remains an open, difficult problem for state-of-the-art ASR systems, motivating further research.
  • The MUSCAT dataset is publicly released on Hugging Face for use in ASR research and benchmarking.

Abstract

The goal of multilingual speech technology is to facilitate seamless communication between individuals speaking different languages, creating the experience as though everyone were a multilingual speaker. To create this experience, speech technology needs to address several challenges: Handling mixed multilingual input, specific vocabulary, and code-switching. However, there is currently no dataset benchmarking this situation. We propose a new benchmark to evaluate current Automatic Speech Recognition (ASR) systems, whether they are able to handle these challenges. The benchmark consists of bilingual discussions on scientific papers between multiple speakers, each conversing in a different language. We provide a standard evaluation framework, beyond Word Error Rate (WER) enabling consistent comparison of ASR performance across languages. Experimental results demonstrate that the proposed dataset is still an open challenge for state-of-the-art ASR systems. The dataset is available in https://huggingface.co/datasets/goodpiku/muscat-eval \\ ewline \Keywords{multilingual, speech recognition, audio segmentation, speaker diarization}