GaMMA: Towards Joint Global-Temporal Music Understanding in Large Multimodal Models

arXiv cs.AI / 5/4/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • GaMMA is a new large multimodal model aimed at broad, end-to-end music understanding by jointly learning musical audio signals and language.
  • The model builds on LLaVA’s streamlined encoder–decoder design for cross-modal learning, and uses mixture-of-experts audio encoders to handle both time-series and non-time-series music tasks under one parameter set.
  • GaMMA is trained using large-scale curated datasets and a progressive pipeline covering pretraining, supervised fine-tuning (SFT), and reinforcement learning (RL).
  • The paper introduces MusicBench, a large human-curated benchmark with 3,739 multiple-choice questions to evaluate temporal and global (non-temporal) music understanding.
  • Experiments report new state-of-the-art results in the music domain, including 79.1% on MuchoMusic, 79.3% on MusicBench-Temporal, and 81.3% on MusicBench-Global.

Abstract

In this paper, we propose GaMMA, a state-of-the-art (SoTA) large multimodal model (LMM) designed to achieve comprehensive musical content understanding. GaMMA inherits the streamlined encoder-decoder design of LLaVA, enabling effective cross-modal learning between music and language. By incorporating audio encoders in a mixture-of-experts manner, GaMMA effectively unifies both time-series and non-time-series music understanding tasks within one set of parameters. Our approach combines carefully curated datasets at scale with a progressive training pipeline, effectively pushing the boundaries of music understanding via pretraining, supervised fine-tuning (SFT), and reinforcement learning (RL). To comprehensively assess both temporal and non-temporal capability of music LMMs, we introduce MusicBench, the largest music-oriented benchmark, comprising 3,739 human-curated multiple-choice questions covering diverse aspects of musical understanding. Extensive experiments demonstrate that GaMMA establishes new SoTA in the music domain, achieving 79.1% accuracy on MuchoMusic, 79.3% on MusicBench-Temporal, and 81.3% on MusicBench-Global, consistently outperforming previous methods.