How Class Ontology and Data Scale Affect Audio Transfer Learning

arXiv cs.LG / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper conducts a rigorous study of audio-to-audio transfer learning by pre-training different model states on ontology-based subsets of AudioSet and then fine-tuning on three downstream computer audition tasks (acoustic scene, bird activity, and speech command recognition).
  • It finds that scaling pre-training data—both by increasing the number of samples and by expanding the number of classes—improves transfer learning performance.
  • The study reports that this scaling benefit is often outweighed by how similar the pre-training data is to the downstream task, which can cause the model to learn sufficiently comparable features.
  • The work frames transfer learning as still having open mechanistic questions and aims to clarify when and why it works in the audio domain specifically.

Abstract

Transfer learning is a crucial concept within deep learning that allows artificial neural networks to benefit from a large pre-training data basis when confronted with a task of limited data. Despite its ubiquitous use and clear benefits, there are still many open questions regarding the inner workings of transfer learning and, in particular, regarding the understanding of when and how well it works. To that extent, we perform a rigorous study focusing on audio-to-audio transfer learning, in which we pre-train various model states on (ontology-based) subsets of AudioSet and fine-tune them on three computer audition tasks, namely acoustic scene recognition, bird activity recognition, and speech command recognition. We report that increasing the number of samples and classes in the pre-training data both have a positive impact on transfer learning. This is, however, generally surpassed by similarity between pre-training and the downstream task, which can lead the model to learn comparable features.
広告