LLMpediaThe first transparent, open encyclopedia generated by LLMs

MagnaTagATune

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Expansion Funnel Raw 1 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted1
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
MagnaTagATune
NameMagnaTagATune
TypeAudio dataset
Created2007
CreatorsThe Centre for Digital Music, Queen Mary University of London; International Community
DomainMusic information retrieval
Size~26,000 clips
LicenseCreative Commons (varied)

MagnaTagATune is a widely used audio dataset for music annotation and retrieval research, developed by the Centre for Digital Music at Queen Mary University of London with community-sourced labels from the MagnaTagATune game. The dataset is frequently cited in studies comparing feature extraction, machine learning, and deep learning approaches across tasks such as multi-label tagging, genre classification, and music similarity. It has been referenced alongside prominent datasets and resources in the fields of music information retrieval and audio signal processing.

Overview

MagnaTagATune was released to support research in music information retrieval, audio classification, and multimedia analysis and is often compared to datasets like the Million Song Dataset, GTZAN, FMA, MUSDL, MUSDB18, MedleyDB, Jamendo, Free Music Archive, and AudioSet. The project involved institutions including Queen Mary University of London, the Centre for Digital Music, and collaborators who designed a gamified annotation platform influenced by projects at MIT, Stanford, and UC Berkeley. The dataset contains tens of thousands of short audio clips sampled from a diverse set of musical works; it has been used in publications alongside evaluations involving IEEE, ACM, ISMIR, and NeurIPS venues. Prominent researchers and groups that have used or cited MagnaTagATune include Trevor Cox, Brian McFee, Emilia Gómez, Jordi Pons, Mark Sandler, Ge Wang, Anssi Klapuri, Simon Dixon, and Xavier Serra.

Dataset Composition and Annotation

The corpus comprises approximately 25,000–27,000 musical clips, each around 29 seconds, drawn from Creative Commons and commercial sources with metadata and listener-provided tags. Annotation was collected via the MagnaTagATune online game and includes multi-label tags covering instrumentation, genre, mood, tempo, and vocal presence. Tags in the dataset often intersect with taxonomies used in datasets like CAL500, MSD (Million Song Dataset), Last.fm, Echonest datasets, and Million Playlist Dataset; labels include instruments (e.g., piano, guitar, violin), genres (e.g., rock, jazz, classical), moods (e.g., happy, melancholic), and production descriptors (e.g., live, acoustic). The annotation process generated both frequent and sparse tags, mirroring annotation characteristics seen in Last.fm folksonomies, Echo Nest analyses, and MusicBrainz metadata efforts. Researchers have used the tag matrix to evaluate multi-label learning algorithms such as binary relevance, classifier chains, and deep neural networks, as in studies drawing comparisons to CIFAR, ImageNet, and LibriSpeech in methodology.

Collection Methodology

Audio clips were sourced and clipped down to fixed durations from a wide range of albums and artists, with provenance and rights managed through Creative Commons and permissioned content, similar to sourcing strategies in Free Music Archive and Jamendo initiatives. The gamified annotation system engaged online players to listen to clips and assign tags, inspired by human-computation efforts like ESP Game and Galaxy Zoo, and paralleled crowd-sourcing approaches by Amazon Mechanical Turk used in other corpora. Quality control included consensus thresholds and agreement measures analogous to inter-annotator agreement practices found in corpus linguistics and image labeling projects at Stanford Vision Lab and MIT CSAIL. Metadata links were reconciled with identifiers from MusicBrainz, Discogs, and ISRC registries where available.

Tasks and Benchmarks

MagnaTagATune has been used to benchmark multi-label tag prediction, audio feature extraction, metric learning for similarity, and transfer learning for tasks like genre recognition and mood estimation. Common baselines include MFCC, chroma, spectral contrast, and learned embeddings from convolutional neural networks and recurrent architectures evaluated in papers at ISMIR, ICASSP, IEEE TASLP, and ACM MM. The dataset has underpinned comparisons among algorithms such as support vector machines, random forests, k-nearest neighbors, convolutional neural networks, residual networks, Siamese networks, and transformer models, often cited in relation to benchmarks like GTZAN, FMA, AudioSet, and MUSDB evaluations. Evaluation metrics include AUC-ROC, mean average precision, precision at K, and F1-score, aligning with evaluation conventions used in ImageNet, COCO, and LibriSpeech studies.

Usage and Impact

MagnaTagATune influenced development of tagging systems, recommender systems, and content-based music retrieval methods used in academic and industrial research at institutions like Spotify, Pandora, Deezer, Google, Amazon, Apple, and Microsoft Research. It provided a reproducible testbed that fostered advances in representation learning, contrastive learning, and supervised multi-label training regimes that were later adapted in studies at Facebook AI Research, DeepMind, and OpenAI. The dataset appears in curricula and tutorials for MIR courses at Queen Mary University of London, Universitat Pompeu Fabra, University of Cambridge, Stanford University, and Harvard University, and continues to be cited in ISMIR proceedings and IEEE journals.

Limitations and Criticisms

Critiques of MagnaTagATune include limited clip length, potential licensing heterogeneity, tag imbalance, and annotation noise stemming from the game-based collection method—points also raised for GTZAN and some Last.fm-derived resources. Concerns about representativeness and cultural bias mirror debates around datasets like ImageNet and AudioSet, while metadata sparsity complicates tasks requiring artist-level or album-level context compared to datasets such as Million Song Dataset and MusicNet. Subsequent work has recommended combining MagnaTagATune with larger, more diverse corpora like FMA, AudioSet, and Free Music Archive to mitigate bias and improve generalization for modern deep learning pipelines.

Category:Audio datasets