Datasets

Reference Annotations

Introduction

Shared data enable researchers to reliably compare their results with others. The Centre for Digital Music (C4DM) has a tradition of providing ground truth data for research, and our chord, onset and segmentation annotations have been used by many researchers in the MIR community. This year, we have focused on extending these resources in scope and quantity.

This web page accompanies a late-breaking paper at the 11th Conference on Music Information Retrieval [1].

Reference Annotations: The Beatles

Note: Please be sure to read the page describing these annotations before use. In particular, the level of confidence we have in the individual annotations is described there, as well as the original CD issue numbers from which we worked.

Chris Harte's PhD thesis (2010) which describes the chord syntax, transcription process and verification process for the Beatles collection can be downloaded here.

Room Impulse Response Data Set

This collection of room impulse responses was measured in the Great Hall, the Octagon, and a classroom at the Mile End campus of Queen Mary, University of London in 2008. The measurements were created using the sine sweep technique [1] with a Genelec 8250A loudspeaker and two microphones, an omnidirectional DPA 4006 and a B-format Soundfield SPS422B.

These IRs are released under the Creative Commons Attribution-Noncommercial-Share-Alike license with attribution to the Centre for Digital Music, Queen Mary, University of London.

Automatic Annotations

In the field of music computing, many methods aim at automatically extracting musical information such as beat times, chords, and keys from raw audio data.

Publishing the automatically extracted data can be useful in several ways: researchers working on similar methods can compare their results with those published; non-specialist observers can examine the state of the art; the published data can be used as input for new research methods (e.g. for music classification).

Singing Voice Audio Dataset

This dataset is for the purpose of the analysis of singing voice. It is our hope that the publication of this dataset will encourage further work into the area of singing voice audio analysis by removing one of the main impediments in this research area - the lack of data (unaccompanied singing).

It contains over 70 original vocal recordings by 28 professional, semi-professional and amateur singers. Singing style is predominantly Chinese Opera but some recordings are Western Opera. All recordings are 44.1 KHz sample rate and have been amplitude normalised.