MDB Drums dataset for automatic drum transcription

by Chih-Wei Wu

Data availability is the key to the success of many machine learning based Music Information Retrieval (MIR) systems. While there are different potential solutions to deal with insufficient data (for instance., semi-supervised learning, data augmentation, unsupervised learning, self-taught learning), the most direct way of tackling this problem is to create more annotated datasets.

Automatic Drum Transcription (ADT) is, similar to most of the MIR tasks, in need of more realistic and diverse datasets. However, the creation process of such datasets is usually difficult for the following reasons: 1) the synchronization between the drum strokes the onset times has to be exact. In previous work, this was done by installing triggers on the drum sets. However, the installation and the recording process also limits the size of the dataset. Another work around solution is to synthesize drum tracks with user defined onset time and drum samples; this will result in drum tracks with perfect ground truth, but the resulting music might be unrealistic and unrepresentative of the real-world music. 2) the variety of the drum sounds has to be high enough to cover a wide range from electronic to acoustic drum sounds. Most of the previous work only uses a small subset of drum sounds (e.g., certain drum machines or a few drum kits), which is not ideal in this regard. 3) the playing techniques can be hard to differentiate, especially on instruments such as snare drum. A majority of existing datasets only contain the annotations of the basic strikes for simplicity.

We try to address the above mentioned difficulties by introducing a new ADT dataset with a semi-automatic creation process.

Why MDB?

The goal of this project is to create a new dataset with minimum effort from the human annotators. In order to achieve this goal, a robust onset detection algorithm to locate the drum events is important. To ensure the robustness of the onset detector, we want the input signal to be as clean as possible. However, we also want to have a signal to be as realistic as possible (i.e., polyphonic mixtures of both melodic and percussive instruments). With these considerations in mind, we decided to avoid collecting a new dataset from scratch but rather work on the existing dataset with desirable properties.

As a result, the MusicDelta subset in the MedleyDB dataset is chosen for its:

  1. Multitrack format. This can potentially increase the robustness of onset detector and facilitate the semi-automatic annotation process. In addition, the multitrack files can be mixed in any arbitrary combination, providing more possibilities for experimentation.
  2. Real-world recordings. This means the recordings are more realistic and closer to the real use cases. Also, the diversity in terms of music genres offers a more representative sample pool.

Dataset creation

The processing flow of the creation of MDB Drum is shown in Fig. 1. First, the songs in the selected dataset are processed with an onset detector. This provides a consistent estimation of the onset locations. Next, the onsets are labeled with their corresponding instrument names (i.e., Hi-hat, Snare Drum, Kick Drum). This step inevitably requires manual annotation from the human experts. Following the manual annotation, a set of automatic checks were implemented to examine the annotations for common errors (e.g., typos, duplicates). Finally, the human experts went through an iterative process  of cross-checking their annotations prior to the release of the dataset.

Flowchart of the dataset creation process

An dataset example is shown in Fig. 2. The original polyphonic mixture (top) appears noisy, and it is difficult to locate the drum events through both listening and visual inspection. The drum-only recording (middle), however, has sharp attacks and short decays in the waveform, providing a cleaner representation for the onset detector. Finally, the detected onsets (bottom), as marked in red, are relatively accurate, which greatly simplifies the process of manual annotation.

One example of the semi-automatic annotation process

Dataset

The resulting dataset contains 23 tracks of real-world music with a diverse distribution of music genres (e.g., rock, disco, grunge, punk, reggae, jazz, funk, latin, country, britpop, to name just a few). The average duration of the tracks is around 54s, and the number of annotated instrument classes is 6 (only major classes) or 21 (with playing techniques). The users may choose to mix the multitracks in any combination (e.g., guitar + drum, bass + drum) due to the multitrack format of the original MedleyDB dataset.

All details can be found in our short paper here.

 

Objective descriptors for the assessment of student music performances

by Amruta Vidwans

Learning a musical instrument is difficult. It needs regular practice, expert advice, and supervision. Even today, musical training is largely driven by interaction between student and a human teacher plus individual practice session at home.

Can technology improve this process and the learning experience? Can an algorithm perform an assessment of a student music performance? If yes, we are one step closer to a truly musically intelligent music tutoring system  that will support students learn their instrument of choice by providing feedback on aspects like rhythmic correctness, note accuracy, etc. An automatic assessment is not only useful to students for their practice sessions but could also help band directors in the auditioning and (pre-)selection process. While there are a few commercial products for practicing instruments, the assessment in these products is usually either trivial or opaque to the user.

The realization of a musically intelligent system for music performance assessment requires knowledge from multiple disciplines such as digital signal processing, machine learning, audio content analysis, musicology, and music psychology. With recent advances in Music Information Retrieval (MIR), noticeable progress has been made in related research topics.

Despite these efforts, identifying a reliable and effective method for assessing music performances remains an unsolved problem. In our study, we explore the effectiveness of various objective descriptors by comparing three sets of features extracted from the audio recording of a music performance, (i) a baseline set with common low-level features (often used but hardly meaningful for this task), (ii) a score-independent set with designed performance features (custom-designed descriptors such as pitch deviation etc., but without knowledge of the musical score), and (iii) a score-based set with designed performance features (taking advantage of the known musical score). The goal is to identify a set of meaningful objective descriptors for the general assessment of student music performances. The data we used covers Alto Saxophone recordings of three years of student auditions (Florida state auditions) rated by experts in the assessment categories of musicality, note accuracy, rhythmic accuracy, and tone quality.

Label: Musicality E1 E2 E3 E4
Correlation (r) 0.19 0.49 0.56 0.58

Our observations (as seen in Table 1) are that, as expected, the baseline features (E1) are not able to capture any qualitative aspects of the music performance so that the regression model mostly fails to predict the expert assessments . Another expected result is that score-based features (E3) are able represent the data generally better than score-independent features (E2) in all categories. The combination of score-independent and score-based features (E4) show some trend to improve results, but the gain remains small, hinting at redundancies between the feature sets. With values between 0.5 and 0.65 for the correlation between the prediction and the human assessments, there is still a long way to go before computers will be able to reliably assess student music performance, but the results show that an automatic assessment is possible to a certain degree.

To learn more, please see the published paper for details.

Header image used with kind permission of Rachel Maness from http://wrongguytoask.blogspot.com/2012/08/woodwinds.html

GTCMT @ ISMIR 2017

It was great to see alumni and current students meet at the International Society for Music Information Retrieval Conference (ISMIR) in Suzhou, China.

Contributions from the group at the conference:

 

 

GTCMT @ ISMIR 2016

The Georgia Tech Center for Music Technology (GTCMT) has shown strong presence at the International Conference for Music Information Retrieval (ISMIR) with students, post-docs, and alumni.

Contributions from the group at the conference:

 

older projects

Assessment of Music Performances
Design and evaluation of features for the characterization of (student) music performances and create models to automatically assess these performances, detect errors, and give instantaneous feedback to the performer.

[show_more color=”#eeb211″]Resources
Source repository: github
Publications:
– Wu, C.-W.; Gururani, S.; Laguna, C.; Pati, A.; Vidwans, A.; Lerch, A., Towards the Objective Assessment of Music Performances, Proceedings of the International Conference on Music Perception and Cognition (ICMPC), San Francisco, 2016
Contributors (current)
Siddharth Kumar Gururani, Chris Laguna, Ashis Pati, Amruta Jayant Vidwans, Chih-Wei Wu
Contributors (past)
Cian O’Brien, Yujia Yan, Ying Zhan

[/show_more]

Automatic Drum Transcription (PhD Project)
Automatic drum transcription in polyphonic mixtures of music using a signal-adaptive NMF-based method.

[show_more color=”#eeb211″]

Resources
Source repository: github
Publications:
– Wu, C.-W.; Lerch, A., On Drum Playing Technique Detection in Polyphonic Mixtures, Proceedings of the International Conference on Music Information Retrieval (ISMIR), New York, 2016
– Wu, C.-W.; Lerch, A., Drum Transcription using Partially Fixed Non-Negative Matrix Factorization With Template Adaptation, in Proceedings of the International Conference on Music Information Retrieval (ISMIR), Malaga, 2015.
– Wu, C.-W.; Lerch, A., Drum Transcription using Partially Fixed Non-Negative Matrix Factorization, Proceedings of the European Signal Processing Conference (EUSIPCO), Nice, 2015.
Contributors
Chih-Wei Wu

[/show_more]

Audio Quality Enhancement (MS Project)
Web application to improve audio quality of low quality recordings (especially for low quality mobile phone recordings). Processing steps include detecting and correcting clipping (distortion), removing noise, normalization of loudness, and equalization. The REPAIR Web App allows users to upload low-quality audio and download the improved audio.

[show_more color=”#eeb211″]

Resources
Web Application: REPAIR Web App
Source repository: github
Publications:
– Laguna, C, Master Project Report: A Web Application for Audio Quality Enhancement, MS Project Report, Georgia Institute of Technology, 2016
– Laguna, C.; Lerch, A., An Efficient Algorithm for Clipping Detection and Declipping Audio, Proceedings of the 141st AES Convention, Los Angeles, 2016
– Laguna, C.; Lerch, A., Client-Side Audio Declipping, Proceedings of the 2nd Web Audio Conference (WAC), Atlanta, 2016
Contributors
Chris Laguna

[/show_more]

Outlier detection in music datasets (Cooperation with Virginia Tech)
Unsupervised detection of anomalies in music datasets.

[show_more color=”#eeb211″]

Resources
Publications:
– Lu, Y.-C.; Wu, C.-W.; Lu, C.T.; Lerch, A., Automatic Outlier Detection in Music Genre Datasets, Proceedings of the International Conference on Music Information Retrieval (ISMIR), New York, 2016
– Lu, Y.-C.; Wu, C.-W.; Lu, C.-T.; Lerch, A., An Unsupervised Approach to Anomaly Detection in Music Datasets, Proceedings of the ACM SIGIR Conference (SIGIR), Pisa, 2016
Contributors
Chih-Wei Wu

[/show_more]

Automatic Practice Logging (Semester Project)
Automatic identification of continuous recordings of musicians practicing their repertoire. The goal is a detailed description of what and where they practiced, which can be used by students and instructors to communicate about the countless hours spent practicing.

[show_more color=”#eeb211″]

Resources
Publications:
– Winters, R. M.; Gururani, S.; Lerch, A., Automatic Practice Logging: Introduction, Dataset & Preliminary Study, Proceedings of the International Conference on Music Information Retrieval (ISMIR), New York, 2016
Source repository: github
Contributors
R. Michael Winters, Siddharth Kumar Gururani

[/show_more]

Machine Listening Module (MS Project)
Machine listening provides a set of data with which music can be synthesized, modified, or sonified. Real time audio feature extraction opens up new worlds for interactive music, improvisation, and generative composition. Promoting the use of machine listening as a compositional tool, this project brings the technique into DIY embedded systems such as the Raspberry Pi, integrating machine listening with analog synthesizers in the eurorack format.

[show_more color=”#eeb211″]

Resources
Source repository: github
Project Report:
– Latina, C., Machine Listening Eurorack Module, MS Project Report, Georgia Institute of Technology, 2016.
Contributors
Chris Latina

[/show_more]

Sample detection in Polyphonic Music
Sampling, the usage of snippets or loops from existing songs or libraries in new music productions or mashups, is a common technique in many music genres. The goal of this project is to design an NMF-based algorithm that is able to detect the presence of a sample of audio in a set of tracks. The sample audio may be pitch shifted or time stretched so the algorithm should ideally be robust against such manipulation.

[show_more color=”#eeb211″]

Resources
Contributors
Siddarth Kumar

[/show_more]

Web Resources for Audio Content Analysis
Online resources for tasks related to music information retrieval and machine learning, including matlab files, a list of datasets, and exercises.

[show_more color=”#eeb211″]

Resources
WWW: AudioContentAnalysis.org
Contributors
Alexander Lerch

[/show_more]

Other Projects

Application of MIR Techniques to Medical Signals
Based on the physionet.org challenge dataset for reducing false alarms in ECG and blood pressure signals, MIR approaches are investigated for the detection of alarm situations in the intensive care unit. The 5 types of alarms asystole, extreme bradycardia, extreme tachycardia, ventricular tachycardia, and ventricular flutter are detected.

[show_more color=”#eeb211″]

Resources
Contributors
Amruta Vidwans

[/show_more]

Real-time speaker annotation in conference settings
Generating a transcript of a conference meeting requires not only the transcription of text but also assigning the text to specific speakers. This system is designed to detect an unknown number of speakers and assign text to these speakers in a real-time scenario.

[show_more color=”#eeb211″]

Resources
Source repository: github
Contributors
Avrosh Kumar

[/show_more]

Application for Vocal Training and Assessment using Real-Time Pitch Tracking
A cross-platform application for vocal training and evaluation Screenshot Vocal Assessmentusing monophonic pitch tracking. The system is designed to take real-time voice input using standard microphones available in most mobile devices. The assessment is carried out in reference to reference vocal lessons based on pitch and timing accuracy. Real-time feedback is provided to the user in the form of a pitch contour plotted against the reference pitch to be sung.

[show_more color=”#eeb211″]

Resources
Source repository: github
Project Report:
– Pati, A., An Application for Vocal Training and Evaluation using Real-time Monophonic Pitch Tracking, Technical Report, Georgia Tech, 2015.
Contributors
Ashis Pati

[/show_more]

Vocopter Singing Game
Vocopter is a mobile game adapted from the classic Copter game. Vocopter allows a playful approach to assess the accuracy of intonation.

[show_more color=”#eeb211″]

Resources
Source repository: github
Contributors
Rithesh Kumar

[/show_more]

Project Riyaaz
Riyaaz is an Urdu word which means devoted practice. The project aims at implementating an app that aids the practice of Indian classical vocal music. It requires the student pass through a curriculum of exercises designed to strengthen their grasp of Swara (tonality, pitch) and Tala (rhythm). The interface provides real-time graphical feedback in order to help improving their skills.

[show_more color=”#eeb211″]

Resources
Contributors
Milap Rane

[/show_more]

Audio-Adaptive Visual Animations of Paintings

Animated_Gif
Original oil on canvas painting: Dusan Malobabic

A painting is an expression frozen in time. It is the imagination of the viewer that paints the untold past and the future of the captured moment. This project is an attempt to induce movements in a painting evoked by sounds or music. The idea is to extract various descriptors from music, for example, onsets and tonal content, and map them to a function to process an image and bring it to life as well enhance the music listening experience.

[show_more color=”#eeb211″]

Resources
Contributors
Avrosh Kumar

[/show_more]

Automatic Audio-Lyrics Alignment
Automatic alignment of song lyrics to audio recordings at the line level. The alignment makes use of voice activity detection, pitch detection, and the detection of repeating structures.

[show_more color=”#eeb211″]

Resources
Contributors
Amruta Vidwans

[/show_more]

Genre-specific Key Profiles
Investigation of differences and commonalities of audio pitch class profiles of different musical genres.

[show_more color=”#eeb211″]

Resources
Publication: O’Brien, C.; Lerch, A., Genre-Specific Key Profiles; Proceedings of the International Computer Music Conference (ICMC), Denton, 2015.
Contributors
Cian O’Brien

[/show_more]

Supervised Feature Learning via Sparse Coding for Music Information Retrieval
Sparse coding allows to learn features from the dataset in an unsupervised way. It is investigated how added supervised training functionality can improve the descriptiveness of the learned features.

[show_more color=”#eeb211″]

Resources:
Thesis: smartech
Contributors
Cian O’Brien

[/show_more]

Real-time Onset Detection
Design of an Onset Detection Algorithm suitable for real-time processing and a low latency live input scenario.

[show_more color=”#eeb211″]

<Contributors
Rithesh Kumar

[/show_more]

Predominant Instrument Recognition in Polyphonic Audio
Identification of a single predominant instrument per audio file using pitch features, timbre features and features extracted from short-time harmonics.

[show_more color=”#eeb211″]

Contributors
Chris Laguna

[/show_more]

Time-Domain Multi-Pitch Detection with Sparse Additive Modeling
Frame-level multi-pitch detection in the time domain with locally periodic kernel functions and sparsity constraints.

[show_more color=”#eeb211″]

Contributors
Yujia Yan

[/show_more]

Identification of live music performance via ambient audio content features
Automatic identification of recordings of live performance as opposed to studio recordings.

[show_more color=”#eeb211″]

Resources
dataset: github
Contributors
Raja Raman

[/show_more]

Wiki tutorial for running SuperCollider on Raspberry Pi
Various tutorial on installation and configuration of SuperCollider on a Raspberry Pi.

[show_more color=”#eeb211″]

Resources
WWW: Embedded Music Page
Contributors
Chris Latina

[/show_more]

Metric Learning for Music Discovery with Source and Target Playlists
Playlist generation for music exploration by defining sets of source songs and target songs and deriving a playlist through metric learning and boundary constraints.

[show_more color=”#eeb211″]

Resources
slides: presentation
Contributors
Ying-Shu Kuo

[/show_more]

Audio Chord Detection Using Deep Learning
Improve audio chord detection by using a Deep Network to extract the tonal features from the audio.

[show_more color=”#eeb211″]

Resources
Publication: Zhou, X.; Lerch, A., Chord Detection Using Deep Learning, in Proceedings of the International Conference on Music Information Retrieval (ISMIR), Malaga, 2015.
Contributors
Xinquan Zhou

[/show_more]