HMM-Based Voice Separation of MIDI Performance

Andrew McLeod, Mark Steedman

Research output: Contribution to journalArticlepeer-review

Abstract

Voice separation is an important component of Music Information Retrieval (MIR). In this paper, we present an HMM which can be used to separate music performance data in the form of MIDI into monophonic voices. It works on two basic principles: that consecutive notes within a single voice will tend to occur on similar pitches, and that there are short (if any) temporal gaps between them. We also present an incremental algorithm which can perform inference on the model efficiently. We show that our approach achieves a significant improvement over existing approaches, when run on a corpus of 78 compositions by J.S. Bach, each of which has been separated into the gold standard voices suggested by the original score. We also show that it can be used to perform voice separation on live MIDI data without an appreciable loss in accuracy. The code for the model described here is available at https://github.com/apmcleod/voice-splitting.
Original languageEnglish
Pages (from-to)17-26
Number of pages10
JournalJournal of New Music Research
Volume45
Issue number1
DOIs
Publication statusPublished - 2016

Fingerprint Dive into the research topics of 'HMM-Based Voice Separation of MIDI Performance'. Together they form a unique fingerprint.

Cite this