Show simple item record

dc.contributor.authorBenetos, Een_US
dc.contributor.authorWeyde, Ten_US
dc.contributor.author16th International Society for Music Information Retrieval Conference (ISMIR)en_US
dc.contributor.editorWiering, Fen_US
dc.contributor.editorMüller, Men_US
dc.date.accessioned2015-12-16T13:48:13Z
dc.date.accessioned2015-12-16T13:59:59Z
dc.date.available2015-06-22en_US
dc.date.issued2015-10-26en_US
dc.date.submitted2015-10-30T12:12:07.745Z
dc.date.submitted2015-12-16T13:57:37.755Z
dc.identifier.urihttp://qmro.qmul.ac.uk/xmlui/handle/123456789/9852
dc.description.abstractIn this paper, an efficient, general-purpose model for multiple instrument polyphonic music transcription is proposed. The model is based on probabilistic latent component analysis and supports the use of sound state spectral templates, which represent the temporal evolution of each note (e.g. attack, sustain, decay). As input, a variable-Q transform (VQT) time-frequency representation is used. Computational efficiency is achieved by supporting the use of pre-extracted and pre-shifted sound state templates. Two variants are presented: without temporal constraints and with hidden Markov model-based constraints controlling the appearance of sound states. Experiments are performed on benchmark transcription datasets: MAPS, TRIOS, MIREX multiF0, and Bach10; results on multi-pitch detection and instrument assignment show that the proposed models outperform the state-of-the-art for multiple-instrument transcription and is more than 20 times faster compared to a previous sound state-based model. We finally show that a VQT representation can lead to improved multi-pitch detection performance compared with constant-Q representations.en_US
dc.format.extent701 - 707 (7)en_US
dc.language.isoenen_US
dc.publisherInternational Society for Music Information Retrievalen_US
dc.relation.replaceshttp://qmro.qmul.ac.uk/xmlui/handle/123456789/9849
dc.relation.replaces123456789/9849
dc.rightshttp://ismir2015.uma.es/articles/131_Paper.pdf
dc.titleAn efficient temporally-constrained probabilistic model for multiple-instrument music transcriptionen_US
dc.typeConference Proceeding
dc.rights.holder© The Author(s) 2015
pubs.author-urlhttp://www.eecs.qmul.ac.uk/~emmanouilb/en_US
pubs.notesNo embargoen_US
pubs.notesOpen access CC-BY paper, no embargoen_US
pubs.publication-statusPublisheden_US
pubs.publisher-urlhttp://www.ismir.net/en_US
dcterms.dateAccepted2015-06-22en_US
qmul.funderA Machine Learning Framework for Audio Analysis and Retrieval::Royal Academy of Engineeringen_US
qmul.funderA Machine Learning Framework for Audio Analysis and Retrieval::Royal Academy of Engineeringen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record