Show simple item record

dc.contributor.authorYcart, Aen_US
dc.contributor.authorBenetos, Een_US
dc.contributor.author18th International Society for Music Information Retrieval Conference (ISMIR 2017)en_US
dc.date.accessioned2017-07-21T09:49:55Z
dc.date.available2017-06-23en_US
dc.date.issued2017-10-23en_US
dc.date.submitted2017-07-16T10:12:00.941Z
dc.identifier.urihttp://qmro.qmul.ac.uk/xmlui/handle/123456789/24946
dc.description.abstractNeural networks, and especially long short-term memory networks (LSTM), have become increasingly popular for sequence modelling, be it in text, speech, or music. In this paper, we investigate the predictive power of simple LSTM networks for polyphonic MIDI sequences, using an empirical approach. Such systems can then be used as a music language model which, combined with an acoustic model, can improve automatic music transcription (AMT) performance. As a first step, we experiment with synthetic MIDI data, and we compare the results obtained in various settings, throughout the training process. In particular, we compare the use of a fixed sample rate against a musically-relevant sample rate. We test this system both on synthetic and real MIDI data. Results are compared in terms of note prediction accuracy. We show that the higher the sample rate is, the better the prediction is, because self transitions are more frequent. We suggest that for AMT, a musically-relevant sample rate is crucial in order to model note transitions, beyond a simple smoothing effect.en_US
dc.format.extent421 - 427 (7)en_US
dc.publisherISMIRen_US
dc.rightsLicensed under a Creative Commons Attribution 4.0 International License (CC BY 4.0). Attribution: Adrien Ycart and Emmanouil Benetos. “A study on LSTM networks for polyphonic music sequence modelling”, 18th International Society for Music Information Retrieval Conference, Suzhou, China, 2017.
dc.titleA study on LSTM networks for polyphonic music sequence modellingen_US
dc.typeConference Proceeding
dc.rights.holder© Adrien Ycart and Emmanouil Benetos.
pubs.author-urlhttp://www.eecs.qmul.ac.uk/~ay304/en_US
pubs.notesNo embargoen_US
pubs.notesConference proceedings are CC-BY, there is no embargoen_US
pubs.publication-statusAccepteden_US
pubs.publisher-urlhttps://ismir2017.smcnus.org/en_US
dcterms.dateAccepted2017-06-23en_US
qmul.funderA Machine Learning Framework for Audio Analysis and Retrieval::Royal Academy of Engineeringen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record