A Comparative Study of Neural Models for Polyphonic Music Sequence Transduction
dc.contributor.author | Ycart, A | en_US |
dc.contributor.author | Stoller, D | en_US |
dc.contributor.author | Benetos, E | en_US |
dc.contributor.author | 20th conference of the International Society for Music Information Retrieval (ISMIR) | en_US |
dc.date.accessioned | 2019-08-16T14:00:47Z | |
dc.date.available | 2019-06-07 | en_US |
dc.date.issued | 2019-11-04 | en_US |
dc.identifier.uri | https://qmro.qmul.ac.uk/xmlui/handle/123456789/59184 | |
dc.description.abstract | Automatic transcription of polyphonic music remains a challenging task in the field of Music Information Retrieval. One under-investigated point is the post-processing of time-pitch posteriograms into binary piano rolls. In this study, we investigate this task using a variety of neural network models and training procedures. We introduce an adversarial framework, that we compare against more traditional training losses. We also propose the use of binary neuron outputs and compare them to the usual real-valued outputs in both training frameworks. This allows us to train networks directly using the F-measure as training objective. We evaluate these methods using two kinds of transduction networks and two different multi-pitch detection systems, and compare the results against baseline note-tracking methods on a dataset of classical piano music. Analysis of results indicates that (1) convolutional models improve results over baseline models, but no improvement is reported for recurrent models; (2) supervised losses are superior to adversarial ones; (3) binary neurons do not improve results; (4) cross-entropy loss results in better or equal performance compared to the F-measure loss. | en_US |
dc.format.extent | 470 - 477 | en_US |
dc.rights | This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/. | |
dc.rights | Attribution 3.0 United States | * |
dc.rights.uri | http://creativecommons.org/licenses/by/3.0/us/ | * |
dc.title | A Comparative Study of Neural Models for Polyphonic Music Sequence Transduction | en_US |
dc.type | Conference Proceeding | |
dc.rights.holder | © The Author(s) 2019 | |
pubs.notes | Not known | en_US |
pubs.publication-status | Accepted | en_US |
pubs.publisher-url | https://ismir2019.ewi.tudelft.nl/ | en_US |
dcterms.dateAccepted | 2019-06-07 | en_US |
rioxxterms.funder | Default funder | en_US |
rioxxterms.identifier.project | Default project | en_US |
qmul.funder | A Machine Learning Framework for Audio Analysis and Retrieval::Royal Academy of Engineering | en_US |
Files in this item
This item appears in the following Collection(s)
Except where otherwise noted, this item's license is described as This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.