Show simple item record

dc.contributor.authorManco, I
dc.contributor.authorBenetos, E
dc.contributor.authorQuinton, E
dc.contributor.authorFazekas, G
dc.contributor.authorInternational Joint Conference on Neural Networks (IJCNN)
dc.date.accessioned2021-05-25T15:07:11Z
dc.date.available2021-04-10
dc.date.available2021-05-25T15:07:11Z
dc.date.issued2021-07-18
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/72068
dc.description.abstractContent-based music information retrieval has seen rapid progress with the adoption of deep learning. Current approaches to high-level music description typically make use of classification models, such as in auto tagging or genre and mood classification. In this work, we propose to address music description via audio captioning, defined as the task of generating a natural language description of music audio content in a human-like manner. To this end, we present the first music audio captioning model, MusCaps, consisting of an encoder-decoder with temporal attention. Our method combines convolutional and recurrent neural network architectures to jointly process audio-text inputs through a multimodal encoder and leverages pre-training on audio data to obtain representations that effectively capture and summarise musical features in the input. Evaluation of the generated captions through automatic metrics shows that our method outperforms a baseline designed for non-music audio captioning. Through an ablation study, we unveil that this performance boost can be mainly attributed to pre-training of the audio encoder, while other design choices – modality fusion, decoding strategy and the use of attention -- contribute only marginally. Our model represents a shift away from classification-based music description and combines tasks requiring both auditory and linguistic understanding to bridge the semantic gap in music information retrieval.en_US
dc.format.extent? - ? (8)
dc.publisherIEEEen_US
dc.titleMusCaps: generating captions for music audioen_US
dc.typeConference Proceedingen_US
pubs.author-urlhttps://ilariamanco.com/en_US
pubs.notesNot knownen_US
pubs.publication-statusAccepteden_US
pubs.publisher-urlhttps://www.ijcnn.org/en_US
dcterms.dateAccepted2021-04-10
qmul.funderUKRI Centre for Doctoral Training in Artificial Intelligence and Music::Engineering and Physical Sciences Research Councilen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record