dc.contributor.author | Liang, J | en_US |
dc.contributor.author | Phan, QH | en_US |
dc.contributor.author | Benetos, E | en_US |
dc.contributor.author | 7th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE) | en_US |
dc.date.accessioned | 2022-10-27T09:26:33Z | |
dc.date.available | 2022-09-15 | en_US |
dc.date.issued | 2022-11-03 | en_US |
dc.identifier.uri | https://qmro.qmul.ac.uk/xmlui/handle/123456789/82109 | |
dc.description.abstract | Everyday sounds cover a considerable range of sound categories in our daily life, yet for certain sound categories it is hard to collect sufficient data. Although existing works have applied few-shot learning paradigms to sound recognition successfully, most of them have not exploited the relationship between labels in audio taxonomies. This work adopts a hierarchical prototypical network to leverage the knowledge rooted in audio taxonomies. Specifically, a VGG-like convolutional neural network is used to extract acoustic features. Prototypical nodes are then calculated in each level of the tree structure. A multi-level loss is obtained by multiplying a weight decay with multiple losses. Experimental results demonstrate our hierarchical prototypical networks not only outperform prototypical networks with no hierarchy information but yield a better result than other state-of-the art algorithms. Our code is available in: https://github.com/JinhuaLiang/HPNs_tagging | en_US |
dc.format.extent | ? - ? (5) | en_US |
dc.subject | everyday sound recognition | en_US |
dc.subject | few-shot learning | en_US |
dc.subject | hierarchical prototypical network | en_US |
dc.title | Leveraging label hierarchies for few-shot everyday sound recognition | en_US |
dc.type | Conference Proceeding | |
pubs.author-url | https://jinhualiang.github.io/ | en_US |
pubs.notes | Not known | en_US |
pubs.publication-status | Accepted | en_US |
pubs.publisher-url | https://dcase.community/workshop2022/ | en_US |
dcterms.dateAccepted | 2022-09-15 | en_US |
rioxxterms.funder | Default funder | en_US |
rioxxterms.identifier.project | Default project | en_US |
qmul.funder | AI for everyday sounds::Engineering and Physical Sciences Research Council | en_US |