Show simple item record

dc.contributor.authorBohdal, O
dc.contributor.authorYang, Y
dc.contributor.authorHospedales, T
dc.contributor.authorNeurIPS
dc.date.accessioned2024-07-16T08:14:00Z
dc.date.available2021-01-01
dc.date.available2024-07-16T08:14:00Z
dc.date.issued2021-01-01
dc.identifier.issn1049-5258
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/98160
dc.description.abstractGradient-based meta-learning and hyperparameter optimization have seen significant progress recently, enabling practical end-to-end training of neural networks together with many hyperparameters. Nevertheless, existing approaches are relatively expensive as they need to compute second-order derivatives and store a longer computational graph. This cost prevents scaling them to larger network architectures. We present EvoGrad, a new approach to meta-learning that draws upon evolutionary techniques to more efficiently compute hypergradients. EvoGrad estimates hypergradient with respect to hyperparameters without calculating second-order gradients, or storing a longer computational graph, leading to significant improvements in efficiency. We evaluate EvoGrad on three substantial recent meta-learning applications, namely cross-domain few-shot learning with feature-wise transformations, noisy label learning with Meta-Weight-Net and low-resource cross-lingual learning with meta representation transformation. The results show that EvoGrad significantly improves efficiency and enables scaling meta-learning to bigger architectures such as from ResNet10 to ResNet34.en_US
dc.format.extent22234 - 22246
dc.titleEvoGrad: Efficient Gradient-Based Meta-Learning and Hyperparameter Optimizationen_US
dc.typeConference Proceedingen_US
pubs.notesNot knownen_US
pubs.publication-statusPublisheden_US
pubs.volume27en_US
dcterms.dateAccepted2021-01-01


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record