Show simple item record

dc.contributor.authorLuo, D
dc.contributor.authorHuang, J
dc.contributor.authorGong, S
dc.contributor.authorJin, H
dc.contributor.authorLiu, Y
dc.date.accessioned2024-06-25T08:08:31Z
dc.date.available2024-06-25T08:08:31Z
dc.date.issued2024-01-01
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/97660
dc.description.abstractAccurate video moment retrieval (VMR) requires universal visual-textual correlations that can handle unknown vocabulary and unseen scenes. However, the learned correlations are likely either biased when derived from a limited amount of moment-text data which is hard to scale up because of the prohibitive annotation cost (fully-supervised), or unreliable when only the video-text pairwise relationships are available without fine-grained temporal annotations (weakly-supervised). Recently, the vision-language models (VLM) demonstrate a new transfer learning paradigm to benefit different vision tasks through the universal visual-textual correlations derived from large-scale vision-language pairwise web data, which has also shown benefits to VMR by fine-tuning in the target domains.In this work, we propose a zero-shot method for adapting generalisable visual-textual priors from arbitrary VLM to facilitate moment-text alignment, without the need for accessing the VMR data. To this end, we devise a conditional feature refinement module to generate boundary-aware visual features conditioned on text queries to enable better moment boundary understanding. Additionally, we design a bottom-up proposal generation strategy that mitigates the impact of domain discrepancies and breaks down complex-query retrieval tasks into individual action retrievals, thereby maximizing the benefits of VLM. Extensive experiments conducted on three VMR benchmark datasets demonstrate the notable performance advantages of our zero-shot algorithm, especially in the novel-word and novel-location out-of-distribution setups.en_US
dc.format.extent5452 - 5461
dc.publisherIEEEen_US
dc.titleZero-Shot Video Moment Retrieval from Frozen Vision-Language Modelsen_US
dc.typeConference Proceedingen_US
dc.rights.holder© 2024 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
dc.identifier.doi10.1109/WACV57701.2024.00538
pubs.notesNot knownen_US
pubs.publication-statusPublisheden_US
rioxxterms.funderDefault funderen_US
rioxxterms.identifier.projectDefault projecten_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record