Show simple item record

dc.contributor.authorWu, Gen_US
dc.contributor.authorZhu, Xen_US
dc.contributor.authorGong, Sen_US
dc.date.accessioned2023-08-31T12:39:59Z
dc.date.issued2020-01-01en_US
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/90325
dc.description.abstractLearning discriminative spatio-temporal representation is the key for solving video re-identification (re-id) challenges. Most existing methods focus on learning appearance features and/or selecting image frames, but ignore optimising the compatibility and interaction of appearance and motion attentive information. To address this limitation, we propose a novel model to learning Spatio-Temporal Associative Representation (STAR). We design local frame-level spatio-temporal association to learn discriminative attentive appearance and short-term motion features, and global video-level spatio-temporal association to form compact and discriminative holistic video representation. We further introduce a pyramid ranking regulariser for facilitating end-to-end model optimisation. Extensive experiments demonstrate the superiority of STAR against state-of-the-art methods on four video re-id benchmarks, including MARS, DukeMTMC-VideoReID, iLIDS-VID and PRID-2011.en_US
dc.titleSpatio-temporal associative representation for video person re-identificationen_US
dc.typeConference Proceeding
pubs.notesNot knownen_US
pubs.publication-statusPublisheden_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record