Show simple item record

dc.contributor.authorXu, M
dc.contributor.authorWang, Y
dc.contributor.authorXu, B
dc.contributor.authorZhang, J
dc.contributor.authorRen, J
dc.contributor.authorHuang, Z
dc.contributor.authorPoslad, S
dc.contributor.authorXu, P
dc.date.accessioned2024-07-12T08:31:22Z
dc.date.available2024-07-12T08:31:22Z
dc.date.issued2023-12-13
dc.identifier.citationMeng Xu, Youchen Wang, Bin Xu, Jun Zhang, Jian Ren, Zhao Huang, Stefan Poslad, Pengfei Xu, A critical analysis of image-based camera pose estimation techniques, Neurocomputing, Volume 570, 2024, 127125, ISSN 0925-2312, https://doi.org/10.1016/j.neucom.2023.127125. (https://www.sciencedirect.com/science/article/pii/S0925231223012481) Abstract: Camera, and associated with its objects within the field of view, localization could benefit many computer vision fields, such as autonomous driving, robot navigation, and augmented reality (AR). After decades of progress, camera localization, also called camera pose estimation could compute the 6DoF pose of objects for a camera in a given image, with respect to different images in a sequence or formats. Structure feature-based localization methods have achieved great success when integrated with image matching or with a coordinate regression stage. Absolute and relative pose regression methods using transfer learning can support end-to-end localization to directly regress a camera pose but achieve a less accurate performance. Despite the rapid development of multiple branches in this area, a comprehensive, in-depth and comparative analysis is lacking to summarize, classify and compare, structure feature-based and regression-based camera localization methods. Existing surveys either focus on larger SLAM (Simultaneous Localization and Mapping) systems or on only part of the camera localization method, lack detailed comparisons and descriptions of the methods or datasets used, neural network designs such as loss designs, and input formats, etc. In this survey, we first introduce specific application areas and the evaluation metrics for camera localization pose according to different sub-tasks (learning-based 2D-2D task, 2D-3D task, and 3D-3D task). Then, we review common methods for structure feature-based camera pose estimation approaches, absolute pose regression and relative pose regression approaches by critically modelling the methods to inspire further improvements in their algorithms such as loss functions, and neural network structures. Furthermore, we summarize what are the popular datasets used for camera localization and compare the quantitative and qualitative results of these methods with detailed performance metrics. Finally, we discuss future research possibilities and applications. Keywords: Camera pose regression; Structure feature-based localization; Absolute pose regression; Relative pose regressionen_US
dc.identifier.issn0925-2312
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/98057
dc.description.abstractCamera, and associated with its objects within the field of view, localization could benefit many computer vision fields, such as autonomous driving, robot navigation, and augmented reality (AR). After decades of progress, camera localization, also called camera pose estimation could compute the 6DoF pose of objects for a camera in a given image, with respect to different images in a sequence or formats. Structure feature-based localization methods have achieved great success when integrated with image matching or with a coordinate regression stage. Absolute and relative pose regression methods using transfer learning can support end-to-end localization to directly regress a camera pose but achieve a less accurate performance. Despite the rapid development of multiple branches in this area, a comprehensive, in-depth and comparative analysis is lacking to summarize, classify and compare, structure feature-based and regression-based camera localization methods. Existing surveys either focus on larger SLAM (Simultaneous Localization and Mapping) systems or on only part of the camera localization method, lack detailed comparisons and descriptions of the methods or datasets used, neural network designs such as loss designs, and input formats, etc. In this survey, we first introduce specific application areas and the evaluation metrics for camera localization pose according to different sub-tasks (learning-based 2D-2D task, 2D-3D task, and 3D-3D task). Then, we review common methods for structure feature-based camera pose estimation approaches, absolute pose regression and relative pose regression approaches by critically modelling the methods to inspire further improvements in their algorithms such as loss functions, and neural network structures. Furthermore, we summarize what are the popular datasets used for camera localization and compare the quantitative and qualitative results of these methods with detailed performance metrics. Finally, we discuss future research possibilities and applications.en_US
dc.publisherElsevieren_US
dc.relation.ispartofNeurocomputing
dc.titleA critical analysis of image-based camera pose estimation techniquesen_US
dc.typeArticleen_US
dc.rights.holder© 2023 Elsevier B.V. All rights reserved.
dc.identifier.doi10.1016/j.neucom.2023.127125
pubs.notesNot knownen_US
pubs.publication-statusPublisheden_US
pubs.volume570en_US
rioxxterms.funderDefault funderen_US
rioxxterms.identifier.projectDefault projecten_US
rioxxterms.funder.projectb215eee3-195d-4c4f-a85d-169a4331c138en_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record