LoCUS: Learning Multiscale 3D-consistent Features from Posed Images

dc.contributor.authorKloepfer, Dominik A.en
dc.contributor.authorCampbell, Dylanen
dc.contributor.authorHenriques, João F.en
dc.date.accessioned2025-06-19T04:31:58Z
dc.date.available2025-06-19T04:31:58Z
dc.date.issued2023-01-15en
dc.description.abstractAn important challenge for autonomous agents such as robots is to maintain a spatially and temporally consistent model of the world. It must be maintained through occlusions, previously-unseen views, and long time horizons (e.g., loop closure and re-identification). It is still an open question how to train such a versatile neural representation without supervision. We start from the idea that the training objective can be framed as a patch retrieval problem: given an image patch in one view of a scene, we would like to retrieve (with high precision and recall) all patches in other views that map to the same real-world location. One drawback is that this objective does not promote reusability of features: by being unique to a scene (achieving perfect precision/recall), a representation will not be useful in the context of other scenes. We find that it is possible to balance retrieval and reusability by constructing the retrieval set carefully, leaving out patches that map to far-away locations. Similarly, we can easily regulate the scale of the learned features (e.g., points, objects, or rooms) by adjusting the spatial tolerance for considering a retrieval to be positive. We optimize for (smooth) Average Precision (AP), in a single unified ranking-based objective. This objective also doubles as a criterion for choosing landmarks or keypoints, as patches with high AP. We show results creating sparse, multi-scale, semantic spatial maps composed of highly identifiable landmarks, with applications in landmark retrieval, localization, semantic segmentation and instance segmentation.en
dc.description.sponsorshipAcknowledgements. We are grateful for funding from EPSRC AIMS CDT EP/S024050/1 (D.K.), Continental AG (D.C.), and the Royal Academy of Engineering (RF/201819/18/163, J.H.).en
dc.description.statusPeer-revieweden
dc.format.extent11en
dc.identifier.isbn9798350307184en
dc.identifier.issn1550-5499en
dc.identifier.otherScopus:85185876179en
dc.identifier.otherARIES:a383154xPUB47146en
dc.identifier.otherORCID:/0000-0002-4717-6850/work/162523125en
dc.identifier.urihttp://www.scopus.com/inward/record.url?scp=85185876179&partnerID=8YFLogxKen
dc.identifier.urihttps://hdl.handle.net/1885/733764478
dc.language.isoenen
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en
dc.relation.ispartofseries2023 IEEE/CVF International Conference on Computer Vision, ICCV 2023en
dc.relation.ispartofseriesProceedings of the IEEE International Conference on Computer Visionen
dc.rightsPublisher Copyright: © 2023 IEEE.en
dc.titleLoCUS: Learning Multiscale 3D-consistent Features from Posed Imagesen
dc.typeConference paperen
local.bibliographicCitation.lastpage16598en
local.bibliographicCitation.startpage16588en
local.contributor.affiliationKloepfer, Dominik A.; University of Oxforden
local.contributor.affiliationCampbell, Dylan; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.contributor.affiliationHenriques, João F.; University of Oxforden
local.identifier.doi10.1109/ICCV51070.2023.01525en
local.identifier.pure3c2bc6cd-317d-43c7-bb09-b039f8f48a19en
local.type.statusPublisheden

Downloads