Unifying Local and Global Multimodal Features for Place Recognition in Aliased and Low-Texture Environments
García-Hernández, Alberto, Giubilato, Riccardo, Strobl, Klaus H., Civera, Javier, Triebel, Rudolph
–arXiv.org Artificial Intelligence
Abstract-- Perceptual aliasing and weak textures pose significant challenges to the task of place recognition, hindering the performance of Simultaneous Localization and Mapping (SLAM) systems. This paper presents a novel model, called UMF (standing for Unifying Local and Global Multimodal Features) that 1) leverages multi-modality by crossattention blocks between vision and LiDAR features, and 2) includes a re-ranking stage that re-orders based on local feature matching the top-k candidates retrieved using a global representation. UMF outperforms significantly previous baselines in those challenging aliased environments. Simultaneous Localization and Mapping (SLAM) has emerged as a central technology in a multitude of industries In this paper we propose a novel multimodal place recognition including autonomous driving [1], [2], automated method that we denote as UMF, standing for Unifying construction [3], and agriculture [4], [5]. Our model leverages and adoption have been accelerated by advancements in deep local and global features from visual and LiDAR data, sensor technologies, including multi-camera setups, RGB-D fusing both modalities via cross-attention mechanisms.
arXiv.org Artificial Intelligence
Mar-20-2024
- Country:
- Europe > Spain
- Aragón (0.14)
- North America > United States
- Texas (0.14)
- Europe > Spain
- Genre:
- Research Report (0.84)
- Industry:
- Food & Agriculture > Agriculture (0.34)
- Information Technology (0.49)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning > Neural Networks (0.69)
- Robots (0.92)
- Vision (1.00)
- Sensing and Signal Processing > Image Processing (0.93)
- Artificial Intelligence
- Information Technology