LiDAR-BIND-T: Improved and Temporally Consistent Sensor Modality Translation and Fusion for Robotic Applications
Balemans, Niels, Anwar, Ali, Steckel, Jan, Mercelis, Siegfried
–arXiv.org Artificial Intelligence
This paper extends LiDAR-BIND, a modular multi-modal fusion framework that binds heterogeneous sensors (radar, sonar) to a LiDAR-defined latent space, with mechanisms that explicitly enforce temporal consistency. We introduce three contributions: (i) temporal embedding similarity that aligns consecutive latent representations, (ii) a motion-aligned transformation loss that matches displacement between predictions and ground truth LiDAR, and (iii) windowed temporal fusion using a specialised temporal module. We further update the model architecture to better preserve spatial structure. Evaluations on radar/sonar-to-LiDAR translation demonstrate improved temporal and spatial coherence, yielding lower absolute trajectory error and better occupancy map accuracy in Cartographer-based SLAM (Simultaneous Localisation and Mapping). We propose different metrics based on the Fréchet Video Motion Distance (FVMD) and a correlation-peak distance metric providing practical temporal quality indicators to evaluate SLAM performance. The proposed temporal LiDAR-BIND, or LiDAR-BIND-T, maintains modular modality fusion while substantially enhancing temporal stability, resulting in improved robustness and performance for downstream SLAM.
arXiv.org Artificial Intelligence
Oct-1-2025
- Country:
- Europe > Belgium
- Flanders > Antwerp Province > Antwerp (0.04)
- North America > United States
- Texas (0.04)
- Europe > Belgium
- Genre:
- Research Report > New Finding (0.93)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning > Neural Networks
- Deep Learning (0.93)
- Representation & Reasoning > Information Fusion (1.00)
- Robots (1.00)
- Vision (1.00)
- Machine Learning > Neural Networks
- Data Science (1.00)
- Sensing and Signal Processing (1.00)
- Artificial Intelligence
- Information Technology