Fusion of Pervasive RF Data with Spatial Images via Vision Transformers for Enhanced Mapping in Smart Cities

Mkrtchyan, Rafayel, Manukyan, Armen, Khachatrian, Hrant, Raptis, Theofanis P.

arXiv.org Artificial Intelligence 

Accurate environment mapping is an important computing task for a wide range of smart city applications, including autonomous navigation, wireless network operations and extended reality environments. On the one hand, conventional smart city mapping techniques, such as satellite imagery, LiDAR scans, and manual annotations, often su ff er from limitations related to cost, accessibility and accuracy. On the other hand, open-source mapping platforms, such as OpenStreetMap, have been widely utilized in artificial intelligence (AI) applications for environment mapping, serving as a source of ground truth. However, human errors and the evolving nature of real-world environments introduce biases that can negatively impact the performance of neural networks trained on such data. In this paper, we present a deep learning-based approach that integrates the DINOv2 architecture to improve building mapping by combining (possibly erroneous) maps from open-source platforms with pervasive radio frequency (RF) data collected from multiple wireless user equipments and base stations. Unlike prior methods, our approach leverages a vision transformer-based architecture to jointly process both RF and map modalities within a unified framework, e ffectively capturing spatial dependencies and structural priors for enhanced mapping accuracy. For the evaluation purposes, we employ a synthetic dataset co-produced by Huawei. To address the challenges associated with real-world data imperfections, we introduce controlled noise to its RF data so as to simulate real-world conditions. Additionally, we develop and train a model that leverages only aggregated path loss information to tackle the mapping problem. We measure the results according to three performance metrics which capture di fferent qualities: (i) The Jaccard index, also known as intersection over union (IoU), (ii) the Hausdor ff distance, and (iii) the Chamfer distance. Our design achieves a macro IoU of 65.3%, significantly surpassing (i) the erroneous maps baseline, which yields 40.1%, (ii) an RF-only method from the literature, which yields 37.3%, and (iii) a non-AI fusion baseline that we designed which yields 42.2%. The comparative evaluation highlights the limitations of relying solely on RF data or on spatial data, as well as the e ff ectiveness that AI can have on fusing data towards enhancing smart city mapping accuracy. Introduction Smart cities, characterized by their pervasive integration of digital technologies [8] and interconnected systems [6], face unique challenges in accurately capturing and updating the physical and dynamic characteristics of urban spaces.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found