Goto

Collaborating Authors

 dbm



Exact Recovery in the Data Block Model

Asadi, Amir R., Davoodi, Akbar, Javadi, Ramin, Parvaresh, Farzad

arXiv.org Machine Learning

Community detection in networks is a fundamental problem in machine learning and statistical inference, with applications in social networks, biological systems, and communication networks. The stochastic block model (SBM) serves as a canonical framework for studying community structure, and exact recovery, identifying the true communities with high probability, is a central theoretical question. While classical results characterize the phase transition for exact recovery based solely on graph connectivity, many real-world networks contain additional data, such as node attributes or labels. In this work, we study exact recovery in the Data Block Model (DBM), an SBM augmented with node-associated data, as formalized by Asadi, Abbe, and Verdú (2017). We introduce the Chernoff--TV divergence and use it to characterize a sharp exact recovery threshold for the DBM. We further provide an efficient algorithm that achieves this threshold, along with a matching converse result showing impossibility below the threshold. Finally, simulations validate our findings and demonstrate the benefits of incorporating vertex data as side information in community detection.


JaGuard: Jamming Correction of GNSS Deviation with Deep Temporal Graphs

Kesić, Ivana, Blatnik, Aljaž, Fortuna, Carolina, Bertalanič, Blaž

arXiv.org Artificial Intelligence

Abstract--Global Navigation Satellite Systems (GNSS) face growing disruption from intentional jamming, undermining availability exactly when reliable positioning and timing are essential. We tackle this challenge by recasting jamming mitigation as a dynamic graph regression problem and propose a Jamming Guardian (JaGuard), a new receiver-centric deep temporal graph network-based method that estimates, and thereby corrects, the receiver's latitude and longitude errors. At each 1 Hz epoch, we model the satellite-receiver scene as a heterogeneous star graph with the receiver as the center node and the tracked satellites as leaves. These satellites have time-varying attributes such as SNR, azimuth, elevation, and latitude/longitude. A single-layer Heterogeneous Graph ConvLSTM (HeteroGCLSTM) fuses one-hop spatial context with short-term temporal dynamics to produce a 2D deviation vector for error mitigation. We evaluate our approach on datasets collected from physical hardware (two different commercial receivers), subjected to controlled conducted RF interference. Interference is introduced with three jammer types: Continuous Wave CW, multi-tone 3 CW, and wideband FM. Each jammer type was exercised at six power levels from 45 to 70 dBm, with 50 repetitions per scenario, including pre-jam, jam, and recovery phases. Compared to strong multivariate time series baselines (TSMixer MLP, uniform CNN, and Seq2Point CNN), our model consistently yields the lowest Mean Absolute Error (MAE) in positional deviation. Under severe jamming at 45 dBm, it achieves an MAE of 3.64-7.74 On mixed-mode datasets that pool all power levels, the MAE is 3.78 cm for GP01 and 4.25 cm for U-blox 10, surpassing Seq2Point, TSMixer, and uniform CNN. A data-efficiency split further shows that with only 10% of the training data, our approach remains clearly ahead, achieving an MAE of about 20 cm versus 36-42 cm for the baselines. Global Navigation Satellite Systems (GNSS) underpin nearly every critical infrastructure, from telecommunications [1] and aviation safety [2], power-grid synchronization [3], emerging drone ecosystems where location privacy and integrity are paramount [4], to autonomous driving [5].






0bb4aec1710521c12ee76289d9440817-Reviews.html

Neural Information Processing Systems

First provide a summary of the paper, and then address the following criteria: Quality, clarity, originality and significance. The paper presents a method for learning layers of representation and for completing missing queries both in input and labels in single procedure unlike some other methods like deep boltzmann machines (DBM). It is a recurrent net following the same operations as DBM with the goal of predicting a subset of inputs from its complement. Parts of paper are badly written, especially model explanation and multi-inference section, nevertheless the paper should be published and I hope the authors will rewrite them. Details: - The procedure is taken from DBM, however other then that, is there a relation between the DBM and this algorithm, or should we just treat the algorithm as one particular function (recurrent net (RNN)) that predicts subset of inputs from its complement?


Semantic-Aware Edge Intelligence for UAV Handover in 6G Networks

Al-Hameed, Aubida A., Qazzaz, Mohammed M. H., Hafeez, Maryam, Zaidi, Syed A.

arXiv.org Artificial Intelligence

6G wireless networks aim to exploit semantic awareness to optimize radio resources. By optimizing the transmission through the lens of the desired goal, the energy consumption of transmissions can also be reduced, and the latency can be improved. To that end, this paper investigates a paradigm in which the capabilities of generative AI (GenAI) on the edge are harnessed for network optimization. In particular, we investigate an Unmanned Aerial Vehicle (UAV) handover framework that takes advantage of GenAI and semantic communication to maintain reliable connectivity. To that end, we propose a framework in which a lightweight MobileBERT language model, fine-tuned using Low-Rank Adaptation (LoRA), is deployed on the UAV. This model processes multi-attribute flight and radio measurements and performs multi-label classification to determine appropriate handover action. Concurrently, the model identifies an appropriate set of contextual "Reason Tags" that elucidate the decision's rationale. Our model, evaluated on a rule-based synthetic dataset of UAV handover scenarios, demonstrates the model's high efficacy in learning these rules, achieving high accuracy in predicting the primary handover decision. The model also shows strong performance in identifying supporting reasons, with an F1 micro-score of approximately 0.9 for reason tags.


DO-EM: Density Operator Expectation Maximization

Vishnu, Adit, Shastry, Abhay, Kashyap, Dhruva, Bhattacharyya, Chiranjib

arXiv.org Machine Learning

Density operators, quantum generalizations of probability distributions, are gaining prominence in machine learning due to their foundational role in quantum computing. Generative modeling based on density operator models (\textbf{DOMs}) is an emerging field, but existing training algorithms -- such as those for the Quantum Boltzmann Machine -- do not scale to real-world data, such as the MNIST dataset. The Expectation-Maximization algorithm has played a fundamental role in enabling scalable training of probabilistic latent variable models on real-world datasets. \textit{In this paper, we develop an Expectation-Maximization framework to learn latent variable models defined through \textbf{DOMs} on classical hardware, with resources comparable to those used for probabilistic models, while scaling to real-world data.} However, designing such an algorithm is nontrivial due to the absence of a well-defined quantum analogue to conditional probability, which complicates the Expectation step. To overcome this, we reformulate the Expectation step as a quantum information projection (QIP) problem and show that the Petz Recovery Map provides a solution under sufficient conditions. Using this formulation, we introduce the Density Operator Expectation Maximization (DO-EM) algorithm -- an iterative Minorant-Maximization procedure that optimizes a quantum evidence lower bound. We show that the \textbf{DO-EM} algorithm ensures non-decreasing log-likelihood across iterations for a broad class of models. Finally, we present Quantum Interleaved Deep Boltzmann Machines (\textbf{QiDBMs}), a \textbf{DOM} that can be trained with the same resources as a DBM. When trained with \textbf{DO-EM} under Contrastive Divergence, a \textbf{QiDBM} outperforms larger classical DBMs in image generation on the MNIST dataset, achieving a 40--60\% reduction in the Fréchet Inception Distance.