Goto

Collaborating Authors

 control message


Causal Model-Based Reinforcement Learning for Sample-Efficient IoT Channel Access

Arun, Aswin, Thomas, Christo Kurisummoottil, Sarvendranath, Rimalpudi, Saad, Walid

arXiv.org Artificial Intelligence

Despite the advantages of multi-agent reinforcement learning (MARL) for wireless use case such as medium access control (MAC), their real-world deployment in Internet of Things (IoT) is hindered by their sample inefficiency. To alleviate this challenge, one can leverage model-based reinforcement learning (MBRL) solutions, however, conventional MBRL approaches rely on black-box models that are not interpretable and cannot reason. In contrast, in this paper, a novel causal model-based MARL framework is developed by leveraging tools from causal learn- ing. In particular, the proposed model can explicitly represent causal dependencies between network variables using structural causal models (SCMs) and attention-based inference networks. Interpretable causal models are then developed to capture how MAC control messages influence observations, how transmission actions determine outcomes, and how channel observations affect rewards. Data augmentation techniques are then used to generate synthetic rollouts using the learned causal model for policy optimization via proximal policy optimization (PPO). Analytical results demonstrate exponential sample complexity gains of causal MBRL over black-box approaches. Extensive simulations demonstrate that, on average, the proposed approach can reduce environment interactions by 58%, and yield faster convergence compared to model-free baselines. The proposed approach inherently is also shown to provide interpretable scheduling decisions via attention-based causal attribution, revealing which network conditions drive the policy. The resulting combination of sample efficiency and interpretability establishes causal MBRL as a practical approach for resource-constrained wireless systems.


Optimizing Multi-Gateway LoRaWAN via Cloud-Edge Collaboration and Knowledge Distillation

Yang, Hong

arXiv.org Artificial Intelligence

For large-scale multi-gateway LoRaWAN networks, this study proposes a cloud-edge collaborative resource allocation and decision-making method based on edge intelligence, HEAT-LDL (HEAT-Local Distill Lyapunov), which realizes collaborative decision-making between gateways and terminal nodes. HEAT-LDL combines the Actor-Critic architecture and the Lyapunov optimization method to achieve intelligent downlink control and gateway load balancing. When the signal quality is good, the network server uses the HEAT algorithm to schedule the terminal nodes. To improve the efficiency of autonomous decision-making of terminal nodes, HEAT-LDL performs cloud-edge knowledge distillation on the HEAT teacher model on the terminal node side. When the downlink decision instruction is lost, the terminal node uses the student model and the edge decider based on prior knowledge and local history to make collaborative autonomous decisions. Simulation experiments show that compared with the optimal results of all compared algorithms, HEAT-LDL improves the packet success rate and energy efficiency by 20.5% and 88.1%, respectively.


A Delay-free Control Method Based On Function Approximation And Broadcast For Robotic Surface And Multiactuator Systems

Zhao, Yuchen

arXiv.org Artificial Intelligence

Robotic surface consisting of many actuators can change shape to perform tasks, such as facilitating human-machine interactions and transporting objects. Increasing the number of actuators can enhance the robot's capacity, but controlling them requires communication bandwidth to increase equally in order to avoid time delays. We propose a novel control method that has constant time delays no matter how many actuators are in the robot. Having a distributed nature, the method first approximates target shapes, then broadcasts the approximation coefficients to the actuators, and relies on themselves to compute the inputs. We build a robotic pin array and measure the time delay as a function of the number of actuators to confirm the system size-independent scaling behavior. The shape-changing ability is achieved based on function approximation algorithms, i.e. discrete cosine transform or matching pursuit. We perform experiments to approximate target shapes and make quantitative comparison with those obtained from standard sequential control method. A good agreement between the experiments and theoretical predictions is achieved, and our method is more efficient in the sense that it requires less control messages to generate shapes with the same accuracy. Our method is also capable of dynamic tasks such as object manipulation.


A Zero Trust Framework for Realization and Defense Against Generative AI Attacks in Power Grid

Munir, Md. Shirajum, Proddatoori, Sravanthi, Muralidhara, Manjushree, Saad, Walid, Han, Zhu, Shetty, Sachin

arXiv.org Artificial Intelligence

Understanding the potential of generative AI (GenAI)-based attacks on the power grid is a fundamental challenge that must be addressed in order to protect the power grid by realizing and validating risk in new attack vectors. In this paper, a novel zero trust framework for a power grid supply chain (PGSC) is proposed. This framework facilitates early detection of potential GenAI-driven attack vectors (e.g., replay and protocol-type attacks), assessment of tail risk-based stability measures, and mitigation of such threats. First, a new zero trust system model of PGSC is designed and formulated as a zero-trust problem that seeks to guarantee for a stable PGSC by realizing and defending against GenAI-driven cyber attacks. Second, in which a domain-specific generative adversarial networks (GAN)-based attack generation mechanism is developed to create a new vulnerability cyberspace for further understanding that threat. Third, tail-based risk realization metrics are developed and implemented for quantifying the extreme risk of a potential attack while leveraging a trust measurement approach for continuous validation. Fourth, an ensemble learning-based bootstrap aggregation scheme is devised to detect the attacks that are generating synthetic identities with convincing user and distributed energy resources device profiles. Experimental results show the efficacy of the proposed zero trust framework that achieves an accuracy of 95.7% on attack vector generation, a risk measure of 9.61% for a 95% stable PGSC, and a 99% confidence in defense against GenAI-driven attack.


Learning Failure-Inducing Models for Testing Software-Defined Networks

Ollando, Raphaël, Shin, Seung Yeob, Briand, Lionel C.

arXiv.org Artificial Intelligence

Software-defined networks (SDN) enable flexible and effective communication systems that are managed by centralized software controllers. However, such a controller can undermine the underlying communication network of an SDN-based system and thus must be carefully tested. When an SDN-based system fails, in order to address such a failure, engineers need to precisely understand the conditions under which it occurs. In this article, we introduce a machine learning-guided fuzzing method, named FuzzSDN, aiming at both (1) generating effective test data leading to failures in SDN-based systems and (2) learning accurate failure-inducing models that characterize conditions under which such system fails. To our knowledge, no existing work simultaneously addresses these two objectives for SDNs. We evaluate FuzzSDN by applying it to systems controlled by two open-source SDN controllers. Further, we compare FuzzSDN with two state-of-the-art methods for fuzzing SDNs and two baselines for learning failure-inducing models. Our results show that (1) compared to the state-of-the-art methods, FuzzSDN generates at least 12 times more failures, within the same time budget, with a controller that is fairly robust to fuzzing and (2) our failure-inducing models have, on average, a precision of 98% and a recall of 86%, significantly outperforming the baselines.


Conflict Mitigation Framework and Conflict Detection in O-RAN Near-RT RIC

Adamczyk, Cezary, Kliks, Adrian

arXiv.org Artificial Intelligence

The steady evolution of the Open RAN concept sheds light on xApps and their potential use cases in O-RANcompliant deployments. There are several areas where xApps can be used that are being widely investigated, but the issue of mitigating conflicts between xApp decisions requires further in-depth investigation. This article defines a conflict mitigation framework (CMF) built into the existing O-RAN architecture; it enables the Conflict Mitigation component in O-RAN's Near- Real-Time RAN Intelligent Controller (Near-RT RIC) to detect and resolve all conflict types defined in the O-RAN Alliance's technical specifications. Methods for detecting each type of conflict are defined, including message flows between Near-RT RIC components. The suitability of the proposed CMF is proven with a simulation of an O-RAN network. Results of the simulation show that enabling the CMF allows balancing the network control capabilities of conflicting xApps to significantly improve network performance, with a small negative impact on its reliability. It is concluded that defining a unified CMF in Near-RT RIC is the first step towards providing a standardized method of conflict detection and resolution in O-RAN environments.


The Emergence of Wireless MAC Protocols with Multi-Agent Reinforcement Learning

Mota, Mateus P., Valcarce, Alvaro, Gorce, Jean-Marie, Hoydis, Jakob

arXiv.org Artificial Intelligence

In this paper, we propose a new framework, exploiting the multi-agent deep deterministic policy gradient (MADDPG) algorithm, to enable a base station (BS) and user equipment (UE) to come up with a medium access control (MAC) protocol in a multiple access scenario. In this framework, the BS and UEs are reinforcement learning (RL) agents that need to learn to cooperate in order to deliver data. The network nodes can exchange control messages to collaborate and deliver data across the network, but without any prior agreement on the meaning of the control messages. In such a framework, the agents have to learn not only the channel access policy, but also the signaling policy. The collaboration between agents is shown to be important, by comparing the proposed algorithm to ablated versions where either the communication between agents or the central critic is removed. The comparison with a contention-free baseline shows that our framework achieves a superior performance in terms of goodput and can effectively be used to learn a new protocol.


How to Deploy AI Inference on the Edge with the LG AIoT Board and AWS IoT Greengrass

#artificialintelligence

With so many cloud applications infused with artificial intelligence (AI) and machine learning (ML) capabilities, AI/ML is being democratized by cloud services. The growth of AI in a wide range of applications demands more purpose-built processors to provide scalable levels of performance, flexibility, and efficiency. The LG AIoT board helps customers accelerate their computer vision and ML journey using Amazon Web Services (AWS). OEMs can now easily incorporate visual intelligence, voice intelligence, and control intelligence into their products. The LG Neural Engine (LNE) in the LG AIoT board offloads the compute requirements of deep learning algorithms to the specially designed processor, which delivers 1 TFLOPS of compute performance.