Goto

Collaborating Authors

 dynamic spectrum access


RFRL Gym: A Reinforcement Learning Testbed for Cognitive Radio Applications

Rosen, Daniel, Rochez, Illa, McIrvin, Caleb, Lee, Joshua, D'Alessandro, Kevin, Wiecek, Max, Hoang, Nhan, Saffarini, Ramzy, Philips, Sam, Jones, Vanessa, Ivey, Will, Harris-Smart, Zavier, Harris-Smart, Zavion, Chin, Zayden, Johnson, Amos, Jones, Alyse M., Headley, William C.

arXiv.org Artificial Intelligence

Radio Frequency Reinforcement Learning (RFRL) is anticipated to be a widely applicable technology in the next generation of wireless communication systems, particularly 6G and next-gen military communications. Given this, our research is focused on developing a tool to promote the development of RFRL techniques that leverage spectrum sensing. In particular, the tool was designed to address two cognitive radio applications, specifically dynamic spectrum access and jamming. In order to train and test reinforcement learning (RL) algorithms for these applications, a simulation environment is necessary to simulate the conditions that an agent will encounter within the Radio Frequency (RF) spectrum. In this paper, such an environment has been developed, herein referred to as the RFRL Gym. Through the RFRL Gym, users can design their own scenarios to model what an RL agent may encounter within the RF spectrum as well as experiment with different spectrum sensing techniques. Additionally, the RFRL Gym is a subclass of OpenAI gym, enabling the use of third-party ML/RL Libraries. We plan to open-source this codebase to enable other researchers to utilize the RFRL Gym to test their own scenarios and RL algorithms, ultimately leading to the advancement of RL research in the wireless communications domain. This paper describes in further detail the components of the Gym, results from example scenarios, and plans for future additions. Index Terms-machine learning, reinforcement learning, wireless communications, dynamic spectrum access, OpenAI gym


Dynamic Spectrum Access using Stochastic Multi-User Bandits

Bande, Meghana, Magesh, Akshayaa, Veeravalli, Venugopal V.

arXiv.org Machine Learning

However, they assume that users have knowledge of the total number of users occupying their channel at any given time. Dynamic spectrum access has emerged to address the problem of spectrum under-utilization caused by treating the frequency On any given channel, we assume that the reward obtained spectrum as a fixed commodity. We study the spectrum is a random variable that is drawn from a distribution that sharing paradigm in which all the users are treated equally i.e., depends on the number of users on the channel. For example, there is no distinction between primary or secondary users. We the instantaneous reward could be the rate achieved by the user model the system as a stochastic multi-user multi-armed bandit on the channel which may decrease due to interference from (MAB) problem [1] where the channels correspond to the other users accessing the channel. The decrease in the reward arms of the bandit similar to the model considered in [2]-[11].


Spectrum Management in Dynamic Spectrum Access: A Deep Reinforcement Learning Approach

#artificialintelligence

Generally, in dynamic spectrum access (DSA) networks, co-operations and centralized control are unavailable and DSA users have to carry out wireless transmissions individually. DSA users have to know other users' behaviors by sensing and analyzing wireless environments, so that DSA users can adjust their parameters properly and carry out effective wireless transmissions. In this thesis, machine learning and deep learning technologies are leveraged in DSA network to enable appropriate and intelligent spectrum managements, including both spectrum access and power allocations. Accordingly, a novel spectrum management framework utilizing deep reinforcement learning is proposed, in which deep reinforcement learning is employed to accurately learn wireless environments and generate optimal spectrum management strategies to adapt to the variations of wireless environments. Due to the model-free nature of reinforcement learning, DSA users only need to directly interact with environments to obtain optimal strategies rather than relying on accurate channel estimations.


Distributive Dynamic Spectrum Access through Deep Reinforcement Learning: A Reservoir Computing Based Approach

Chang, Hao-Hsuan, Song, Hao, Yi, Yang, Zhang, Jianzhong, He, Haibo, Liu, Lingjia

arXiv.org Machine Learning

Dynamic spectrum access (DSA) is regarded as an effective and efficient technology to share radio spectrum among different networks. As a secondary user (SU), a DSA device will face two critical problems: avoiding causing harmful interference to primary users (PUs), and conducting effective interference coordination with other secondary users. These two problems become even more challenging for a distributed DSA network where there is no centralized controllers for SUs. In this paper, we investigate communication strategies of a distributive DSA network under the presence of spectrum sensing errors. To be specific, we apply the powerful machine learning tool, deep reinforcement learning (DRL), for SUs to learn "appropriate" spectrum access strategies in a distributed fashion assuming NO knowledge of the underlying system statistics. Furthermore, a special type of recurrent neural network (RNN), called the reservoir computing (RC), is utilized to realize DRL by taking advantage of the underlying temporal correlation of the DSA network. Using the introduced machine learning-based strategy, SUs could make spectrum access decisions distributedly relying only on their own current and past spectrum sensing outcomes. Through extensive experiments, our results suggest that the RC-based spectrum access strategy can help the SU to significantly reduce the chances of collision with PUs and other SUs. We also show that our scheme outperforms the myopic method which assumes the knowledge of system statistics, and converges faster than the Q-learning method when the number of channels is large.