dynamic spectrum access
RFRL Gym: A Reinforcement Learning Testbed for Cognitive Radio Applications
Rosen, Daniel, Rochez, Illa, McIrvin, Caleb, Lee, Joshua, D'Alessandro, Kevin, Wiecek, Max, Hoang, Nhan, Saffarini, Ramzy, Philips, Sam, Jones, Vanessa, Ivey, Will, Harris-Smart, Zavier, Harris-Smart, Zavion, Chin, Zayden, Johnson, Amos, Jones, Alyse M., Headley, William C.
Radio Frequency Reinforcement Learning (RFRL) is anticipated to be a widely applicable technology in the next generation of wireless communication systems, particularly 6G and next-gen military communications. Given this, our research is focused on developing a tool to promote the development of RFRL techniques that leverage spectrum sensing. In particular, the tool was designed to address two cognitive radio applications, specifically dynamic spectrum access and jamming. In order to train and test reinforcement learning (RL) algorithms for these applications, a simulation environment is necessary to simulate the conditions that an agent will encounter within the Radio Frequency (RF) spectrum. In this paper, such an environment has been developed, herein referred to as the RFRL Gym. Through the RFRL Gym, users can design their own scenarios to model what an RL agent may encounter within the RF spectrum as well as experiment with different spectrum sensing techniques. Additionally, the RFRL Gym is a subclass of OpenAI gym, enabling the use of third-party ML/RL Libraries. We plan to open-source this codebase to enable other researchers to utilize the RFRL Gym to test their own scenarios and RL algorithms, ultimately leading to the advancement of RL research in the wireless communications domain. This paper describes in further detail the components of the Gym, results from example scenarios, and plans for future additions. Index Terms-machine learning, reinforcement learning, wireless communications, dynamic spectrum access, OpenAI gym
- North America > United States > Virginia (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > Croatia > Primorje-Gorski Kotar County > Rijeka (0.04)
- Asia > China (0.04)
- Leisure & Entertainment (0.88)
- Media > Radio (0.54)
Dynamic Spectrum Access using Stochastic Multi-User Bandits
Bande, Meghana, Magesh, Akshayaa, Veeravalli, Venugopal V.
However, they assume that users have knowledge of the total number of users occupying their channel at any given time. Dynamic spectrum access has emerged to address the problem of spectrum under-utilization caused by treating the frequency On any given channel, we assume that the reward obtained spectrum as a fixed commodity. We study the spectrum is a random variable that is drawn from a distribution that sharing paradigm in which all the users are treated equally i.e., depends on the number of users on the channel. For example, there is no distinction between primary or secondary users. We the instantaneous reward could be the rate achieved by the user model the system as a stochastic multi-user multi-armed bandit on the channel which may decrease due to interference from (MAB) problem [1] where the channels correspond to the other users accessing the channel. The decrease in the reward arms of the bandit similar to the model considered in [2]-[11].
- North America > United States > Illinois > Champaign County > Champaign (0.05)
- North America > United States > New York (0.04)
- North America > United States > New Jersey > Somerset County > Bridgewater (0.04)
- Information Technology > Artificial Intelligence (0.94)
- Information Technology > Data Science > Data Mining > Big Data (0.51)
Spectrum Management in Dynamic Spectrum Access: A Deep Reinforcement Learning Approach
Generally, in dynamic spectrum access (DSA) networks, co-operations and centralized control are unavailable and DSA users have to carry out wireless transmissions individually. DSA users have to know other users' behaviors by sensing and analyzing wireless environments, so that DSA users can adjust their parameters properly and carry out effective wireless transmissions. In this thesis, machine learning and deep learning technologies are leveraged in DSA network to enable appropriate and intelligent spectrum managements, including both spectrum access and power allocations. Accordingly, a novel spectrum management framework utilizing deep reinforcement learning is proposed, in which deep reinforcement learning is employed to accurately learn wireless environments and generate optimal spectrum management strategies to adapt to the variations of wireless environments. Due to the model-free nature of reinforcement learning, DSA users only need to directly interact with environments to obtain optimal strategies rather than relying on accurate channel estimations.
- Telecommunications (1.00)
- Information Technology > Networks (1.00)
Distributive Dynamic Spectrum Access through Deep Reinforcement Learning: A Reservoir Computing Based Approach
Chang, Hao-Hsuan, Song, Hao, Yi, Yang, Zhang, Jianzhong, He, Haibo, Liu, Lingjia
Dynamic spectrum access (DSA) is regarded as an effective and efficient technology to share radio spectrum among different networks. As a secondary user (SU), a DSA device will face two critical problems: avoiding causing harmful interference to primary users (PUs), and conducting effective interference coordination with other secondary users. These two problems become even more challenging for a distributed DSA network where there is no centralized controllers for SUs. In this paper, we investigate communication strategies of a distributive DSA network under the presence of spectrum sensing errors. To be specific, we apply the powerful machine learning tool, deep reinforcement learning (DRL), for SUs to learn "appropriate" spectrum access strategies in a distributed fashion assuming NO knowledge of the underlying system statistics. Furthermore, a special type of recurrent neural network (RNN), called the reservoir computing (RC), is utilized to realize DRL by taking advantage of the underlying temporal correlation of the DSA network. Using the introduced machine learning-based strategy, SUs could make spectrum access decisions distributedly relying only on their own current and past spectrum sensing outcomes. Through extensive experiments, our results suggest that the RC-based spectrum access strategy can help the SU to significantly reduce the chances of collision with PUs and other SUs. We also show that our scheme outperforms the myopic method which assumes the knowledge of system statistics, and converges faster than the Q-learning method when the number of channels is large.
- Asia (0.93)
- North America > United States > Virginia (0.14)
- Telecommunications (1.00)
- Health & Medicine > Therapeutic Area (0.57)
- Government > Regional Government > North America Government > United States Government (0.46)
- Information Technology > Networks (0.46)